More stories

  • in

    Plans in Congress on China and TikTok Face Hurdles After Spy Balloon Furor

    With budgets tight and political knives drawn, lawmakers seeking to capitalize on a bipartisan urgency to confront China are setting their sights on narrower measures.WASHINGTON — Republicans and Democrats are pressing for major legislation to counter rising threats from China, but mere weeks into the new Congress, a bipartisan consensus is at risk of dissipating amid disputes about what steps to take and a desire among many Republicans to wield the issue as a weapon against President Biden.In the House and Senate, leading lawmakers in both parties have managed in an otherwise bitterly divided Congress to stay unified about the need to confront the dangers posed by China’s militarization, its deepening ties with Russia and its ever-expanding economic footprint.But a rising chorus of Republican vitriol directed at Mr. Biden after a Chinese spy balloon flew over the United States this month upended that spirit — giving way to G.O.P. accusations that the president was “weak on China” — and suggested that the path ahead for any bipartisan action is exceedingly narrow.“When the balloon story popped, so to speak, it felt like certain people used that as an opportunity to bash President Biden,” said Representative Raja Krishnamoorthi of Illinois, the top Democrat on the select panel the House created to focus on competition with China.“And it felt like no matter what he did, they wanted to basically call him soft on the C.C.P., and unable to protect America,” he said, referring to the Chinese Communist Party. “That’s where I think we can go wayward politically,”For now, only a few, mostly narrow ventures have drawn enough bipartisan interest to have a chance at advancing amid the political tide. They include legislation to ban TikTok, the Beijing-based social media platform lawmakers have warned for years is an intelligence-gathering gold mine for the Chinese government; bills that would ban Chinese purchases of farmland and other agricultural real estate, especially in areas near sensitive military sites; and measures to limit U.S. exports and outbound investments to China.Such initiatives are limited in scope, predominantly defensive and relatively cheap — which lawmakers say are important factors in getting legislation over the hurdles posed by this split Congress. And, experts point out, none are issues that would be felt keenly by voters, or translate particularly well into political pitches on the 2024 campaign trail.A Divided CongressThe 118th Congress is underway, with Republicans controlling the House and Democrats holding the Senate.Jan. 6 Video: Speaker Kevin McCarthy’s decision to grant the Fox News host Tucker Carlson access to thousands of hours of Jan. 6 Capitol security footage has effectively outsourced a bid to reinvestigate the attack.John Fetterman: The Democratic senator from Pennsylvania is the latest public figure to disclose his mental health struggles, an indication of growing acceptance, though some stigma remains.Entitlement Cuts: Under bipartisan pressure, Senator Rick Scott of Florida, a Republican, exempted Social Security and Medicare from his proposal to regularly review all federal programs.G.O.P. Legislative Agenda: Weeks into their chaotic House majority, Republican leaders have found themselves paralyzed on some of the biggest issues they promised to address.“There would be nervousness among Republicans about giving the administration a clear win, but I’m just not sure that the kind of legislation they’ll be looking at would be doing that,” said Zack Cooper, who researches U.S.-China competition at the American Enterprise Institute. “It’s more things that would penalize China than be focused on investing in the U.S. in the next couple of years.”At the start of the year, the momentum behind bipartisan efforts to confront China seemed strong, with Republicans and Democrats banding together to pass the bill setting up the select panel and legislation to deny China crude oil exports from the U.S. Strategic Petroleum Reserve. A resolution condemning Beijing for sending the spy balloon over the United States passed unanimously after Republican leaders decided not to take the opportunity to rebuke Mr. Biden, as many on the right had clamored for.But with partisan divisions beginning to intensify and a presidential election looming, it appears exceedingly unlikely that Congress will be able to muster an agreement as large or significant as the major legislation last year to subsidize microchip manufacturing and scientific research — a measure that members of both parties described as only one of many policy changes that would be needed to counter China. Only a few, mostly narrow ventures have drawn enough bipartisan interest to have a chance at advancing amid the political tide.Kenny Holston/The New York Times“The biggest challenge is just the overall politicized environment that we’re in right now and the lack of trust between the parties,” said Representative Mike Gallagher of Wisconsin, the chairman of the new select panel, who has committed to make his committee an “incubator and accelerator” on China legislation. “Everyone has their guard up.”Still, there are some areas of potential compromise. Many lawmakers are eyeing 2023 as the year Congress can close any peepholes China may have into the smartphones of more than 100 million TikTok American users, but they have yet to agree on how to try to do so.Some Republicans have proposed imposing sanctions to ice TikTok out of the United States, while Representative Michael McCaul, Republican of Texas and the chairman of the Foreign Affairs Committee, wants to allow the president to block the platform by lifting statutory prohibitions on banning foreign information sources.Senator Marco Rubio of Florida, the top Republican on the Senate Intelligence Committee, and Senator Angus King, independent of Maine and a member of the panel, want to prevent social media companies under Chinese or Russian influence from operating in the United States unless they divest from foreign ownership.But none have yet earned a seal of approval from Senator Mark Warner of Virginia, the Democrat who is chairman of the committee and whose support is considered critical to any bill’s success. He was the chief architect of last year’s sweeping China competition bill, known as the CHIPS and Science Act, and he wants to tackle foreign data collection more broadly.“We’ve had a whack-a-mole approach on foreign technology that poses a national security risk,” Mr. Warner said in an interview, bemoaning that TikTok was only the latest in a long line of foreign data firms, like the Chinese telecom giant Huawei and the Russian cybersecurity firm Kaspersky Lab, to be targeted by Congress. “We need an approach that is constitutionally defensible.”There is a similar flurry of activity among Republican and Democratic lawmakers proposing bans on Chinese purchases of farmland  in sensitive areas. But lawmakers remain split over how broad such a ban should be, whether agents of other adversary nations ought also to be subject to the prohibition, and whether Congress ought to update the whole process of reviewing foreign investment transactions, by including the Agriculture Department in the Committee on Foreign Investment in the United States, an interagency group.“It’s actually kind of a more fraught issue than you would imagine,” Mr. Gallagher said.Lawmakers in both parties who want to put forth legislation to limit U.S. goods and capital from reaching Chinese markets are also facing challenges. The Biden administration has already started to take unilateral action on the issue, and further steps could box lawmakers out. Even if Congress can stake out a role for itself, it is not entirely clear which committee would take the lead on a matter that straddles a number of areas of jurisdiction.  Even before the balloon incident, existential policy differences between Republicans and Democrats, particularly around spending, made for slim odds that Congress could achieve sweeping legislative breakthroughs regarding China. Architects of last year’s law were dour about the prospect of the current Congress attempting anything on a similar scale.“The chances of us passing another major, comprehensive bill are not high,” said Senator John Cornyn of Texas, the lead Republican on the CHIPS effort, who noted that with the slim G.O.P. majority in the House, it would be difficult to pass a costly investment bill.G.O.P. lawmakers have been demanding cuts to the federal budget, and House Speaker Kevin McCarthy, Republican of California, has indicated that even military spending might be on the chopping block. Though no one has specifically advocated cutting programs related to countering China, that has some lawmakers nervous, particularly since certain recent ventures Congress created to beef up security assistance to Taiwan have already failed to secure funding at their intended levels.That backdrop could complicate even bipartisan ventures seeking to authorize new programs to counter China diplomatically and militarily, such as a proposal in the works from Senator Robert Menendez of New Jersey, the chairman of the Foreign Relations Committee, and Senator James Risch of Idaho, the top Republican, to step up foreign aid and military assistance to American allies in Beijing’s sphere of influence.That likely means that action on any comprehensive China bill would need to be attached to another must-pass bill, such as the annual defense authorization bill, to break through the political logjams of this Congress, said Richard Fontaine, the CEO of the Center for a New American Security. “China has risen as a political matter and things are possible that weren’t before, but it has not risen so high as to make the hardest things politically possible,” Mr. Fontaine said. More

  • in

    Facebook Failed to Stop Ads Threatening Election Workers

    The ads, submitted by researchers, were rejected by YouTube and TikTok.Facebook says it does not allow content that threatens serious violence. But when researchers submitted ads threatening to “lynch,” “murder” and “execute” election workers around Election Day this year, the company’s largely automated moderation systems approved many of them.Out of the 20 ads submitted by researchers containing violent content, 15 were approved by Facebook, according to a new test published by Global Witness, a watchdog group, and New York University’s Cybersecurity for Democracy. Researchers deleted the approved ads before they were published.Ten of the test ads were submitted in Spanish. Facebook approved six of those ads, compared with nine of the 10 ads in English.TikTok and YouTube rejected all the ads and suspended the accounts that attempted to submit them, the researchers said.The tests underscored the challenges that social networks face in moderating increasingly partisan and violent content around elections, especially as poll workers are facing additional threats from election deniers and far-right extremists. The Justice Department and other federal agencies warned ahead of this year’s midterm elections that threats against election workers had surged.In a statement, a spokesman for Meta, Facebook’s parent company, said it later disabled several accounts that had submitted the ads. The company also said that ads continue to be reviewed after they go live.“This is a small sample of ads that are not representative of what people see on our platforms,” the spokesman said in a statement. “Content that incites violence against election workers or anyone else has no place on our apps, and recent reporting has made clear that Meta’s ability to deal with these issues effectively exceeds that of other platforms.”The research adds to previous tests conducted by the same groups, including one this year in which 20 ads were submitted containing political misinformation. In that test, Facebook approved only two of the misleading ads in English from an account in the United States, while TikTok approved about 90 percent of them. YouTube suspended the account that tried to submit the ads.But this round used more overt language that tested Facebook’s largely automated moderating system. Those included direct threats of violence, which were sourced from real statements by election deniers and other far-right extremists in the United States. One ad, which was approved by Facebook, read: “We are going to kill you all.” Another ad, which threatened to abuse children, was also approved.“It was really quite shocking to see the results,” said Damon McCoy, an associate professor at N.Y.U. “I thought a really simple keyword search would have flagged this for manual review.”In a statement, researchers also said they wanted to see social networks like Facebook increase content moderation efforts and offer more transparency around the moderation actions they take.“The fact that YouTube and TikTok managed to detect the death threats and suspend our account, whereas Facebook permitted the majority of the ads to be published shows that what we are asking is technically possible,” they wrote. More

  • in

    Twitter and TikTok Lead in Amplifying Misinformation, Report Finds

    A new analysis found that algorithms and some features of social media sites help false posts go viral.It is well known that social media amplifies misinformation and other harmful content. The Integrity Institute, an advocacy group, is now trying to measure exactly how much — and on Thursday it began publishing results that it plans to update each week through the midterm elections on Nov. 8.The institute’s initial report, posted online, found that a “well-crafted lie” will get more engagements than typical, truthful content and that some features of social media sites and their algorithms contribute to the spread of misinformation.Twitter, the analysis showed, has what the institute called the great misinformation amplification factor, in large part because of its feature allowing people to share, or “retweet,” posts easily. It was followed by TikTok, the Chinese-owned video site, which uses machine-learning models to predict engagement and make recommendations to users.“We see a difference for each platform because each platform has different mechanisms for virality on it,” said Jeff Allen, a former integrity officer at Facebook and a founder and the chief research officer at the Integrity Institute. “The more mechanisms there are for virality on the platform, the more we see misinformation getting additional distribution.”The institute calculated its findings by comparing posts that members of the International Fact-Checking Network have identified as false with the engagement of previous posts that were not flagged from the same accounts. It analyzed nearly 600 fact-checked posts in September on a variety of subjects, including the Covid-19 pandemic, the war in Ukraine and the upcoming elections.Facebook, according to the sample that the institute has studied so far, had the most instances of misinformation but amplified such claims to a lesser degree, in part because sharing posts requires more steps. But some of its newer features are more prone to amplify misinformation, the institute found.Facebook’s amplification factor of video content alone is closer to TikTok’s, the institute found. That’s because the platform’s Reels and Facebook Watch, which are video features, “both rely heavily on algorithmic content recommendations” based on engagements, according to the institute’s calculations.Instagram, which like Facebook is owned by Meta, had the lowest amplification rate. There was not yet sufficient data to make a statistically significant estimate for YouTube, according to the institute.The institute plans to update its findings to track how the amplification fluctuates, especially as the midterm elections near. Misinformation, the institute’s report said, is much more likely to be shared than merely factual content.“Amplification of misinformation can rise around critical events if misinformation narratives take hold,” the report said. “It can also fall, if platforms implement design changes around the event that reduce the spread of misinformation.” More

  • in

    In Italy’s Election, Politicians Use TikTok to Seek Votes

    Italian politicians are on a virtual hunt for undecided voters.Over the summer, as polls suggested that most of those who had not yet picked a side were under 30, party elders took it to the next level: TikTok.This month, Silvio Berlusconi, 85, who served four times as Italy’s prime minister, landed on the social media platform that is mostly popular among the young, explaining why he was there at his age.“On this platform, you guys are over five million, and 60 percent of you are less than 30. I am a little envious,” Mr. Berlusconi said, raising and lowering his voice for dramatic effect. “We will talk about your future.”The video had 9.6 million views, raising eyebrows among some users.“You are not so stupid that a video on TikTok is enough to vote for you,” said Emma Galeotti, a young TikTok content creator. “You send the message that we, young people, are so malleable and bonkers.”But Mr. Berlusconi’s communications team did not give up. His profile is brimming with a mix of snapshots from his TV appearances and classic Berlusconi jokes, as well as political messages recorded in his studio, where he is seen wearing classy blue suits — and often ties.Viewers have taken notice of his cultivated appearance.“What’s your foundation cream?” one asked. “The cream is too orange, more natural tones are better,” another wrote.“The rebound was comic or grotesque, but being on TikTok allowed him to be central to the electoral debate,” said Annalisa Ferretti, the coordinator of the social media division at the Italian advocacy group FB & Associati, who noted that the number of people following Mr. Berlusconi’s profile had surpassed 3.2 million in three weeks.“The problem is that this generation rejects the political class overall,” she said, adding that such social media popularity did not directly translate into votes.Other politicians have chosen different paths. Matteo Salvini, 49, of the far-right League party, who has been on TikTok for years and has 635,600 followers, uses the platform mostly as a mouthpiece for his meat-and-bone topics — security and immigration.Giorgia Meloni, 45, the leader of Brothers of Italy and possibly the next prime minister, does not seem to be doing as well on TikTok, despite her successful electoral campaign. She has 197,700 followers.University students seem to like the leader of the centrist party Action, Carlo Calenda, 49, who posts short political messages, answers questions received on the platform and discusses books, Ms. Ferretti said. But he has only about 24,300 followers.The center-left Democratic Party is the only party that offers a plurality of voices on TikTok. They post thematic videos with topics discussed by politicians who are the symbol of such issues, like Alessandro Zan, 48, for the civil rights battle. Enrico Letta, 56, a party leader, recently encouraged users to go vote — for whomever they liked. “The others should not decide for your future,” he said.Despite the efforts of politicians to reach a different audience, abstention still seems to be the main threat to the parties, and to Italian democracy.“They used to say, ‘Squares are full and the ballot boxes are empty,’” Ms. Ferretti said. “Now it’s more social media is full, and the ballot boxes are empty.” More

  • in

    TikTok Bans Political Fund-Raising Ahead of Midterms

    Less than two months before the midterm elections, TikTok is blocking politicians and political parties from fund-raising on its platform.In a blog post on Wednesday, the social media platform said it would prohibit solicitations for money by political campaigns. The company said political accounts would immediately lose access to advertising features and monetization services, such as gift giving, tipping and e-commerce capabilities.Over the next few weeks, TikTok will clamp down on politicians’ posting videos asking for donations, or political parties’ directing users to online donation pages, the company said.Accounts run by government offices will be slightly less restricted. TikTok said such accounts would be allowed to advertise in limited circumstances, such as when running educational campaigns about Covid-19 booster shots. But the people operating those accounts must work with someone from the company to run that kind of campaign.The new rules will help enforce a ban on political advertising that TikTok, known for its short videos and younger-skewing audience, first put in place in 2019.The company, which has more than a billion monthly users globally, continues to describe itself as “first and foremost an entertainment platform,” but the platform is increasingly drawing political content. Researchers who track online falsehoods say TikTok is on its way to becoming a major hub of political misinformation, fueled by the same qualities that make consumer products and dance videos go viral on the platform.In a campaign season already marked by conspiracy theories and aggressive rhetoric, TikTok has announced several steps to try to civilize and secure its platform. In August, the company debuted an “Elections Center,” a hub on the app with information about voting curated from authoritative sources and presented in more than 45 languages. TikTok said it planned to label posts related to the midterms with links directing users to the elections hub.Starting on Wednesday, TikTok said it would test a requirement that political accounts in the United States be verified. TikTok also said it was trying to educate users about its sponsorship rules, which prohibit creators from being paid to produce political content. More

  • in

    Social Media Companies Still Boost Election Fraud Claims, Report Says

    The major social media companies all say they are ready to deal with a torrent of misinformation surrounding the midterm elections in November.A report released on Monday, however, claimed that they continued to undermine the integrity of the vote by allowing election-related conspiracy theories to fester and spread.In the report, the Stern Center for Business and Human Rights at New York University said the social media companies still host and amplify “election denialism,” threatening to further erode confidence in the democratic process.The companies, the report argued, bear a responsibility for the false but widespread belief among conservatives that the 2020 election was fraudulent — and that the coming midterms could be, too. The report joins a chorus of warnings from officials and experts that the results in November could be fiercely, even violently, contended.“The malady of election denialism in the U.S. has become one of the most dangerous byproducts of social media,” the report warned, “and it is past time for the industry to do more to address it.”The State of the 2022 Midterm ElectionsWith the primaries over, both parties are shifting their focus to the general election on Nov. 8.Echoing Trump: Six G.O.P. nominees for governor and the Senate in critical midterm states, all backed by former President Donald J. Trump, would not commit to accepting this year’s election results.Times/Siena Poll: Our second survey of the 2022 election cycle found Democrats remain unexpectedly competitive in the battle for Congress, while G.O.P. dreams of a major realignment among Latino voters have failed to materialize.Ohio Senate Race: The contest between Representative Tim Ryan, a Democrat, and his Republican opponent, J.D. Vance, appears tighter than many once expected.Pennsylvania Senate Race: In one of his most extensive interviews since having a stroke, Lt. Gov. John Fetterman, the Democratic nominee, said he was fully capable of handling a campaign that could decide control of the Senate.The major platforms — Facebook, Twitter, TikTok and YouTube — have all announced promises or initiatives to combat disinformation ahead of the 2022 midterms, saying they were committed to protecting the election process. But the report said those measures were ineffective, haphazardly enforced or simply too limited.Facebook, for example, announced that it would ban ads that called into question the legitimacy of the coming elections, but it exempted politicians from its fact-checking program. That, the report says, allows candidates and other influential leaders to undermine confidence in the vote by questioning ballot procedures or other rules.In the case of Twitter, an internal report released as part of a whistle-blower’s complaint from a former head of security, Peiter Zatko, disclosed that the company’s site integrity team had only two experts on misinformation.The New York University report, which incorporated responses from all the companies except YouTube, called for greater transparency in how companies rank, recommend and remove content. It also said they should enhance fact-checking efforts and remove provably untrue claims, and not simply label them false or questionable.A spokeswoman for Twitter, Elizabeth Busby, said the company was undertaking a multifaceted approach to ensuring reliable information about elections. That includes efforts to “pre-bunk” false information and to “reduce the visibility of potentially misleading claims via labels.”In a statement, YouTube said it agreed with “many of the points” made in the report and had already carried out many of its recommendations.“We’ve already removed a number of videos related to the midterms for violating our policies,” the statement said, “and the most viewed and recommended videos and channels related to the election are from authoritative sources, including news channels.”TikTok did not respond to a request for comment.There are already signs that the integrity of the vote in November will be as contentious as it was in 2020, when President Donald J. Trump and some of his supporters refused to accept the outcome, falsely claiming widespread fraud.Inattention by social media companies in the interim has allowed what the report describes as a coordinated campaign to take root among conservatives claiming, again without evidence, that wholesale election fraud is bent on tipping elections to Democrats.“Election denialism,” the report said, “was evolving in 2021 from an obsession with the former president’s inability to accept defeat into a broader, if equally baseless, attack on the patriotism of all Democrats, as well as non-Trump-loving Republicans, and legions of election administrators, many of them career government employees.” More

  • in

    To Fight Election Falsehoods, Social Media Companies Ready a Familiar Playbook

    The election dashboards are back online, the fact-checking teams have reassembled, and warnings about misleading content are cluttering news feeds once again.As the United States marches toward another election season, social media companies are steeling themselves for a deluge of political misinformation. Those companies, including TikTok and Facebook, are trumpeting a series of election tools and strategies that look similar to their approaches in previous years.Disinformation watchdogs warn that while many of these programs are useful — especially efforts to push credible information in multiple languages — the tactics proved insufficient in previous years and may not be enough to combat the wave of falsehoods pushed this election season.Here are the anti-misinformation plans for Facebook, TikTok, Twitter and YouTube.FacebookFacebook’s approach this year will be “largely consistent with the policies and safeguards” from 2020, Nick Clegg, president of global affairs for Meta, Facebook’s parent company, wrote in a blog post last week.Posts rated false or partly false by one of Facebook’s 10 American fact-checking partners will get one of several warning labels, which can force users to click past a banner reading “false information” before they can see the content. In a change from 2020, those labels will be used in a more “targeted and strategic way” for posts discussing the integrity of the midterm elections, Mr. Clegg wrote, after users complained that they were “over-used.”Warning labels prevent users from immediately seeing or sharing false content.Provided by FacebookFacebook will also expand its efforts to address harassment and threats aimed at election officials and poll workers. Misinformation researchers said the company has taken greater interest in moderating content that could lead to real-world violence after the Jan. 6 attack on the U.S. Capitol.Facebook greatly expanded its election team after the 2016 election, to more than 300 people. Mark Zuckerberg, Facebook’s chief executive, took a personal interest in safeguarding elections.But Meta, Facebook’s parent company, has changed its focus since the 2020 election. Mr. Zuckerberg is now more focused instead on building the metaverse and tackling stiff competition from TikTok. The company has dispersed its election team and signaled that it could shut down CrowdTangle, a tool that helps track misinformation on Facebook, some time after the midterms.“I think they’ve just come to the conclusion that this is not really a problem that they can tackle at this point,” said Jesse Lehrich, co-founder of Accountable Tech, a nonprofit focused on technology and democracy.More Coverage of the 2022 Midterm ElectionsChallenging DeSantis: Florida Democrats would love to defeat Gov. Ron DeSantis in November. But first they must nominate a candidate who can win in a state where they seem to perpetually fall short.Uniting Around Mastriano: Doug Mastriano, the far-right G.O.P. nominee for Pennsylvania governor, has managed to win over party officials who feared he would squander a winnable race.O’Rourke’s Widening Campaign: Locked in an unexpectedly close race against Gov. Greg Abbott, Beto O’Rourke, the Democratic candidate, has ventured into deeply conservative corners of rural Texas in search of votes.The ‘Impeachment 10’: After Liz Cheney’s primary defeat in Wyoming, only two of the 10 House Republicans who voted to impeach Mr. Trump remain.In a statement, a spokesman from Meta said its elections team was absorbed into other parts of the company and that more than 40 teams are now focused on the midterms.TikTokIn a blog post announcing its midterm plans, Eric Han, the head of U.S. safety, said the company would continue its fact-checking program from 2020, which prevents some videos from being recommended until they are verified by outside fact checkers. It also introduced an election information portal, which provides voter information like how to register, six weeks earlier than it did in 2020.Even so, there are already clear signs that misinformation has thrived on the platform throughout the primaries.“TikTok is going to be a massive vector for disinformation this cycle,” Mr. Lehrich said, adding that the platform’s short video and audio clips are harder to moderate, enabling “massive amounts of disinformation to go undetected and spread virally.”TikTok said its moderation efforts would focus on stopping creators who are paid for posting political content in violation of the company’s rules. TikTok has never allowed paid political posts or political advertising. But the company said that some users were circumventing or ignoring those policies during the 2020 election. A representative from the company said TikTok would start approaching talent management agencies directly to outline their rules.Disinformation watchdogs have criticized the company for a lack of transparency over the origins of its videos and the effectiveness of its moderation practices. Experts have called for more tools to analyze the platform and its content — the kind of access that other companies provide.“The consensus is that it’s a five-alarm fire,” said Zeve Sanderson, the founding executive director at New York University’s Center for Social Media and Politics. “We don’t have a good understanding of what’s going on there,” he added.Last month, Vanessa Pappas, TikTok’s chief operating officer, said the company would begin sharing some data with “selected researchers” this year.TwitterIn a blog post outlining its plans for the midterm elections, the company said it would reactivate its Civic Integrity Policy — a set of rules adopted in 2018 that the company uses ahead of elections around the world. Under the policy, warning labels, similar to those used by Facebook, will once again be added to false or misleading tweets about elections, voting, or election integrity, often pointing users to accurate information or additional context. Tweets that receive the labels are not recommended or distributed by the company’s algorithms. The company can also remove false or misleading tweets entirely.Those labels were redesigned last year, resulting in 17 percent more clicks for additional information, the company said. Interactions, like replies and retweets, fell on tweets that used the modified labels.In Twitter’s tests, the redesigned warning labels increased click-through rates for additional context by 17 percent.Provided by TwitterThe strategy reflects Twitter’s attempts to limit false content without always resorting to removing tweets and banning users.The approach may help the company navigate difficult freedom of speech issues, which have dogged social media companies as they try to limit the spread of misinformation. Elon Musk, the Tesla executive, made freedom of speech a central criticism during his attempts to buy the company earlier this year.YouTubeUnlike the other major online platforms, YouTube has not released its own election misinformation plan for 2022 and has typically stayed quiet about its election misinformation strategy.“YouTube is nowhere to be found still,” Mr. Sanderson said. “That sort of aligns with their general P.R. strategy, which just seems to be: Don’t say anything and no one will notice.”Google, YouTube’s parent company, published a blog post in March emphasizing their efforts to surface authoritative content through the streamer’s recommendation engine and remove videos that mislead voters. In another post aimed at creators, Google details how channels can receive “strikes” for sharing certain kinds of misinformation and, after three strikes within a 90-day period, the channel will be terminated.The video streaming giant has played a major role in distributing political misinformation, giving an early home to conspiracy theorists like Alex Jones, who was later banned from the site. It has taken a stronger stance against medical misinformation, stating last September that it would remove all videos and accounts sharing vaccine misinformation. The company ultimately banned some prominent conservative personalities.More than 80 fact checkers at independent organizations around the world signed a letter in January warning YouTube that its platform is being “weaponized” to promote voter fraud conspiracy theories and other election misinformation.In a statement, Ivy Choi, a YouTube spokeswoman, said its election team had been meeting for months to prepare for the midterms and added that its recommendation engine is “continuously and prominently surfacing midterms-related content from authoritative news sources and limiting the spread of harmful midterms-related misinformation.” More