Sympathy, and Job Offers, for Twitter’s Misinformation Experts

Author: Yuvi November 28, 2022

Sympathy, and Job Offers, for Twitter's Misinformation Experts

In the weeks since Elon Musk took over Twitter, dozens of people responsible for keeping dangerous or inaccurate material in check on the service have posted on LinkedIn that they resigned or lost their jobs. Their statements have drawn a flood of condolences — and attempts to recruit them.

Overtures arrived from rival tech services, retailers, consulting firms, government contractors and other organizations that want to use the former Twitter employees — and those recently let go by Meta and the payments platform Stripe — to track and combat false and toxic information on the internet. .

Ania Smith, the chief executive of TaskRabbit, the Ikea-owned marketplace for gig workers, commented on a former Twitter employee’s post this month that she should consider applying for a product director role, working in part on trust and safety tools.

“The war for talent has really been exceptional in the last 24 months in tech,” Ms. Smith said in an interview. “So when we see layoffs happening, whether it’s at Twitter or Meta or other companies, it’s definitely an opportunity to go after some of the very high-caliber talent we know they hire.”

She added that making users feel safe on the TaskRabbit platform was a key component of her company’s success.

“We can’t really continue growing without investing in a trust and safety team,” she said.

The threats posed by conspiracy theories, misleadingly manipulated media, hate speech, child abuse, fraud and other online harms have been studied for years by academic researchers, think tanks and government analysts. But increasingly, companies in and outside the tech industry see that abuse as a potentially costly liability, especially as more work is conducted online and regulators and clients push for stronger guardrails.

On LinkedIn, under posts eulogizing Twitter’s work on elections and content moderation, comments promoted openings at TikTok (threat researcher), DoorDash (community policy manager) and Twitch (trust and security incident manager). Managers at other companies solicited suggestions for names to add to recruiting databases. Google, Reddit, Microsoft, Discord and ActiveFence — a four-year-old company that said last year that it had raised $100 million and that it could scan more than three million sources of malicious chatter in every language — also have job postings.

The trust and safety field barely existed a decade ago, and the talent pool is still small, said Lisa Kaplan, the founder of Alethea, a company that uses early-detection technology to help protect clients against disinformation campaigns. The three-year-old company has 35 employees; Ms. Kaplan said she hoped to add 23 more by mid-2023 and was trying to recruit former Twitter employees.

Disinformation, she said, is like “the new malware” — a “digital reality that is ultimately going to impact every company.” Clients that once employed armed guards to stand outside data rooms, and then built online firewalls to block hackers, are now calling firms like Alethea for backup when, for example, coordinated influence campaigns target public perception of their brand and threaten their stock price, Ms. . Kaplan said.

“Anyone can do this — it’s fast, cheap and easy,” she said. “As more actors get into the practice of weaponizing information, either for financial, reputational, political or ideological gain, you’re going to see more targets. This market is emerging because the threat has risen and the consequences have become more real.

Disinformation became widely recognized as a significant problem in 2016, said John Kelly, who was an academic researcher at Columbia, Harvard and Oxford before founding Graphika, a social media analysis firm, in 2013. The company’s employees are known as “the cartographers of the internet age” for their work building detailed maps of social media for clients such as Pinterest and Meta.

Graphika’s focus, initially on mining digital marketing insights, has steadily shifted toward topics such as disinformation campaigns coordinated by foreigners, extremist narratives and climate misinformation. The transition, which began in 2016 with the discovery of Russian influence operations targeting the US presidential election, intensified with the onslaught of Covid-19 conspiracy theories during the pandemic, Mr. Kelly said.

“The problems have spilled out of the political arena and become a Fortune 500 problem,” he said. “The range of online harms has expanded, and the range of people doing online harm has expanded.”

Efforts to tackle misinformation and disinformation have included research initiatives from top-tier universities and policy institutes, media literacy campaigns and initiatives to repopulate news deserts with local journalism outfits.

Many social media platforms have set up internal teams to address the problem or outsourced content moderation work to large companies such as Accenture, according to a July report from the geopolitical think tank German Marshall Fund. In September, Google completed its $5.4 billion acquisition of Mandiant, an 18-year-old company that tracks online influence activities as well as offering other cybersecurity services.

A growing group of start-ups, many of which rely on artificial intelligence to root out and decode online narratives, conduct similar exercises, often for clients in corporate America.

Alethea wrapped up a $10 million fund-raising round in October. Also last month, Spotify said it bought the five-year-old Irish company Kinzen, citing its grasp on “the complexity of analyzing audio content in hundreds of languages ​​and dialects, and the challenges in effectively evaluating the nuance and intent of that content.” ” (Months earlier, Spotify found itself trying to quell an uproar over accusations that its star podcast host, Joe Rogan, was spreading vaccine misinformation.) Amazon’s Alexa Fund participated in a $24 million funding round last winter for five-year-old Logically, which uses artificial intelligence to identify misinformation and disinformation on topics such as climate change and Covid-19.

“Along with all the fantastic aspects of the web come new problems like bias, misinformation and offensive content to name a few,” Biz Stone, a Twitter co-founder, wrote on a crowdfunding page last year for Factmata, another AI-fueled disinformation. To install defense operation. “It can be confusing and difficult to cut through to the trusted, truthful information.”

The businesses are hiring across a broad spectrum of trust and safety roles despite a host of recent layoff announcements.

Companies have courted people expert at recognizing content posted by child abusers or human traffickers, as well as former military counterterrorism agents with advanced degrees in law, political science and engineering. Moderators, many of whom work as contractors, are also in demand.

Mounir Ibrahim, the vice president of public affairs and impact for Truepic, a tech company specializing in image and digital content authenticity, said many early clients were banks and insurance companies that relied more and more on digital transactions.

“We are at an inflection point of the modern internet right now,” he said. “We are facing a tsunami of generative and synthetic material that is going to hit our computer screens very, very soon — not just images and videos, but text, code, audio, everything under the sun. And this is going to have tremendous effects on not just disinformation but brand integrity, the financial tech world, on the insurance world and across nearly every vertical that is now digitally transforming on the heels of Covid.”

Truepic was featured with companies such as Zignal Labs and Memetica in the German Marshall Fund report about disinformation-defense start-ups. Anya Schiffrin, the lead author and a senior lecturer at Columbia’s School of International and Public Affairs, said future regulation of disinformation and other malicious content could lead to more jobs in the trust and security space.

She said regulators around the European Union were already hiring people to help carry out the new Digital Services Act, which requires internet platforms to combat misinformation and restrict certain online ads.

“I’m really tired of these really rich companies saying that it’s too expensive — it’s a cost of doing business, not an extra, add-on luxury,” Ms. Schiffrin said. “If you can’t provide accurate, quality information to your customers, then you’re not a going concern.”

Author: Yuvi

My name is Yuvi, I work as Sub Editor at

28 November, 2022, 10:15 pm

News Cinema on twitter News Cinema on facebook

Monday, 28th November 2022

Latest Web Stories

More Stories
‘Stop, look more friendly…. look motivated’: What JLo really said to Ben Affleck at the Grammys
Charles Bronson could be free from prison in months as he confirms in letter to his ‘long lost son’
WATCH: Shocking Video Shows Building Collapsing Like House Of Cards After Turkey Earthquake
Reserve Bank of Australia Governor Philip Lowe steps outside Sydney house ahead of February meeting
Parents can now have eyes in back of their watches as new app monitors where their children are’s spy reveals ALL from inside the Grammys’ most exclusive parties
Previous Chinese Balloon Incursions Initially Went Undetected
Half a million British homes have worse wifi reception than on Everest, analysis shows
The Most Romantic Outdoor Places in Delhi for the Perfect V-Day Surprise
WA: Witness reports spotting GIANT bull shark day before 16-year-old girl killed
Toxic Fumes Are Released From Burning Train That Derailed in Ohio
Biden’s State of the Union Prep: No Acronyms and Tricks to Conquer a Stutter
Biden’s State of the Union Prep: No Acronyms and Tricks to Conquer a Stutter
Green alternatives to burial and cremation could be on the way
Homeowners share shocking photos of half-finished building work after construction firm went bust
Brooke Boney clashes with Alice Springs business owner behind Facebook page highlighting crime wave
The Nets Wished Upon Stars. Their Dreams Haven’t Come True.
The Capitol fence returns for the State of the Union address
Police Review Board Says Violence During Protests Shows Lack of Training
Lucky Powerball player could win $747 MILLION as jackpot goes up for grabs TONIGHT game abysmal odds
Seat by seat, scientists reveal your risk of dying on an airplane if it crashes
Staring Down Bankruptcy, Bed Bath & Beyond Says It Will Sell Stock
After ANOTHER near miss at a US airport, how can Americans trust their plane journey will be safe?
Long Covid rates could be slashed in half if people ate healthy and exercised
TJ Holmes buys ring and Tiffany chain ‘for Amy Robach’
On Second Thought, St. Andrews Steps Back From Remodel by Swilcan Bridge
Listen to AI-generated Donald Trump read ‘The Three Little Pigs’
Amid Criticism, Elite Crime Teams Dwindled. Then the cities brought them back.
Vox Media Is Raising $100 Million From Penske Media
Queen Letizia and King Felipe of Spain jet off to Angola
The Australian home borrowers who face a 65 percent surge in their mortgage repayments in 2023
Astronauts Relax as Cosmonauts Prep for Cargo Mission
On A Winding Path, Carlos Beltran Will Return To Mets
Police Officer Who Shot a Man in the Back Is Charged With Assault
Police Officer Who Shot a Man in the Back Is Charged With Assault
NFL ‘secrets’ revealed by girlfriend of 49ers star Fred Warner ahead of Super Bowl weekend
Education Issues Vault to Top of the GOP’s Presidential Race
Family who left car in Heathrow’s official zone return to discover vehicle could be written off
Google launches its ChatGPT competitor called Bard
Pair Is Charged With Plotting to ‘Destroy Baltimore’ by Attacking Electrical Grid
Jennifer Lopez seemingly snaps at Ben Affleck at Grammy awards
Breakfast Sandwiches and Fruit Cups Are Recalled Over Listeria Risk