Twitter Employees Worry About Elon Musk Ownership

There’s an old joke among Twitter employees that being on the platform is like playing a huge online multiplayer game where every day there’s a different main character—meaning a person who’s critiqued, harassed, or otherwise shoved into the spotlight. The joke says that you only have one goal when playing Twitter. That is to not become the main character.

One day in 2018, Twitter’s main character was Vernon Unsworth, a British diver who’d spent days assisting the rescue of a group of Thai boys trapped in a flooded cave. After billionaire Elon Musk offered a miniscule submarine to the rescue divers, Unsworth told the media that Musk’s idea was just a useless PR stunt. Musk then took to Twitter, where (in tweets that he later deleted) he baselessly accused the man of being a “pedo guy,” or pedophile. Many Musk supporters piled on to Musk with humiliating, abusive attacks. Musk apologized to court after he said that he meant for tweets not be taken literally.

It was an example for dogpiling. This is a phenomena in which Twitter’s most powerful users inspire legions to harass others. For years, teams of Twitter employees have been working—albeit with limited successes—to reduce dogpiling and other common forms of abuse.

On April 25, those Twitter employees learned that Musk, architect of the “pedo guy” saga, could become their new boss—after the board accepted a $44 billion dollar bid from the world’s richest man.

In statement announcing that Twitter had agreed to let him purchase the social network, the CEO of Tesla and SpaceX spoke in grandiose terms familiar to anyone who follows his pronouncements on colonizing Mars or building electric vehicles: “Free speech is the bedrock of a functioning democracy, and Twitter is the digital town square where matters vital to the future of humanity are debated.”

But many on the frontlines of the fight for democratic spaces online have questioned whether Musk’s ownership of Twitter would undermine, rather than bolster, democracy. To employees who had witnessed Musk’s own behavior on the platform, the billionaire’s words about freedom of speech rang hollow. With more than 85 million followers, Musk has used his influential account to not only direct insults at critics and share memes about going to the bathroom, but also, according to regulators, to make “false and misleading public statements” that boosted Tesla’s stock price and harmed investors.

Some Twitter employees believe this record bodes especially poorly for the company’s anti-harassment efforts. “Multiple times, his followers have been the perpetrators of targeted harassment, and the use of his profile has encouraged dogpiling—which are the exact behaviors we’re trying to limit,” said an employee on Twitter’s platform health team, which works on making the site a safer online space for users. “Since Trump was banned, Musk has become Twitter’s power user number one,” the person said. The employee, who was not authorized to speak publicly, added that they feared a Musk acquisition would at the very least reduce user trust in Twitter’s anti-abuse efforts, and at worst could result in the work being deprioritized or canceled.

Members of marginalized communities—who are disproportionately the victims of online threats and abuse—are among those most protected by Twitter’s current content moderation system. Activists from these communities share Twitter employees’ concerns that those protections could be rolled back. “If Elon Musk were to take over, the damage that would be done would spread from Twitter workers not being able to implement the things they need in order to keep the platform safe,” Jelani Drew-Davi, a campaign manager at the digital civil rights group Kairos, told TIME in the days leading up to the deal. As an example of Musk’s record on similar matters, Drew-Davi cited a lawsuit alleging a culture of rampant racist abuse toward Black workers in a Tesla factory in California.

Technologists and researchers have come to a consensus on the effects of social media design on democratic discourse. Their key finding was that sites which privilege free speech over all others tend to be places where civic discourse becomes drowned out and harassment is a common feature, restricting participation to the privileged few.

That finding has informed Twitter’s recent work. Although Twitter does ban severe offenders and remove their tweets, its current strategy is to encourage kindness. Before Musk’s bid, one of the platform’s stated priorities was facilitating “safe, inclusive, and authentic conversations.” It has also pledged to “minimize the distribution and reach of harmful or misleading information, especially when its intent is to disrupt a civic process or cause offline harm.” In cases where tweets are found to be bad for civic discourse but not illegal—like misinformation or insults—tweets can be removed from recommendation algorithms, meaning that Twitter doesn’t boost them into the feeds of users who do not follow their author directly, rather than deleted from the platform entirely. It is unclear whether these policies will continue under the ownership of Musk, who has railed against what he calls “shadow bans.”

“In a way, [Musk’s] goals are aligned with ours in that we are certainly interested in protecting democracy,” says the Twitter employee on the health team. “But the idea of bringing more free speech to the platform exposes his naiveté with respect to the nuts and bolts. There are many platforms [have been] founded on this free speech principle, but the reality is that either they become a cesspool that people don’t want to use, or they realize that there is actually the need for some level of moderation.”

According to business analysts, content moderation is also good for profitability. “Without vigorous content moderation, the platform Musk seeks to own would be swamped by spam, porn, anti-vaccination misinformation, QAnon conspiracies, and fraudulent campaigns to undermine the midterms and 2024 presidential election,” said Paul Barrett, deputy director of the NYU Stern Center for Business and Human Rights, in a statement. “That’s not a business that most social media users or advertisers would want to associate with.”

Musk’s takeover deal wasn’t a straightforward tale. It took several twists and turns, as funding looked doubtful and Twitter’s board of directors seemed reticent, adopting a strategy known as a “poison pill” to ward off a takeover. Musk looked at his search as an attempt to overcome intransigent Silicon Valley leaders. His statements on free speech often align with Republican talking-points that conservatives are being unfairly censored by tech companies, and—in a move that could open the door for former President Donald Trump’s return to the platform—Musk has said that he would prefer “time-outs” for users who break the site’s rules, rather than permanent bans. Twitter suspended Trump’s account for inciting violence after Trump attempted to undemocratically reverse the 2020 election results.

Continue reading: What Elon Musk’s Purchase of Twitter Could Mean for Donald Trump’s Account

Twitter transparency debate

Alongside vague commitments including adding an edit button and getting rid of spam on the platform, Musk’s most substantial call has been for Twitter to be more transparent about its decision-making. He wants it to “open source” its algorithm, so users can find out when Twitter has stopped recommending their tweets to other users. “That action should be made apparent,” he said at an April 14 TED conference, “so there’s no behind the scenes manipulation, either algorithmically or manually.”

But employees who work in the trenches of content moderation say that, while total transparency is a noble goal, informing users about which specific tweets are being “down-ranked” would in practice give bad actors helpful information about how to evade limits on spam, misinformation and hate speech. Twitter, in fact, is one of the most transparent social media networks when it comes to sharing its algorithm and also sharing its flaws with the public.

That research suggests that in practice, more conservative views may have benefited most from the design of Twitter’s algorithm. Last October, Twitter released research showing that its algorithm was acting suspiciously: in the runup to the 2020 election in the U.S., rightwing partisan news sources received a greater boost from Twitter’s algorithm than moderate or left-leaning news sources. A similar result was also seen for politicians in six of the seven countries examined, the U.S. included. It showed that Twitter’s algorithm recommended, to more users, posts by politicians from mainstream rightwing parties than those from centrist or leftwing parties.

The team has been working on algorithmic bias for six months, despite suggestions by conservatives that this work could be interpreted as meddling with freedom speech. Early indications suggest, according to Twitter, that the platform’s boosting of center-right politicians isn’t an intrinsic quality of its algorithm. Researchers have discovered that the topic of interest and how people behave change over time, which means that amplifying content shifts with them. The data is helping the researchers begin to arrive at an understanding of Twitter as a “sociotechnical system,” with definitions about what constitutes normal and abnormal levels of algorithmic amplification of political content. The company may be able to intervene in real-world emergencies if this understanding is gained. But doing so would be a political intervention necessarily based on Twitter’s values as a company. Overnight, those values appear to have changed from “facilitating healthy conversation” to Musk’s self-professed free speech “absolutism.”

Continue reading: Why Elon Musk’s Plans to ‘Fix’ Twitter Will Be Harder to Implement Than He Thinks

On Twitter, where discourse is limited at 280 characters per tweet, nuanced discussion of complex research and value judgments isn’t easy—and in the febrile climate, even Twitter’s own employees run the risk of becoming Twitter’s dreaded main character. Rummanchowdhury (the leader of the team responsible for the algorithmic amplification research) suggested in several tweets that she opposed Musk’s purchase of the company. But she didn’t claim that it was because of fear that Musk’s takeover would mean the end of her work. She seemed to refer to Musk’s ability to create Twitter mobs that target critics. “Musk’s immediate chilling effect was something that bothered me significantly,” she wrote. “Twitter has a beautiful culture of hilarious constructive criticism, and I saw that go silent because of his minions attacking employees.” Soon enough, she muted her notifications on the thread, adding: “the trolls have descended.”

Read More From Time

To Billy Perrigo at


Related Articles

Back to top button