Skip to main content

X hiring content moderators for a new 'Trust and Safety' center in Austin

Elon Musk at a speaking event holding a microphone

X, formerly known at Twitter, is building its content moderation team back up, but not completely.

According to Bloomberg, the company is hiring 100 employees for a "Trust and Safety center of excellence" in Austin, TX. After Elon Musk took over the platform in 2022, he laid off around 80 percent of the company and gutted the content moderation team. The new team will focus on enforcing the platform's rules around violence and hate speech, and specifically combatting CSE (Child Sexual Exploitation).

The timing of this report comes days before X CEO Linda Yaccarino is scheduled to testify before the Senate Judiciary Committee for a hearing on protecting children online. On Friday, Jan. 26, X published a blog post outlining its work to "tackle Child Sexual Exploitation" on the platform.

As a self-professed "free-speech absolutist," Musk bought the platform to create a digital town square by rolling back policies and efforts which he believed hindered free speech. However, this soon turned X into an even more toxic place, which caused big ticket advertisers to leave the platform and X's valuation to plummet.

Is the development of a content moderation headquarters the sign of positive change for the platform? According to 2020 NYU Stern report, Twitter used to have around 1,500 content moderators, so a team of 100 is peanuts compared to what it used to be. No doubt, Yaccarino will have more to say about it at the hearing on Jan. 31.



from Mashable https://ift.tt/JTFZvca
via IFTTT

Comments

Popular posts from this blog

Instagram accidentally reinstated Pornhub’s banned account

After years of on-and-off temporary suspensions, Instagram permanently banned Pornhub’s account in September. Then, for a short period of time this weekend, the account was reinstated. By Tuesday, it was permanently banned again. “This was done in error,” an Instagram spokesperson told TechCrunch. “As we’ve said previously, we permanently disabled this Instagram account for repeatedly violating our policies.” Instagram’s content guidelines prohibit  nudity and sexual solicitation . A Pornhub spokesperson told TechCrunch, though, that they believe the adult streaming platform’s account did not violate any guidelines. Instagram has not commented on the exact reasoning for the ban, or which policies the account violated. It’s worrying from a moderation perspective if a permanently banned Instagram account can accidentally get switched back on. Pornhub told TechCrunch that its account even received a notice from Instagram, stating that its ban had been a mistake (that message itse...

Watch Aidy Bryant *completely* lose it as 'SNL' roasts political pundits

On Saturday Night Live , there are breaks and then there's whatever happened here. The Season 45 premiere featured a sketch that was meant to expose the empty noisemaking of political punditry on TV. But part of the joke involved a series of quick costume changes, and some weirdness during one of those switches led to a complete and total breakdown. Aidy Bryant, the segment's host, couldn't take it. She manages to keep it together until what appears to be an accidental wide shot exposes some of the magic as we see a woman who's probably a member of the SNL wardrobe crew fiddling with Aidy's costume. Read more... More about Saturday Night Live , Aidy Bryant , Entertainment , and Movies Tv Shows from Mashable https://ift.tt/2okrAOq via IFTTT

California Gov. Newsom vetoes bill SB 1047 that aims to prevent AI disasters

California Gov. Gavin Newsom has vetoed bill SB 1047, which aims to prevent bad actors from using AI to cause "critical harm" to humans. The California state assembly passed the legislation by a margin of 41-9 on August 28, but several organizations including the Chamber of Commerce had urged Newsom to veto the bill . In his veto message on Sept. 29, Newsom said the bill is "well-intentioned" but "does not take into account whether an Al system is deployed in high-risk environments, involves critical decision-making or the use of sensitive data. Instead, the bill applies stringent standards to even the most basic functions - so long as a large system deploys it."  SB 1047 would have made the developers of AI models liable for adopting safety protocols that would stop catastrophic uses of their technology. That includes preventive measures such as testing and outside risk assessment, as well as an "emergency stop" that would completely shut down...