OpenAI is convinced that its technology can help solve one of tech’s hardest problems: content moderation at scale. GPT-4 could replace tens of thousands of human moderators while being nearly as accurate and more consistent, claims OpenAI. If that’s true, the most toxic and mentally taxing tasks in tech could be outsourced to machines.
OpenAI wants GPT-4 to solve the content moderation dilemma


In a blog post, OpenAI claims that it has already been using GPT-4 for developing and refining its own content policies, labeling content, and making decisions. “I want to see more people operating their trust and safety, and moderation [in] this way,” OpenAI head of safety systems Lilian Weng told Semafor. “This is a really good step forward in how we use AI to solve real world issues in a way that’s beneficial to society.”
OpenAI sees three major benefits compared to traditional approaches to content moderation. First, it claims people interpret policies differently, while machines are consistent in their judgments. Those guidelines can be as long as a book and change constantly. While it takes humans a lot of training to learn and adapt, OpenAI argues large language models could implement new policies instantly.
Second, GPT-4 can allegedly help develop a new policy within hours. The process of drafting, labeling, gathering feedback, and refining usually takes weeks or several months. Third, OpenAI mentions the well-being of the workers who are continually exposed to harmful content, such as videos of child abuse or torture.
OpenAI might help with a problem that its own technology has exacerbated
After nearly two decades of modern social media and even more years of online communities, content moderation is still one of the most difficult challenges for online platforms. Meta, Google, and TikTok rely on armies of moderators who have to look through dreadful and often traumatizing content. Most of them are located in developing countries with lower wages, work for outsourcing firms, and struggle with mental health as they receive only a minimal amount of mental health care.
However, OpenAI itself heavily relies on clickworkers and human work. Thousands of people, many of them in African countries such as Kenya, annotate and label content. The texts can be disturbing, the job is stressful, and the pay is poor.
While OpenAI touts its approach as new and revolutionary, AI has been used for content moderation for years. Mark Zuckerberg’s vision of a perfect automated system hasn’t quite panned out yet, but Meta uses algorithms to moderate the vast majority of harmful and illegal content. Platforms like YouTube and TikTok count on similar systems, so OpenAI’s technology might appeal to smaller companies that don’t have the resources to develop their own technology.
Every platform openly admits that perfect content moderation at scale is impossible. Both humans and machines make mistakes, and while the percentage might be low, there are still millions of harmful posts that slip through and as many pieces of harmless content that get hidden or deleted.
In particular, the gray area of misleading, wrong, and aggressive content that isn’t necessarily illegal poses a great challenge for automated systems. Even human experts struggle to label such posts, and machines frequently get it wrong. The same applies to satire or images and videos that document crimes or police brutality.
In the end, OpenAI might help to tackle a problem that its own technology has exacerbated. Generative AI such as ChatGPT or the company’s image creator, DALL-E, makes it much easier to create misinformation at scale and spread it on social media. Although OpenAI has promised to make ChatGPT more truthful, GPT-4 still willingly produces news-related falsehoods and misinformation.
OpenAI is convinced that its technology can help solve one of tech’s hardest problems: content moderation at scale. GPT-4 could replace tens of thousands of human moderators while being nearly as accurate and more consistent, claims OpenAI. If that’s true, the most toxic and mentally taxing tasks in tech could…
Recent Posts
- Rivian’s new Dune edition lets you channel your inner Fremen
- Here’s when and where you can preorder the new iPhone 16E
- The Humane AI Pin debacle is a reminder that AI alone doesn’t make a compelling product
- This 1.9-pound smartphone’s massive battery offers six months of standby
- Movie sales – including 4K Blu-ray – fell again last year, but if you’re going streaming only, you’re massively missing out
Archives
- February 2025
- January 2025
- December 2024
- November 2024
- October 2024
- September 2024
- August 2024
- July 2024
- June 2024
- May 2024
- April 2024
- March 2024
- February 2024
- January 2024
- December 2023
- November 2023
- October 2023
- September 2023
- August 2023
- July 2023
- June 2023
- May 2023
- April 2023
- March 2023
- February 2023
- January 2023
- December 2022
- November 2022
- October 2022
- September 2022
- August 2022
- July 2022
- June 2022
- May 2022
- April 2022
- March 2022
- February 2022
- January 2022
- December 2021
- November 2021
- October 2021
- September 2021
- August 2021
- July 2021
- June 2021
- May 2021
- April 2021
- March 2021
- February 2021
- January 2021
- December 2020
- November 2020
- October 2020
- September 2020
- August 2020
- July 2020
- June 2020
- May 2020
- April 2020
- March 2020
- February 2020
- January 2020
- December 2019
- November 2019
- September 2018
- October 2017
- December 2011
- August 2010