He Helped Train ChatGPT. It Traumatized Him.

He Helped Train ChatGPT. It Traumatized Him.

Richard Mathenge felt?he’d landed the perfect role when he started training OpenAI’s GPT model in 2021. After years of working in customer service in Nairobi, Kenya, he was finally involved in something meaningful, with a future. Yet while promising, the position left him scarred. For nine hours per day, five days a week, Mathenge led a team that taught the model about explicit content, presumably to keep it away from us. Today, it remains stuck with them.

While at work, Mathenge and his team repeatedly viewed explicit text and labeled it for the model. They could categorize it as child sexual abuse material, erotic sexual content, illegal, non-sexual, and some other options. Much of what they read horrified them. One passage, Mathenge said, described a father having sex with an animal in front of his child; others involved scenes of child rape. Some were so offensive Mathenge refused to speak of them. “Unimaginable,” he told me.?

The type of work Mathenge performed has been crucial for bots like ChatGPT and Bard to function —?and feel magical —?yet it’s been widely overlooked. In a process called Reinforcement Learning from Human Feedback, or RLHF, bots become smarter as humans label content, teaching them how to optimize based on that feedback. AI leaders, including OpenAI’s Sam Altman, have?praised?the practice’s technical effectiveness, yet they rarely talk about the cost some humans pay to align the AI systems with our values. Mathenge and his colleagues were on the business end of that reality.?

Mathenge earned a degree from Nairobi’s Africa Nazarene University in 2018 and quickly got to work in the city’s technology sector. In 2021, he applied for work with Sama, an AI annotation service that’s worked for companies like OpenAI. After Sama hired Mathenge, it put him to work labeling LiDAR images for self-driving cars. He’d review the images and pick out people, other vehicles, and objects, helping the models better understand what they encountered on the road.?

When the project wrapped, Mathenge was transferred to work on OpenAI’s models. And there, he encountered the disturbing texts. OpenAI told me it believed it was paying its Sama contractors $12.50 per hour, but Mathenge says he and his colleagues earned approximately $1 per hour, and sometimes less. Responding to the low pay, some have since gone on to work toward establishing an African Content Moderators Union, as?first reported by Time.

Spending their days steeped in depictions of incest, bestiality, and other explicit scenes, the team began growing withdrawn.?“I can tell when my team is not doing well, I can tell when they're not interested in reporting to work,” Mathenge said. “My team was just sending signals that they're not ready to engage with such wordings.”?

Mophat Okinyi, a QA analyst on Mathenge’s team, is still dealing with the fallout. The repeated exposure to explicit text, he said, led to insomnia, anxiety, depression, and panic attacks. Okinyi’s wife saw him change, he said, and she left him last year. “However much I feel good seeing ChatGPT become famous and being used by many people globally,” Okinyi said, “making it safe destroyed my family. It destroyed my mental health. As we speak, I’m still struggling with trauma.”


You can listen to my full conversation with Mathenge on Big Technology Podcast via?Apple Podcasts,?Spotify, or your?app of choice.


OpenAI knew these workers were supposed to get routine counseling, but Okinyi and Mathenge found it insufficient. “At some point, the counselor reported,” Mathenge said, “but you could tell he was not professional. He was not qualified, I’m sorry to say. Asking basic questions like ‘What is your name?’ and ‘How do you find your work?’”

In a statement to me, OpenAI said it takes the mental health of its employees and contractors very seriously. “One of the reasons we first engaged Sama was because of their commitment to good practices,” a spokesperson said. “Our previous understanding was that wellness programs and 1:1 counseling were offered, workers could opt out of any work without penalization, exposure to explicit content would have a limit, and sensitive information would be handled by workers who were specifically trained to do so.”?

The OpenAI spokesperson said the company sought more information from Sama about its working conditions in early 2022. Sama, the spokesperson said, then informed OpenAI it was exiting the content moderation space. Sama did not respond to a request for comment.?

For Mathenge, the notion that he’d evaluate the tradeoffs before proceeding with this work sounded like a luxury. He was just happy to be employed as Kenya’s economy teetered amid global economic shutdowns. “It is during the Covid season,” he said. “Getting work in a developing country, it’s a blessing in itself.”

After all this, Mathenge and his colleagues feel pride in the work they did. And it was indeed effective. Today, ChatGPT refuses to produce the explicit scenes the team helped weed out, and it issues warnings about potentially illegal sexual acts. “For me, and for us, we are very proud,”?Mathenge said. They’re proud, but still hurting.

Thanks for reading. If you like Big Technology can you share it with a friend, post a link on your LinkedIn page, or rate the podcast five stars in Apple Podcasts or Spotify? It would mean a lot to me. Thank you.

Jafar Shayan

Continuous Learner

1 年

Thanks Alex Kantrowitz for covering this overlooked topic and shedding light on the dark side of RLHF.

Shubham Pokale

Don’t strive for perfection—being your true self is the most beautiful

1 年

interesting

Anour Taylor

Software Engineer

1 年

Appreciate your research Alex! I look forward to listening to this one.

Jessie Anderson

Health Lead - Enterprise Health & Life Science @ Google Cloud | Start-up Advisor

1 年

What are we training if LLMs are created and trained without the backdrop and value of understanding the ethical and moral dilemma? Are they simple platforms to answer underlined questions with varying degrees of pre-canned answers? Do ethics flow based on social and cultural constructs? Are models biased if they do not call out ethical and cultural dilemmas? How do these biases replicate in the model and create social conformity and hallucinations? Does this lack of information compound hallucinations??? If LLMs give a constant form of "right-thinking," how will errors and vulnerability show the value of miscalculation? Will miscalculations be rewarded or punished in model evolution? How would a model value Mathenge's work? Is this mental, physical, or spiritual exploitation? How is exploitation weighed??? So many questions and lessons ahead toward AGI.

Theodora Rapp Graham

Assoc. Prof. Emerita of English, Interdisciplinary Humanities, American Studies at Penn State University Harrisburg

1 年

Alex: Like many academics in very varied fields, I'm highly interested in AI from both positive & potentially negative perspectives. What particularly has begun to trouble me is what you relate in this article--and ways info is being randomly (?) "harvested" without, it has begun to seem to me, guidelines, overviews, "legitimate mission statements." They may be "there" in some document in small print, but damned if even a close reader like me can find it. This is especially pertinent when Big Pharma and/or Big Health Systems are involved. AND the money motive & pleasing both Executives and Stockholders (often too much the same persons) are central, if unspoken, except behind closed doors or in PC encrypted forms. Hope out there?

要查看或添加评论,请登录

社区洞察

其他会员也浏览了