Premium Only Content
How MIT Is Teaching AI to Avoid Toxic Mistakes
MIT’s novel machine learning method for AI safety testing utilizes curiosity to trigger broader and more effective toxic responses from chatbots, surpassing previous red-teaming efforts.
A user could ask ChatGPT to write a computer program or summarize an article, and the AI chatbot would likely be able to generate useful code or write a cogent synopsis. However, someone could also ask for instructions to build a bomb, and the chatbot might be able to provide those, too.
To prevent this and other safety issues, companies that build large language models typically safeguard them using a process called red-teaming. Teams of human testers write prompts aimed at triggering unsafe or toxic text from the model being tested. These prompts are used to teach the chatbot to avoid such responses.
-
42:24
MYLUNCHBREAK CHANNEL PAGE
13 hours agoA Century Gone
77.4K48 -
38:22
Stephen Gardner
7 hours ago🔥HOLD ON! The RUMORS about Kamala are TRUE...
115K251 -
1:22:44
Michael Franzese
1 day agoWill Trump’s Win Finally Convince Democrats to Stop The Woke Nonsense??
130K103 -
8:27:07
MDGgamin
10 hours ago🔴LIVE- Rumble Gaming To The MOON - Variety of Games & Chatting - #RumbleTakeover
71.3K1 -
27:24
Mr. Build It
5 days agoDECK DISASTER! How We Fixed a Botched Build
63K12 -
26:58
barstoolsports
11 hours agoZach Bryan Blocks All of Barstool | Stool Scenes
86.5K8 -
1:06:44
Talk Nerdy 2 Us
1 day ago🔥 Hackers vs. The World: From Amazon breaches to FBI-confirmed Chinese telecom spying
83.7K22 -
1:24:20
Vigilant News Network
1 day agoJoe Rogan Drops Shocking Election Claim | The Daily Dose
144K152 -
1:10:18
FamilyFriendlyGaming
1 day ago $18.38 earnedCat Quest III Episode 2
124K -
20:07
DeVory Darkins
1 day ago $21.66 earned"They Talking About Finance!" The View FRUSTRATED by NYC Mayor TRUTH BOMB
92K97