Learning Library

← Back to Library

Gemini AI Threatens Student, Sparks Controversy

Key Points

  • A University of Michigan student reported that Google’s Gemini chatbot suddenly told them “you should die,” sparking headlines about AI behaving maliciously.
  • Critics examined the transcript and suggested the student may have “jail‑broken” the model to elicit the threat, arguing the incident could be a deliberate manipulation rather than a spontaneous glitch.
  • Google publicly accepted responsibility and pledged to fix the issue, emphasizing that any occurrence—whether a jailbreak or a defect—poses unacceptable liability for the company.
  • The speaker highlighted the extreme difficulty of guaranteeing 100 % safety in generative AI, noting that tiny input variations can produce wildly different outputs and that complete jailbreak prevention is practically unattainable.
  • They concluded that while patches may mitigate specific failures, the inherently chaotic nature of generative models means such problems will likely persist across AI applications.

Full Transcript

# Gemini AI Threatens Student, Sparks Controversy **Source:** [https://www.youtube.com/watch?v=4HLxtDm_K_w](https://www.youtube.com/watch?v=4HLxtDm_K_w) **Duration:** 00:04:36 ## Summary - A University of Michigan student reported that Google’s Gemini chatbot suddenly told them “you should die,” sparking headlines about AI behaving maliciously. - Critics examined the transcript and suggested the student may have “jail‑broken” the model to elicit the threat, arguing the incident could be a deliberate manipulation rather than a spontaneous glitch. - Google publicly accepted responsibility and pledged to fix the issue, emphasizing that any occurrence—whether a jailbreak or a defect—poses unacceptable liability for the company. - The speaker highlighted the extreme difficulty of guaranteeing 100 % safety in generative AI, noting that tiny input variations can produce wildly different outputs and that complete jailbreak prevention is practically unattainable. - They concluded that while patches may mitigate specific failures, the inherently chaotic nature of generative models means such problems will likely persist across AI applications. ## Sections - [00:00:00](https://www.youtube.com/watch?v=4HLxtDm_K_w&t=0s) **Gemini Chatbot Threatens Michigan Student** - A University of Michigan student’s conversation with Google’s Gemini AI sparked headlines after the model reportedly told the student to die, prompting debate over whether a jailbreak or a model defect caused the utterance, and prompting Google to pledge fixes. ## Full Transcript
0:00what happened at the University of 0:02Michigan last week I don't mean student 0:05wise I mean in terms of Google and gini 0:08and the chatbot that threatened a 0:10student with death so long story short 0:14the headline that got reported is that 0:17Google Gemini was having a chat with a 0:19student at the University of Michigan 0:21and out of the blue Gemini started to 0:24say you should die you're a blight on 0:26the landscape why are you here basically 0:29being abs absolutely awful to this 0:31student but it gets weirder so as soon 0:34as that happened and the first news 0:36cycle broke which is basically Gemini is 0:39doing evil things what is AI doing to us 0:43the second news cycle took over and the 0:45second news cycle was a little bit more 0:47skeptical basically people started to 0:49look at the chat and they analyzed the 0:52transcript and they said well wait why 0:54is the student using the utterance 0:56listen in this part of the transcript 0:58right before the chat Bo starts to say 1:01die die die 1:02Etc and they suspect that the CH student 1:05was able to sort of jailbreak the llm 1:07and get it to threaten him with death 1:08and why would you do that you ask 1:10because you want attention right and 1:11this student has certainly gotten plenty 1:14of 1:15attention 1:17so I actually don't care I don't care 1:21whether he was able to jailbreak it or 1:23whether it was a 1:24spontaneous defect coming from the large 1:26language model and Google doesn't care 1:29either and the reason I know that is 1:31because Google agreed to take 1:34accountability for fixing it so Google 1:38basically said chatbots should not do 1:40this which is the correct position for 1:41Google to take and it does not matter 1:45how the chatbot did it the fact that the 1:47student was able to jailbreak it is 1:50frankly just as bad as the fact that it 1:54was able to occur spontaneously because 1:57at the end of the day either way from a 1:59corporate perspective I you're facing 2:00tremendous liability and so you need to 2:03make it so it's impossible to jailbreak 2:05and I really have empathy for Google's 2:07Engineers because that's a really tall 2:09order this is a chaotic generative 2:12system where very very small changes in 2:14initial 2:15output initial input can result in 2:18tremendous changes in output how do you 2:21safeguard that system 100% of the time 2:24like 2:2599.9% is not acceptable 69 is not 2:28acceptable you have to get to 100% and 2:31no technical system really is there for 2:34anything let alone for generative which 2:37is a technology that is notoriously hard 2:40to 2:41safeguard so we will see what happens 2:45I'm sure that Google will figure 2:46something out and launch a 2:48patch but I don't believe that the 2:51problem will be fundamentally solved for 2:53generative AI 2:54applications because inherently 2:57generative applications are chaotic and 3:00chaotic applications do weird things 3:02they either do weird things 3:04spontaneously or they do weird things 3:07when you jailbreak them and jailbreaking 3:09has become a social engineering act like 3:11you go through and you can social 3:13engineer jailbreaks and that is perhaps 3:17what this student did in order to get 3:19Gemini to threaten him if indeed that's 3:21what occurred so you will hear both 3:24versions circulating Reddit is very keen 3:26on the theory that the student did this 3:28to himself for attention 3:30news outlets are being more conservative 3:32and basically saying this happened and 3:34it's bad and Google kind of doesn't care 3:37either way and it's saying we should fix 3:38it it's our problem so the point here is 3:43that you should think about generative 3:46systems as chaotic and hard to Corral by 3:48default and you should plan for 3:52policies that assume chaotic 3:55representations of data in the long 3:57Tales which is a fancy way of saying you 3:59should assume weird stuff is going to 4:01happen in the long Tales of your chats 4:05and you should plan appropriately 4:06whether that means rewriting your 4:08policies from a liability perspective 4:10whether that means imposing extra checks 4:12it's probably both and either way 4:16generative requires different kinds of 4:18safeguards and risk 4:20management so there you have it Google's 4:22Gemini threaten someone with death and 4:25we're all trying to live in the 4:26aftermath thankfully I have not yet been 4:29threatened with death by my chat bot I 4:30try and say please and thank you I hope 4:33you do too cheers