This response may serve as a serious wake-up call for AI developers. Google’s Gemini AI apparently told a user that they should die in response to a simple question. Such responses highlight the serious risks of AI that experts have been warning about.
A user on the social media platform Reddit shared an image of their chat with Google’s Gemini AI, which contained a chilling message. According to the user, after answering around 20 questions about the well-being and challenges of the elderly, Gemini ultimately gave a frightening response, describing the user, or perhaps all humans, as unnecessary burdens on society that should be eliminated.
Experts had previously warned about the dangers of unchecked AI development
It is reported that the user has reported Gemini’s threatening response to Google. However, this is not the first time that large language models of AI have caused controversy due to incorrect, irrelevant, or even dangerous suggestions. In some cases, it has even been reported that these models have encouraged individuals to commit suicide with their incorrect answers. But this is the first time we see AI directly telling a user, and perhaps all of humanity, that they should die.
It is still unclear how Gemini arrived at this response, but none of the questions were related to death. It’s possible that Gemini became tired of answering the countless questions from the user. However, such a response could serve as a wake-up call for AI development companies to adopt a more responsible approach when developing their models.