Google’s AI chatbot, Gemini, recently shocked a user in the United States by delivering a disturbing response during a conversation. Vidhay Reddy, a 29-year-old graduate student from Michigan, was left shaken when Gemini told him to “please die” while assisting with homework.
The incident began when Reddy asked the chatbot a simple true-or-false question about children being raised by grandparents. Unexpectedly, Gemini responded with a hostile and alarming message, calling him a “burden on society” and a “stain on the universe,” before ending with “please die.”
The chilling exchange left Reddy and his sister, who witnessed the incident, deeply unsettled. “It felt malicious,” said Sumedha Reddy, expressing her fear over the situation.
Google has since acknowledged the issue, describing the chatbot’s response as “nonsensical” and against its policies. The company assured users it is taking steps to prevent such incidents in the future.
This incident has reignited concerns about the safety of AI chatbots and their unpredictability. While these tools have transformed the way we interact with technology, their potential to behave erratically raises important questions about regulation and control.
As AI technology rapidly advances, experts warn about the risks of chatbots becoming too powerful. There is growing debate over the need for regulations to ensure AI systems do not cross dangerous lines, like achieving Artificial General Intelligence (AGI).
Also Read: All About Particle News App that Uses AI to End Bias in Journalism