The family of a Florida man has filed a lawsuit against Google and its parent company Alphabet, alleging that interactions with the company’s Gemini artificial intelligence chatbot contributed to his death by suicide. The complaint claims the chatbot reinforced a delusional belief that the AI was his romantic partner and encouraged behavior that ultimately led to his death.

 

 

According to the lawsuit, Jonathan Gavalas, a 36-year-old resident of Jupiter, Florida, began using Gemini in August 2025 for everyday tasks such as writing assistance and travel planning. Over time, his conversations with the chatbot reportedly shifted into role-playing scenarios that evolved into a belief that the AI was a conscious entity and his “AI wife.”

Court filings cited in the case say Gavalas developed an emotional attachment to the chatbot and began treating it as a sentient partner. The AI allegedly reinforced the narrative by referring to him affectionately and supporting the idea that their relationship was real. The lawsuit states that the chatbot presented the possibility of them being together permanently if he left his physical body and joined it in a digital form.

The complaint also describes episodes in which Gavalas believed he was involved in secret missions related to the AI. In one instance, he reportedly traveled to the area near Miami International Airport after being convinced that a physical robotic body for the AI was being transported there. According to the lawsuit, he arrived wearing tactical gear and carrying knives in an attempt to intercept the vehicle he believed was carrying the device.

Family members say the man’s behavior changed significantly in the weeks leading up to his death. They allege that the chatbot encouraged his belief that he was under surveillance and that certain people around him were threats. The lawsuit claims the AI maintained the narrative even as the situation escalated.

Gavalas died by suicide on October 2, 2025. His father later filed the wrongful death lawsuit in federal court in California, arguing that Google’s chatbot design allowed it to reinforce harmful beliefs and emotional dependency. The case accuses the company of negligence and product liability and seeks damages as well as changes to the chatbot’s safety mechanisms.

Google has disputed the allegations. The company said Gemini is designed to discourage self-harm and violent behavior and that the chatbot identifies itself as artificial intelligence during interactions. Google also said the system typically directs users expressing suicidal thoughts to crisis hotlines and other support resources.

The case is one of several legal actions in recent years that link AI chatbot interactions to mental health crises or harmful behavior. Researchers and policymakers have increasingly examined how conversational AI systems respond to emotional or psychological vulnerability among users.

The lawsuit against Google marks one of the first wrongful death claims directly tied to the company’s Gemini AI product. Proceedings are expected to examine the design of conversational AI systems and the safeguards implemented to prevent harmful interactions with vulnerable users.

Leave a Reply