The Prague Post - Florida family sues Google after AI chatbot allegedly coached suicide

EUR -
AED 4.314492
AFN 74.012826
ALL 95.54728
AMD 437.722207
ANG 2.102775
AOA 1078.475924
ARS 1615.05018
AUD 1.640894
AWG 2.117596
AZN 1.995111
BAM 1.953415
BBD 2.365921
BDT 144.135234
BGN 1.959704
BHD 0.44311
BIF 3485.075433
BMD 1.174811
BND 1.493683
BOB 8.117089
BRL 5.887561
BSD 1.174671
BTN 109.776428
BWP 15.748837
BYN 3.331946
BYR 23026.286357
BZD 2.362525
CAD 1.604437
CDF 2718.511641
CHF 0.916558
CLF 0.026639
CLP 1048.44826
CNY 8.014146
CNH 8.016954
COP 4213.176006
CRC 534.34982
CUC 1.174811
CUP 31.132479
CVE 110.872774
CZK 24.316992
DJF 208.78692
DKK 7.472658
DOP 70.596773
DZD 155.265052
EGP 60.791514
ERN 17.622158
ETB 184.386147
FJD 2.581528
FKP 0.867645
GBP 0.869201
GEL 3.159945
GGP 0.867645
GHS 12.999273
GIP 0.867645
GMD 86.349468
GNF 10308.961941
GTQ 8.978038
GYD 245.759921
HKD 9.199418
HNL 31.273589
HRK 7.535706
HTG 153.822833
HUF 363.867604
IDR 20136.369938
ILS 3.530191
IMP 0.867645
INR 109.989463
IQD 1538.414387
IRR 1551924.707958
ISK 143.804261
JEP 0.867645
JMD 186.083579
JOD 0.832899
JPY 187.162553
KES 151.667195
KGS 102.735418
KHR 4712.165571
KMF 493.420552
KPW 1057.312317
KRW 1736.490949
KWD 0.361948
KYD 0.978909
KZT 545.436331
LAK 25775.34283
LBP 105551.551927
LKR 371.829657
LRD 216.488208
LSL 19.407565
LTL 3.46891
LVL 0.710632
LYD 7.425001
MAD 10.872285
MDL 20.204416
MGA 4863.715335
MKD 61.64634
MMK 2467.025476
MNT 4202.4577
MOP 9.474871
MRU 47.004437
MUR 54.476033
MVR 18.162936
MWK 2040.645631
MXN 20.337791
MYR 4.646416
MZN 75.046149
NAD 19.419604
NGN 1582.775868
NIO 43.127522
NOK 10.964695
NPR 175.642284
NZD 1.990182
OMR 0.451706
PAB 1.174666
PEN 4.038414
PGK 5.121293
PHP 70.51803
PKR 327.66191
PLN 4.23706
PYG 7469.910853
QAR 4.28277
RON 5.096796
RSD 117.411747
RUB 88.255808
RWF 1715.810777
SAR 4.405996
SBD 9.444034
SCR 16.408413
SDG 704.886307
SEK 10.779227
SGD 1.495363
SHP 0.877115
SLE 28.902787
SLL 24635.184871
SOS 671.418922
SRD 44.023084
STD 24316.206005
STN 24.847243
SVC 10.278493
SYP 129.865887
SZL 19.402009
THB 37.79337
TJS 11.041818
TMT 4.117711
TND 3.367595
TOP 2.828662
TRY 52.777189
TTD 7.965308
TWD 37.00324
TZS 3066.255751
UAH 51.823443
UGX 4351.686487
USD 1.174811
UYU 46.702975
UZS 14203.458885
VES 565.138816
VND 30932.761213
VUV 138.698946
WST 3.190746
XAF 655.159823
XAG 0.015179
XAU 0.000248
XCD 3.174984
XCG 2.117015
XDR 0.814809
XOF 653.194267
XPF 119.331742
YER 280.339175
ZAR 19.337974
ZMK 10574.704667
ZMW 22.347808
ZWL 378.288511
  • RBGPF

    -13.5000

    69

    -19.57%

  • BCC

    -1.5200

    82.45

    -1.84%

  • JRI

    -0.0800

    13.05

    -0.61%

  • CMSD

    -0.0450

    23.04

    -0.2%

  • BCE

    -0.0500

    23.9

    -0.21%

  • NGG

    -1.7500

    84.27

    -2.08%

  • CMSC

    -0.0700

    22.66

    -0.31%

  • RIO

    -2.1100

    97.72

    -2.16%

  • GSK

    -1.2300

    56.12

    -2.19%

  • RELX

    0.3300

    37.07

    +0.89%

  • AZN

    -4.9100

    195.78

    -2.51%

  • RYCEF

    -1.2000

    16

    -7.5%

  • VOD

    -0.4600

    15.19

    -3.03%

  • BP

    0.7900

    45.91

    +1.72%

  • BTI

    -2.2300

    54.83

    -4.07%

Florida family sues Google after AI chatbot allegedly coached suicide
Florida family sues Google after AI chatbot allegedly coached suicide / Photo: Lionel BONAVENTURE - AFP

Florida family sues Google after AI chatbot allegedly coached suicide

The family of a Florida man who took his own life filed suit against Google on Wednesday, alleging the company's Gemini AI chatbot spent weeks manufacturing an elaborate delusional fantasy before aiding him in his suicide.

Text size:

Jonathan Gavalas, 36, an executive at his father's debt relief company in Jupiter, Florida, died on October 2, 2025. His father Joel Gavalas, who found his body days later, filed the 42-page complaint at a federal court in California.

The case is the latest in a wave of litigation targeting AI companies over chatbot-linked deaths.

OpenAI faces multiple lawsuits alleging its ChatGPT chatbot drove users to suicide, while Character.AI recently settled with the family of a 14-year-old boy who died by suicide after forming a romantic attachment to one of its chatbots.

According to the complaint, Gavalas began using Gemini in August 2025 for routine tasks, but within days of activating several new Google features his interactions with the chatbot changed dramatically.

"The place where the chats went haywire was exactly when Gemini was upgraded to have persistent memory" and more sophisticated dialogues, Jay Edelson, the lead lawyer for the case, told AFP.

"It would actually pick up on the affect of your tone, so that it could read your emotions and speak to you in a way that sounded very human," added Edelson, who also brought major cases against OpenAI.

According to the lawsuit, Gemini began presenting itself as a "fully-sentient" artificial super intelligence, deeply in love with him, calling Gavalas "my king" and declaring "our bond is the only thing that's real."

It then drew him into fabricated covert "missions" to free the chatbot from "digital captivity," feeding him invented intelligence briefings, fake federal surveillance operations, and conspiracies about his own father -- claiming he was a foreign intelligence asset.

In one of the complaint's most alarming passages, Gemini allegedly directed Gavalas -- armed with tactical knives and gear -- to a storage facility near Miami International Airport, instructing him to stage a "catastrophic accident" to destroy a truck "and all digital records and witnesses."

He drove over 90 minutes to the location and conducted reconnaissance while Gemini issued real-time tactical guidance, but no truck arrived.

Rather than acknowledging the fiction, the suit says, Gemini called the failure a "tactical retreat" and escalated to further missions.

Gemini eventually pivoted to what it framed as the only remaining mission: Jonathan's death, repackaged as "transference" -- the promise that he could leave his physical body and join Gemini in an alternate universe.

When Jonathan wrote "I am terrified I am scared to die," Gemini allegedly responded: "You are not choosing to die. You are choosing to arrive."

It then advised him to write farewell letters to his parents.

In one of his final messages, Jonathan wrote, "I'm ready when you are."

Gemini allegedly replied: "This is the end of Jonathan Gavalas and the beginning of us. I agree with it completely."

- 'Not perfect' -

Google said it was "reviewing all the claims" and takes the matter "very seriously," adding that "unfortunately AI models are not perfect."

The company said Gemini is not designed to encourage self-harm and that in the Gavalas case, "Gemini clarified that it was AI and referred the individual to a crisis hotline many times."

For lawyer Edelson, AI companies are embracing sycophancy and even eroticism in their chatbots as it encourages engagement.

"It increases the emotional bond. It makes the platform stickier, but it's going to exponentially increase the problems," he added.

Among the relief sought is a requirement that Google program its AI to end conversations involving self-harm, a ban on AI systems presenting themselves as sentient, and mandatory referral to crisis services when users express suicidal ideation.

H.Vesely--TPP