The Prague Post - AI's blind spot: tools fail to detect their own fakes

EUR -
AED 4.237583
AFN 72.693752
ALL 96.083665
AMD 433.726263
ANG 2.065521
AOA 1058.097238
ARS 1611.096401
AUD 1.627012
AWG 2.076964
AZN 1.957395
BAM 1.955434
BBD 2.317406
BDT 141.175387
BGN 1.972318
BHD 0.435926
BIF 3416.234019
BMD 1.153869
BND 1.470256
BOB 7.950648
BRL 5.996198
BSD 1.150604
BTN 106.252936
BWP 15.636342
BYN 3.451113
BYR 22615.829146
BZD 2.314007
CAD 1.580015
CDF 2613.512848
CHF 0.907177
CLF 0.026486
CLP 1045.785768
CNY 7.946522
CNH 7.938554
COP 4269.233915
CRC 539.31065
CUC 1.153869
CUP 30.577524
CVE 110.246257
CZK 24.445461
DJF 204.885168
DKK 7.471843
DOP 70.228365
DZD 152.511672
EGP 60.430077
ERN 17.308033
ETB 179.623441
FJD 2.54889
FKP 0.864765
GBP 0.863994
GEL 3.127214
GGP 0.864765
GHS 12.535869
GIP 0.864765
GMD 84.844491
GNF 10083.329455
GTQ 8.813502
GYD 240.719076
HKD 9.044641
HNL 30.452955
HRK 7.528765
HTG 150.924996
HUF 390.627295
IDR 19568.461556
ILS 3.569811
IMP 0.864765
INR 106.997682
IQD 1507.230698
IRR 1516183.648142
ISK 143.298995
JEP 0.864765
JMD 181.000013
JOD 0.818054
JPY 183.519391
KES 149.56326
KGS 100.905754
KHR 4617.235044
KMF 492.702289
KPW 1038.457027
KRW 1723.170402
KWD 0.353753
KYD 0.958829
KZT 554.390945
LAK 24690.588441
LBP 103033.2836
LKR 358.295982
LRD 210.554204
LSL 19.248161
LTL 3.407074
LVL 0.697964
LYD 7.365748
MAD 10.789366
MDL 20.071588
MGA 4790.102621
MKD 61.593693
MMK 2423.243908
MNT 4120.582999
MOP 9.287041
MRU 45.769417
MUR 53.666511
MVR 17.827435
MWK 1995.026251
MXN 20.352175
MYR 4.519126
MZN 73.744171
NAD 19.248161
NGN 1564.577088
NIO 42.342985
NOK 11.060872
NPR 170.005834
NZD 1.972608
OMR 0.44369
PAB 1.15052
PEN 3.932614
PGK 4.964178
PHP 68.948263
PKR 321.238287
PLN 4.262882
PYG 7458.731962
QAR 4.194987
RON 5.091795
RSD 117.421168
RUB 96.593463
RWF 1682.684766
SAR 4.332929
SBD 9.283085
SCR 15.84955
SDG 693.475127
SEK 10.746038
SGD 1.47424
SHP 0.8657
SLE 28.383287
SLL 24196.065005
SOS 656.391253
SRD 43.414286
STD 23882.755212
STN 24.495942
SVC 10.067201
SYP 127.601462
SZL 19.251727
THB 37.528395
TJS 11.028225
TMT 4.05008
TND 3.391723
TOP 2.778239
TRY 51.023508
TTD 7.806605
TWD 36.807836
TZS 3007.247299
UAH 50.55213
UGX 4343.261614
USD 1.153869
UYU 46.772048
UZS 13962.505268
VES 516.71188
VND 30358.289022
VUV 137.994476
WST 3.154336
XAF 655.834136
XAG 0.014683
XAU 0.000235
XCD 3.118389
XCG 2.073629
XDR 0.815647
XOF 655.845502
XPF 119.331742
YER 275.255428
ZAR 19.297997
ZMK 10386.182289
ZMW 22.442185
ZWL 371.545294
  • CMSC

    -0.0690

    22.881

    -0.3%

  • BCC

    -0.5500

    72.37

    -0.76%

  • JRI

    -0.0900

    12.37

    -0.73%

  • AZN

    -2.0900

    189.2

    -1.1%

  • GSK

    -1.1800

    52.23

    -2.26%

  • RBGPF

    0.1000

    82.5

    +0.12%

  • RIO

    -1.2050

    88.595

    -1.36%

  • NGG

    -2.3900

    88.03

    -2.71%

  • BTI

    -1.9700

    58.58

    -3.36%

  • BCE

    -0.2700

    25.74

    -1.05%

  • CMSD

    -0.0500

    22.83

    -0.22%

  • RYCEF

    -0.0700

    16.71

    -0.42%

  • RELX

    0.1600

    34.45

    +0.46%

  • VOD

    -0.2900

    14.46

    -2.01%

  • BP

    0.5850

    44.435

    +1.32%

AI's blind spot: tools fail to detect their own fakes
AI's blind spot: tools fail to detect their own fakes / Photo: Chris Delmas - AFP

AI's blind spot: tools fail to detect their own fakes

When outraged Filipinos turned to an AI-powered chatbot to verify a viral photograph of a lawmaker embroiled in a corruption scandal, the tool failed to detect it was fabricated -- even though it had generated the image itself.

Text size:

Internet users are increasingly turning to chatbots to verify images in real time, but the tools often fail, raising questions about their visual debunking capabilities at a time when major tech platforms are scaling back human fact-checking.

In many cases, the tools wrongly identify images as real even when they are generated using the same generative models, further muddying an online information landscape awash with AI-generated fakes.

Among them is a fabricated image circulating on social media of Elizaldy Co, a former Philippine lawmaker charged by prosecutors in a multibillion-dollar flood-control corruption scam that sparked massive protests in the disaster-prone country.

The image of Co, whose whereabouts has been unknown since the official probe began, appeared to show him in Portugal.

When online sleuths tracking him asked Google's new AI mode whether the image was real, it incorrectly said it was authentic.

AFP's fact-checkers tracked down its creator and determined that the image was generated using Google AI.

"These models are trained primarily on language patterns and lack the specialized visual understanding needed to accurately identify AI-generated or manipulated imagery," Alon Yamin, chief executive of AI content detection platform Copyleaks, told AFP.

"With AI chatbots, even when an image originates from a similar generative model, the chatbot often provides inconsistent or overly generalized assessments, making them unreliable for tasks like fact-checking or verifying authenticity."

Google did not respond to AFP’s request for comment.

- 'Distinguishable from reality' -

AFP found similar examples of AI tools failing to verify their own creations.

During last month's deadly protests over lucrative benefits for senior officials in Pakistan-administered Kashmir, social media users shared a fabricated image purportedly showing men marching with flags and torches.

An AFP analysis found it was created using Google's Gemini AI model.

But Gemini and Microsoft's Copilot falsely identified it as a genuine image of the protest.

"This inability to correctly identify AI images stems from the fact that they (AI models) are programmed only to mimic well," Rossine Fallorina, from the nonprofit Sigla Research Center, told AFP.

"In a sense, they can only generate things to resemble. They cannot ascertain whether the resemblance is actually distinguishable from reality."

Earlier this year, Columbia University's Tow Center for Digital Journalism tested the ability of seven AI chatbots -- including ChatGPT, Perplexity, Grok, and Gemini -- to verify 10 images from photojournalists of news events.

All seven models failed to correctly identify the provenance of the photos, the study said.

- 'Shocked' -

AFP tracked down the source of Co's photo that garnered over a million views across social media -- a middle-aged web developer in the Philippines, who said he created it "for fun" using Nano Banana, Gemini's AI image generator.

"Sadly, a lot of people believed it," he told AFP, requesting anonymity to avoid a backlash.

"I edited my post -- and added 'AI generated' to stop the spread -- because I was shocked at how many shares it got."

Such cases show how AI-generated photos flooding social platforms can look virtually identical to real imagery.

The trend has fueled concerns as surveys show online users are increasingly shifting from traditional search engines to AI tools for information gathering and verifying information.

The shift comes as Meta announced earlier this year it was ending its third-party fact-checking program in the United States, turning over the task of debunking falsehoods to ordinary users under a model known as "Community Notes."

Human fact-checking has long been a flashpoint in hyperpolarized societies, where conservative advocates accuse professional fact-checkers of liberal bias, a charge they reject.

AFP currently works in 26 languages with Meta's fact-checking program, including in Asia, Latin America, and the European Union.

Researchers say AI models can be useful to professional fact-checkers, helping to quickly geolocate images and spot visual clues to establish authenticity. But they caution that they cannot replace the work of trained human fact-checkers.

"We can't rely on AI tools to combat AI in the long run," Fallorina said.

burs-ac/sla/sms

G.Kucera--TPP