The China Mail - AI's blind spot: tools fail to detect their own fakes

USD -
AED 3.67315
AFN 63.503991
ALL 83.375041
AMD 377.180403
ANG 1.790083
AOA 917.000367
ARS 1383.990604
AUD 1.452433
AWG 1.8
AZN 1.70397
BAM 1.69972
BBD 2.014322
BDT 122.712716
BGN 1.709309
BHD 0.377349
BIF 2968.5
BMD 1
BND 1.28787
BOB 6.936019
BRL 5.255304
BSD 1.000117
BTN 94.794201
BWP 13.787919
BYN 2.976987
BYR 19600
BZD 2.011341
CAD 1.38995
CDF 2282.50392
CHF 0.798523
CLF 0.023433
CLP 925.260396
CNY 6.91185
CNH 6.92017
COP 3680.29
CRC 464.427092
CUC 1
CUP 26.5
CVE 96.12504
CZK 21.309304
DJF 177.720393
DKK 6.492704
DOP 59.72504
DZD 133.275765
EGP 52.642155
ERN 15
ETB 156.62504
EUR 0.866104
FJD 2.260391
FKP 0.75231
GBP 0.75375
GEL 2.680391
GGP 0.75231
GHS 10.97039
GIP 0.75231
GMD 73.503851
GNF 8780.000355
GTQ 7.653901
GYD 209.354875
HKD 7.82605
HNL 26.510388
HRK 6.545204
HTG 131.099243
HUF 338.020388
IDR 16990.8
ILS 3.13762
IMP 0.75231
INR 94.864204
IQD 1310
IRR 1313250.000352
ISK 124.760386
JEP 0.75231
JMD 157.422697
JOD 0.70904
JPY 160.29904
KES 129.903801
KGS 87.450384
KHR 4012.00035
KMF 428.00035
KPW 899.886996
KRW 1508.00035
KWD 0.30791
KYD 0.833446
KZT 483.490125
LAK 21900.000349
LBP 89550.000349
LKR 315.037957
LRD 183.625039
LSL 17.160381
LTL 2.95274
LVL 0.60489
LYD 6.375039
MAD 9.344504
MDL 17.566669
MGA 4175.000347
MKD 53.384435
MMK 2102.490525
MNT 3571.507434
MOP 8.069509
MRU 40.120379
MUR 46.770378
MVR 15.450378
MWK 1737.000345
MXN 18.121104
MYR 3.924039
MZN 63.950377
NAD 17.160377
NGN 1383.460377
NIO 36.720377
NOK 9.70286
NPR 151.667079
NZD 1.740645
OMR 0.385081
PAB 1.000109
PEN 3.459504
PGK 4.309039
PHP 60.550375
PKR 279.203701
PLN 3.72275
PYG 6538.855961
QAR 3.65325
RON 4.427304
RSD 101.818038
RUB 81.419514
RWF 1461
SAR 3.752351
SBD 8.042037
SCR 14.429246
SDG 601.000339
SEK 9.47367
SGD 1.292804
SHP 0.750259
SLE 24.550371
SLL 20969.510825
SOS 571.503662
SRD 37.601038
STD 20697.981008
STN 21.35
SVC 8.75063
SYP 111.824334
SZL 17.160369
THB 32.860369
TJS 9.556069
TMT 3.5
TND 2.926038
TOP 2.40776
TRY 44.433404
TTD 6.795201
TWD 32.044404
TZS 2576.487038
UAH 43.837189
UGX 3725.687866
UYU 40.481115
UZS 12205.000334
VES 467.928355
VND 26337.5
VUV 119.756335
WST 2.77551
XAF 570.070221
XAG 0.014291
XAU 0.000222
XCD 2.70255
XCG 1.802452
XDR 0.706792
XOF 568.000332
XPF 104.103591
YER 238.603589
ZAR 17.119995
ZMK 9001.203584
ZMW 18.826586
ZWL 321.999592
  • RBGPF

    -13.5000

    69

    -19.57%

  • GSK

    -0.1000

    53.84

    -0.19%

  • BCE

    -0.2200

    25.25

    -0.87%

  • NGG

    -0.4800

    81.92

    -0.59%

  • CMSD

    -0.0900

    22.66

    -0.4%

  • RELX

    -0.1000

    31.97

    -0.31%

  • BTI

    0.3749

    57.8

    +0.65%

  • RYCEF

    -0.5900

    14.65

    -4.03%

  • CMSC

    -0.0500

    22.77

    -0.22%

  • RIO

    0.8500

    86.64

    +0.98%

  • BP

    0.5100

    46.68

    +1.09%

  • JRI

    -0.2700

    11.8

    -2.29%

  • BCC

    0.1400

    74.43

    +0.19%

  • AZN

    5.0200

    188.42

    +2.66%

  • VOD

    -0.1400

    14.49

    -0.97%

AI's blind spot: tools fail to detect their own fakes
AI's blind spot: tools fail to detect their own fakes / Photo: © AFP

AI's blind spot: tools fail to detect their own fakes

When outraged Filipinos turned to an AI-powered chatbot to verify a viral photograph of a lawmaker embroiled in a corruption scandal, the tool failed to detect it was fabricated -- even though it had generated the image itself.

Text size:

Internet users are increasingly turning to chatbots to verify images in real time, but the tools often fail, raising questions about their visual debunking capabilities at a time when major tech platforms are scaling back human fact-checking.

In many cases, the tools wrongly identify images as real even when they are generated using the same generative models, further muddying an online information landscape awash with AI-generated fakes.

Among them is a fabricated image circulating on social media of Elizaldy Co, a former Philippine lawmaker charged by prosecutors in a multibillion-dollar flood-control corruption scam that sparked massive protests in the disaster-prone country.

The image of Co, whose whereabouts has been unknown since the official probe began, appeared to show him in Portugal.

When online sleuths tracking him asked Google's new AI mode whether the image was real, it incorrectly said it was authentic.

AFP's fact-checkers tracked down its creator and determined that the image was generated using Google AI.

"These models are trained primarily on language patterns and lack the specialized visual understanding needed to accurately identify AI-generated or manipulated imagery," Alon Yamin, chief executive of AI content detection platform Copyleaks, told AFP.

"With AI chatbots, even when an image originates from a similar generative model, the chatbot often provides inconsistent or overly generalized assessments, making them unreliable for tasks like fact-checking or verifying authenticity."

Google did not respond to AFP’s request for comment.

- 'Distinguishable from reality' -

AFP found similar examples of AI tools failing to verify their own creations.

During last month's deadly protests over lucrative benefits for senior officials in Pakistan-administered Kashmir, social media users shared a fabricated image purportedly showing men marching with flags and torches.

An AFP analysis found it was created using Google's Gemini AI model.

But Gemini and Microsoft's Copilot falsely identified it as a genuine image of the protest.

"This inability to correctly identify AI images stems from the fact that they (AI models) are programmed only to mimic well," Rossine Fallorina, from the nonprofit Sigla Research Center, told AFP.

"In a sense, they can only generate things to resemble. They cannot ascertain whether the resemblance is actually distinguishable from reality."

Earlier this year, Columbia University's Tow Center for Digital Journalism tested the ability of seven AI chatbots -- including ChatGPT, Perplexity, Grok, and Gemini -- to verify 10 images from photojournalists of news events.

All seven models failed to correctly identify the provenance of the photos, the study said.

- 'Shocked' -

AFP tracked down the source of Co's photo that garnered over a million views across social media -- a middle-aged web developer in the Philippines, who said he created it "for fun" using Nano Banana, Gemini's AI image generator.

"Sadly, a lot of people believed it," he told AFP, requesting anonymity to avoid a backlash.

"I edited my post -- and added 'AI generated' to stop the spread -- because I was shocked at how many shares it got."

Such cases show how AI-generated photos flooding social platforms can look virtually identical to real imagery.

The trend has fueled concerns as surveys show online users are increasingly shifting from traditional search engines to AI tools for information gathering and verifying information.

The shift comes as Meta announced earlier this year it was ending its third-party fact-checking program in the United States, turning over the task of debunking falsehoods to ordinary users under a model known as "Community Notes."

Human fact-checking has long been a flashpoint in hyperpolarized societies, where conservative advocates accuse professional fact-checkers of liberal bias, a charge they reject.

AFP currently works in 26 languages with Meta's fact-checking program, including in Asia, Latin America, and the European Union.

Researchers say AI models can be useful to professional fact-checkers, helping to quickly geolocate images and spot visual clues to establish authenticity. But they caution that they cannot replace the work of trained human fact-checkers.

"We can't rely on AI tools to combat AI in the long run," Fallorina said.

burs-ac/sla/sms

E.Lau--ThChM