The China Mail - 'Tool for grifters': AI deepfakes push bogus sexual cures

USD -
AED 3.673102
AFN 62.493319
ALL 81.650049
AMD 368.780249
ANG 1.79046
AOA 917.999616
ARS 1391.440285
AUD 1.38485
AWG 1.8025
AZN 1.697591
BAM 1.670681
BBD 2.014496
BDT 122.776371
BGN 1.66992
BHD 0.377299
BIF 2975
BMD 1
BND 1.273528
BOB 6.911397
BRL 5.004602
BSD 1.000201
BTN 95.835344
BWP 14.087599
BYN 2.794335
BYR 19600
BZD 2.011549
CAD 1.37225
CDF 2244.99985
CHF 0.783702
CLF 0.022735
CLP 894.791543
CNY 6.785151
CNH 6.78612
COP 3789.73
CRC 454.512452
CUC 1
CUP 26.5
CVE 94.702891
CZK 20.8311
DJF 177.719843
DKK 6.404399
DOP 59.700677
DZD 132.447999
EGP 52.867303
ERN 15
ETB 157.449907
EUR 0.85697
FJD 2.191597
FKP 0.739691
GBP 0.746365
GEL 2.680049
GGP 0.739691
GHS 11.409837
GIP 0.739691
GMD 72.499865
GNF 8779.999965
GTQ 7.630738
GYD 209.246802
HKD 7.83251
HNL 26.620617
HRK 6.458898
HTG 130.972363
HUF 306.545501
IDR 17535.3
ILS 2.902601
IMP 0.739691
INR 95.7091
IQD 1310
IRR 1315000.000078
ISK 123.180086
JEP 0.739691
JMD 158.141561
JOD 0.708994
JPY 158.39103
KES 129.250112
KGS 87.450082
KHR 4011.999726
KMF 421.999959
KPW 899.97066
KRW 1493.490202
KWD 0.30849
KYD 0.833543
KZT 473.448852
LAK 21955.000133
LBP 90063.841638
LKR 325.320759
LRD 183.250142
LSL 16.4899
LTL 2.95274
LVL 0.60489
LYD 6.330395
MAD 9.20875
MDL 17.192645
MGA 4177.499513
MKD 52.834798
MMK 2099.865061
MNT 3580.130218
MOP 8.069362
MRU 39.98999
MUR 46.895264
MVR 15.401208
MWK 1741.000482
MXN 17.225302
MYR 3.931505
MZN 63.910286
NAD 16.489493
NGN 1369.370618
NIO 36.714995
NOK 9.233501
NPR 153.332792
NZD 1.691475
OMR 0.384492
PAB 1.000184
PEN 3.44698
PGK 4.193011
PHP 61.460973
PKR 278.591881
PLN 3.636395
PYG 6094.852476
QAR 3.645502
RON 4.456702
RSD 100.601025
RUB 73.24798
RWF 1461
SAR 3.707824
SBD 8.016136
SCR 13.867581
SDG 600.503741
SEK 9.369043
SGD 1.2756
SHP 0.746601
SLE 24.650261
SLL 20969.502105
SOS 571.497017
SRD 37.206963
STD 20697.981008
STN 21.25
SVC 8.751249
SYP 110.528733
SZL 16.490133
THB 32.420153
TJS 9.346574
TMT 3.5
TND 2.888037
TOP 2.40776
TRY 45.464801
TTD 6.790867
TWD 31.544499
TZS 2595.000031
UAH 43.968225
UGX 3740.52909
UYU 39.831211
UZS 12044.999697
VES 510.148815
VND 26345
VUV 118.077659
WST 2.708521
XAF 560.318959
XAG 0.011986
XAU 0.000215
XCD 2.70255
XCG 1.802565
XDR 0.694969
XOF 557.509472
XPF 102.624995
YER 238.649788
ZAR 16.455495
ZMK 9001.134371
ZMW 18.82781
ZWL 321.999592
  • RBGPF

    -0.2100

    60.79

    -0.35%

  • CMSD

    0.0400

    23.6

    +0.17%

  • CMSC

    0.0898

    23.14

    +0.39%

  • NGG

    0.4500

    87.43

    +0.51%

  • JRI

    0.0100

    13.14

    +0.08%

  • BCC

    2.4200

    69.4

    +3.49%

  • BCE

    -0.2000

    24.19

    -0.83%

  • RIO

    -2.4500

    109.59

    -2.24%

  • GSK

    -0.0300

    50.96

    -0.06%

  • RELX

    -0.1600

    31.46

    -0.51%

  • BTI

    1.3500

    66.7

    +2.02%

  • RYCEF

    -0.0700

    15.93

    -0.44%

  • BP

    -0.0200

    44.12

    -0.05%

  • VOD

    -0.0300

    15.48

    -0.19%

  • AZN

    -2.7600

    184.96

    -1.49%

'Tool for grifters': AI deepfakes push bogus sexual cures
'Tool for grifters': AI deepfakes push bogus sexual cures / Photo: © AFP

'Tool for grifters': AI deepfakes push bogus sexual cures

Holding an oversized carrot, a brawny, shirtless man promotes a supplement he claims can enlarge male genitalia -- one of countless AI-generated videos on TikTok peddling unproven sexual treatments.

Text size:

The rise of generative AI has made it easy -- and financially lucrative -- to mass-produce such videos with minimal human oversight, often featuring fake celebrity endorsements of bogus and potentially harmful products.

In some TikTok videos, carrots are used as a euphemism for male genitalia, apparently to evade content moderation policing sexually explicit language.

"You would notice that your carrot has grown up," the muscled man says in a robotic voice in one video, directing users to an online purchase link.

"This product will change your life," the man adds, claiming without evidence that the herbs used as ingredients boost testosterone and send energy levels "through the roof."

The video appears to be AI-generated, according to a deepfake detection service recently launched by the Bay Area-headquartered firm Resemble AI, which shared its results with AFP.

"As seen in this example, misleading AI-generated content is being used to market supplements with exaggerated or unverified claims, potentially putting consumers' health at risk," Zohaib Ahmed, Resemble AI's chief executive and co-founder, told AFP.

"We're seeing AI-generated content weaponized to spread false information."

- 'Cheap way' -

The trend underscores how rapid advances in artificial intelligence have fueled what researchers call an AI dystopia, a deception-filled online universe designed to manipulate unsuspecting users into buying dubious products.

They include everything from unverified -- and in some cases, potentially harmful -- dietary supplements to weight loss products and sexual remedies.

"AI is a useful tool for grifters looking to create large volumes of content slop for a low cost," misinformation researcher Abbie Richards told AFP.

"It's a cheap way to produce advertisements," she added.

Alexios Mantzarlis, director of the Security, Trust, and Safety Initiative at Cornell Tech, has observed a surge of "AI doctor" avatars and audio tracks on TikTok that promote questionable sexual remedies.

Some of these videos, many with millions of views, peddle testosterone-boosting concoctions made from ingredients such as lemon, ginger and garlic.

More troublingly, rapidly evolving AI tools have enabled the creation of deepfakes impersonating celebrities such as actress Amanda Seyfried and actor Robert De Niro.

"Your husband can't get it up?" Anthony Fauci, former director of the National Institute of Allergy and Infectious Diseases, appears to ask in a TikTok video promoting a prostate supplement.

But the clip is a deepfake, using Fauci's likeness.

- 'Pernicious' -

Many manipulated videos are created from existing ones, modified with AI-generated voices and lip-synced to match what the altered voice says.

"The impersonation videos are particularly pernicious as they further degrade our ability to discern authentic accounts online," Mantzarlis said.

Last year, Mantzarlis discovered hundreds of ads on YouTube featuring deepfakes of celebrities -- including Arnold Schwarzenegger, Sylvester Stallone, and Mike Tyson -- promoting supplements branded as erectile dysfunction cures.

The rapid pace of generating short-form AI videos means that even when tech platforms remove questionable content, near-identical versions quickly reappear -- turning moderation into a game of whack-a-mole.

Researchers say this creates unique challenges for policing AI-generated content, requiring novel solutions and more sophisticated detection tools.

AFP's fact checkers have repeatedly debunked scam ads on Facebook promoting treatments -- including erectile dysfunction cures -- that use fake endorsements by Ben Carson, a neurosurgeon and former US cabinet member.

Yet many users still consider the endorsements legitimate, illustrating the appeal of deepfakes.

"Scammy affiliate marketing schemes and questionable sex supplements have existed for as long as the internet and before," Mantzarlis said.

"As with every other bad thing online, generative AI has made this abuse vector cheaper and quicker to deploy at scale."

V.Liu--ThChM