The China Mail - Firms and researchers at odds over superhuman AI

USD -
AED 3.672504
AFN 69.456103
ALL 84.764831
AMD 381.290295
ANG 1.789623
AOA 916.000367
ARS 1179.376574
AUD 1.538935
AWG 1.8025
AZN 1.70397
BAM 1.692527
BBD 2.010212
BDT 121.665008
BGN 1.696633
BHD 0.375579
BIF 2964.389252
BMD 1
BND 1.278698
BOB 6.879841
BRL 5.543904
BSD 0.99563
BTN 85.673489
BWP 13.382372
BYN 3.258189
BYR 19600
BZD 1.999913
CAD 1.35865
CDF 2877.000362
CHF 0.812438
CLF 0.024131
CLP 926.026567
CNY 7.181604
CNH 7.18941
COP 4135.519882
CRC 501.838951
CUC 1
CUP 26.5
CVE 95.422093
CZK 21.500904
DJF 177.292199
DKK 6.45704
DOP 58.803167
DZD 130.034183
EGP 49.707931
ERN 15
ETB 134.317771
EUR 0.865404
FJD 2.24825
FKP 0.736781
GBP 0.737708
GEL 2.740391
GGP 0.736781
GHS 10.254857
GIP 0.736781
GMD 70.503851
GNF 8627.060707
GTQ 7.650902
GYD 208.299078
HKD 7.849415
HNL 25.985029
HRK 6.522704
HTG 130.569859
HUF 348.50504
IDR 16299.3
ILS 3.620404
IMP 0.736781
INR 86.184504
IQD 1304.227424
IRR 42100.000352
ISK 124.650386
JEP 0.736781
JMD 159.404613
JOD 0.70904
JPY 144.10604
KES 128.631388
KGS 87.450384
KHR 3992.038423
KMF 426.503794
KPW 899.999993
KRW 1367.140383
KWD 0.30622
KYD 0.829648
KZT 510.665917
LAK 21481.545584
LBP 89206.525031
LKR 298.109126
LRD 199.125957
LSL 17.917528
LTL 2.95274
LVL 0.60489
LYD 5.439834
MAD 9.103111
MDL 17.04989
MGA 4495.694691
MKD 53.251698
MMK 2099.702644
MNT 3581.705956
MOP 8.049154
MRU 39.525767
MUR 45.510378
MVR 15.405039
MWK 1726.364069
MXN 18.95075
MYR 4.245504
MZN 63.950377
NAD 17.917528
NGN 1542.440377
NIO 36.640561
NOK 9.912804
NPR 137.077582
NZD 1.661972
OMR 0.384259
PAB 0.99563
PEN 3.593613
PGK 4.159058
PHP 56.090375
PKR 282.254944
PLN 3.698316
PYG 7944.268963
QAR 3.631864
RON 4.350504
RSD 101.423565
RUB 79.779066
RWF 1437.670373
SAR 3.753593
SBD 8.347391
SCR 14.210372
SDG 600.503676
SEK 9.483995
SGD 1.281904
SHP 0.785843
SLE 22.050371
SLL 20969.503664
SOS 568.99312
SRD 37.528038
STD 20697.981008
SVC 8.711869
SYP 13001.852669
SZL 17.905759
THB 32.405038
TJS 10.055644
TMT 3.5
TND 2.945956
TOP 2.342104
TRY 39.40328
TTD 6.751763
TWD 29.520367
TZS 2573.66622
UAH 41.29791
UGX 3587.901865
UYU 40.932889
UZS 12650.253126
VES 102.167038
VND 26075
VUV 119.102168
WST 2.619186
XAF 567.657825
XAG 0.027532
XAU 0.000291
XCD 2.70255
XDR 0.705984
XOF 567.657825
XPF 103.206265
YER 243.350363
ZAR 17.92535
ZMK 9001.203587
ZMW 24.069058
ZWL 321.999592
  • CMSC

    0.0900

    22.314

    +0.4%

  • CMSD

    0.0250

    22.285

    +0.11%

  • RBGPF

    0.0000

    69.04

    0%

  • SCS

    0.0400

    10.74

    +0.37%

  • RELX

    0.0300

    53

    +0.06%

  • RIO

    -0.1400

    59.33

    -0.24%

  • GSK

    0.1300

    41.45

    +0.31%

  • NGG

    0.2700

    71.48

    +0.38%

  • BP

    0.1750

    30.4

    +0.58%

  • BTI

    0.7150

    48.215

    +1.48%

  • BCC

    0.7900

    91.02

    +0.87%

  • JRI

    0.0200

    13.13

    +0.15%

  • VOD

    0.0100

    9.85

    +0.1%

  • BCE

    -0.0600

    22.445

    -0.27%

  • RYCEF

    0.1000

    12

    +0.83%

  • AZN

    -0.1200

    73.71

    -0.16%

Firms and researchers at odds over superhuman AI
Firms and researchers at odds over superhuman AI / Photo: © AFP/File

Firms and researchers at odds over superhuman AI

Hype is growing from leaders of major AI companies that "strong" computer intelligence will imminently outstrip humans, but many researchers in the field see the claims as marketing spin.

Text size:

The belief that human-or-better intelligence -- often called "artificial general intelligence" (AGI) -- will emerge from current machine-learning techniques fuels hypotheses for the future ranging from machine-delivered hyperabundance to human extinction.

"Systems that start to point to AGI are coming into view," OpenAI chief Sam Altman wrote in a blog post last month. Anthropic's Dario Amodei has said the milestone "could come as early as 2026".

Such predictions help justify the hundreds of billions of dollars being poured into computing hardware and the energy supplies to run it.

Others, though are more sceptical.

Meta's chief AI scientist Yann LeCun told AFP last month that "we are not going to get to human-level AI by just scaling up LLMs" -- the large language models behind current systems like ChatGPT or Claude.

LeCun's view appears backed by a majority of academics in the field.

Over three-quarters of respondents to a recent survey by the US-based Association for the Advancement of Artificial Intelligence (AAAI) agreed that "scaling up current approaches" was unlikely to produce AGI.

- 'Genie out of the bottle' -

Some academics believe that many of the companies' claims, which bosses have at times flanked with warnings about AGI's dangers for mankind, are a strategy to capture attention.

Businesses have "made these big investments, and they have to pay off," said Kristian Kersting, a leading researcher at the Technical University of Darmstadt in Germany and AAAI member.

"They just say, 'this is so dangerous that only I can operate it, in fact I myself am afraid but we've already let the genie out of the bottle, so I'm going to sacrifice myself on your behalf -- but then you're dependent on me'."

Scepticism among academic researchers is not total, with prominent figures like Nobel-winning physicist Geoffrey Hinton or 2018 Turing Prize winner Yoshua Bengio warning about dangers from powerful AI.

"It's a bit like Goethe's 'The Sorcerer's Apprentice', you have something you suddenly can't control any more," Kersting said -- referring to a poem in which a would-be sorcerer loses control of a broom he has enchanted to do his chores.

A similar, more recent thought experiment is the "paperclip maximiser".

This imagined AI would pursue its goal of making paperclips so single-mindedly that it would turn Earth and ultimately all matter in the universe into paperclips or paperclip-making machines -- having first got rid of human beings that it judged might hinder its progress by switching it off.

While not "evil" as such, the maximiser would fall fatally short on what thinkers in the field call "alignment" of AI with human objectives and values.

Kersting said he "can understand" such fears -- while suggesting that "human intelligence, its diversity and quality is so outstanding that it will take a long time, if ever" for computers to match it.

He is far more concerned with near-term harms from already-existing AI, such as discrimination in cases where it interacts with humans.

- 'Biggest thing ever' -

The apparently stark gulf in outlook between academics and AI industry leaders may simply reflect people's attitudes as they pick a career path, suggested Sean O hEigeartaigh, director of the AI: Futures and Responsibility programme at Britain's Cambridge University.

"If you are very optimistic about how powerful the present techniques are, you're probably more likely to go and work at one of the companies that's putting a lot of resource into trying to make it happen," he said.

Even if Altman and Amodei may be "quite optimistic" about rapid timescales and AGI emerges much later, "we should be thinking about this and taking it seriously, because it would be the biggest thing that would ever happen," O hEigeartaigh added.

"If it were anything else... a chance that aliens would arrive by 2030 or that there'd be another giant pandemic or something, we'd put some time into planning for it".

The challenge can lie in communicating these ideas to politicians and the public.

Talk of super-AI "does instantly create this sort of immune reaction... it sounds like science fiction," O hEigeartaigh said.

A.Sun--ThChM