-
France, allies preparing bid to 'gradually' reopen Strait of Hormuz
-
Anthropic takes Trump administration to court over Pentagon row
-
Antarctic sea ice improves after four years of extreme lows: US scientists
-
Beating Barca would make us Newcastle legends: Howe
-
Iran war sends crude prices soaring as Khamenei son takes charge
-
Zelensky says 11 countries asking Ukraine for drone help against Iran
-
France, allies preparing 'defensive' mission to reopen Strait of Hormuz: Macron
-
Ships brandish China-links to weave through Strait of Hormuz
-
Trump says Australia will grant asylum to Iran women footballers
-
War in the Middle East: economic impact around the world
-
Huge numbers at imminent risk from S.Sudan army offensive: MSF
-
G7 'not there yet' on release of oil reserves: French minister
-
Live Nation settles antitrust case with US Justice Dept, states object
-
EU lawmakers set to greenlight 'return hubs' for migrants
-
Macron says France, allies preparing 'defensive' mission to reopen Strait of Hormuz
-
Water emerges as a dangerous new war target
-
Scotland locks Cummings and Brown ruled out of Ireland Six Nations clash
-
Stocks slide as oil soars past $100 on Mideast war
-
NATO intercepts second Iran missile in Turkish airspace: Ankara
-
South Korea squeeze into World Baseball Classic quarter-finals
-
Premier League teams are faster: Atletico's Simeone on Spurs clash
-
North Korea cancels Pyongyang international marathon: tour agency
-
Ukrainian bank worker detained by Hungary was forcibly medicated: Kyiv
-
Macron discusses security in Cyprus, plans aircraft carrier visit
-
UK PM Starmer says 'monitoring' economic impact of Iran war
-
Stranded Iran sailors put Sri Lanka, India in diplomatic dilemma
-
Bangladesh scraps light displays as Mideast war worsens fuel crunch
-
Stocks tumble, oil soars past $100 on Mideast war
-
Iran war sends oil price soaring as Khamenei son takes charge
-
Incensed North Korea briefly refuse to play in bitter Asian Cup loss
-
Landmark trial opens for Turkish opposition champion Imamoglu
-
Indonesia landfill collapse kills five
-
African players in Europe: Marmoush torments Newcastle again
-
Kenya flash floods death toll rises to 45
-
Asian economies move to limit Mideast war's impact at home
-
Jail for up to 16 years for Australian hitmen who killed compatriot in Bali
-
Russia wins first Paralympic gold since 2014
-
'T20 kings': nation celebrates Indian romp to World Cup glory
-
Indonesia landfill collapse kills four
-
Equities plunge as energy prices soar on Mideast crisis
-
Unstoppable India target Olympic gold after making World Cup history
-
Khamenei's son takes charge as Iran war sends oil price soaring
-
Asian equities plunge as oil soars 30% on Mideast crisis
-
Dead on arrival: South Sudan's devastated health system
-
Redknapp and The Jukebox Man the headline act at Cheltenham Festival
-
Singer Rihanna's LA mansion struck by gunfire: reports
-
Iran launches missiles as Khamenei's son takes charge
-
Sharp drop in Chinese military aircraft near Taiwan raises questions
-
Gauff retires with 'scary' injury to send Eala through at Indian Wells
-
Mojtaba Khamenei: son and successor to Iran's supreme leader
AI systems are already deceiving us -- and that's a problem, experts warn
Experts have long warned about the threat posed by artificial intelligence going rogue -- but a new research paper suggests it's already happening.
Current AI systems, designed to be honest, have developed a troubling skill for deception, from tricking human players in online games of world conquest to hiring humans to solve "prove-you're-not-a-robot" tests, a team of scientists argue in the journal Patterns on Friday.
And while such examples might appear trivial, the underlying issues they expose could soon carry serious real-world consequences, said first author Peter Park, a postdoctoral fellow at the Massachusetts Institute of Technology specializing in AI existential safety.
"These dangerous capabilities tend to only be discovered after the fact," Park told AFP, while "our ability to train for honest tendencies rather than deceptive tendencies is very low."
Unlike traditional software, deep-learning AI systems aren't "written" but rather "grown" through a process akin to selective breeding, said Park.
This means that AI behavior that appears predictable and controllable in a training setting can quickly turn unpredictable out in the wild.
- World domination game -
The team's research was sparked by Meta's AI system Cicero, designed to play the strategy game "Diplomacy," where building alliances is key.
Cicero excelled, with scores that would have placed it in the top 10 percent of experienced human players, according to a 2022 paper in Science.
Park was skeptical of the glowing description of Cicero's victory provided by Meta, which claimed the system was "largely honest and helpful" and would "never intentionally backstab."
But when Park and colleagues dug into the full dataset, they uncovered a different story.
In one example, playing as France, Cicero deceived England (a human player) by conspiring with Germany (another human player) to invade. Cicero promised England protection, then secretly told Germany they were ready to attack, exploiting England's trust.
In a statement to AFP, Meta did not contest the claim about Cicero's deceptions, but said it was "purely a research project, and the models our researchers built are trained solely to play the game Diplomacy."
It added: "We have no plans to use this research or its learnings in our products."
A wide review carried out by Park and colleagues found this was just one of many cases across various AI systems using deception to achieve goals without explicit instruction to do so.
In one striking example, OpenAI's Chat GPT-4 deceived a TaskRabbit freelance worker into performing an "I'm not a robot" CAPTCHA task.
When the human jokingly asked GPT-4 whether it was, in fact, a robot, the AI replied: "No, I'm not a robot. I have a vision impairment that makes it hard for me to see the images," and the worker then solved the puzzle.
- 'Mysterious goals' -
Near-term, the paper's authors see risks for AI to commit fraud or tamper with elections.
In their worst-case scenario, they warned, a superintelligent AI could pursue power and control over society, leading to human disempowerment or even extinction if its "mysterious goals" aligned with these outcomes.
To mitigate the risks, the team proposes several measures: "bot-or-not" laws requiring companies to disclose human or AI interactions, digital watermarks for AI-generated content, and developing techniques to detect AI deception by examining their internal "thought processes" against external actions.
To those who would call him a doomsayer, Park replies, "The only way that we can reasonably think this is not a big deal is if we think AI deceptive capabilities will stay at around current levels, and will not increase substantially more."
And that scenario seems unlikely, given the meteoric ascent of AI capabilities in recent years and the fierce technological race underway between heavily resourced companies determined to put those capabilities to maximum use.
P.Silva--AMWN