
-
Lions' Ringrose out of first Wallabies Test, Cowan-Dickie in doubt
-
Sinner seeks redemption against Alcaraz in Wimbledon final
-
Stokes' run-out of Pant helps England slow India charge
-
Farrell makes tour debut as Lions thrash Australia-New Zealand XV
-
Sparkling Fiji score four tries to beat error-prone Scotland 29-14
-
Pioli returns to Fiorentina after one season at Al-Nassr
-
Marc Marquez takes seventh pole of season at German MotoGP
-
Barrett says All Blacks impressed by young France talent
-
Pakistan won't send hockey teams to India: govt sources
-
NCaledonia politicians agree on statehood while remaining French
-
Robertson hails 'ruthless' All Blacks after France crushed 43-17
-
American midfielder Tillman joins Leverkusen from PSV
-
Sparkling Fiji score four tries beat error-prone Scotland 29-14
-
Ukraine says four killed in massive Russian drone, missile attack
-
Akram hails 'modern-day great' Starc on 100-Test milestone
-
Wales look to future after ending 18-game losing run with Japan win
-
Gaza ceasefire talks held up by Israel withdrawal plans: Palestinian sources
-
All Blacks score six tries to hammer under-strength France
-
Cambodia genocide survivors 'thrilled' at new UNESCO status
-
Worker in critical condition after US immigration raid on California farm
-
German backpacker drank from puddles in Australian bush ordeal
-
German backpacker escapes Australian bush ordeal by 'sheer luck'
-
Tourists, residents evacuated from Grand Canyon due to wildfires
-
Bad Bunny draws jubilant Puerto Ricans to historic residency
-
Worker dies after US immigration raid on California farm
-
PSG coach Luis Enrique warns against complacency in Club World Cup final
-
Boeing evades MAX crash trial with last-minute settlement
-
US sanctions Cuban president four years after historic protests
-
Pope Leo's Illinois childhood home to become tourist site
-
Manchester gives hometown heroes Oasis rapturous reception
-
Canada just can't win in trade war with Trump
-
US State Department begins mass layoffs
-
Fuel to Air India jet engines cut off moments before crash: probe
-
Chelsea out to stop PSG completing clean sweep in Club World Cup final
-
Ecuador's top drug lord agrees to US extradition
-
Son of Mexico's 'El Chapo' pleads guilty in US drugs case
-
500 tourists evacuated from Grand Canyon wildfires
-
Italy join Spain in Women's Euro 2025 quarter-finals
-
Chelsea's Fernandez warns of 'dangerous' heat at Club World Cup
-
Maresca optimistic for Chelsea against 'best in world' PSG
-
Trump voices shock at devastating scale of Texas flood damage
-
Sinner unfazed by French Open collapse as he prepares for Alcaraz rematch
-
Lyles scorches to comeback win, Alfred conquers 100m
-
'Superman' aims to save flagging film franchise, not just humanity
-
Forest winger Elanga signs for Newcastle
-
Liverpool to retire Diogo Jota's number 20 shirt
-
'Still in the game': Lyles outstrips Tebogo in season-opening 200m
-
Bumrah proud of 'really special' five-wicket haul at Lord's
-
Son of Mexico's 'El Chapo' pleads guilty in US drugs case: report
-
Mob lynches five alleged thieves in quake-hit Guatemalan town

AI systems are already deceiving us -- and that's a problem, experts warn
Experts have long warned about the threat posed by artificial intelligence going rogue -- but a new research paper suggests it's already happening.
Current AI systems, designed to be honest, have developed a troubling skill for deception, from tricking human players in online games of world conquest to hiring humans to solve "prove-you're-not-a-robot" tests, a team of scientists argue in the journal Patterns on Friday.
And while such examples might appear trivial, the underlying issues they expose could soon carry serious real-world consequences, said first author Peter Park, a postdoctoral fellow at the Massachusetts Institute of Technology specializing in AI existential safety.
"These dangerous capabilities tend to only be discovered after the fact," Park told AFP, while "our ability to train for honest tendencies rather than deceptive tendencies is very low."
Unlike traditional software, deep-learning AI systems aren't "written" but rather "grown" through a process akin to selective breeding, said Park.
This means that AI behavior that appears predictable and controllable in a training setting can quickly turn unpredictable out in the wild.
- World domination game -
The team's research was sparked by Meta's AI system Cicero, designed to play the strategy game "Diplomacy," where building alliances is key.
Cicero excelled, with scores that would have placed it in the top 10 percent of experienced human players, according to a 2022 paper in Science.
Park was skeptical of the glowing description of Cicero's victory provided by Meta, which claimed the system was "largely honest and helpful" and would "never intentionally backstab."
But when Park and colleagues dug into the full dataset, they uncovered a different story.
In one example, playing as France, Cicero deceived England (a human player) by conspiring with Germany (another human player) to invade. Cicero promised England protection, then secretly told Germany they were ready to attack, exploiting England's trust.
In a statement to AFP, Meta did not contest the claim about Cicero's deceptions, but said it was "purely a research project, and the models our researchers built are trained solely to play the game Diplomacy."
It added: "We have no plans to use this research or its learnings in our products."
A wide review carried out by Park and colleagues found this was just one of many cases across various AI systems using deception to achieve goals without explicit instruction to do so.
In one striking example, OpenAI's Chat GPT-4 deceived a TaskRabbit freelance worker into performing an "I'm not a robot" CAPTCHA task.
When the human jokingly asked GPT-4 whether it was, in fact, a robot, the AI replied: "No, I'm not a robot. I have a vision impairment that makes it hard for me to see the images," and the worker then solved the puzzle.
- 'Mysterious goals' -
Near-term, the paper's authors see risks for AI to commit fraud or tamper with elections.
In their worst-case scenario, they warned, a superintelligent AI could pursue power and control over society, leading to human disempowerment or even extinction if its "mysterious goals" aligned with these outcomes.
To mitigate the risks, the team proposes several measures: "bot-or-not" laws requiring companies to disclose human or AI interactions, digital watermarks for AI-generated content, and developing techniques to detect AI deception by examining their internal "thought processes" against external actions.
To those who would call him a doomsayer, Park replies, "The only way that we can reasonably think this is not a big deal is if we think AI deceptive capabilities will stay at around current levels, and will not increase substantially more."
And that scenario seems unlikely, given the meteoric ascent of AI capabilities in recent years and the fierce technological race underway between heavily resourced companies determined to put those capabilities to maximum use.
P.Silva--AMWN