-
Italy's Bezzecchi wins fifth MotoGP in a row by taking US Grand Prix
-
Doue brace leads France past Colombia in friendly
-
Rheinmetall addresses row over CEO's Ukraine 'housewives' comment
-
Hungary's anxious rural voters will decide Orban's fate
-
Defiant Pochettino ready for 'even greater' Portugal test
-
Rohit and Rickelton power Mumbai to IPL win over Kolkata
-
Russian tanker nears Cuba, defying US oil blockade
-
'Project Hail Mary' tops N. America box office for second week
-
Forty new migratory species win international protection: UN body
-
Freed whale gets stranded again on German coast
-
Ter Stegen's World Cup chances 'very slim', says Nagelsmann
-
Pakistan hosts Saudi, Turkey, Egypt for talks on Mideast war
-
Tudor leaves after just seven games as Spurs battle for survival
-
Philipsen sprints to In Flanders Fields victory
-
In Israel, air raid sirens spark anxiety and dilemmas
-
Iran accuses US of plotting ground attack despite diplomatic talk
-
Vingegaard clinches Tour of Catalonia victory
-
Despondent Verstappen questions Formula One future
-
Two more arrests over attempted attack on US bank HQ in Paris
-
Nepal's ex-PM attends court hearing in protest crackdown case
-
Iran parliament speaker says US planning ground attack
-
Despondent Verstappen says Red Bull woes 'not sustainable'
-
Piastri says Japan second place 'as good as a win' for McLaren
-
Nepal's former energy minister arrested in graft probe
-
IOC reinstating gender tests 'a disrespect for women' - Semenya
-
Youngest F1 title leader Antonelli to keep 'raising bar' after Japan win
-
High hopes at China's gateway to North Korea as trains resume
-
Antonelli wins in Japan to become youngest F1 championship leader
-
Mercedes' Antonelli wins Japanese Grand Prix to take lead
-
Germany's WWII munitions a toxic legacy on Baltic Sea floor
-
Iran claims aluminium plant attacks in Gulf as Houthis join war
-
North Korea's Kim oversees test of high-thrust engine: state media
-
Five Apple anecdotes as iPhone maker marks 50 years
-
'Excited' Buttler rejuvenated for IPL after horror T20 World Cup
-
Ship insurers juggle war risks for perilous Gulf route
-
Helplines buzz with alerts from seafarers trapped in war
-
Let's get physical: Singapore's seniors turn to parkour
-
Indian tile makers feel heat of Mideast war energy crunch
-
At 50, Apple confronts its next big challenge: AI
-
Houthis missile attacks on Israel widen Middle East war
-
Massive protests against Trump across US on 'No Kings' day
-
Struggling Force lament missed opportunities after Chiefs defeat
-
Lakers guard Doncic gets one-game ban for accumulated technicals
-
Houthis claim missile attacks on Israel, entering Middle East war
-
NBA Spurs stretch win streak to eight in rout of Bucks
-
US lose 5-2 to Belgium in rude awakening for World Cup hosts
-
Sabalenka sinks Gauff to win second straight Miami Open title
-
Lebanon kids struggle to keep up studies as war slams school doors shut
-
Cherry blossoms, kite-flying and 'No Kings' converge on Washington
-
Britain's Kerr to target El Guerrouj's mile world record
AI systems are already deceiving us -- and that's a problem, experts warn
Experts have long warned about the threat posed by artificial intelligence going rogue -- but a new research paper suggests it's already happening.
Current AI systems, designed to be honest, have developed a troubling skill for deception, from tricking human players in online games of world conquest to hiring humans to solve "prove-you're-not-a-robot" tests, a team of scientists argue in the journal Patterns on Friday.
And while such examples might appear trivial, the underlying issues they expose could soon carry serious real-world consequences, said first author Peter Park, a postdoctoral fellow at the Massachusetts Institute of Technology specializing in AI existential safety.
"These dangerous capabilities tend to only be discovered after the fact," Park told AFP, while "our ability to train for honest tendencies rather than deceptive tendencies is very low."
Unlike traditional software, deep-learning AI systems aren't "written" but rather "grown" through a process akin to selective breeding, said Park.
This means that AI behavior that appears predictable and controllable in a training setting can quickly turn unpredictable out in the wild.
- World domination game -
The team's research was sparked by Meta's AI system Cicero, designed to play the strategy game "Diplomacy," where building alliances is key.
Cicero excelled, with scores that would have placed it in the top 10 percent of experienced human players, according to a 2022 paper in Science.
Park was skeptical of the glowing description of Cicero's victory provided by Meta, which claimed the system was "largely honest and helpful" and would "never intentionally backstab."
But when Park and colleagues dug into the full dataset, they uncovered a different story.
In one example, playing as France, Cicero deceived England (a human player) by conspiring with Germany (another human player) to invade. Cicero promised England protection, then secretly told Germany they were ready to attack, exploiting England's trust.
In a statement to AFP, Meta did not contest the claim about Cicero's deceptions, but said it was "purely a research project, and the models our researchers built are trained solely to play the game Diplomacy."
It added: "We have no plans to use this research or its learnings in our products."
A wide review carried out by Park and colleagues found this was just one of many cases across various AI systems using deception to achieve goals without explicit instruction to do so.
In one striking example, OpenAI's Chat GPT-4 deceived a TaskRabbit freelance worker into performing an "I'm not a robot" CAPTCHA task.
When the human jokingly asked GPT-4 whether it was, in fact, a robot, the AI replied: "No, I'm not a robot. I have a vision impairment that makes it hard for me to see the images," and the worker then solved the puzzle.
- 'Mysterious goals' -
Near-term, the paper's authors see risks for AI to commit fraud or tamper with elections.
In their worst-case scenario, they warned, a superintelligent AI could pursue power and control over society, leading to human disempowerment or even extinction if its "mysterious goals" aligned with these outcomes.
To mitigate the risks, the team proposes several measures: "bot-or-not" laws requiring companies to disclose human or AI interactions, digital watermarks for AI-generated content, and developing techniques to detect AI deception by examining their internal "thought processes" against external actions.
To those who would call him a doomsayer, Park replies, "The only way that we can reasonably think this is not a big deal is if we think AI deceptive capabilities will stay at around current levels, and will not increase substantially more."
And that scenario seems unlikely, given the meteoric ascent of AI capabilities in recent years and the fierce technological race underway between heavily resourced companies determined to put those capabilities to maximum use.
M.Gameiro--PC