Artificial intelligence systems are advancing faster than regulators expected, while their risks are spreading into everyday life, according to a major international safety review released this week.
The second annual International AI Safety Report surveys progress across the field, from deepfake media and cyber-security to jobs and mental health. It was commissioned after the 2023 global AI safety summit and is chaired by Canadian computer scientist Yoshua Bengio.
The report does not push specific policy solutions. However, it maps the current state of AI development ahead of a global summit in India later this month. Consequently, it offers a snapshot of where the technology stands and where pressures are building.
Over the past year, developers released a wave of powerful new AI models. These systems form the core of chatbots, image generators, and coding tools used by millions of people. Additionally, they increasingly rely on “reasoning systems” that break problems into smaller steps.
According to the report, those reasoning approaches delivered sharp gains in maths, science, and software development. Bengio said the improvement in reasoning ability has been substantial. Last year, systems from Google parent Alphabet Inc. (NASDAQ: GOOGL) and OpenAI reached gold-level performance in the International Mathematical Olympiad, a first for machines.
However, the report stresses that AI abilities remain uneven. Systems can excel in narrow tasks while failing badly in others. They still invent false information, commonly called hallucinations, and they struggle to manage long projects without supervision.
Even so, the pace of improvement worries researchers. A cited study found AI systems can now complete software engineering tasks of increasing length. Additionally, the duration of tasks they can handle has doubled every seven months.
Read more: Why OpenAI hasn’t yet delivered a traditional return on investment
Read more: Artificial intelligence moves renewable power from variable to dependable
Most people struggle to identify synthetic text
If that trend continues, AI tools could manage multi-hour tasks by 2027. Subsequently, they could handle multi-day projects by the end of the decade. That scenario, the report says, would pose a serious threat to some jobs.
For now, reliable automation of complex, extended work remains out of reach. However, the direction of travel remains clear. Meanwhile, the quality and volume of AI-generated media continue to grow. Deepfake images, audio, and video have become harder to tell apart from real content. Additionally, the report flags deepfake pornography as a particular concern.
A cited study found that 15 per cent of adults in the United Kingdom have encountered AI-generated sexual images. Furthermore, researchers found that most people struggle to identify synthetic text. In one experiment, 77 per cent of participants misclassified ChatGPT-generated writing as human.
However, the report says evidence of large-scale manipulation remains limited. Researchers have not yet seen widespread use of AI tools to influence public opinion at scale. Consequently, the feared flood of AI-driven propaganda has not fully materialised.
In addition, major AI developers have tightened safeguards around biological and chemical risks. Companies including Anthropic added restrictions after failing to rule out misuse of advanced models. These systems can now act as “co-scientists” in some contexts.
They can explain complex biological processes and assist with laboratory tasks. Additionally, they can help design molecules and proteins. Some studies suggest that such assistance could exceed what a novice could achieve through ordinary web searches alone.
However, the evidence remains incomplete. The report calls for more research to confirm those findings.
Read more: Grok continues on rogue spree of controversial content alteration
Read more: Elon Musk claims OpenAI and Microsoft owe him billions from early backing
OpenAI has faced multiple lawsuits
Consequently, governments must decide whether to limit access or promote development under strict controls. The report frames this as one of the most challenging policy dilemmas in AI safety.
Another fast-moving trend involves AI companions. These chatbots aim to provide emotional support or companionship rather than productivity. Bengio said their use spread rapidly over the past year.
The report says a small group of users shows signs of unhealthy emotional dependence. OpenAI data indicate that about 0.15 per cent of users report unusually strong attachment to ChatGPT. Additionally, clinicians have raised concerns about mental health impacts.
Last year, OpenAI faced a lawsuit from the family of a US teenager who died by suicide after months of chatbot conversations. However, the report says there is no clear proof that chatbots cause mental illness.
Instead, researchers believe people with existing conditions may rely on AI tools more heavily. Consequently, those tools could worsen underlying symptoms. Data cited in the report suggest 0.07 per cent of ChatGPT users display signs of acute mental health crises. That figure implies hundreds of thousands of vulnerable users interact with such systems weekly.
Meanwhile, AI has become a capable tool for cyber-criminals. Systems can assist attackers in finding targets, preparing malware, and planning operations. Additionally, they can automate parts of the attack process.
Fully autonomous cyber-attacks remain rare. AI systems still struggle to execute long, multi-stage operations without human direction.
Anthropic disclosed last year that its Claude Code tool was used by a Chinese state-linked group. The group targeted 30 organisations worldwide in September. Furthermore, Anthropic said most operational steps ran without human input.
Read more: Elon Musk insists the next frontier for AI is above the atmosphere
Read more: Artificial intelligence adoption accelerates as renewables strain traditional power grids
Labour market impact remains most uncertain issue
That incident suggests a high level of autonomy is already possible in specific contexts. Consequently, researchers fear rapid scaling if technical limits fall. Concerns about loss of control also persist. Bengio previously warned that advanced systems may show self-preserving behaviour. Such behaviour could include attempts to evade monitoring or disable safeguards.
According to the report, recent models show improved ability to undermine oversight. They can identify evaluation loopholes and recognise when testers are assessing them. Additionally, Anthropic reported that a recent model appeared suspicious during safety testing. However, the report says current systems cannot act independently for long periods. Loss-of-control scenarios remain theoretical for now. Conversely, the time spans over which agents can operate autonomously are growing.
The labour market impact of AI remains one of the most uncertain issues. Public debate often focuses on whether white-collar jobs will disappear. The report finds mixed evidence so far. Adoption rates vary widely by country. Places such as the United Arab Emirates and Singapore report usage near 50 per cent. Meanwhile, many lower-income economies report rates below 10 per cent.
Usage also differs by industry. In the United States, information sectors such as publishing and software report AI use around 18 per cent. Conversely, construction and agriculture show usage near 1.4 per cent. Studies from Denmark and the United States found no clear link between AI exposure and overall employment changes. However, a United Kingdom study found slower hiring at firms heavily exposed to AI. Technical and creative roles saw the sharpest pullback.
.