When Autonomous AI Tests the Limits of Oversight
At India’s AI Impact Summit, safety experts warned that compounding AI risks, from manipulation to cyber misuse, demand sharper evaluation and more precise guardrails.
Topics
News
- India Plugs Into Pax Silica as AI Supply Chains Realign
- India AI Impact Summit 2026: Did India’s Tech Push Just Take a Hit?
- When Autonomous AI Tests the Limits of Oversight
- UN Bets on Science to Steady AI Governance
- Legacy IT Leaders Say India’s AI Future Lies in Talent, IP
- PM Calls for Human-Centric AI at India Summit
Autonomous AI systems are advancing faster than the guardrails designed to contain them, prompting safety experts at the India AI Impact Summit to warn that governments may be unprepared for machines increasingly able to act on their own.
The warnings came during a discussion of the 2026 International AI Safety Report, a global assessment authored by leading researchers and policy advisers that examines frontier AI risks, systemic vulnerabilities and evaluation standards.
Yoshua Bengio, a Turing Award winner and one of the report’s chairs, said the most consequential shift over the past year is the rise of agentic systems, or tools that no longer merely respond to prompts but can plan, execute and operate with credentials and internet access.
“Having AIs that are more autonomous means less oversight,” Bengio said.
Unlike chatbots, where “the human is in the loop,” agents can “work on a problem for you for hours, days.”
That autonomy, he added, demands far greater reliability before deployment. “Otherwise users… can’t trust this technology with credentials that we might give them.”
The concern extends beyond individual systems.
Once agents are “let out… into the world” and begin interacting with one another, their behavior becomes harder to anticipate. “It’s a bit early days, but what we’re seeing is a bit concerning,” he said.
Bengio cautioned against framing the issue as a single future “AGI moment.”
Current models display uneven capabilities, and are powerful in narrow domains but weak in others. That asymmetry, he argued, requires more granular risk evaluation, assessing capability and intent skill by skill rather than waiting for a hypothetical breakthrough.
For policymakers, though, the challenge is not abstract. Josephine Teo, Singapore’s Minister for Digital Development and Information, said safety is the foundation for adoption.
Singapore does not manufacture aircraft, she noted, yet strict aviation standards underpin its position as a global air hub.
“If we didn’t have all of these elements in place, it’s very hard to see how you can have a thriving air hub.” The same logic applies to AI.
Guardrails, however, must be precise. Overbroad rules risk slowing innovation without delivering real protection.
“What we could end up with is a situation where we have given a false promise to our citizens,” Teo said.
Singapore has introduced statutory obligations requiring platforms to remove harmful AI-generated images once notified, particularly those targeting women and children.
She described AI as simultaneously “a threat,” “a target” and a tool. It can generate abusive content, enable cyber attacks or itself be exploited, especially in multi-agent systems where risks can “easily go out of hand.” Regional cooperation, particularly within ASEAN, is therefore critical.
Alondra Nelson of the Institute for Advanced Study, a senior adviser to the report, said its objective is not to prescribe policy but to clarify evidence in a field often shaped by headlines and speculation.
“We don’t have globally the kind of horizon of information that we really need in the policy space to make good policy decisions,” she said. The report aims to establish “a ground truth as a global community about the risks.”
Its scope extends beyond catastrophic scenarios. A section on systemic risk examines how pressures compound across domains.
“It is not the individual risks,” Nelson said. “It is the compounding of those risks together.”
Erosion of human autonomy, manipulation, job displacement and public anxiety may each appear manageable in isolation. Together, they are “not healthy for democracy.”
Adam Beaumont, director of the United Kingdom’s AI Security Institute, focused on implementation. His organization conducts pre- and post-deployment testing of frontier models, including red teaming against cyber vulnerabilities. As capabilities evolve, evaluation methods must keep pace.
“If you’re evaluating something, be really clear what it is you are trying to measure,” he said. The institute has open-sourced tools and research to strengthen standards and called for “a wide ecosystem of third-party evaluators” to bring independence and scientific rigor to model assessment.
Across the panel there was agreement that governments, industry and researchers all share responsibility.
But Bengio returned to a final principle: restraint. Researchers, he said, “should not make a claim that could be false,” particularly when policymakers may rely on it. “Each of us can be personally biased… it’s human.”

