- Australia’s December 2025 National AI Plan abandoned the 10 mandatory guardrails for high-risk AI that former Industry Minister Ed Husic had proposed in September 2024.
- The government shifted to a “technology-neutral” approach that relies on updating existing laws — the Privacy Act, Consumer Law, and sector-specific regulations — rather than introducing AI-specific legislation.
- The Productivity Commission estimated AI could add AU$116 billion to Australia’s economy over the next decade, strengthening the case against prescriptive regulation.
- A new AI Safety Institute launched in early 2026 with AU$29.9 million in funding to test AI systems and recommend targeted reforms where existing laws fall short.
What Happened
On December 2, 2025, Australia’s federal government released its National AI Plan, replacing a September 2024 proposal for mandatory AI guardrails with a voluntary, industry-led approach. The plan marked a significant policy reversal from the framework outlined by then-Industry Minister Ed Husic.
Husic’s original proposal identified 10 mandatory guardrails covering accountability, risk management, data governance, testing protocols, human oversight, transparency, contestability, supply chain visibility, record keeping, and conformity assessments. These requirements would have applied to high-risk AI deployments in sectors such as healthcare, law enforcement, and government services, as well as to all general-purpose AI models regardless of their intended application.
The National AI Plan discarded this entire framework in favor of what the government described as a “regulate as necessary but as little as possible” philosophy.
Why It Matters
The reversal reflects a broader tension between economic opportunity and regulatory caution that governments worldwide are navigating. Three factors drove the shift in Australia’s position.
First, the Productivity Commission found that AI could contribute AU$116 billion — roughly AU$4,400 per capita — to Australia’s economy over the next decade. The Commission’s analysis concluded that mandatory rules could stifle innovation and weaken international competitiveness at a critical moment for AI adoption.
Second, lobbying from the Digital Industry Group (DIGI), which represents Apple, Google, Meta, and Microsoft in Australia, argued that existing laws already provided adequate safeguards and that AI-specific regulation would create compliance burdens without proportionate safety benefits.
Third, a change in ministerial leadership reshaped the policy direction. Minister Tim Ayres, who succeeded Husic, aligned more closely with industry arguments for a light-touch approach. Husic had argued that patchwork approaches create unpredictability and regulatory gaps, advocating instead for a comprehensive AI Act modeled on the EU framework.
Technical Details
Rather than creating new AI-specific obligations, the National AI Plan directs existing regulators to identify and close gaps in their current legislative frameworks. The Privacy Act, Australian Consumer Law, workplace health and safety regulations, and sector-specific rules in healthcare, finance, and telecommunications are expected to serve as the primary regulatory tools for governing AI systems.
The government established an AI Safety Institute in early 2026, funded with AU$29.9 million. The Institute is tasked with testing AI systems, conducting risk assessments, and performing regulatory gap analyses using a structured methodology to recommend targeted reforms only where existing laws prove demonstrably insufficient.
The Plan also includes commitments to develop voluntary AI standards in collaboration with Standards Australia and to participate in international AI governance forums including the OECD AI Policy Observatory. However, none of these measures carry binding obligations, enforcement mechanisms, or compliance deadlines for AI operators.
Who’s Affected
AI developers and deployers operating in Australia face a substantially lighter compliance environment than their counterparts in the EU or South Korea. Companies that had begun preparing for the mandatory guardrails framework can set aside those specific compliance programs, though they remain subject to existing sectoral regulations.
Civil society organizations and AI ethics researchers have criticized the reversal. Consumer advocacy groups argue that voluntary measures are insufficient for high-risk applications such as automated decision-making in government services, algorithmic hiring tools, and insurance underwriting systems where errors can cause significant individual harm.
What’s Next
The AI Safety Institute will publish its first regulatory gap analysis in 2026, which could recommend targeted legislative changes in sectors where existing laws demonstrably fail to address AI-specific risks. Australia’s approach remains subject to revision — a change in government or a high-profile AI incident could prompt reconsideration of mandatory guardrails, particularly if major trading partners such as the EU or UK impose cross-border compliance requirements that affect Australian businesses.
Related Reading
- White House Releases National Policy Framework for AI, Setting Federal Guardrails
- UK AI Policy: Pro-Innovation Approach With AI Bill Expected in 2026
- Decoding the 2026 White House AI Blueprint: Federal AI Policy Takes Shape
- Pentagon Labels Anthropic a Supply-Chain Risk Over Refusal to Remove AI Safety Guardrails
- California sets its own AI rules for state contractors, pushing back against federal policy