Reviewing politics
and culture since 1913

AI regulation does not stifle innovation

Achieving balance between human potential and machines isn’t just possible – it’s necessary.

By Tim Clement-Jones

Ever since co-founding the All-Party Parliamentary Group on AI nine years ago, still ably administered by the Big Innovation Centre, I’ve been deeply involved in debating and advising on the implications of artificial intelligence. My optimism about AI’s potential remains strong – from helping identify new Parkinson’s treatments to DeepMind’s protein structure predictions that could transform drug discovery and personalised medicine.

Yet this technology is unlike anything we’ve seen before. It’s potentially more autonomous, with greater impact on human creativity and employment, and more opaque in its decision-making processes.

The conventional wisdom that regulation stifles innovation needs turning on its head. As AI becomes more powerful and pervasive, appropriate regulation isn’t just about restricting harmful practices – it’s key to driving widespread adoption and sustainable growth. Many potential AI adopters are hesitating not due to technological limitations but uncertainties about liability, ethical boundaries and public acceptance. Clear regulatory frameworks addressing algorithmic bias, data privacy and decision transparency can actually accelerate adoption by providing clarity and confidence.

Different jurisdictions are adopting varied approaches. The European Union’s AI Act, with its risk-based framework, started coming into effect this year. Singapore has established comprehensive AI governance through its model AI governance framework. Even China regulates public-facing generative AI models with fairly heavy inspection regimes.

Subscribe to the New Statesman today for only £1 a week.

The UK’s approach has been more cautious. The previous government held the AI Safety Summit at Bletchley Park and established the AI Safety Institute (now inexplicably renamed the AI Security Institute), but with no regulatory teeth. The current government has committed to binding regulation for companies developing the most powerful AI models, though progress remains slower than hoped. Notably, 60 countries – including Saudi Arabia and the UAE, but not Britain or the US – signed the Paris AI Action Summit declaration in February this year, committing to ensuring AI is “open, inclusive, transparent, ethical, safe, secure and trustworthy”.

Select and enter your email address Your weekly guide to the best writing on ideas, politics, books and culture every Saturday. The best way to sign up for The Saturday Read is via saturdayread.substack.com The New Statesman's quick and essential guide to the news and politics of the day. The best way to sign up for Morning Call is via morningcall.substack.com
Visit our privacy Policy for more information about our services, how Progressive Media Investments may use, process and share your personal data, including information on your rights in respect of your personal data and how you can unsubscribe from future marketing communications.
THANK YOU

Several critical issues demand urgent attention. Intellectual property: the use of copyrighted material for training large language models without licensing has sparked substantial litigation and, in the UK, unprecedented parliamentary debate. Governments need to act decisively to ensure creative works aren’t ingested into generative AI models without return to rights-holders, with transparency duties on developers.

Digital citizenship: we must equip citizens for the AI age, ensuring they understand how their data is used and AI’s ethical implications. Beyond the UAE, Finland and Estonia, few governments are taking this seriously enough.

International convergence: despite differing regulatory regimes, we need developers to collaborate and commercialise innovations globally while ensuring consumer trust in common international ethical and safety standards.

Well-designed regulation can be a catalyst for AI adoption and innovation. Just as environmental regulations spurred cleaner technologies, AI regulations focusing on explainability and fairness could push developers toward more sophisticated, responsible systems.

The goal isn’t whether to regulate AI, but how to regulate it promoting both innovation and responsibility. We need principles-based rather than overly prescriptive regulation, assessing risk and emphasising transparency and accountability without stifling creativity.

Achieving the balance between human potential and machine innovation isn’t just possible – it’s necessary as we step into an increasingly AI-driven world. That’s what we must make a reality.

This article first appeared in our Spotlight on Technology supplement, of 13 June 2025.

Content from our partners
Lives stuck in limbo
Rare Diseases: Closing the translation gap
Clinical leadership can drive better rare disease care