If the early days of AI were marked by a laissez-faire approach, 2026 will be the year of intense battles over regulation and who, if anyone, gets to control it.
I think back to the early Internet, specifically the 1992-1994 period, when high-speed connections were sparse and the web still mostly described something spiders built.
Does hands off work?
It’s the kind of get-out-of-the-way regulation that the current US Administration is proposing, though the reasoning is less about unfettered access and more out of concern for global competition.
AI is not necessarily about the people. It’s likely that long-term concern revolves around the coming AI arms race; the country with the most powerful AI may control information and attack vectors that affect access to information and critical infrastructure (not to mention who controls the robot army).
Caught in this mix are, in the US at least, 50 state governments that are probably more interested in regulating AI to protect their citizens from things like AI bias, AI misinformation, and even AI access to systems that would be better handled by humans. There’s also some concern about protecting jobs from AI.
From the time the Internet started functioning (in the late 1980s) to the introduction of the World Wide Web (arguably 1993), it took years for even this milquetoast Telecom Act to arrive, and decades more for anyone to realize that we might also need to look at information regulation. While the EU has moved quickly to enact regulations that protect users and their data, the US has, aside from the aging Child Online Protection Act (COPA), virtually no federal Internet regulations and few toothy state-level controls.
The rapid, unprecedented growth of AI (see “AI Time“) and its spread into every sector of our lives has, though, inspired some panicky work at the US state level and, perhaps, equally panicky pushback at the US federal level. To wit, the White House’s latest executive order seeks to eliminate and block any and all state-level AI regulation.
Executive inaction
On some level, though it pains me to say this, I agree with the White House: the US cannot afford to fall behind its chief global adversary, China, in the AI race. We’re currently ahead, but that’s changing fast. One report noted that China’s “open-source LLMs’ global share” had, in just a year, grown from 1.2% to just 30%.
The threat of Western tools being overrun by China is not spin. It’s a real possibility. An equally real possibility, though, is that no one fully understands the potential harm of vast, global use of AI. Having no regulatory framework means that we’re basically throwing up our hands and seeing what happens. The US White House’s light-touch approach is not going to be enough.
Worse yet, its insistence on focusing on anti-bias efforts in AI will prove a harmful distraction. Withholding funds for states that don’t comply will only serve to slow down AI development and create even more unreliable models.
AI is bigger than all of us
What’s needed is a compromise regulation that consistently works across states and even reaches our global partners, say, in the EU. It requires decisions about what’s best for humanity and to ensure that we maintain a balance of control between us and China. It means using an examination of our recent Information Age past and the growth of the Internet to inform our approach to new information and intelligence technology that is moving at triple time. It means avoiding past mistakes and getting ahead of the unintended consequences of an AI future.
The squabbling between state and federal officials is unhelpful. Just as the local governments’ trying to regulate with a patchwork of their own legislation is sure to fail, not because the regulation is bad, but because it won’t hold up with technology that effortlessly crosses every known border.
The lack of rational, coherent discussion on this topic is not just frustrating; it’s dangerous. AI development will continue apace, no matter what we do. If the world doesn’t wrap its arms around thoughtful regulation, including checks, balances, and control, AI will end up no less dangerous than unregulated nuclear power.

The best business laptops for all budgets
Follow TechRadar on Google News and add us as a preferred source to get our expert news, reviews, and opinion in your feeds. Make sure to click the Follow button!
And of course you can also follow TechRadar on TikTok for news, reviews, unboxings in video form, and get regular updates from us on WhatsApp too.
https://cdn.mos.cms.futurecdn.net/V8c4bCpcf4yGamvzKBQHn-1920-80.jpg
Source link
lance.ulanoff@futurenet.com (Lance Ulanoff)




