Close Menu
Earth & BeyondEarth & Beyond

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    Maya Hawke Announces New Album and Spring Tour

    Transfer rumors, news: Bruno Fernandes plans Man United future talks

    Critical Role Campaign 4 using West Marches style keeps things fresh

    Facebook X (Twitter) Instagram
    Earth & BeyondEarth & Beyond
    YouTube
    Subscribe
    • Home
    • Business
    • Entertainment
    • Gaming
    • Health
    • Lifestyle
    • Sports
    • Technology
    • Trending & Viral News
    Earth & BeyondEarth & Beyond
    Subscribe
    You are at:Home»Technology»Let 2026 be the year the world comes together for AI safety
    Technology

    Let 2026 be the year the world comes together for AI safety

    Earth & BeyondBy Earth & BeyondJanuary 5, 2026005 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Email
    Let 2026 be the year the world comes together for AI safety
    Share
    Facebook Twitter LinkedIn Pinterest Email

    Two children sitting on a carpet looking at a tablet device

    AI technologies need to be safe and transparent. There are few, if any, benefits from being outside efforts to achieve this. Credit: George Chan/Getty

    You don’t need to be an oracle to know that the coming year will see further advances in artificial intelligence, as updated and new models, publications and patents continue their inexorable rise. If current trends are a reliable guide, many countries will also be enacting more AI-related laws and regulations. In 2023, at least 30 such laws were passed around the world, according to the Artificial Intelligence Index Report 2025, produced by researchers at Stanford University in California. The following year saw another 40.

    Over the past couple of years, AI lawmaking has been busiest in the East Asia and Pacific region, in Europe and in individual US states. Between them, US states passed 82 AI-related bills in 2024. But there are some notable cold spots, too: there has been relatively little activity in low and lower-middle-income countries (see ‘AI policy trends’). Meanwhile, the US federal government is bucking the trend by cancelling AI policy work and challenging state-level AI laws.

    AI policy trends. Line chart showing the proportion of developed, developing and least developed countries with national AI strategies from 2017 to 2023. A greater proportion of developed countries have AI policies compared with developing countries with the least developed countries (LDCs) lagging farthest behind.

    Source: UNCTAD based on data from the 2024 AI Index report.

    This must be the year that more lower-income countries start regulating AI technologies, and that the United States is persuaded of the dangers of its approach. The country is one of the biggest markets for AI technologies, and people around the world are using models developed mainly by US companies. All nations need AI laws and policies, regardless of their position on the spectrum of producers and consumers. It’s impossible to imagine the technologies used in energy, food production, pharmaceuticals or communications being outside the ambit of safety regulation. The same should be true of AI.

    There is a growing international consensus. The authorities in China, for example, are taking AI regulation extremely seriously, as are those of many European countries. Most of the rules of the European Union’s AI Act are expected to come into force in August. In 2024, the African Union published continent-wide guidance for AI policymaking. There are also moves to establish a global organization for cooperation on AI, possibly through the United Nations.

    A wide spectrum of national and regional laws and regulations are in place or under development. Some countries, for example, are looking to ban ‘deepfake’ videos. This should be a universal goal. Companies should also provide details of the data used to train models, and need to ensure that copyright is respected in the training process. The overriding ambition must be to achieve regulations similar to those governing other general-purpose technologies. AI developers — most of which are companies (see ‘Model industry’) — need to transparently explain how their products work, demonstrate that their models have been produced through legal means, and show that the technology is safe and that there is accountability for risks and harm. Transparency is also needed from researchers, more of whom need to publish their models in the peer-reviewed literature.

    Model industry. Line chart showing the number notable AI models produced by industry, industry and academia, industry and government and academia form 2003 to 2005. Industry dominates the development of AI models. The top-three contributors from 2014 to 2024 have been Google (187 notable models), Meta (82) and Microsoft (39).

    Source: 2025 Stanford AI Index report.

    According to UNCTAD, the UN trade-policy agency for low- and middle-income countries, two-thirds of high-income countries and 30% of middle-income countries had AI policies or strategies in place at the end of 2023, but little more than 10% of the lowest-income countries did. These nations need to be supported in their AI-regulatory efforts.

    There is also a need to engage with the United States. On taking office, President Donald Trump cancelled a programme set up by the previous administration through which the National Institute of Standards and Technology had started to scope out AI standards with technology companies. In December, an executive order was issued forbidding state laws that conflict with White House AI policy.

    The leaders of some technology companies seem to be satisfied with this direction of travel. But others know that good regulations give companies consistency in standards and allow them to plan predictably for the long term. Light-touch regulation, or no regulation at all, is in neither their nor their customers’ interests.

    Civil servants working at regulatory agencies know that better regulation is needed, both to protect people, particularly children, from harm and to reassure consumers. They can see that there is considerable public anxiety about the risks of AI, fuelled, in part, by companies’ stated ambition to work towards artificial general intelligence. And they can hear the voices of those in the AI research community, including some of the field’s pioneers, who are warning of possible existential risks from uncontrollable AI.

    Officials in the Trump administration have said that regulation will risk the United States losing the AI race with China. But, as Nature’s news team has reported (Nature 648, 503–505; 2025), China is exploring an alternative path to innovation. Its AI companies are creating innovative products, using technologies that are more open than those of US counterparts, and working to nationwide regulations that require more disclosure.

    Technology companies know that their business models rely on public cooperation, particularly when it comes to access to data. That cooperation will evaporate if people lose confidence that their data are safe and being used responsibly, or learn that AI products are harming people.

    AI is potentially a transformative technology. However, we don’t know how that will manifest, or what impact it will have. Many countries are rightly being cautious and assessing risks, but more coherence is needed in policymaking. Nations should work together to design policies that not only enable development, but also incorporate guardrails. Let 2026 be the year everyone agrees on that.

    safety World Year
    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleBitcoin eyes longest daily winning streak in 3 months
    Next Article Venezuela live updates: interim president offers to ‘collaborate’ with US after Trump warns of further strikes | Venezuela
    Earth & Beyond
    • Website

    Related Posts

    Smoke Rises Over Big Cypress National Preserve

    March 3, 2026

    What Is That Mysterious Metallic Device US Chief Design Officer Joe Gebbia Is Using?

    March 3, 2026

    Middle East crisis: Will Iran play at World Cup co-hosted by USA? How conflict affects sport | Football News

    March 3, 2026
    Leave A Reply Cancel Reply

    Latest Post

    If you do 5 things, you’re more indecisive than most—what to do instead

    UK ministers launch investigation into blaze that shut Heathrow

    The SEC Resets Its Crypto Relationship

    How MLB plans to grow Ohtani, Dodger fandom in Japan into billions for league

    Stay In Touch
    • YouTube
    Latest Reviews

    Smoke Rises Over Big Cypress National Preserve

    By Earth & BeyondMarch 3, 2026

    What Is That Mysterious Metallic Device US Chief Design Officer Joe Gebbia Is Using?

    By Earth & BeyondMarch 3, 2026

    Middle East crisis: Will Iran play at World Cup co-hosted by USA? How conflict affects sport | Football News

    By Earth & BeyondMarch 3, 2026

    Subscribe to Updates

    Get the latest tech news from FooBar about tech, design and biz.

    Most Popular

    Blackpink Share New Song “Jump” Amid Deadline World Tour: Watch the Video

    July 13, 202544 Views

    Bitcoin in the bush – crypto mining brings power to rural areas

    March 25, 202513 Views

    Honor of Kings breaks esports attendance Guinness World Record 

    November 10, 202511 Views
    Our Picks

    Maya Hawke Announces New Album and Spring Tour

    Transfer rumors, news: Bruno Fernandes plans Man United future talks

    Critical Role Campaign 4 using West Marches style keeps things fresh

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    © 2026 Earth & Beyond.
    • About Us
    • Contact Us
    • Privacy Policy
    • Terms and Conditions
    • Disclaimer

    Type above and press Enter to search. Press Esc to cancel.

    Newsletter Signup

    Subscribe to our weekly newsletter below and never miss the latest product or an exclusive offer.

    Enter your email address

    Thanks, I’m not interested