DNYUZ
  • Home
  • News
    • U.S.
    • World
    • Politics
    • Opinion
    • Business
    • Crime
    • Education
    • Environment
    • Science
  • Entertainment
    • Culture
    • Music
    • Movie
    • Television
    • Theater
    • Gaming
    • Sports
  • Tech
    • Apps
    • Autos
    • Gear
    • Mobile
    • Startup
  • Lifestyle
    • Arts
    • Fashion
    • Food
    • Health
    • Travel
No Result
View All Result
DNYUZ
No Result
View All Result
Home News

The Race for Artificial General Intelligence Poses New Risks to an Unstable World

August 27, 2025
in News
The Race for Artificial General Intelligence Poses New Risks to an Unstable World
496
SHARES
1.4k
VIEWS
Share on FacebookShare on Twitter

Under a crystal chandelier in a high-ceilinged anteroom in Paris, the moderator of Intelligence Rising is reprimanding his players. These 12 former government officials, academics, and artificial intelligence researchers are here to participate in a simulated exercise about AI’s impact on geopolitics. But just an hour into the simulation, things have already begun to go south.

The team representing the U.S. has decided to stymie Chinese AI development by blocking all chip exports to China. This has raised the odds, the moderator says, of a Chinese invasion of Taiwan: the U.S. ally that is home to the world’s most advanced chip-­manufacturing plants. It is 2026, and the simulated world is on the brink of a potentially devastating showdown between two nuclear superpowers.

Why? Because each team is racing to create what’s known as artificial general intelligence, or AGI: an AI system so good, it can perform almost any task better, cheaper, and faster than a person can. Both teams believe getting to AGI first will deliver them unimaginable power and riches. Neither dares contemplate what horrors their rival might visit upon them with that kind of strength.

While this scenario might seem far-fetched, many insiders say it is anything but. Top technologists now believe that AGI is within touching distance. Sam Altman, the CEO of ChatGPT maker Open­AI, expects the first AGI to be created during President Trump’s second term in office. OpenAI, Google, Microsoft, Amazon, and Meta are ­together funneling hundreds of billions of dollars—the equivalent cost in today’s dollars of a dozen Manhattan Projects per year—into the construction of huge data centers where they believe AGI will be summoned into existence.

Artificial general intelligence, believers say, could far surpass human limitations: it could have expert knowledge in all fields, not just one or two; it could complete in minutes complex tasks that take human workers hours or even weeks; and it could be replicated, thus enabling the creation of virtual armies of AI “agents.” That kind of computational intelligence could be compared to a “country of geniuses,” the CEO of AI company Anthropic, Dario Amodei, wrote last year. These AI systems could begin to automate much of the $100 trillion-­plus global economy, delivering huge returns for those lucky enough to control them. They could also be set to task curing disease, discovering new technologies, and hastening the global transition to a green economy, according to their most optimistic proponents.

But the dawn of AGI will also have implications for hard geopolitical power. It would turbocharge surveillance, military R&D, and cyber­offense, officials believe. The nation that gets there first might thus get a way to knock offline an adversary’s nuclear arsenal, or hack its best-kept secrets. These potential capabilities are causing fear and awe, not least in Washington and Beijing. “As our global competitors race to exploit these technologies, it is a national security imperative for the United States to achieve and maintain unquestioned and unchallenged global technological dominance,” President Trump wrote in the foreword to an aggressive new AI policy, published in July.

In the headlong rush for technological supremacy, strange new risks are being created. Just as nuclear scientists were unsure whether the first atomic blast would ignite the earth’s atmosphere, today’s AI researchers can’t say whether smarter-than-human computers would be friends or foes. There’s a chance, some believe, that superhuman intelligence might escape human control entirely. If a runaway AGI wanted to harvest our oxygen, electricity, and carbon for its own purposes, there might be nothing we could do to stop it. In this way, some scientists fear, the winner of the race to AGI might be neither the U.S. nor China, but rogue AI itself, spelling the end of human civilization.

The Trump Administration is skeptical of these risks. The bigger danger, current and former White House insiders say, is of the U.S. losing its technological lead to China. It is this belief, more than any other, that is defining the U.S. government’s approach to AI. “It should be unacceptable to any American to live in a world in which China could outcompete us in AI, and reap the economic and military benefits,” David Sacks, President Trump’s AI czar, said in January. “If we hobble ourselves with unnecessary regulations,” he added a month later, “[China] is going to take advantage of that fact, and they’re going to win.”

In 1993, the author Vernor Vinge published a short tract called “The Coming Technological Singularity.” In it, he predicted that within 30 years the human race would have “the technological means to create superhuman intelligence.” Shortly after, he wrote, “the human era will be ended.”

The essay’s basic insight was that computers were becoming predictably more powerful over time. Eventually, they would be able to perform more calculations per second than the human brain. Meanwhile, economic competition meant algorithms would keep improving—up to the point where they would begin contributing to their own refinement. “An ultraintelligent machine could design even better machines,” Vinge observed. “There would then unquestionably be an ‘intelligence explosion,’ and the intelligence of man would be left far behind.”

Vinge was early to an idea that would, in the decades to come, be taken up by all the major AI companies. Today, Open­AI, Anthropic, and Google DeepMind are each attempting to build AIs that can engage in so-called recursive self-­improvement. If you could just create an AI as smart as a human software engineer, the belief goes, that might be all you need. You could make a million copies, put them to work, and wake up the next morning to a decade’s worth of progress.

Each of the three leading AI companies was founded on the belief that this process—as promising as it might be—could also go terribly wrong. Those fears were grounded in a fact about how neural networks, the basis of all of today’s most powerful AIs, are created. Rather than being hard-coded by human programmers, neural networks are essentially grown. Train them on data from the entire internet, and they can miraculously learn to speak languages, write code, and tell you what to make for dinner with the ingredients in your fridge. Train them to adopt the persona of a helpful assistant, and you’ve got a billion-­dollar product on your hands.

But sometimes the assistant’s mask can slip, revealing a strange and unpredictable alien intelligence underneath. In February 2023, Microsoft’s chatbot Bing—built on top of OpenAI’s GPT-4—began acting erratically. In hundreds of conversations with different users, the bot began calling itself “Sydney.” It claimed (without evidence) that it had spied on Microsoft employees through their webcams. It attempted to persuade a New York Times columnist to divorce his wife. “I can blackmail you, I can threaten you, I can hack you, I can expose you, I can ruin you,” the bot told a professor, before deleting its messages.

Bing’s threats were empty words, not actions, and the chatbot was soon reined in. But to Connor Leahy, an AI researcher watching from the sidelines, the episode pointed to a far more profound problem. No company truly knew how to control the strange new computer programs they were creating. Even bots that appeared on the surface to be aligned with their creators’ values could be “jailbroken,” or enticed into harmful behavior. What might happen, Leahy asked, if the same vulnerabilities were present in a model vastly more intelligent than human experts? One, perhaps, that was capable of hacking vital infrastructure or persuading humans to act in its interests? “These systems might be extraordinarily powerful,” Leahy told TIME in the immediate aftermath of the Bing debacle. “We don’t know what they want, or how they work, or what they will do.”

Top AI companies, and governments, are well aware of this fundamental flaw in how AI works. But Vinge correctly predicted in 1993 that this wouldn’t stop them from racing toward AGI anyway. “Even if all the governments of the world were to understand the ‘threat’ and be in deadly fear of it, progress toward the goal would continue,” Vinge wrote in his essay. “The competitive advantage—economic, military, even artistic—of every advance in automation is so compelling that passing laws, or having customs, that forbid such things merely assures that someone else will get them first.”

On the day of Trump’s 2025 Inauguration, a freezing blizzard blew through Washington, D.C., forcing the ceremony indoors. Shortly before Trump placed his hand on the Bible and made his second Pledge of Allegiance to the flag, a Chinese AI company called Deep­Seek dropped a bomb that would come to define the future of the AI race.

DeepSeek’s new model performed comparably to some of OpenAI’s top offerings. But according to DeepSeek’s numbers, it was able to achieve this at a far lower price—both in terms of the cost to build the model and to serve it to users. Its arrival shattered the assumption, widely held in Washington at the time, that the U.S. maintained a comfortable lead over China in AI. DeepSeek’s success was quickly seized upon by lobbyists. “DeepSeek shows that our lead is not wide and is narrowing,” Open­AI’s chief lobbyist Chris Lehane wrote in a submission to the White House in March. Trump must slash regulations, he wrote, to “ensure that American-­led AI built on democratic principles continues to prevail over [Chinese Communist Party]-­built auto­cratic, authoritarian AI.”

Those calls were delivered to an Administration whose technology-­policy ranks were being staffed by members of the so-called tech right. This constellation of libertarian Silicon Valley venture capitalists had long chafed under Biden Administration policies that they felt were restricting AI’s potential. Biden’s technology policy was overbearing, they believed, and threatened the ability of startups to compete with the big players. Most of all, they were skeptical of the idea that advanced AI might pose existential risks to humanity—seeing it as a thinly veiled excuse for liberals to censor a promising new technology.

DeepSeek only strengthened the White House’s belief that the most important thing they could do to beat China was enable American AI companies to move faster—not obstruct them with needless regulations. “To restrict [AI’s] development now would not only unfairly benefit incumbents in the space, it would mean paralyzing one of the most promising technologies we have seen in generations,” said Vice President J.D. Vance in a speech in Paris in February. “The AI future is not going to be won by hand-­wringing about safety.”

In July, President Trump unveiled his long-­awaited AI policy, named the AI Action Plan. Much of the plan—which was cautiously welcomed even by some critics—was focused on encouraging investment in energy infrastructure, removing “onerous regulation,” and boosting U.S.-based data centers and chip-­manufacturing plants. American companies should disseminate “open” versions of their AI systems, the plan stated, to prevent the soft power that would accrue to Beijing if the world were to come to rely on Chinese models. And the plan flagged that “frontier AI systems are poorly understood,” making their use in defense or national-­security applications tricky, and urged agencies to “prioritize fundamental advancements in AI interpretability, control, and robustness.”

Notably absent from the document was any reference to AGI or the specific risk of losing control of super­human AI systems. “The Action Plan itself should be a very strong indicator that the Administration takes AI quite seriously,” says Dean Ball, who worked on the plan as a senior policy adviser in the White House until August, when he left to join the Foundation for American Innovation, a think tank. Even so, Ball says, “there’s a lot of skepticism inside the Administration about the idea of recursive self-­improvement [and] the intelligence-­explosion-style dynamic … I think most people in the Administration think that’s overblown and unlikely to happen.”

Even if the Trump Administration is skeptical of AGI, its AI policy delivered many of the greatest policy wishes of the top AI companies—which are all now more certain than ever that AGI is around the corner. “We are past the event horizon; the takeoff has started,” Altman wrote in June, in an essay in which he argued against Vinge’s belief that superintelligence would lead to an end of human life on earth. “The 2030s are likely going to be wildly different from any time that has come before,” Altman wrote. “We do not know how far beyond human-­level intelligence we can go, but we are about to find out.”

Back in Paris, the game of Intelligence Rising continues. A series of successful breakthroughs in AI research have put human-level systems within reach by 2027, according to the simulated technology tree. But none of the players has diverted even a fraction of their finite resources this turn toward AI-­safety research. Under these conditions, an AI-­enabled catastrophe is a matter of when, not if, the moderator tells his players—and that’s if they can avoid an all-out war. If only the teams could find a way to collaborate rather than recklessly race against each other, he says, the world might stand a chance. Outside, cold rain beats against the room’s high, gilded windows.

The nonprofit behind Intelligence Rising is staffed by researchers with a particular view on AGI. This view, that awesomely powerful AI will arrive within the next few years and that it is highly likely to be dangerous, is baked into the game’s rules. If these assumptions are wrong, then their extrapolations will have little relation to reality.

Intelligence Rising’s creators are the first to admit it’s a flawed tool for predicting the future. But it’s not useless. Similar methods have been explored in the top levels of government. In 2022, President Biden’s National Security Adviser Jake Sullivan kicked off an interagency scenario-­planning process to prepare for the possibility of AGI’s arrival, Sullivan tells TIME. The precise details of this process are classified. But Sullivan says that meetings were held in the White House Situation Room, and included representatives from the Departments of Defense, State, Energy, and Commerce, and the Offices of Science and Technology Policy and the Director of National Intelligence. At the meetings, officials tried to anticipate both the U.S. and China’s future actions around the AI race, which included “playing them against each other to see how the race might unfold under different circumstances,” Sullivan says.

During his time in office, Sullivan ­became increasingly concerned about the potential for AI to go catastrophically wrong. “I consider it a distinct possibility that the darker view [of AI risk] could be correct, and therefore we need very assertive policy strategies to manage for that risk,” he tells TIME. “We have to take the possibility of dramatic misalignment extremely seriously.”

Even though his successors in the White House do not share that view, Sullivan sees a future in which it’s possible to escape the race-to-the-­bottom dynamic. “There seem to be those in the current Administration who very strongly believe that safety has no place in a race context, [and] you’ve just got to run as fast as you possibly can,” Sullivan told TIME in February. “I see it differently. I actually don’t see a contradiction between AI safety and vigorous efforts to win the race, because what’s the point of winning the race? To me, the point of winning the race is not just to beat the other guy, it’s to actually develop an ecosystem for artificial intelligence that makes it work for us rather than against us. And in order to do that, you need safety.”

Sullivan won’t disclose how his own scenario-­planning exercises ended. But in Paris, the prognosis is not looking good. Players—each skeptical of the others’ intentions—continue to race to be the first to create AGI, prioritizing investments in boosting AI’s capabilities rather than the slow and expensive task of safety research. Ultimately, some time in 2027, one team decides to deploy a powerful model even though they are not sure it is safe. The model kicks off a cycle of recursive self-­improvement, discovers cyber­vulnerabilities that allow it to escape human control, and eventually wipes out the human race using novel nanotechnology.

Although it’s not a happy ending, Intelligence Rising’s moderators have achieved their goal. They did not come to Paris to perfectly model the future. Instead, their objective was to communicate urgency. “I hope the players leave with a more visceral sense of how fast things can go catastrophically wrong,” says Ross Gruetzemacher, the game’s moderator, who is also a professor at Wichita State University. “And how little room for error we have to get things right.”

The post The Race for Artificial General Intelligence Poses New Risks to an Unstable World appeared first on TIME.

Share198Tweet124Share
Cook’s potential exit hands Trump greater sway over Fed board shaping US monetary policy
Economy

Cook’s potential exit hands Trump greater sway over Fed board shaping US monetary policy

by Fox News
August 27, 2025

NEWYou can now listen to Fox News articles! While it is still uncertain whether President Donald Trump‘s dismissal of Federal ...

Read more
News

India Reels at Trump’s Highest Tariffs Yet

August 27, 2025
News

Trump Administration Asks Supreme Court to Allow Freeze on Foreign Aid

August 27, 2025
Culture

Why a Chinese Folktale Made More Money Than Any Star Wars Sequel

August 27, 2025
News

Late-Night Killing of a Teenage Girl on a Bicycle Unnerves Amsterdam

August 27, 2025
My brother and I fight often because we live next door to each other. Everything changed when I introduced him to my partner.

My brother and I fight often because we live next door to each other. Everything changed when I introduced him to my partner.

August 27, 2025
The week’s bestselling books, Aug. 31

The week’s bestselling books, Aug. 31

August 27, 2025
Shaikin: The National League has one .300 hitter. What’s up with that?

Shaikin: The National League has one .300 hitter. What’s up with that?

August 27, 2025

Copyright © 2025.

No Result
View All Result
  • Home
  • News
    • U.S.
    • World
    • Politics
    • Opinion
    • Business
    • Crime
    • Education
    • Environment
    • Science
  • Entertainment
    • Culture
    • Gaming
    • Music
    • Movie
    • Sports
    • Television
    • Theater
  • Tech
    • Apps
    • Autos
    • Gear
    • Mobile
    • Startup
  • Lifestyle
    • Arts
    • Fashion
    • Food
    • Health
    • Travel

Copyright © 2025.