Authored by Joseph Lord via The Epoch Times (emphasis ZeroHedge)
Former Google CEO Eric Schmidt said that autonomous artificial intelligence (AI) is coming—and that it could pose an existential threat to humanity.
“We’re soon going to be able to have computers running on their own, deciding what they want to do,” Schmidt, who has long raised alarm about both the dangers and the benefits AI poses to humanity, said during a Dec. 15 appearance on ABC’s “This Week.”
“That’s a dangerous point: When the system can self improve, we need to seriously think about unplugging it,” Schmidt said.
Schmidt is far from the first tech leader to raise these concerns.
The rise of consumer AI products like ChatGPT has been unprecedented in the past two years, with major improvements to the language-based model. Other AI models have become increasingly adept at creating visual art, photographs, and full-length videos that are nearly indistinguishable from reality in many cases.
For some, the technology calls to mind the “Terminator” series, which centers on a dystopian future where AI takes over the planet, leading to apocalyptic results.
For all the fears that ChatGPT and similar platforms have raised, consumer AI services available today still fall into a category experts would consider “dumb AI.” These AI are trained on a massive set of data, but lack consciousness, sentience, or the ability to behave autonomously.
Schmidt and other experts are not particularly worried about these systems.
Rather, they’re concerned about more advanced AI, known in the tech world as “artificial general intelligence” (AGI), describing far more complex AI systems that could have sentience and, by extension, could develop conscious motives independent from and potentially dangerous to human interests.
Schmidt said no such systems exist today yet, and we’re rapidly moving toward a new, in-between type of AI: one lacking the sentience that would define an AGI, and still able to act autonomously in fields like research and weaponry.
“I’ve done this for 50 years. I’ve never seen innovation at this scale,” Schmidt said of the rapid developments in AI complexity.
Schmidt said that more developed AI would have many benefits to humanity—and could have just as many “bad things like weapons and cyber attacks.”
The Challenge
The challenge, Schmidt said, is multifaceted.
At a core level, he repeated a common sentiment among tech leaders: if autonomous AGI-like systems are inevitable, it will require massive cooperation among both corporate interests and governments internationally to avoid potentially devastating consequences.
That’s easier said than done. AI provides U.S. competitors like China, Russia, and Iran with a potential leg-up over the United States that would be difficult to achieve otherwise.
Within the tech industry as well, there’s currently massive competition among major corporations—Google, Microsoft, and others—to outcompete rivals, a situation that raises inherent risks of improper security protocols for dealing with a rogue AI, Schmidt said.
“The competition is so fierce, there’s a concern that one of the companies will decide to omit the [safety] steps and then somehow release something that really does some harm,” Schmidt said. Such harms would only become evident after the fact, he said.
The challenge is greater on the international stage, where adversarial nations are likely to see the new technology as revolutionary for their efforts to challenge U.S. global hegemony and expand their own influence.
“The Chinese are clever, and they understand the power of a new kind of intelligence for their industrial might, their military might, and their surveillance system,” Schmidt said.
That’s a bit of a catch-22 for U.S. leaders in the field, who find themselves forced to balance existential concerns for humanity with the potential for the United States to fall behind its adversaries, which could be catastrophic to global stability.
In the worst case, such systems could be used to engineer crippling biological and nuclear weapons, particularly by terror groups like ISIS.
For this reason, Schmidt said, it’s absolutely crucial that the United States continue to innovate in the field, and ultimately maintain technological dominance over China and other adversarial states and groups.
Industry Leaders Demand Regulation
Regulation of the field remains insufficient, Schmidt said. But he expects that governments’ focus on enhancing safeguards around the tech will accelerate dramatically in the coming years.
Asked by anchor George Stephanopoulos if governments were doing enough to regulate it, Schmidt replied, “Not yet, but they will, because they’ll have to.”
Despite some initial interest in the field—hearings, legislative proposals, and other initiatives—emerging during the current 118th Congress, this session seems to be on track to end without any major legislation related to AI.
President-elect Donald Trump, for his part, has warned of the vast risks posed by AI, saying during an appearance on Logan Paul’s “Impaulsive” podcast that it’s “really powerful stuff.”
He also spoke of the need to maintain competitiveness with adversaries.
“It brings with it difficulty, but we have to be at the forefront,” Trump said. “It’s going to happen, and if it’s going to happen, we have to take the lead over China. China’s the primary threat.”
Schmidt’s takes on both the benefits and the challenges of the technology aligns with other industry reactions.
In June 2024, OpenAI and Google employees signed a letter warning of “serious risks” posed by AI, and calling for greater government oversight of the field.
Elon Musk has put forward similar warnings, saying that Google is seeking to create a “digital God” through its DeepMind AI program.
In August, these concerns intensified after it was discovered that an AI took autonomous action to avoid being shut down—raising fears that humanity is already losing control over its creation as governments remain inactive.
Sourced from ZeroHedge
••••
The Liberty Beacon Project is now expanding at a near exponential rate, and for this we are grateful and excited! But we must also be practical. For 7 years we have not asked for any donations, and have built this project with our own funds as we grew. We are now experiencing ever increasing growing pains due to the large number of websites and projects we represent. So we have just installed donation buttons on our websites and ask that you consider this when you visit them. Nothing is too small. We thank you for all your support and your considerations … (TLB)
••••
Comment Policy: As a privately owned web site, we reserve the right to remove comments that contain spam, advertising, vulgarity, threats of violence, racism, or personal/abusive attacks on other users. This also applies to trolling, the use of more than one alias, or just intentional mischief. Enforcement of this policy is at the discretion of this websites administrators. Repeat offenders may be blocked or permanently banned without prior warning.
••••
Disclaimer: TLB websites contain copyrighted material the use of which has not always been specifically authorized by the copyright owner. We are making such material available to our readers under the provisions of “fair use” in an effort to advance a better understanding of political, health, economic and social issues. The material on this site is distributed without profit to those who have expressed a prior interest in receiving it for research and educational purposes. If you wish to use copyrighted material for purposes other than “fair use” you must request permission from the copyright owner.
••••
Disclaimer: The information and opinions shared are for informational purposes only including, but not limited to, text, graphics, images and other material are not intended as medical advice or instruction. Nothing mentioned is intended to be a substitute for professional medical advice, diagnosis or treatment.
Disclaimer: The views and opinions expressed in this article are those of the author and do not necessarily reflect the official policy or position of The Liberty Beacon Project.






Leave a Reply