Tuesday, March 10, 2026

Humanity Crossed A Threshold, And Most Of Us Scrolled Past It

  I have not updated this blog since last week on purpose. My objective here is to give food for thoughts, not to comment on current events. 

 For all I can see, the Iran war may be one of the greatest blunders in recent history. The Trump Administration, pushed by Israel, believed that after decapitation, Iran would fall and become an easy pray. The opposite happened. Senseless murder of a religious leader and true terrorism with the willful mass slaughter of little girls had the exact opposite effect of galvanizing a country around its government. But how on earth could even the most delusional people not understand that this would be the case? 

  The result now is that Trump is in a bind. Announce "Mission accomplished", end the war and he risks a fatal blow to the credibility of the US, plus of course the wrath of Netanyahu and the consequential blow-back of being compromised. Continue the war, and the relentless rise of the price of oil will plunge the world in a deep recession and compromise his chance in November. What to do? 

  He certainly would like to negotiate a ceasefire but how can you do that with people you kill as they sit as the table? The Iranians rightfully said: Enough!    

  So instead of focusing on these miserable events which are entirely due to arrogance and hubris, let's take a step back and look at what's really happening in the background.

  If the Ukraine war was the first war of the drones, the Iran war will be the first war of AI. Together, drones and AI are de-multiplying lethality but also the cost of waging war. Slowly, war is becoming almost exclusively economic. And if that's the case, then obviously the US cannot be left in charge of the world's currency. Changes are coming!

Authored by Kay Rubacek via The Epoch Times,

Something happened last week that most people scrolled past.

Two Amazon data centers in the United Arab Emirates were struck during Iran’s retaliation for U.S. military action. Another facility in Bahrain was reportedly damaged after a drone landed nearby. The earlier strikes that triggered the retaliation were said to have used AI-assisted targeting systems.

It was a brief moment in the news cycle, quickly overtaken by the next political story. But the implications are difficult to ignore.

Artificial intelligence has now crossed into active geopolitical conflict.

The infrastructure that powers the digital world—the same systems that store family photos, run businesses, and answer questions on our phones—has become strategic wartime infrastructure. Algorithms woven quietly into civilian technology are now helping guide decisions about where weapons land.

Humanity crossed a threshold, and most of us scrolled past it.

But we know from history that major technological shifts rarely announce themselves with a single dramatic moment. They appear first as signals in small news items, policy disputes, unexplained departures by insiders.

Another signal appeared almost at the same time.

The federal government recently removed the artificial intelligence systems developed by Anthropic from its networks. Shortly afterward, OpenAI stepped in with a defense agreement of its own.

The public does not know the full story behind the change. We do not know exactly what demands were made behind closed doors, what ethical guardrails were contested, or why one of the world’s leading AI companies was suddenly pushed out of federal systems.

But the episode itself is another signal.

And yet another signal has been appearing quietly inside the AI industry itself: the departure of safety researchers.

Over the past several years, numerous high-profile researchers tasked with studying the risks and safety of advanced AI systems have left their posts at leading companies and research labs. Many of these departures have come with little public explanation.

Those researchers rarely describe the internal debates they witnessed. Few are in a position to do so.

But patterns like this matter. When the people closest to a powerful technology begin stepping away quietly, it often means they have seen tensions the public has not yet been invited to examine.

History has seen moments like this before.

In the early 1940s, scientists working on what became the Manhattan Project realized they were building something unprecedented. Some raised concerns about what the technology might mean once it left the laboratory. But those debates happened largely behind closed doors. The public understood the stakes only after the technology had already been used.

Artificial intelligence may be unfolding along a similar pattern. We are seeing the signals now—researchers leaving, governments disputing ethical guardrails, and AI systems appearing inside real geopolitical conflict.

Yet the public conversation about artificial intelligence is still shaped by a set of assumptions that make these signals harder to recognize.

Misconception #1: AI Is ‘Just a Tool’

This analogy is comforting. We imagine AI the way we imagine a calculator or a word processor—machines that perform tasks efficiently while remaining firmly under human control.

Tools can become strategic assets in war. But they do not generate their own outputs in ways their creators sometimes struggle to explain, nor do they require constant negotiation over the ethical boundaries of their behavior.

Modern AI systems are not programmed line by line in the traditional sense. They are trained on vast datasets and learn patterns within that data. Their behavior emerges from statistical relationships rather than explicit instructions. AI researchers describe these systems as “grown,” not built. And that makes them fundamentally different from the tools we are used to controlling.

Misconception #2: AI Is Neutral

AI systems are trained on human-generated information. That information reflects human biases, historical conflicts, and uneven representation.

When an AI system generates an answer, it synthesizes patterns it absorbed from that material.

AI has developed fluent language skills that can create the illusion of objectivity. But confident language is not the same as truth.

The recent disputes between governments and AI companies illustrate this clearly. Debates over surveillance limits or autonomous weapons are not simply technical questions. They are moral ones. Guardrails exist precisely because the systems themselves are not neutral.

Misconception #3: Humans Fully Control AI

Traditional software behaves according to explicit instructions written by programmers.

Modern AI systems operate differently. Their outputs are probabilistic, generated through layers of learned relationships inside the model.

Developers are now using AI systems to build AI systems and to manage other AI systems. They are using AI to write code that in the past they would have written themselves, and it’s happening so fast that they cannot monitor or even understand every line of code being generated by systems that do not sleep.

Control, in this environment, is not a switch. It is more like a moving boundary that no one has ever seen before, and the language to even define it is still in its infancy.

Misconception #4: The Experts Know Where This Is Going

In most scientific fields, experts disagree within a fairly narrow range. In artificial intelligence, the range of opinion is unusually wide.

Some researchers believe AI will revolutionize medicine and scientific discovery. Others warn the technology could produce serious societal disruption if development outruns human wisdom.

Among those raising such concerns is Geoffrey Hinton, a Nobel Prize winner and one of the foundational figures of modern AI research.

That range of opinion does not prove disaster is coming. But it does reveal that even the people building these systems do not fully agree on where they lead.

Artificial intelligence is integrating rapidly into the systems that shape modern life—communication, commerce, national security, and governance.

We are seeing signals across all of these domains. We can see clearly that AI is shaping our future whether we like it or not. The question is whether we will recognize the signals in time to understand what is unfolding, or whether we will wait, as societies often do, until the consequences make the signals impossible to ignore.

No comments:

Post a Comment

Humanity Crossed A Threshold, And Most Of Us Scrolled Past It

  I have not updated this blog since last week on purpose. My objective here is to give food for thoughts, not to comment on current events....