Friday, April 3, 2026

Amazon + Anthropic = Massive replacement of blue/white collar employees by robots/autonomous agents

Last update: Monday 4/14/26 3:42am
Amazon recently announced its intentions to replace a high percentage of its blue collar staff with robots. In the same time frame, Anthropic, Amazon's genAI partner, reported that large enterprises will be able realize substantial increases in their profits by replacing substantial segments of their white collar staff with Anthropic's autonomous agents. 

Full text  HERE


 1. Bezos | 2. Amodei | 3. Pentagon (Appendix) 

Introduction
A previous version of this note was framed as the third bad partnership that Anthropic, a corporate idealist, had made with pragmatists: First Sam Altman at OpenAI, then Secretary Hegseth at the Pentagon, and now Jeff Bezos at Amazon. 

The first two partnerships ended badly, and the third also seemed to be headed towards another unhappy ending. However, this time Anthropic seemed to be enjoying its partnership AND the likelihood that its autonomous agents will replace millions of white collar employees of large enterprises into unemployment and financial ruin. Anthropic’s enjoyment of its surging market value suggested that Anthropic had betrayed its corporate ideals, that Anthropic had become corporate pragmatist.

But this interpretation was inconsistent with Anthropic’s simultaneous, deliberate sabotage of its partnership with the Pentagon, a sabotage that the editor of this blog carefully examined in his previous blog note, 'REVISED: Pentagon vs. Anthropic -- Trump fires 'Mad Dog' Dario"

Dr. Amodei is still an idealist who would sooner forfeit lucrative contracts, e.g., with the Pentagon, than deliberately betray his core ideals. But perhaps his core ideals have been curtailed by pragmatic corporate considerations?

Dr. Amodei is also a brilliant man who was educated at three of the best universities on the planet: Caltech, Stanford, and Princeton. So what accounts for his glaring spots of ignorance? 

This report attempts to answer both of these questions. But given that Amazon is only driven by pragmatic corporate objectives, the report will examine Amazon's rationale for its massive deployment of robots and Anthropic's rationale for its deployment of autonomous agents separately. And it will also offer a plausible systemic explanation for Dr. Amodei’s blind spots

1) Jeff Bezos, Amazon founder and pragmatist

(a) Bezos, the benign patron, bought the Washington Post in 2013

Bezos saved the Washington Post from bankruptcy. The Post hadn't found the survival strategy that the New York Times, Wall Street Journal, and other major papers worked out against the internet threat. Bezos made substantial investments, gave editorial staff substantial autonomy — behaved as a benign patron.
  • ''Washington Post closes sale to Amazon founder Jeff Bezos", Washington Post, 10/1/2019

(b) Then Bezos became one of the pragmatic billionaires, changing his course to accommodate President Trump
Among many other things, he blocked the Washington Post's endorsement of Kamala Harris in late October, 2024; he pledged $1 million to the inauguration committee; and he paid $40 million for Melania Trump's documentary.
  • "Amazon Has Found the Easiest Way to Influence the Trump Administration", New Republic, 1/7/25
And he recently fired 30 percent of the staff at the Post:
  • "Jeff Bezos’ Washington Post Cuts 30% of Staff, Laying Off More Than 300 Employees Including Amazon Beat Reporter", Variety, 2/4/26

(c) Amazon always focused on genAI models as revenue generators for Amazon Web Services (AWS) 

The editor of this blog is unaware of Amazon ever making "visionary" promises to develop artificial general intelligence, super intelligence, or any other kind of artificial intelligence entity based on large language models. Amazon invented the cloud services in the early 2000s when Bezos was CEO, and has maintained this focus under Andy Jassy, its current CEO.  

Indeed, Amazon quickly developed a new component of AWS called "Bedrock" in April 2023 (GeekWire, 3/22/26) as a site where developers could display and develop a wide range of models. 
Here's a copy of Amazon's current description:

  • "Amazon Bedrock powers generative AI for more than 100,000 organizations worldwide—from startups to global enterprises across every industry. It provides the proven infrastructure and comprehensive capabilities to confidently build applications and agents that work in production with the flexibility, enterprise security, and proven scalability you need to innovate boldly and deliver AI that drives real business impact."
d) AWS provides substantially higher returns on Amazon's investments than Amazon's Retail Services 
Amazon's SEC report for fiscal 2024 (2/6/25) includes two retail components -- North America and International -- plus AWS. Here's a direct quote from the report:
  • "Operating income increased to $80.0 billion in 2025, compared with $68.6 billion in 2024..
    -- North America segment operating income was $29.6 billion, compared with operating income of $25.0 billion in 2024
    -- International segment operating income was $4.7 billion, compared with an operating income of $3.8 billion in 2024
    -- AWS segment operating income was $45.6 billion, compared with operating income of $39.8 billion in 2024".
Note that AWS generated $45.6 billion which is more than North America's $29.6 billion plus International's $4.7 billion = $34.3 billion. What's far more impressive is the relatively small AWS sales required to generate this much larger income. Here's another quote from the SEC report"
  • Net sales increased 12% to $716.9 billion in 2025, compared with $638.0 billion in 2024. Excluding the $4.4 billion favorable impact from year-over-year changes in foreign exchange rates throughout the year, net sales increased 12% compared with 2024.
    -- North America segment sales increased 10% year-over-year to $426.3 billion.
    -- International segment sales increased 13% year-over-year to $161.9 billion, or increased 10% excluding changes in foreign exchange rates.
    -- AWS segment sales increased 20% year-over-year to $128.7 billion
AWS sales of $128.7 billion was much smaller than the combined sales from North America $426.3 billion plus International $161.9 billion = 588.2 billion. In other words, the retail segments sold four times as much as AWS, but yielded substantially less operating income $34.3 billion than the AWS $45.6 billion. 

Given this huge disparity, the most logical opportunity for substantial improvement would be in the blue collar sectors of the Amazon’s retail segments. This means replacing a substantial percentage of its humans employees with robots. That's Amazon's rationale for its improvement plans that will be discussed in a subsequent section (below).

Question 1: Do Amazon’s leaders believe the old economic maxim that whenever new technology wipes out old jobs over here, it creates new jobs somewhere over there?
Answer 1: Yes, they probably do.

Question 2: Do they believe that the employees they replace with robots will get those new jobs somewhere over there?
Answer 2: No, they probably don’t, nor have economists ever claimed that they would. Economists have merely shrugged and said that the new jobs would offer higher pay, probably to other people, but that society as a whole would be better off. 

The really bad news is that this old notion probably won’t be true when widespread replacement of human workers comes within three to five years, as per Amazon’s plan, rather than over the course of decades, as happened with most prior disruptive technologies. 

If this pessimistic assessment is validated, will Amazon be prepared tto compensate the staff it replaces with robots with paid retraining or financial compensations? Probably, that's the logical pragmatic response, but only if all other employers who replace staff with robots are also required by law to do so.

d) Amazon's Plans
This section aims to presents some details as to how many human employees Amazon will replace with robots and within what time frame. A New Times report that was based on leaked documents covered all bases. So here are that report's key findings:

-- "Amazon’s Plan to Replace 600,000 Workers With Robots Just Leaked",  NY Times, 10/29/25

-- "Now, interviews and a cache of internal strategy documents viewed by The New York Times reveal that Amazon executives believe the company is on the cusp of its next big workplace shift: replacing more than half a million jobs with robots"

-- ''Executives told Amazon’s board last year that they hoped robotic automation would allow the company to continue to avoid adding to its U.S. work force in the coming years, even though they expect to sell twice as many products by 2033. That would translate to more than 600,000 people whom Amazon didn’t need to hire."

-- "At facilities designed for superfast deliveries, Amazon is trying to create warehouses that employ few humans at all. And documents show that Amazon’s robotics team has an ultimate goal to automate 75 percent of its operations."

-- "Amazon is so convinced this automated future is around the corner that it has started developing plans to mitigate the fallout in communities that may lose jobs. Documents show the company has considered building an image as a “good corporate citizen” through greater participation in community events such as parades and Toys for Tots."

-- "The documents contemplate avoiding using terms like “automation” and “A.I.” when discussing robotics, and instead use terms like “advanced technology” or replace the word “robot” with “cobot,” which implies collaboration with humans."

-- "Amazon’s plans could have profound impact on blue-collar jobs throughout the country and serve as a model for other companies like Walmart, the nation’s largest private employer, and UPS."

-- “Nobody else has the same incentive as Amazon to find the way to automate,” said Daron Acemoglu, a professor at the Massachusetts Institute of Technology who studies automation and won the Nobel Prize in economic science last year. “Once they work out how to do this profitably, it will spread to others, too.”

-- "There are concerns automation could affect people of color particularly hard because Amazon’s warehouse workers are about three times as likely as a typical American worker to be Black."


2) Dr. Dario Amodei, PhD, Anthropic Founder/CEO and Idealist

The section will argue that the blind spots in Dr. Dario Amadei’s knowledge was the result of systemic flaws in our college and university curricula. Nobody can know everything, so we will be talking about blind spots that can substantially undermine a person’s effectiveness.


The editor of this blog drafted an earlier version of this discussion in an effort to understand/explain the underlying causes of Anthropic's failed partnership with the Pentagon. A highly condensed description of that failure can be found in the Pentagon Appendix
 to this report. His original data driven report flatly contradicted the unfounded emotional "support" that Dr. Amodei still receives from the vast majority of his colleagues in Silicon Valley.

(a) Rumsfeld's paradigm
At a press briefing in 2002, Donald Rumsfeld, Secretary of Defense, came up with a clever formulation. 
  • Known knowns — There are important things that we need to know, and we know them. Rumsfeld called these the “known knowns”. For example, students will not be admitted to a course unless they can demonstrate that they have already acquired a certain body of knowledge, the course prerequisites.

  • Known unknowns — There are important things that we know that we should know, but we don’t know them yet. Rumsfeld called these things “known unknowns”. For example, instructors usually provide a syllabus that specifies the knowledge students do not know at the beginning of a course but should know by the end of their course. However, students who do not acquire an understanding of most of this specified knowledge by the end of the course will fail the course because they did not convert their known unknowns into known knowns.

  • Unknown unknowns — Rumsfeld identified a third category of knowledge, important things that we should know but we don’t even know the names of such knowledge yet. This is the stuff of what some people call “black swans”.
    Most freshmen don't know much about anything; it's all a big bag of unknown unknowns. So who decides what should be taught? How should this knowledge should be be divided into courses? And when should those courses be taught? These decisions are made by the faculty, of course.

    But colleges and universities long ago learned not to take the aspirations of entering freshmen too seriously. Young students often change their minds. Indeed, changing one's mind after making wrong choices is one of the golden prerogatives of youth that should not be denied.

    Therefore the curriculum should enable the youngest students to change their aspirations without incurring substantial penalties, e.g., added tuition or extra time to graduate. Faculty strive to identify a broad range of knowledge that every college educated graduate should acquire. Courses that satisfy these general requirements are usually offered to freshmen and sophomores. Most of the courses in selected majors are offered to juniors and seniors.

    But the process is more complex and surprisingly so. The youngest students not only learn from the faculty; they learn from each other and they learn from older students, especially from juniors and seniors in other majors than they have been considering.


    Personal anecdote:
    When the editor was a sophomore engineer, he became friends with a junior, a sociology major who already seemed to know more sociology than the professors, but who also knew less than 5 percent as much math as he did. He teased his friend to tell him about some great discovery made by sociologists using kiddy math. His friend's response: read Emil Durkheim's late 19th century classic study "Suicide".

    Suicide is perhaps the most personal decision a person could make, so one would expect that rates of suicide only varied by differences in people's personal psychologies. But using grade school arithmetic, Durkheim found that the rates of four different kinds of suicide also varied by social conditions, e.g., marital status (unmarried), oppressive conditions (prisons), social disruption (extreme gain or loss of wealth), and altruism (military suicide in defense of others).

    About fifteen years after reading Durkheim, the editor submitted his PhD dissertation, a sociometric examination of the communication networks among planners and engineers. Most of his references were to previous studies in social psychology
(b) Dr. Amodei fell through the cracks  
We begin with a quick review of Dr. Amodei's stellar academic background. We note his enrollment at the highly selective California Institute of Technology ("Caltech") in 2001 and his one year pause in his undergraduate studies before he transferred to Stanford University in the fall of 2004. Then we offer a careful analysis of a compelling anti-war op-ed that he published a few months before he left Caltech. 

Enrollment at Caltech 
Dario Amodei was admitted in the Fall 2001. Readers should understand that his admission was a
personal milestone because Caltech is arguably the nation's most selective top tier institution of higher education. Its current total enrollment of 900 students is less than half the size of Harvard's freshman class. The math SAT scores of 50 percent of its freshmen range between 780 and 799. Twenty five percent score a perfect 800.

According to Clay, 9/13/24, Amodei left Caltech at the end of his sophomore year to take a position at Applied Minds as a Research Intern, July 2003 to September 2003; then he worked at Schlumberger as a Geophysicist, February 2004 to September 2004. He entered Stanford in the Fall 2023 semester and graduated in 2006 with a BS in Physics.

Anti-war OpEd (pdf)
Readers should click the above link to verify every assertion in this section. Having read many of Dr. Amodei's convoluted essays on Anthropic's website, the editor of this blog was shocked to find that 19 year old Dario wrote this fiery call to activism with such stunning clarity.

Dario had challenged his intended readers, his fellow students, to recognize their responsibility as the nation's future leaders in science and technology:
  • "All the great idols of Caltech, from Richard Feynman to Linus Pauling, have understood the need to be citizen-scientists, to contribute their analytical skills to the enormous forum that is our democracy. We who seek to emulate them scientifically should also do so politically. We should never let ourselves be reduced to amoral technicians who run the machines of war as casually as we do our computations."
He had challenged his fellow students to give their full attention to a war that made no sense to him. And, yes, he expressed painful disappointment at his failure to get them to breakaway from their nerdy pursuits to join him in the loudest possible protests. He was so angered by their unshakeable apathy that he dropped out of Caltech. Wow!!!

Now comes the analysis. Back then, the editor of this blog was about twenty years older than Dr. Amodei is today. So his perspective was uncertain. Attacking Hussein in response to Ben Ladin's attacks on the World Trade Center and the Pentagon made no sense. And it was widely known that President Bush had been holding a personal grudge against Hussein ever since Hussein organized an attempted assassination of his father, President George H. W. Bush, on a visit to Kuwait in 1993 after he had left office.


Nevertheless, Secretary of State Colin Powell's enthusiastic support for this response in his speech before the United Nations caused editor to nod in silent agreement, just as it commanded similar silent agreement among millions of other Americans.

Newly elected President George W. Bush was counting on this reaction. That's why he and his colleagues concocted a complex set of bold faced lies that convinced his unsuspecting Secretary of State that Hussein had a nuclear arsenal, which Hussein definitely did not have. Many months after our successful "conquest" and our continued failure to find these nuclear weapons of mass destruction (WMDs), the truth came out. Powell was humiliated into retirement as millions of hitherto silent Americans roared their anger at this outrageous presidential deception.

Hussein did have WMDs, but they were't nuclear. He had been slaughtering thousands of Iranians and Kurds for years with chemical and biological WMDs, chemical WMD equipment that he had purchased from western companies. According to Wikipedia, "
On September 22, 1980, Iraq staged an all-out war on Iran from ground, air, and sea and came to occupy a vast part of Iranian territory .... February 1984 to the end of the war [1991], chemical weapons were used extensively.All told, 52% of Iraq's international chemical weapon equipment was of German origin."

Hussein had pretended to have nuclear WMDs in order to deter Iranians hordes from swarming over the border in retaliation. 

The CIA was in on the deception of Secretary Powell. Indeed, the CIA Director provided the most convincing confirmation of the president's lies. Of course the CIA knew all about Hussein's chemical WMDs and their western manufacturers. Nevertheless, young Dario naively included the following CIA denial in his op-ed: "A CIA report commented that Saddam was unlikely to use weapons of mass destruction or seek collusion with terrorist groups unless he was attacked.
"  ... Rubbish.


So where was young Dario when the lies the Bush administration had told to Secretary Powell were publicized? He was at Stanford. There is no evidence that he joined the now loudly protesting angry millions all over the U.S. who would ultimately declare these "Forever Wars" to be wasteful tragedies. Young Dario had evidently become what he had condemned a few years earlier, a silent nerd, focused on his studies. 

Dr. Amodei stayed in his self-absorbed, apathetic nerd mode until 2024, when he suddenly volunteered his company’s services to the Biden administration in a contract that was extended by the Trump administration in 2025. Of course, he missed all of the disturbing public data about the CIA‘s launching of most of the drones and missiles during the Obama years that killed too many people in the “Forever Wars”. 

So he was shocked, dismayed to learn that the CIA had killed 75 people as part of the Trump administration’s operation to kidnap Venezuela’s President Maduro. He wanted to withdraw from government service immediately. Unfortunately, federal contractors are not allowed to resign from contracts, even if they entered into the contracts voluntarily. He had no choice but to figure out a way to get the Pentagon to fire his company, which he did. Then he graciously accepted the resounding, uninformed applause for his "heroism" from his colleagues in Silicon Valley.


Falling through the cracks
Had young Dario entered Stanford as a freshman, he would have encountered many students who were as outraged as he was by the forthcoming assault on Iraq, including some super smart aspiring young scientists, like himself. Instead he dropped in as a junior and missed the frequent interactions students have with other students 
in the general courses they take as freshmen and sophomores who had a wide range of aspirations 

So what did he miss? Had he started at Stanford instead of Caltech, he would have been on Stanford's campus when large scale anti-war demonstrations erupted. According to Stanford Magazine
  • "On March 5, more than 500 students and faculty converged on the Quad to protest and to attend teach-ins offered by 20 professors from 10 departments. Biologist Robert Sapolsky spoke about “Evolution of Aggression and Warfare,” anthropologist Carol Delaney recalled a trip to Iraq, and political scientist Terry Karl, ’70, MA ’76, PhD ’82, addressed the links between “Oil and the War in Iraq.”"
In other words, young Dario would have been one of over 500 other freshmen and sophomores whose anti-war sentiments were as powerful as his own. It would have been unlikely that he would have formed a life-long self-image as the lonely defender of high ideals among an apathetic majority

When the truth came out, he would have found good company among other students who were outraged by the CIA's lies. Twenty years later he would not have been surprised by the CIA's involvement in killing so many Venezuelans during the kidnapping of President Maduro. Hopefully, he would not have volunteered to provide tech support for the CIA because he would have learned what many of us merely relearned: The CIA lies, it lies to our adversaries, and it has lied to America. But that's what espionage agencies are paid to do, they lie and they kill.

With a bit of luck, he might have encountered some super smart juniors and seniors who would have blunted his snobbish belief that the only real sciences are the physical sciences -- physics, chemistry, biology, and combinations thereof. For example, he might have learned that psychologists have discovered surprising limitations on human capacity to deal with complexity in real-time. Aware of these scientific findings, he would not have specified impossible conditions on the use of his models in the Pentagon's battlefield operations. Better still, he would have volunteered his company's models to the State Department, instead of the Pentagon. 

Instead, young Dario has grown up to become the Dr. Amodei we see today, the ultra idealist who sincerely believes that he carries a lonely burden, that anyone who disagrees with him must not be telling the truth.


Science snobbery cuts both ways
Unfortunately, this is a two way street. Colleges and universities have irrefutable evidence that most students who use genAI chatbots earn lower grades because they don't learn as much as they should, i.e, they don't convert enough known unknowns into known knowns by their courses' ends. More generally, chatbots impair their student users’ capacities to engage in critical thinking. 

Worse still, their faculty know that large language models models are not based on any underlying cognitive science, no theory of intelligence that can explain high intelligence in tiny brained crows and also account for the different brain to body mass ratios in creatures as varied as crows, elephants, dolphins, humans, and killer whales. 

This awareness is especially acute among faculty in the cognitive sciences. Cognitive scientists have no idea how or why large language models work. Only Anthropic, OpenAI, Google, and other developers can identify the internal changes that must be made to their models to substantially reduce this harm.

Perhaps faculty aversion to genAI models was also intensified by DARPA's unfortunately premature negative assessment of this technology in this YouTube video nine years ago, wherein its director flatly asserted that LLM's have virtually no capability for logical/deductive reasoning.  Fortunately, DARPA has since reversed this assessment and launched impressive projects that fully exploit the logical/deductive capabilities of LLM's. Two of these projects will be noted in later sections of these notes.

Reactions vary, but some colleges and universities have been as zealous and as counterproductive as Dr. Amodei's ultra idealism. Their faculty are deeply disturbed by the tangible harm these baffling collections of non-scientific software are doing to their students, so their reactions have been almost medieval

They have banned chatbots from their campuses, penalized students who used them, tortured repeated users with extended inquisitions, and excommunicated, a/k/a expelled, students who continued to consort with these evils. 


(c) Unemployed white collar workers
The editor failed to recognize the real significance of Anthropic's initial announcement of its open source Multiple Context Protocol (MCP) back in November 2024.
  • "Introducing the Model Context Protocol", Anthropic 11/25/24
In the following months, Anthropic used the MCP to produce reliable autonomous agents. Recognizing their likely harm to white collar employees, its CEO, Dr. Dario Amodei, issued a public warning/prediction that its MCP would cause millions of white collar employees to lose their jobs.
  • "Why this leading AI CEO is warning the tech could cause mass unemployment", CNN, 5/29/25
Whereas the Pentagon's CTO had merely muttered "holy shit" to a Senate committee, the editor found himself screaming Why!!! Why!!! Why!!! over and over again. If readers direct their browsers to search for the words "channeling our collective efforts" in Anthropic's manifesto, they will find the following paragraph. (Note: red print and boldface were added by the editor.)
  • "Anthropic’s role will be to provide as much evidence as possible that AI safety techniques cannot prevent serious or catastrophic safety risks from advanced AI, and to sound the alarm so that the world’s institutions can channel collective effort towards preventing the development of dangerous AIs. If we’re in a “near-pessimistic” scenario, this could instead involve channeling our collective efforts towards AI safety research and halting AI progress in the meantime. Indications that we are in a pessimistic or near-pessimistic scenario may be sudden and hard to spot. We should therefore always act under the assumption that we still may be in such a scenario unless we have sufficient evidence that we are not"

If Anthropic still believes what it published in its manifesto, it should "halt its AI progress" ... halt its sales??? ... what sales ???... its MCP is a free, open source, free-roaming, genAI animal, available on the Internet to any large corporation that wants to use it. 

At the very least Anthropic should have removed its promise to halt the production of dangerous genAI products from its manifesto. (Note: Interested readers can find a copy of the text of the manifesto as it appeared on Anthropic's website on 3/22/26  HERE)

Deleting its obsolete promises from its manifesto would not be enough. Anthropic should have invested its considerable skills into proactive partnerships with universities and other training operations that might identify the likely new skills required for the new job opportunities. 
  • For example, as reported in TechCrunch, M.I.T. and some other universities have responded to dramatic drops in enrollments in their computer science programs by offering joint degrees with other majors wherein genAI skills could be applied.
    -- "The great computer science exodus (and where students are going instead)", TechCrunch, 2/15/26
Tech-savvy youngsters were aware that Big Tech companies were already announcing substantial layoffs; but the open source MCP would enable all big corporations to make the same kinds of massive staff reductions. 


Here's a reason why joint degrees make sense:
  • Today's large Language models (LLMs) produce good summaries. They also have good deductive skills so they can help researchers to suss out defects in the logic of their hypotheses; but current models are lousy at inference, even Claude.

  • On the other hand data scientists providing support for researchers in just about every field have found that 70 to 80 precent of their effort is devoted to cleaning up the data. This is tedious, time consuming work and seldom requires ingenuity. 

  • Only the most highly talented and highly skilled students will obtain joint degrees that will empower them to create customized autonomous agents for their projects that will perform all of the tedious data preparation, thereby providing the students, now researchers, more time to focus on the concepts they are trying to creatively elucidate. In other words, researchers will do what agents can't; agents will do what would otherwise waste the researchers' time.

  • But colleges and universities that offer data science degrees will have an easier path. Their data science majors could be required to take courses that taught them how to use the MCP to develop autonomous agents. 

    Students obtaining degrees in other fields could be given additional statistics courses that introduced them to the data cleaning process and alerted them to the advantages of working with an appropriately trained data scientist who could develop the kinds of autonomous agents that would be most useful to them
Unfortunately, reality is far more cruel. Had Anthropic merely proposed to replace the lowest level tech employees in large corporations, that might have been regarded as the inevitable consequence of the over-hiring of young techs that Big Tech firms have done in recent years in their race to become the surviving winners in the genAI wars.

But young techs are at most 10 percent of the lowest level white collar staff employed by all large corporations. If only 1,000 large corporations replaced 1,000 low level white collar staff, one million jobs will be lost. Why would Anthropic want to use its AI technology for such harmful purposes?


(e) Mythos -- more pragmatic limits?
As the editor was finishing the previous sections of this report, Anthropic announced a preview version of Mythos, its newest and most powerful model. This should have been great news for the editor because he thinks Anthropic's models, which he accesses through its chatbot Claude, are by far the smartest models on the market. A great model can do many things, but Anthropic immediately focused on one important application -- the model's capacity to find and fix flaws in software, especially flaws in the infrastructure of banks and Big Tech. 

An accident leaked the the code for the new model, but not its weights, onto the Internet. Anthropic immediately formed a consortium of large enterprises called Project Glasswing. Although Anthropic's Website only lists nine members, Fortune Magazine, 4/7/26, reported that "Anthropic is also providing access to roughly 40 more organizations responsible for building or maintaining critical software infrastructure, allowing them to use the model to scan and secure both their own systems and open-source code." ... therefore about 50 members

The 50 invited members are large enterprises, e.g., banks and Big Tech firms, that might be among the most vulnerable to hackers misusing the model. Membership in Glasswing is supposed to give its members a head start on fixing the flaws in their software infrastructures before hackers can exploit these flaws.

Members will be given extensive access to Anthropic's models via APIs on Amazon's AWS, Microsoft's Azure, and Google's cloud. In other words, Anthropic should anticipate heavy usage and substantial income from these activities. And it raises questions:
  • China's Deep Seek "distilled" OpenAI's GPT models via extensive access. Given the leaked code plus extensive access, all of the wealthy members of the consortium should be able to distill a model having comparable effectiveness to Mythos.  

  • The editor of this blog is far more concerned by the possibility that sovereign hacker groups from China, North Korea, and Russia will gain extensive access to Mythos -- through bribery or blackmail -- of personnel on the staff of one of the consortium's 50 members. Then our adversaries will distill models that are comparable to Anthropic's most powerful model ... models that can hack large enterprise infrastructure ...

    ... But Mythos can also be useful across a wide range of military and intelligence applications, like the applications discussed in previous sections of this report:  the CIA's kidnapping of President Maduro and the Pentagon's agentic targeting of adversary units on battlefields -- applications which Anthropic renounced ... for America ... But when our adversaries distill models comparable to Mythos, will Anthropic once again deny all responsibility .... and will Silicon Valley loudly cheer ... or is anything really a secret when you share it with 50 organizations?

    Indeed, no one needs access to Anthropic's most powerful model to discover flaws in infrastructure. That challenge was met by the winners of a DARPA/ARPA-H Challenge last summer 2025, as noted in the following bullet. 

  • "Challenge showcases AI’s power to secure America’s health care", ARPA-H, 9/4/25 ... "At DEF CON 33, ARPA-H joined the Defense Advanced Research Projects Agency (DARPA) to announce the winners of the AI Cyber Challenge (AIxCC), a two-year competition to develop AI-enabled software that automatically identifies and patches vulnerabilities in the source code that underpins critical infrastructure." .... Hmmm  ... sounds like Anthropic's new Mythos
    -- The first academic assessment of the winning software was recently posted on ArXiv, February, 2026
    -- The winning software can be found on GitHub (It's open source)
    -- Anthropic was one of the corporate sponsors of this challenge, so its representatives were aware of the winners' impressive achievements.

In summary, the manner of Anthropic's announcement of Mythos caused many recipients of the announcement to conclude that Mythos was the only way to detect/fix flaws in their infrastructure. Indeed, Treasury Secretary Bessent and Fed Chair Powell warned bank CEOs about the risks of Anthropic's model, Reuters, 4/9/26 ... risks that are, ironically, offset by its ability find and fix risks.

Given DARPA's decades of experience sponsoring highly effective, high tech breakthroughs many, perhaps most members of the consortium might have preferred to use the open source code produced by the challenge winners. 

Mythos' real advantage was superior responses to a wide range of civilian and military challenges plus a defense against hackers in one package. But now that our Chinese, Russian, and North Korean adversaries have had access to the ARPA-H open source code for almost a year, they will add that code to their extensive array of hacking tools, then target Mythos and everyone who uses Mythos for any purpose whatsoever.


(d) Recommendations
Based on the findings in this report, the editor of this blog offers a few recommendations to the California Institute of Technology (Caltech), to Anthropic, and to Anthropic's vociferous supporters in Silicon Valley, all of which the same message: broaden your fields of vision.

-- First, the editor appeals to DARPA:
As per your
mission statementAmerica looks to you "to create and prevent technological surprise for our national security". Ever since your founding in 1957, you have not let us down. In your development process, you deliberately push prospective tech options to extremes in order to identify the boundaries of their effectiveness. Of course, you usually don't publicize these deliberate failures.

Nevertheless, your public denunciation of the logical/deductive powers of large language models in a YouTube video nine years ago, broke your usual silence about failed technologies. Indeed, the rapid evolution of LLM's since then  changed your mind, so much so that you have already achieved considerable success in a few projects that utilize these capabilities. 

Accordingly, the editor of this blog strongly recommends that you issue a public revision of your assessment of LLMs that will assure America's science and tech communities that it is well worth their while to reconsider these capabilities and to pay close attention to your own impressive use of LLMs. 


-- Here's the the editor's recommendation for large banks and other large enterprises 

CLARA: Compositional Learning-And-Reasoning for AI Complex Systems Engineering
This project promises to become one of DARPA's most important projects ever. It implements notions of software as provable theorems that were first introduced by Edsger W. Dijkstra, Tony Hoare, and Haskell Curry &  William Alvin Howard, but were only implemented within the cyber community because of prior sky high implementation costs.The costs of failed security were high enough for cyber clients, e.g., clients who would otherwise incur massive financial losses if their security failed, so the higher development costs were cost-effective.

When DARPA releases the the free open source technologies that emerge from CLARA, the costs of implementing guaranteed cyber safety should be drastically reduced for all clients. 
In other words, Mythos today, but CLARA not too long thereafter.

Here's DARPA's own description of the project and its goals, published on 2/10/26.

/-----/-----/

"Disruption Opportunity
 
Today, the dominant industry approach to artificial intelligence (AI) is to tack specialized automated reasoning (AR) components onto a large language model (LLM) or other similar machine learning (ML) system. These ML-centric systems typically have weak assurance; the “tack-on” approach is an importantly limited way of providing assurance or safeguards.

The Compositional Learning-And-Reasoning for AI Complex Systems Engineering (CLARA) fundamental research program is designed to tightly integrate AR and ML components to create high-assurance AI — which is expected to scale even to complex systems of systems. Integrating the two different branches of AI will provide the speed and flexibility of ML with verifiability based on AR proofs that have strong logical explainability and computational tractability.

In more detail, CLARA is anticipated to create powerful methods for the hierarchical, fine-grained, highly transparent composition of important kinds of ML and AR components, including Bayesian, neural nets, and logic programs. 

CLARA aims to create a theory-driven algorithmic, highly reusable, scalable foundation for high assurance plus broad applicability, useful for many crucial defense and commercial realms which may include, but is not limited to: 

  • Kill web, supply chain & logistics, and wargaming
  • Autonomous and command & control
  • Medical, financial, and legal
  • Science and tech design"
/-----/-----/

-- Here's editor's recommendation to the California Institute of Technology

Please 
broaden your curriculum for freshmen and sophomores asap so that your students not only learn how to produce tools based on physics, chemistry, and biology, but also learn the fundamental capabilities and limitations of the humans who will use these tools, capabilities and limitations identified by the human sciences -- sociology, anthropology, psychology, economics, etc.

Or to put it bluntly:
Stop converting the most brilliant minds of our youngest generations into cyclops, one-eyed monsters whose tunnel vision makes them incapable of perceiving the complex, three dimensional relationships between the tools they develop and the human users of these tools. 


-- Now on to the Silicon Valley of the Cyclops. 
Readers should recall that young Dario's fiery Anti-war OpEd (pdf) back in the spring 2002 semester had challenged his intended readers, his fellow students at Caltech, to recognize their responsibility as the nation's future leaders in science and technology, as followers in the footsteps of "citizen-scientists" Linus Pauling and Richard Feynman.

Linus Pauling and Richard Feynman were not cyclops. Neither acquired their undergrad degrees from from Caltech. Both had benefited from undergrad curricula that gave them exposure to other perspectives.

Dr. Feynman
 was so curious about other branches of knowledge that he describes himself as a "nuisance" when he was a grad student at Princeton where his intense curiosity about other fields of knowledge drove him to badger the professors in philosophy courses with his off the wall questions. 

But let's skip to Dr. Feynman's most famous act of public service: his discovery of what really caused the space shuttle Challenger to breakup. The editor asked Claude, Anthropic's chatbot, for a description. Here are he most relevant parts of Claude's summary:

"The Space Shuttle Challenger broke apart 73 seconds after launch on January 28, 1986, killing all seven crew members. President Reagan appointed the Rogers Commission to investigate, and Richard Feynman was one of the commissioners.


Feynman quickly became frustrated with the commission’s pace and its deference to NASA. He conducted his own investigation, talking directly to engineers at Morton Thiokol, the contractor that built the solid rocket boosters, and to lower-level NASA engineers.


On a televised hearing he demonstrated in ten seconds, with a glass of ice water, what NASA had failed to acknowledge: the O-rings couldn’t seal properly in cold temperatures. That’s what killed the crew.


But his deeper contribution was documenting a systematic disconnect between NASA’s engineers and its management. The engineers estimated the probability of shuttle failure at roughly 1 in 100. Management claimed it was 1 in 100,000 — a thousand times safer than the engineers believed.


The engineers who knew the shuttle was dangerous were not being heard by the managers who decided whether to launch. Feynman concluded: “For a successful technology, reality must take precedence over public relations, for Nature cannot be fooled.“​​​​​​​​​​​​​​​​


Of course Feynman, the brilliant physicist, discovered the physical phenomenon involved in the breakup. But Dr. Feynman was not a cyclop. That's why he had conversations with the humans involved, the managers and the engineers. That's why he perceived the real cause of the catastrophe: the failure of the managers to take the experienced engineers' estimates of failure seriously. 

Now come back to the recent past, to the final months of the Biden administration when Dr. Amodei had volunteered Anthropic's technical support for federal agencies in a contract that the Biden administration signed and the Trump administration extended. Readers in Silicon Valley, please note that Dr. Amodei volunteered Anthropic's support. Neither administration commandeered Anthropic's support via the Defense Production Act or some other legal mechanism.

The contract stipulated that Anthropic's autonomous agent software could not be used autonomously. The software could only be used to recommend drone and missile launches against an adversary.  The final decision to launch the drones/missiles had to be made by a human handler

. Amodei insisted with cyclopian certainty that there was literally only one possible interpretation of this restriction. The handler had to read  and understand the agent's recommendation that was, like all precise machine learning patterns, based on many factors

How it possible for Dr. Amodei not know was that it would be impossible for any human handler to understand the agent's recommendation in real time if the pattern was defined by more than four factors. Dr. Amodei had either never heard of or did not believe Cowan's finding about "The magical number 4 in short-term memory: a reconsideration of mental storage capacity", N. Cowan, Pub Med/NIH, February, 2001. Unlike the brilliant Dr. Feynman, Dr. Cyclop was absolutely certain that no human limitations could be involved. The reader is once again referred to the Pentagon appendix to this note for an alternative, common sense interpretation of the contract's stipulation.

Dr. Linus Pauling, the greatest chemist since Lavoisier, was such a dedicated and effective "citizen-scientist" that he won a Nobel Peace Prize for his efforts. The editor asked Claude to provide him with an overview of how Dr. Pauling labored to mobilize the world's scientists to raise their voices in support for the partial nuclear test ban treaty between the United States and the Soviet Union. Here is Claude's succinct summary:

/-----/-----/-----



"Pauling joined Einstein’s Emergency Committee of Atomic Scientists in 1946 — the year after Hiroshima. That’s when his activism began. Throughout the 1950s he marshaled evidence, denounced nuclear testing, publicly debated Edward Teller, circulated petitions, led demonstrations, gave innumerable speeches, and wrote a book.  The US State Department revoked his passport in 1952 for his activism. He was called a communist sympathizer. Life magazine called his Nobel Peace Prize “A Weird Insult from Norway.” He was effectively forced out of Caltech.

In 1957-1958 he and his wife collected 11,021 signatures from scientists in 49 countries and presented the petition to the United Nations. 

The Partial Nuclear Test Ban Treaty came into force on October 10, 1963 — the same day Pauling was awarded the Nobel Peace Prize. 
 

That’s seventeen years of sustained activism — from 1946 to 1963. Seventeen years of personal sacrifice, professional ostracism, government harassment, public ridicule, and relentless work. He lost his passport. He lost his job. He was called a traitor. He kept going."

/-----/-----/-----

By stark contrast, the young Dario was active during the Spring semester of his sophomore year at Caltech. He took a year off. Then he entered Stanford in the Fall 2004 semester as a junior. He was no longer a fiery activist. He had became a silent apathetic nerd, focusing on his science, just like his fellow students at Caltech whom he had denounced one year earlier.

His activism did not reemerge until 2024 when he volunteered Anthropic's technical support to federal agencies during the Biden administration. Pauling's seventeen years vs. Amodei's one semester (2002) plus about 14 months (November 2024 to March 2025).



Pentagon Appendix
The following segments were copied from the editor's extensive examination of the broken relationship between the Pentagon and Anthropic. They are copied here for the reader's convenient reference. 

The segments are presented in a different order than in the original report.  The first describes the Pentagon's common sense interpretation of the meaning of the requirement in Anthropic's contract that human handlers, not Anthropic's agents,  make the ultimate decision to launch drones and/or missiles that would kill adversaries. Then it presents what the editor has determined to be Dr. Amodei's interpretation of a human handler making the final decision, an interpretation that widely acclaimed psychological research has determined to be impossible for humans.

A. The Pentagon's interpretation of the Anthropic's. contract that humans make the final decision to kill enemy troops
  • Secretary Hegseth's marching orders
    Secretary Hegseth would not have inserted the agents into live battlefield operations unless it was judged to be highly successful during the extensive simulation tests.

    Nor would he have taken the human operators out of the loop.
    He would probably have given orders to the battlefield handlers to review the agent's recommendations. Then they should feel free to reject the recommendations if they determined that the agent's recommendations were incorrect. 

    -- For example, human handlers would check the viability the agent's recommendations. Handlers might notice that the agents recommended using more drones than the handlers' units currently possessed or the agent did not include the handlers' recent estimates of the number of noncombatants within firing range. Such mishaps might occur because the handlers had not updated these parameters in the data fed to the agents. The handlers would resubmit the corrected parameters so the agents could revise their recommendations. 

  • Murphy’s Law
    The Pentagon's operators of the drones and missiles were not just there to approve the agent's recommendations. They were also there as human backups in case the agents went offline due to hardware, network failures or whatever mishap. The Pentagon always has backups for its operations; it even has backups for its backups for its most critical ops.

    If the data was still available, the handlers would focus on the four factors they understood and make their own recommendations. Their recommendations would not be as effective as the agents, but doing nothing is always an option in real time operations, an option that can lead to disastrous results. 

  • More decision factors, less human understanding
    It was highly likely that Anthropic's autonomous agents would be far more effective than the Pentagon's current human controlled tech because the agents considered far more factors in its decisions.

    But the maximum number of factors that humans could consider at the same time was estimated to be no more than 9, as was reported "The Magical Number Seven, Plus or Minus Two: Some Limits on our Capacity for Processing Information", George A. Miller, Classics in the History of Psychology, 1956.

    More recently, that rough estimated maximum was reduced from 9  to 4 factors as reported in "The magical number 4 in short-term memory: a reconsideration of mental storage capacity", N. Cowan, Pub Med/NIH, February, 2001

    This point is crucial. If Dr. Amodei's expectation required humans to process more than 4 factors at a time in real time, then he was requiring human handlers to exceed human capabilities.  This was an impossible expectation, so his so-called contract specification was a fraud. The contract was invalid. Had the Pentagon known that this was his expectation, the Pentagon would not have signed the contract. 
B. Dr. Amodei's unique interpretation
  • Dr. Amodei seems to think that this situation is akin to a worker who has a colleague who is a much better writer. The colleague can write a page of the clearest prose in less than an hour, whereas the worker might require a day or two to produce something that isn't as well written. But the worker is be able to read and understand the colleague's clearly written prose in a few minutes.

    So too, an agent can identify the many factors that best identify enemy targets plus non-combatants in an area and how best to respond to the enemy's threats. Identifying the factors is the hard part. Surely the human handler will be able to understand the agent's recommended best response within a few moments, right? Absolutely wrong if the human handler must understand a recommendation in real time that involves more than 4 factors. Dr. Amodei is either ignorant of Miller's historic finding and Cowan's lower, more accurate assessment .. or he has no respect for psychological research.

No comments:

Post a Comment

Your comments will be greatly appreciated ... Or just click the "Like" button above the comments section if you enjoyed this blog note.