Why having “humans in the loop” in an AI war is an illusion

The availability of artificial intelligence for use in warfare is at the center of a legal battle between Anthropic and the Pentagon. This debate has become urgent, with AI playing a bigger role than ever before in the current conflict with Iran. AI is no longer just helping humans analyze intelligence. It is now an active player—generating targets in real time, controlling and coordinating missile interceptions, and guiding lethal swarms of autonomous drones.

Most of the public conversation regarding the use of AI-driven autonomous lethal weapons centers on how much humans should remain “in the loop.” Under the Pentagon’s current guidelines, human oversight supposedly provides accountability, context, and nuance while reducing the risk of hacking.

AI systems are opaque “black boxes”

But the debate over “humans in the loop” is a comforting distraction. The immediate danger is not that machines will act without human oversight; it is that human overseers have no idea what the machines are actually “thinking.” The Pentagon’s guidelines are fundamentally flawed because they rest on the dangerous assumption that humans understand how AI systems work.

Having studied intentions in the human brain for decades and in AI systems more recently, I can attest that state-of-the-art AI systems are essentially “black boxes.” We know the inputs and outputs, but the artificial “brain” processing them remains opaque. Even their creators cannot fully interpret them or understand how they work. And when AIs do provide reasons, they are not always trustworthy.

The illusion of human oversight in autonomous systems

In the debate over human oversight, a fundamental question is going unasked: Can we understand what an AI system intends to do before it acts?

Imagine an autonomous drone tasked with destroying an enemy munitions factory. The automated command and control system determines that the optimal target is a munitions storage building. It reports a 92% probability of mission success because secondary explosions of the munitions in the building will thoroughly destroy the facility. A human operator reviews the legitimate military objective, sees the high success rate, and approves the strike.

But what the operator does not know is that the AI system’s calculation included a hidden factor: Beyond devastating the munitions factory, the secondary explosions would also severely damage a nearby children’s hospital. The emergency response would then focus on the hospital, ensuring the factory burns down. To the AI, maximizing disruption in this way meets its given objective. But to a human, it is potentially committing a war crime by violating the rules regarding civilian life. 

Keeping a human in the loop may not provide the safeguard people imagine, because the human cannot know the AI’s intention before it acts. Advanced AI systems do not simply execute instructions; they interpret them. If operators fail to define their objectives carefully enough—a highly likely scenario in high-pressure situations—the “black box” system could be doing exactly what it was told and still not acting as humans intended.

This “intention gap” between AI systems and human operators is precisely why we hesitate to deploy frontier black-box AI in civilian health care or air traffic control, and why its integration into the workplace remains fraught—yet we are rushing to deploy it on the battlefield.

To make matters worse, if one side in a conflict deploys fully autonomous weapons, which operate at machine speed and scale, the pressure to remain competitive would push the other side to rely on such weapons too. This means the use of increasingly autonomous—and opaque—AI decision-making in war is only likely to grow.

The solution: Advance the science of AI intentions

The science of AI must comprise both building highly capable AI technology and understanding how this technology works. Huge advances have been made in developing and building more capable models, driven by record investments—forecast by Gartner to grow to around $2.5 trillion in 2026 alone. In contrast, the investment in understanding how the technology works has been minuscule.

We need a massive paradigm shift. Engineers are building increasingly capable systems. But understanding how these systems work is not just an engineering problem—it requires an interdisciplinary effort. We must build the tools to characterize, measure, and intervene in the intentions of AI agents before they act. We need to map the internal pathways of the neural networks that drive these agents so that we can build a true causal understanding of their decision-making, moving beyond merely observing inputs and outputs. 

A promising way forward is to combine techniques from mechanistic interpretability (breaking neural networks down into human-understandable components) with insights, tools, and models from the neuroscience of intentions. Another idea is to develop transparent, interpretable “auditor” AIs designed to monitor the behavior and emergent goals of more capable black-box systems in real time.  

Developing a better understanding of how AI functions will enable us to rely on AI systems for mission-critical applications. It will also make it easier to build more efficient, more capable, and safer systems.

Colleagues and I are exploring how ideas from neuroscience, cognitive science, and philosophy—fields that study how intentions arise in human decision-making—might help us understand the intentions of artificial systems. We must prioritize these kinds of interdisciplinary efforts, including collaborations between academia, government, and industry.

However, we need more than just academic exploration. The tech industry—and the philanthropists funding AI alignment, which strives to encode human values and goals into these models—must direct substantial investments toward interdisciplinary interpretability research. Furthermore, as the Pentagon pursues increasingly autonomous systems, Congress must mandate rigorous testing of AI systems’ intentions, not just their performance.

Until we achieve that, human oversight over AI may be more illusion than safeguard.

Uri Maoz is a cognitive and computational neuroscientist specializing in how the brain transforms intentions into actions. A professor at Chapman University with appointments at UCLA and Caltech, he leads an interdisciplinary initiative focused on understanding and measuring intentions in artificial intelligence systems (ai-intentions.org).

Stem Cell Editing Programs the Immune System to Make Own Therapeutic Proteins

For pathogens like HIV, malaria, and rapidly evolving influenza strains, coaxing the immune system to produce the rare, highly potent antibodies needed for protection has long been a scientific bottleneck. Vaccines can train B cells to evolve such broadly neutralizing antibodies, but only under ideal conditions—and only in a small fraction of people. Even attempts to genetically edit mature B cells produced responses that faded as the cells died out.

A team at the Rockefeller University has now taken a more upstream approach: programming hematopoietic stem and progenitor cells (HSPCs)—the source of all B lymphocytes—to carry permanent genetic instructions for therapeutic antibodies or other proteins. Because the immune system naturally amplifies rare, useful cells after vaccination, even a tiny number of edited stem cells can seed a durable, boostable immune response.

“The immune system is inefficient in that it produces a vast quantity of cells to protect itself,” said Harald Hartweger, a research assistant professor in Michel Nussenzweig’s Laboratory of Molecular Immunology. “We wanted to take advantage of the immune system’s ability to amplify useful, rare cells.”

The study, published in Science and titled “B lymphocyte protein factories produced by hematopoietic stem cell gene editing,” demonstrates that CRISPR‑edited HSPCs can mature into B cells that express engineered antibodies upon vaccination. A standard vaccination then acts as the trigger: antigen exposure drives those edited B cells to expand, differentiate into plasma cells, and secrete high titers of the inserted antibody that last long-term.

According to the paper, as few as ~7,000 edited HSPCs were enough to generate “high titers of long‑lasting protective or therapeutic antibodies and/or cargo proteins.” In mice engineered to produce a broadly neutralizing influenza antibody, this response was strong enough to protect against an otherwise lethal viral infection.

The platform proved unexpectedly versatile. Edited B cells could also secrete non‑antibody proteins, pointing to potential applications in genetic diseases. And by mixing HSPCs engineered with different antibody instructions, the researchers created immune systems capable of producing multiple antibodies simultaneously, an approach that could limit viral escape in HIV or other rapidly mutating pathogens. Human HSPCs edited using the same strategy produced functional human B cells in an immunodeficient mouse model, offering an early sign of translational feasibility.

“Our goal is to permanently impact the genome with a single injection, so that the body can make proteins of interest,” Hartweger said. “That protein could be an antibody that’s universally protective against HIV or influenza, but it could also be any therapeutic protein.”

The team is now moving toward preclinical testing in non‑human primates to evaluate protection against HIV and exploring whether similar strategies could be applied to T cells. The broader vision is a generalizable, long‑term protein‑production platform, one that could support treatments for infectious disease, protein deficiencies, autoimmunity, metabolic disorders, and cancer, according to Hartweger.

As Nussenzweig puts it, “The present study proposes a workaround for the antibody problem—a way of getting around the possibility that we may never get to a universal HIV vaccine, while still providing a promising, long‑lasting solution.”

The post Stem Cell Editing Programs the Immune System to Make Own Therapeutic Proteins appeared first on GEN – Genetic Engineering and Biotechnology News.

STAT+: HaloMD’s legal win highlights the difficulty of challenging arbitration decisions

Arbitration decisions, it turns out, are like cockroaches. They’re very hard to kill. 

It’s a long held truism in the legal world, and it was underscored this week when a federal judge shot down a health insurer’s lawsuit challenging No Surprises Act arbitration decisions. The ruling doesn’t bode well for other pending lawsuits challenging awards doled out under the 2020 law’s arbitration process, known as independent dispute resolution.  

“You can’t second guess the arbitrators,” said Chris Deacon, a health policy consultant and former lawyer. “That’s the whole point of arbitration.”

Continue to STAT+ to read the full story…

Trump taps former public health leader Erica Schwartz to run CDC

President Trump nominated Erica Schwartz on Thursday to be director of the Centers for Disease Control and Prevention, tapping a former public health leader for a position that has been filled mostly on a part-time or interim basis during the second Trump administration.

Schwartz was deputy surgeon general during the first Trump administration and spent much of her career in health roles in the U.S. military.

Read the rest…

A pancreatic cancer breakthrough, and new hope for an off-the-shelf CAR-T treatment

On this week’s episode of the Readout LOUD: a pancreatic cancer breakthrough and new hope for an off-the-shelf CAR-T treatment in lymphoma. 

Your favorite biotech podcasting crew is back to full strength this week, and we’re bringing you two newsy guest interviews. First, we’ll talk with Allogene Therapeutics Chief Medical Officer Zach Roberts about new study results that bolster the company’s efforts to develop an off-the-shelf CAR-T therapy for B-cell lymphoma, a type of blood cancer.

Read the rest…