The OODA loop—for observe, orient, decide, act—is a framework to understand
decision-making in adversarial situations. We apply the same framework to
artificial intelligence agents, who have to make their decisions with
untrustworthy observations and orientation. To solve this problem, we need new
systems of input, processing, and output integrity.
Many decades ago, U.S. Air Force Colonel John Boyd introduced the concept of the
“OODA loop,” for Observe, Orient, Decide, and Act. These are the four steps of
real-time continuous decision-making. Boyd developed it for fighter pilots, but
it’s long been applied in artificial intelligence (AI) and robotics. An AI
agent, like a pilot, executes the loop over and over, accomplishing its goals
iteratively within an ever-changing environment. This is Anthropic’s definition:
“Agents are models using tools in a loop.”...
Tag - integrity
Think of the Web as a digital territory with its own social contract. In 2014,
Tim Berners-Lee called for a “Magna Carta for the Web” to restore the balance of
power between individuals and institutions. This mirrors the original charter’s
purpose: ensuring that those who occupy a territory have a meaningful stake in
its governance.
Web 3.0—the distributed, decentralized Web of tomorrow—is finally poised to
change the Internet’s dynamic by returning ownership to data creators. This will
change many things about what’s often described as the “CIA triad” of ...
In this input integrity attack against an AI system, researchers were able to
fool AIOps tools:
> AIOps refers to the use of LLM-based agents to gather and analyze application
> telemetry, including system logs, performance metrics, traces, and alerts, to
> detect problems and then suggest or carry out corrective actions. The likes of
> Cisco have deployed AIops in a conversational interface that admins can use to
> prompt for information about system performance. Some AIOps tools can respond
> to such queries by automatically implementing fixes, or suggesting scripts
> that can address issues...
Here’s an interesting story about a failure being introduced by LLM-written
code. Specifically, the LLM was doing some code refactoring, and when it moved a
chunk of code from one file to another it changed a “break” to a “continue.”
That turned an error logging statement into an infinite loop, which crashed the
system.
This is an integrity failure. Specifically, it’s a failure of processing
integrity. And while we can think of particular patches that alleviate this
exact failure, the larger problem is much harder to solve.
Davi Ottenheimer ...