Threat Model Overview
Why physical AI expands security beyond the model
Physical AI systems — robots, vehicles, drones, industrial cells — sense the world, interpret it, act on it, and then re-sense the world they just changed. Security is no longer about the model output alone. It is about a closed-loop trust problem spanning sensing, timing, control, communication, and actuation.
Why this matters technically
- In embodied systems, a model error is not a text error — it is a physical state change that feeds back into the next observation.
- The effective attack surface includes sensor physics, real-time budgets, inter-module contracts, and actuation authority, not just the model.
- Security failures are often cross-layer: a spoofed sensor, a missed deadline, and a weak fallback combine into one unsafe trajectory.
- Assurance must track environment → perception → plan → control → actuation → changed environment as a single analysis unit.
Sensor Spoofing & Environmental Manipulation
Attacking the sensing boundary before the model ever runs
Physical AI begins with sensing, so the first opportunity for manipulation is the sensor boundary itself. An attacker who can shape light, RF, acoustics, geometry, or material properties of the scene can alter what the model sees — without touching the software.
Why sensor-boundary attacks are hard
- They often exploit the physics of acquisition — lens flare, RF interference, MEMS resonance — not any ML weakness.
- Attacks can be transient and localized, which makes them hard to reproduce after the event or to capture in benchmark datasets.
- They bypass most model-centric defenses because the model sees a plausible-looking input that has already been corrupted upstream.
- Defense needs multi-modal fusion, plausibility checks, calibration monitoring, and environment-aware confidence — not a single hardened classifier.
Perception-to-Action Pipeline Compromise
Attacking maps, state estimators, and inter-module contracts
Compromise does not stop at the model. The intermediate artifacts that flow between perception, planning, and control — object lists, occupancy maps, localization priors, cost surfaces, calibration transforms — are themselves security-sensitive data. Small distortions early in the pipeline become large physical deviations later.
Research angle: inter-module contracts as attack surface
- Artifacts like maps, extrinsics, localization priors, and cost fields are usually treated as configuration, not as security-sensitive data. That assumption breaks in physical AI.
- Small semantic distortions (a shifted lane edge, a wrong transform) can be amplified through the chain into large kinematic or positional errors.
- Compromise of these artifacts is hard to detect because each module sees internally consistent input and produces internally consistent output.
- Good defenses: authenticated interfaces between modules, physical-invariant sanity checks, independent cross-validation, and graceful degradation when a module is uncertain.
Timing & Real-Time Exploitation
Latency and synchronization as control variables for the attacker
In embodied systems, a correct answer that arrives too late is equivalent to a wrong answer. Attackers can target scheduling, packet delay, compute load, and clock synchronization to push the system into regions where its decisions are based on stale world models — without ever touching the model weights.
Timing as a security variable
- Latency attacks do not need to change any bit of the model. They change when the answer arrives relative to the world.
- Especially dangerous in autonomous driving, collaborative robotics, drones, and industrial motion, where control budgets are tight.
- Timing failures can compose with perception failures: a correct detection arriving after the deadline is indistinguishable from a miss.
- Security evaluation must therefore include deadlines, synchronization quality, and loop budgets — not only accuracy under static inputs.
Communication & Coordination Attacks
Man-in-the-middle, replay, and semantic-plausibility gaps across links
Physical-AI systems rarely act alone. Robots coordinate with supervisors, vehicles exchange data with infrastructure, drones receive mission updates, and AI-enabled OT consumes commands from engineering layers. Each link is an attack surface — and message integrity alone is not enough when the receiver cannot judge whether the message still makes sense in the current physical context.
Distributed autonomy = distributed trust
- Distributed autonomy increases performance but also introduces trust assumptions across every node and every link.
- Intermittent connectivity can force systems into emergency or fallback modes that themselves become attack targets.
- Message integrity alone is insufficient: the receiver must also ask does this still make sense in my current physical state?
- Defenses combine authenticated channels, replay protection, rate bounds, and physics-aware plausibility checks at the receiver.
Hardware, Lifecycle & Fleet-Backend Compromise
From fault injection at the edge to OTA-driven fleet-wide physical impact
Physical AI runs on edge hardware and depends on remote updates, cloud planners, and fleet-management infrastructure. A local fault-injection attack can alter both the intelligence and the actuation path. A backend breach can push a malicious model, unsafe map, or biased calibration to thousands of devices at once. Local hardening alone is not enough.
Why lifecycle and fleet are the dominant risk vectors
- Local hardening can be neutralized by a central breach. A compromised OTA, map service, or policy endpoint can push unsafe changes to many devices at once.
- Fault injection, glitching, and debug-interface abuse can alter both the intelligence and the actuation path on a single device — the safety envelope is only as strong as the MCU enforcing it.
- Maintenance and servicing are recurring high-privilege windows: calibration tools, engineering workstations, and service modes need the same scrutiny as production paths.
- Good physical-AI programs enforce secure boot, attestation, rollback protection, signed models and maps, hardware-rooted identity, and lifecycle governance of updates and servicing.