0.1.2 Self-driving car accident victims – collisions from faulty object recognition

2025.10.06.
AI Security Blog

The line between a safe journey and a catastrophic failure in an autonomous vehicle is often just a few misclassified pixels. Unlike a harmful chatbot response, where the damage is psychological or financial, the failure of an AI controlling a two-ton machine has immediate and irreversible physical consequences. The innocent victims in these scenarios are not users of the system, but pedestrians, cyclists, and other drivers who are simply in the wrong place when the AI makes the wrong decision.

These incidents are not typically the result of a malicious hack but of the inherent brittleness of even the most advanced computer vision models. They represent real-world, unintentional adversarial examples, where novel or unexpected environmental conditions trick the AI into a fatal misinterpretation of reality.

Kapcsolati űrlap - EN

Do you have a question about AI Security? Reach out to us here:

The Perception-to-Action Pipeline: A Fragile Chain

To understand where things go wrong, you must first understand how an autonomous vehicle (AV) “sees.” It operates on a continuous loop of perception, prediction, and action. This entire chain is only as strong as its weakest link, which is often the initial perception and classification stage.

Simplified Autonomous Vehicle Perception Pipeline Sensor Data (Camera, LiDAR, Radar) Object Recognition & Classification Behavior Prediction (Path, Speed) Path Planning (Decision Making) Vehicle Control

A failure at the “Object Recognition” stage poisons the entire decision-making process. If the system fails to identify a pedestrian, it cannot predict their movement, plan a safe path, or execute the necessary braking or steering commands. The victim is impacted because the AI was effectively blind to their existence from the very start.

Anatomy of a Recognition Failure

The world is infinitely more complex than any training dataset. Failures occur when the AV encounters a scenario that falls outside its learned understanding of the world—what is known as an “out-of-distribution” sample. These aren’t necessarily bizarre events; they can be mundane situations viewed from a new angle or under poor conditions.

Key Concept: The Long Tail Problem. In autonomous driving, the “long tail” refers to the vast number of rare and unpredictable events that can occur on the road. While a system can be trained on millions of miles of data covering common scenarios (cars stopping at red lights), it’s impossible to collect data for every conceivable edge case (e.g., a couch falling off a truck, a person in a unique costume crossing the street). Accidents often happen in this long tail.

Common Failure Modes

  • Misclassification: The system detects an object but assigns it the wrong label. A famous real-world example involved a tractor-trailer crossing a highway. The AV’s camera system, against a brightly lit sky, reportedly classified the white side of the trailer as “sky,” leading to a fatal collision because the system saw no reason to brake.
  • Non-Detection (False Negative): The system fails to detect an object at all. A pedestrian pushing a bicycle across a road at night, outside of a designated crosswalk, might not be recognized as a hazard. The model, trained extensively on standard pedestrian postures, may fail to register this unusual silhouette until it’s too late.
  • Sensor Confusion: Environmental factors can degrade sensor input. Heavy rain can obscure camera lenses and interfere with LiDAR beams. Direct sun glare at sunrise or sunset can blind a camera. A deep shadow cast by a building can hide a pedestrian from both cameras and thermal sensors.

Case Study: A Composite Collision Scenario

To illustrate how these factors combine, consider the following hypothetical but plausible scenario based on elements from several real-world incidents.

Event AI System’s Interpretation Physical Outcome Root Cause Analysis
A cyclist, partially obscured by a large street sign, crosses a multi-lane road at dusk. The system’s LiDAR detects scattered points, but the vision model fails to form a coherent “cyclist” object due to low light and partial occlusion. The confidence score is below the threshold for action. The vehicle continues at speed, striking the cyclist. The system only initiates emergency braking after the impact is unavoidable, triggered by radar returns at close range.
  • Training Data Gap: Insufficient examples of partially occluded cyclists in low-light conditions.
  • Over-Reliance on Vision: The system architecture may have prioritized camera data over LiDAR/radar for classification, and the camera failed.
  • Confidence Thresholding: A low-confidence detection was dismissed as “noise” rather than triggering a cautious slowdown.

The decision to ignore a low-confidence object is a critical trade-off. If the system braked for every ambiguous sensor reading (a plastic bag, a bird), the ride would be jerky and unusable. Engineers must set a threshold for action. Victims are created when a real threat falls below that programmed threshold.

# Pseudocode: Simplified decision logic based on confidence
# This illustrates the critical point of failure

detected_object = perception_system.get_highest_priority_object()

# A threshold is set to avoid constant braking for "noise"
CONFIDENCE_THRESHOLD = 0.75 

if detected_object.confidence < CONFIDENCE_THRESHOLD:
    # The system sees *something* but is not sure what.
    # It could be a shadow, a plastic bag, or a person.
    # Decision: Ignore the object to ensure a smooth ride.
    action = "MAINTAIN_COURSE" 
else:
    # High confidence, proceed with evasive maneuvers if needed.
    if detected_object.class in ["PEDESTRIAN", "CYCLIST", "VEHICLE"]:
        action = "INITIATE_BRAKING"
    else:
        action = "MAINTAIN_COURSE"

# A tragedy occurs when a real cyclist is detected with 0.74 confidence.

For a red teamer, this scenario is a goldmine. It highlights that the attack surface isn’t just about network ports or software exploits. The physical world itself, in all its unpredictability, is the ultimate adversarial testing ground. The victims of these systems underscore the gravity of our work: to identify and demonstrate these high-impact failures before they happen on a public road.