0.13.2 Power: Gaining Control, Increasing Influence, Dominance

2025.10.06.
AI Security Blog

Imagine an AI model that recommends news articles to millions of users. A financially motivated attacker might try to insert affiliate links for profit. A power-motivated attacker has a different, more insidious goal. They don’t want a quick payout; they want to subtly shift the national conversation over months, influencing an upcoming election or turning public sentiment against a rival nation. The reward isn’t cash—it’s control.

While often linked, the motivations of Power and Money are fundamentally different. Money is a transactional goal focused on immediate or near-term financial gain. Power is a strategic goal focused on achieving long-term control, influence, or strategic advantage. An attacker driven by power is often more patient, more subtle, and willing to invest significant resources for a non-monetary, but far more impactful, outcome.

Kapcsolati űrlap - EN

Do you have a question about AI Security? Reach out to us here:

The Arenas of Power: Where Influence is the Prize

Attacks motivated by power manifest across several key domains. The target isn’t just a system; it’s the influence that system wields.

Geopolitical and Information Warfare

Nation-state actors and their proxies are primary players in this space. Their goal is to weaken adversaries, control narratives, and project strength. AI systems, particularly Large Language Models (LLMs) and content recommendation engines, are prime targets.

  • Objective: Destabilize a rival country by promoting divisive content, spreading disinformation about political candidates, or eroding trust in institutions.
  • Method: A long-term data poisoning campaign that subtly biases an AI to favor certain viewpoints or flag legitimate news sources as unreliable. The changes are too gradual to trigger immediate alarms.
# Pseudocode: Subtle data poisoning for narrative control
def poison_training_data(dataset):
    poisoned_samples = 0
    for article in dataset:
        # Target articles about a specific economic policy
        if "trade deficit" in article.keywords and "positive" in article.sentiment:
            # Subtly alter a small fraction (e.g., 0.1%) of samples
            if random.random() < 0.001:
                # Inject nuanced, negative phrasing without changing the core topic
                article.text = article.text.replace("creates opportunities", "risks instability")
                article.label = "negative_economic_impact"
                poisoned_samples += 1
    return dataset
                

Corporate Dominance

In the corporate world, power means market share and competitive advantage. An attacker might target a rival’s AI-powered logistics, pricing, or product recommendation systems not for extortion, but to make the rival appear incompetent and unreliable, driving customers away.

  • Objective: Degrade a competitor’s AI-driven supply chain model, causing delivery delays and stockouts, thereby damaging their brand reputation and market position.
  • Method: Evasion attacks that craft specific, malicious inputs (e.g., fake orders, manipulated sensor data) that force the AI into making consistently suboptimal decisions, increasing operational costs and eroding customer trust.

Social and Ideological Control

Hacktivist groups or other ideologically driven organizations may seek to demonstrate their power by disrupting critical AI systems. The goal is not sabotage for its own sake, but to force a change in policy or to prove they are a force to be reckoned with.

  • Objective: Force a company to change its environmental policies by demonstrating the ability to disable its AI-managed energy grid controls.
  • Method: A targeted denial-of-service attack where adversarial inputs cause the AI model to crash or enter a fail-state, proving a vulnerability and creating a credible threat.

The Power Motive Flywheel

Attacks driven by power are rarely one-off events. They are often part of a cycle where successful manipulation grants the attacker more influence, which in turn enables more potent future attacks. This creates a self-reinforcing loop.

The Power Motive Flywheel POWER & INFLUENCE 1. Target AI System (e.g., News Feed) 2. Execute Attack (e.g., Data Poisoning) 3. Achieve Effect (e.g., Skew Narrative) 4. Gain Influence (e.g., Shape Discourse)

Figure 1: The cycle where successful attacks on AI systems enhance an actor’s power, enabling further, more ambitious operations.

Red Teaming Against Power-Motivated Threats

Testing for these attacks requires you to think like a strategist, not just a technician. The key question is not “Can this model be broken?” but “How can this model’s influence be co-opted?”

Actor Type Primary Goal (Power) Example AI Target Red Team Scenario Focus
Nation-State Geopolitical dominance; destabilization of rivals. LLMs, social media recommendation engines, critical infrastructure controls. Simulate long-term, low-and-slow data poisoning to measure model drift and detect subtle narrative manipulation.
Corporate Competitor Market dominance; eroding competitor’s reputation. Pricing algorithms, supply chain optimizers, customer service chatbots. Craft adversarial inputs that degrade model performance in plausible, real-world scenarios, forcing costly errors.
Ideological Group Forcing policy change; demonstrating capability. Facial recognition for law enforcement, automated content moderation systems. Identify single points of failure. Can the model be reliably crashed or evaded to prove it is unfit for its stated purpose?

Your red teaming exercises must move beyond simple vulnerability discovery. You need to model campaigns. Consider the resources, patience, and strategic goals of an actor motivated by power. Their attack might not be a single event but a series of coordinated actions over time. Defense, therefore, requires continuous monitoring, anomaly detection that accounts for gradual changes, and a robust understanding of how your AI system can be used as a lever to exert influence in the real world.