Microsoft 'Whisper Leak' Attack Can Spy on Encrypted AI Chats

Microsoft Unveils 'Whisper Leak' Side-Channel Attack Capable of Identifying AI Chat Topics in Encrypted Traffic

MEDIUM
November 9, 2025
5m read
VulnerabilityThreat IntelligenceCloud Security

Related Entities

Organizations

Other

Whisper LeakOpenAI Mistral AI xAI DeepSeek

Full Report

Executive Summary

Microsoft has disclosed a new side-channel attack, named "Whisper Leak," that can compromise the privacy of AI chatbot sessions, even when protected by TLS encryption. The attack allows a passive network adversary to infer the topic of a user's conversation with a Large Language Model (LLM) by analyzing traffic patterns. The technique exploits the unique packet size and timing sequences generated by LLMs in 'streaming mode.' Microsoft's research demonstrated high accuracy in identifying conversation topics against major AI platforms, including those from OpenAI, Mistral AI, xAI, and DeepSeek. Following responsible disclosure, these providers have deployed mitigations. However, the discovery highlights a significant and previously underestimated privacy risk inherent in the design of popular AI services, posing a threat to individuals and enterprises using them for sensitive communications.


Vulnerability Details

The Whisper Leak attack is a traffic analysis-based side-channel vulnerability. It does not break the encryption itself but rather exploits metadata—specifically, the size and inter-arrival time of encrypted packets. The vulnerability arises from the 'streaming' nature of LLM responses, where the model generates and sends the answer token-by-token (word-by-word).

An attacker positioned to observe the network traffic (e.g., an ISP, a malicious Wi-Fi operator, or a nation-state actor) can perform the following steps:

  1. Capture Traffic: The adversary passively captures the encrypted TLS packets exchanged between the user and the LLM service.
  2. Extract Features: The attacker extracts a sequence of features from the traffic flow, primarily the size of each data packet and the time delay between them.
  3. Create a Fingerprint: This sequence of sizes and timings creates a unique 'fingerprint' for a given response. Since different topics and prompts elicit different response structures, these fingerprints can be distinct.
  4. Train a Classifier: The attacker builds a machine learning model, training it on traffic captures from known prompts and topics. This model learns to associate specific traffic fingerprints with specific conversation subjects.
  5. Infer Topics: Once trained, the model can be used to classify new, unknown encrypted conversations and infer their topics with high probability.

Microsoft's proof-of-concept achieved accuracy scores exceeding 98% in distinguishing between different topics, demonstrating the attack's viability.

Affected Systems

The attack is effective against LLMs that use a streaming response mechanism. The research specifically confirmed its effectiveness against models from:

  • OpenAI
  • Mistral AI
  • xAI
  • DeepSeek

Given the architectural similarities, it is highly probable that other streaming LLM services are also susceptible to this type of analysis.

Exploitation Status

This is a novel attack method disclosed by security researchers. There is no evidence of in-the-wild exploitation. However, the low barrier to entry for a passive network adversary means this technique could be adopted by threat actors. Major providers (OpenAI, Mistral, xAI) have already implemented mitigations after being notified by Microsoft.

Impact Assessment

The primary impact of Whisper Leak is a severe loss of privacy. While the exact content of the conversation remains encrypted, an attacker can determine the subject matter. This could be used to:

  • Identify Dissidents: A nation-state actor could monitor network traffic to identify individuals researching or discussing politically sensitive topics.
  • Corporate Espionage: An attacker could determine if a competitor's employees are using AI to research a new product, patent, or merger and acquisition strategy.
  • Target Individuals: Information about a user's interest in topics like 'financial trouble,' 'gambling addiction,' or specific medical conditions could be used for blackmail, targeted phishing, or social engineering.

For industries like healthcare, law, and finance, where AI is being integrated to handle sensitive client data, this vulnerability poses a significant compliance and ethical risk.

Cyber Observables for Detection

Detecting a passive Whisper Leak attack is extremely difficult, as the attacker is only observing traffic, not modifying it. Detection would focus on identifying the data collection phase.

Type Value Description
network_traffic_pattern Sustained traffic capture from specific IP ranges Large-scale, non-intrusive packet capture targeting IP ranges of known LLM providers.
api_endpoint api.openai.com, api.mistral.ai Monitoring for large volumes of metadata collection associated with traffic to and from these endpoints.
log_source NetFlow, IPFIX, sFlow Analysis of flow records might reveal unusual patterns of observation by specific network nodes.

Detection Methods

Defending against this is more effective through mitigation than detection. However, organizations could use internal traffic analysis to baseline their own LLM usage. This aligns with D3FEND's D3-NTA: Network Traffic Analysis. By establishing a baseline of normal traffic patterns to AI services, it might be possible to detect external analysis if it involves any active probing, though this is unlikely for a purely passive attack.

Remediation Steps

The primary mitigation for this type of side-channel attack is traffic shaping, which is implemented by the service provider. This involves adding padding to packets to obscure their true size and introducing random delays to mask the timing.

  • Padding: The LLM provider can add random amounts of padding data to each packet, making the packet sizes less predictable. This makes it difficult for an attacker's classifier to rely on size as a feature. This is a form of D3FEND's D3-ACH: Application Configuration Hardening.
  • Randomized Timing: Introducing small, random delays between sending packets can disrupt the timing-based fingerprints that the attack relies on.

As an end-user or enterprise, the best course of action is to:

  1. Use Reputable Providers: Ensure your chosen AI provider is aware of this research and has implemented mitigations. Microsoft confirmed that OpenAI, Mistral, and xAI have already done so.
  2. VPN/Proxy Usage: While not a complete solution, routing traffic through a trusted VPN can obscure the traffic from a local network adversary or ISP, though the VPN provider itself could still perform the attack.
  3. Assume Limited Privacy: Users should be educated that even with encryption, metadata can leak information. Avoid discussing highly sensitive or confidential topics on public AI platforms until these mitigations are universally adopted and proven effective.

Timeline of Events

1
November 9, 2025
This article was published

MITRE ATT&CK Mitigations

While the attack works on encrypted data, using a trusted VPN can shift the adversary from the local network/ISP to the VPN provider, adding a layer of defense.

Mapped D3FEND Techniques:

LLM providers can mitigate this by implementing traffic shaping (padding and random timing) on the server side.

Mapped D3FEND Techniques:

D3FEND Defensive Countermeasures

While the Whisper Leak attack itself uses Network Traffic Analysis, defenders can employ the same technique for defensive purposes. Organizations should establish a baseline of network traffic patterns for all communications with external AI services. By monitoring the volume, frequency, and metadata of traffic from internal users to providers like OpenAI and Mistral, security teams can create a profile of 'normal' usage. While this won't stop a passive eavesdropper, it can help detect anomalous data flows that might indicate a more active phase of an attack or large-scale data collection. For example, an internal system suddenly sending an abnormally high volume of queries to an LLM could be a sign of an automated script attempting to generate training data for a Whisper Leak-style classifier. This proactive monitoring provides visibility and a foundation for detecting deviations from expected behavior.

This countermeasure is primarily for the AI service providers. To defend against Whisper Leak, providers like OpenAI and xAI must harden their application servers to perform traffic shaping. This involves two key actions: 1) Packet Padding: Adding a variable amount of random data to each outbound packet to normalize their sizes. This obfuscates the true payload size, which is a critical feature for the attack's machine learning model. 2) Jitter Injection: Introducing small, random delays between the transmission of packets in a streaming response. This disrupts the inter-arrival time sequence, another key feature for the classifier. By manipulating these two traffic characteristics, the provider effectively adds noise to the data, significantly degrading the accuracy of any side-channel analysis and rendering the Whisper Leak attack impractical. Enterprises consuming these services should seek assurance from their vendors that such hardening is in place.

Sources & References

Breaking News – Cyber Threats – 2025-11-08 12:00 PST
XLoggs (xloggs.com) November 8, 2025
AI chat privacy at risk: Microsoft details Whisper Leak side-channel attack
Security Affairs (securityaffairs.com) November 9, 2025

Article Author

Jason Gomes

Jason Gomes

• Cybersecurity Practitioner

Cybersecurity professional with over 10 years of specialized experience in security operations, threat intelligence, incident response, and security automation. Expertise spans SOAR/XSOAR orchestration, threat intelligence platforms, SIEM/UEBA analytics, and building cyber fusion centers. Background includes technical enablement, solution architecture for enterprise and government clients, and implementing security automation workflows across IR, TIP, and SOC use cases.

Threat Intelligence & AnalysisSecurity Orchestration (SOAR/XSOAR)Incident Response & Digital ForensicsSecurity Operations Center (SOC)SIEM & Security AnalyticsCyber Fusion & Threat SharingSecurity Automation & IntegrationManaged Detection & Response (MDR)

Tags

Side-Channel AttackWhisper LeakAI SecurityLLMPrivacyEncryptionMicrosoft

📢 Share This Article

Help others stay informed about cybersecurity threats

Continue Reading