Hiding in Plain Sight: Secure Communication with Language Models

Author: Denis Avetisyan


A new approach to steganography leverages the power of large language models to conceal information with provable security guarantees.

Existing provable secure steganography schemes exhibit varying efficiencies in information utilization, a critical factor as maximizing this rate directly impacts the amount of data concealed within a carrier signal without raising suspicion.
Existing provable secure steganography schemes exhibit varying efficiencies in information utilization, a critical factor as maximizing this rate directly impacts the amount of data concealed within a carrier signal without raising suspicion.

This paper introduces a list decoding-based steganography scheme for large language models, optimizing embedding capacity and entropy utilization.

Existing provably secure steganography (PSS) schemes struggle to balance security with practical embedding capacity, particularly when integrated with modern large language models (LLMs). This work, ‘Provably Secure Steganography Based on List Decoding’, introduces a novel approach leveraging list decoding to maximize information hiding within LLM-generated text. By maintaining a set of plausible secret messages rather than seeking a single correct solution, our scheme substantially improves entropy utilization and achievable capacity. Does this list-decoding strategy represent a fundamental shift in the design of secure communication systems for the age of LLMs?


The Illusion of Concealment

Historically, the art of concealing messages – steganography – has frequently depended on making imperceptible changes to carrier texts or images. These alterations, however subtle, introduce statistical anomalies that increasingly sophisticated analytical tools can detect. Traditional methods, like modifying the least significant bits of an image or employing synonym substitution in text, create detectable patterns distinct from natural communication. This inherent vulnerability stems from the very act of hiding information; any deviation from typical language or image characteristics acts as a flag, signaling the potential presence of a covert message. Consequently, while seemingly innocuous to casual observation, these techniques are often susceptible to scrutiny from even moderately advanced detection algorithms, necessitating a shift toward more nuanced and adaptive concealment strategies.

The advent of powerful language models introduces a fascinating paradox to the field of covert communication. While these models can generate text so convincingly human that hidden messages could be embedded within seemingly innocuous content, their very sophistication also provides tools for detecting such manipulations. A cleverly disguised message, once undetectable to human scrutiny or simple algorithms, now faces scrutiny from systems trained to recognize subtle anomalies in linguistic patterns, stylistic inconsistencies, or semantic deviations. This necessitates a shift in steganographic techniques; rather than relying on minor alterations, future methods must focus on exploiting the inherent probabilistic nature of language models themselves, crafting messages that appear not just plausible, but statistically likely, within the model’s generated output – a delicate balance between concealment and natural language generation.

Truly secure communication in the digital age demands a departure from traditional steganography, which often relies on detectable alterations to existing data. Instead, research focuses on embedding information directly within the process of natural language generation itself. This approach leverages the inherent complexity and nuance of language models to create messages that are not hidden within text, but rather generated by the model with covert data woven into the statistical probabilities guiding word choice. The goal isn’t to disguise a message, but to make the very act of communication a carrier of concealed data, rendering detection significantly more difficult as distinguishing between intentional concealment and the natural stochasticity of language becomes a formidable challenge. This method effectively transforms language models into both the sender and the encryption key, offering a potential pathway to genuinely secure and undetectable communication.

Our embedding process iteratively filters candidate lists based on language model mappings and prefix/suffix matching, enabling efficient and uniquely decodable embedding of a secret message within the stegotext.
Our embedding process iteratively filters candidate lists based on language model mappings and prefix/suffix matching, enabling efficient and uniquely decodable embedding of a secret message within the stegotext.

Hiding in Plain Sight: A Probabilistic Approach

This steganographic method leverages the probabilistic nature of Large Language Model (LLM) text generation to conceal messages. Unlike traditional steganography which embeds data within existing, fixed content, this approach operates during the creation of the carrier text. A secret message is encoded by influencing the LLM’s output distribution during text generation, guiding it towards sequences that represent the hidden data. This is achieved by subtly biasing the LLM’s token selection probabilities, ensuring the generated text both conveys a plausible meaning and encodes the intended message. The system does not modify existing text, but rather generates new text specifically designed to contain the embedded information.

Unlike traditional steganographic techniques that rely on direct character substitution – where each letter of the secret message maps to a specific replacement character in the cover text – this method employs a non-instantaneous code. This means a single character of the hidden message can be represented by a sequence of characters within the generated text. This approach increases security by obscuring the direct relationship between the message and the cover text, making simple frequency analysis ineffective. The encoding process considers multiple possible sequences for each message character, and the selection of the final sequence is determined by a key and the context of previously encoded characters, adding a layer of complexity beyond simple one-to-one mappings.

The proposed steganographic method employs a ‘Candidate List’ approach to enhance decoding robustness and embedding capacity. Rather than relying on a single, definitive embedded message, the system maintains a list of plausible messages that satisfy the encoding criteria. This allows for successful message recovery even if some portions of the generated text are altered or lost during transmission, as the decoder can select the most probable message from the Candidate List based on remaining valid data. This technique demonstrably increases embedding capacity-the amount of information concealed within the carrier text-compared to existing steganographic methods that require exact matches for decoding, and mitigates the impact of noise or adversarial modifications to the generated text.

Our method's information utilization rate varies with secret bit length and candidate list size ([NN]) across three large language models-Llama3, Mistral3, and Qwen2.
Our method’s information utilization rate varies with secret bit length and candidate list size ([NN]) across three large language models-Llama3, Mistral3, and Qwen2.

The List Decoder: A Matter of Probabilities

List decoding is utilized as a method for efficiently identifying the correct message from a set of candidate messages. Unlike traditional decoding which aims to find a single best match, list decoding maintains a list of the most probable candidates throughout the decoding process. This approach mitigates the impact of errors introduced during transmission or encoding by considering multiple possibilities, ultimately increasing the probability of accurate information retrieval. The algorithm systematically evaluates each candidate, reducing the list size with each step until the correct message is identified or a predetermined list size is reached, offering a balance between computational complexity and decoding reliability.

The Alias Method is implemented to optimize the sampling phase of the decoding process. This technique precomputes and stores probabilities in a manner that allows for efficient selection of candidate messages, avoiding the computational overhead associated with traditional random sampling. By restructuring the probability distribution, the Alias Method achieves a consistent runtime performance comparable to Random Sampling, while significantly accelerating the overall decoding speed and improving the scalability of the scheme to larger datasets. This precomputation allows for constant-time sampling once completed, leading to a predictable and optimized decoding process.

To ensure accurate retrieval and mitigate decoding ambiguities, a ‘Validation Suffix’ is incorporated into each encoded message. This suffix functions as a unique identifier, guaranteeing that only one valid decoding path exists for each message, thereby preventing misinterpretation. Rigorous testing demonstrates this implementation achieves a decoding error rate of less than 8.695 x 10-7, indicating a highly reliable and precise decoding process. The validation suffix effectively minimizes the probability of incorrect message recovery during the decoding phase.

A System Built to Endure: Security and Performance

A rigorous security analysis has confirmed the method’s robust defense against a spectrum of potential attacks, affirming its capacity to safeguard concealed messages. This evaluation encompassed both known cryptographic attacks and newly devised strategies specifically targeting information hiding schemes. Results indicate a high degree of resistance to common threats, such as statistical analysis and modification attacks, due to the intricate embedding process and the method’s inherent ability to diffuse message alterations. The scheme’s resilience stems from its careful manipulation of data redundancy and its effective dispersion of concealed information throughout the carrier text, making it exceptionally difficult for adversaries to detect or extract the hidden content without compromising the integrity of the original message. This demonstrated security is crucial for applications demanding confidentiality and data protection.

A critical measure of any secure communication scheme is its reliability – specifically, the proportion of messages successfully and accurately retrieved. This system’s demonstrated ‘Success Rate’ consistently achieves high levels of message recovery across diverse testing scenarios, indicating a robust and practical design. Rigorous evaluation involved transmitting numerous concealed messages under varying conditions, including simulated noise and adversarial attempts to disrupt decoding. The consistently high percentage of correctly decoded messages confirms the scheme’s effectiveness not just in theoretical simulations, but also in potentially challenging real-world applications, establishing its dependability for secure data transmission and storage. This level of reliability is paramount for applications where data integrity and confidentiality are non-negotiable, such as financial transactions, medical records, and sensitive governmental communications.

This data concealment scheme distinguishes itself through efficient utilization of entropy within the generated text, a critical factor in both security and data capacity. By maximizing the entropy utilization rate, the method effectively packs more information into each character of the concealed message, thereby achieving a significantly higher embedding capacity compared to current state-of-the-art techniques. This isn’t simply about hiding more data, however; increased entropy inherently enhances security, as a higher degree of randomness makes the concealed message more resistant to statistical analysis and various attack vectors. The scheme’s success lies in its ability to approach the theoretical limits of information density, creating a robust and efficient system for secure communication and data storage.

The pursuit of provably secure steganography, as demonstrated in this work leveraging list decoding, feels predictably optimistic. One anticipates the inevitable edge cases production will unearth. The paper details maximizing entropy utilization and embedding capacity – elegant concepts, certainly. However, it’s a reasonable expectation that real-world data distributions will introduce unforeseen complexities, forcing adjustments to the theoretical bounds. As Vinton Cerf aptly stated, “Any sufficiently advanced technology is indistinguishable from magic… until it breaks.” The claim of ‘provable’ security, while mathematically sound within the defined parameters, will invariably face the pragmatic realities of implementation and adversarial attacks. It’s a novel approach, undoubtedly, but experience suggests that tomorrow’s security is simply today’s unresolved bug.

What Comes Next?

This pursuit of ‘provably secure’ steganography, built atop the shifting sands of large language models, feels… optimistic. It’s a lovely theoretical exercise, increasing embedding capacity and squeezing entropy, but production systems have a habit of discovering attack vectors the theorists missed. If a system crashes consistently, at least it’s predictable; security by obscurity is still security, after a fashion. The real question isn’t whether this list decoding scheme can be secure, but how quickly it will succumb to the inevitable adversarial pressure. It’s a new problem, sure, but the underlying principle remains: hiding data is hard, and anyone claiming otherwise is either selling something or hasn’t seen enough data.

The focus on LLMs is also… fleeting. The model landscape changes quarterly. This scheme is intricately tied to the specifics of list decoding within these models. A new architecture, a different training paradigm, and much of this work will be relegated to the footnotes of digital history. It’s a beautifully crafted solution to a problem that’s already evolving. We don’t write code – we leave notes for digital archaeologists.

Perhaps the most fruitful avenue isn’t greater security, but a more honest accounting of risk. Instead of chasing perfect concealment, future work might explore quantifiable undetectability – a probabilistic assessment of how long a hidden message will remain so. After all, ‘cloud-native’ just means the same mess, just more expensive, and the same applies to security promises.


Original article: https://arxiv.org/pdf/2604.21394.pdf

Contact the author: https://www.linkedin.com/in/avetisyan/

See also:

2026-04-24 12:05