The Tell-Tale Fade: How AI Text Reveals Its Origins

Author: Denis Avetisyan


A new analysis reveals a predictable pattern in how artificial intelligence generates text, offering a robust method for distinguishing it from human writing.

Analysis of text generated by GPT-J-6B on the EvoBench dataset reveals a characteristic pattern of temporal dynamics, where artificial text consistently exhibits a steeper decay in both derivative and local standard deviation features-highlighted by pronounced separation in the latter half of the sequence-compared to human-authored text, suggesting a fundamental difference in the statistical properties of generated versus natural language content.
Analysis of text generated by GPT-J-6B on the EvoBench dataset reveals a characteristic pattern of temporal dynamics, where artificial text consistently exhibits a steeper decay in both derivative and local standard deviation features-highlighted by pronounced separation in the latter half of the sequence-compared to human-authored text, suggesting a fundamental difference in the statistical properties of generated versus natural language content.

Late-stage volatility decay – a reduction in statistical variability at the end of generated sequences – serves as a reliable signature for AI-generated text detection.

Despite advances in identifying AI-generated text, current zero-shot detection methods often overlook crucial temporal dynamics inherent in how these sequences are created. Our work, ‘When AI Settles Down: Late-Stage Stability as a Signature of AI-Generated Text Detection’, reveals a phenomenon we term ‘Late-Stage Volatility Decay’: a characteristic stabilization of token-level statistical fluctuations as AI text generation progresses. This divergence from the sustained variability of human writing-most pronounced in the latter half of sequences-forms the basis for two novel detection features. Could analyzing this ‘settling down’ of AI offer a robust and complementary approach to reliably distinguishing machine-authored content?


The Erosion of Textual Authenticity: A Mathematical Imperative

The advent of large language models has ushered in an era where the distinction between human-written and machine-generated text is becoming increasingly nebulous. These models, trained on massive datasets of text and code, now possess the capacity to generate content that mimics human writing styles with remarkable fidelity – crafting everything from compelling narratives and informative articles to nuanced poetry and even functional code. This capability isn’t simply about replicating grammar and syntax; it extends to adopting specific tones, perspectives, and even the subtle idiosyncrasies that characterize individual authors. Consequently, determining the true origin of a text – whether penned by a person or produced by an algorithm – presents a significant and evolving challenge, impacting fields ranging from academic integrity and journalism to content creation and digital security. The sophistication of these models suggests a future where authorship itself may require redefinition, as the lines between human creativity and artificial intelligence continue to blur.

As artificial intelligence models generate increasingly sophisticated text, conventional methods for distinguishing between human and machine authorship are proving inadequate. Existing detection tools, often relying on identifying predictable patterns or stylistic inconsistencies, are easily bypassed by the nuanced outputs of modern large language models. This escalating arms race between generation and detection necessitates a shift towards more robust solutions – techniques that move beyond surface-level analysis and delve into the underlying statistical properties of text. Researchers are now exploring methods that focus on identifying subtle anomalies in linguistic choices, perplexity, and burstiness – characteristics difficult for AI to perfectly replicate – but maintaining accuracy and avoiding false positives remains a significant challenge. The development of truly reliable detection systems is crucial, not only for academic integrity and content authenticity, but also for safeguarding against the spread of misinformation and maintaining trust in online information.

Detecting AI-generated text isn’t simply about identifying grammatical errors or nonsensical phrasing; the true difficulty resides in pinpointing statistical deviations from authentic human writing. Large language models generate text by predicting the most probable sequence of words, and while increasingly sophisticated, this process leaves subtle fingerprints in the text’s statistical properties – patterns in word choice, sentence structure, and even punctuation usage. However, natural language is inherently complex and variable; human writing itself contains a wide range of stylistic choices and statistical fluctuations. This inherent noise makes isolating the telltale signs of machine generation exceptionally challenging, demanding detection methods that can differentiate between the genuine randomness of human expression and the algorithmic patterns embedded within AI-authored content. The task, therefore, requires a nuanced understanding of linguistic statistics and the development of algorithms capable of discerning exceedingly subtle anomalies within a sea of natural variation.

Analysis of frontier models using MIRDGE reveals a consistent decay in log probability during late-stage generation, even for reasoning-focused models like DeepSeek-R1 and GPT-o3-mini.
Analysis of frontier models using MIRDGE reveals a consistent decay in log probability during late-stage generation, even for reasoning-focused models like DeepSeek-R1 and GPT-o3-mini.

Log Probability as a Determinant: Quantifying Confidence

Log probability sequences form the core of our analytical method, quantifying the confidence assigned to predicted tokens by a language model. Specifically, each token generated receives a log probability score – the natural logarithm of the probability assigned to that token given the preceding text. log(P(token_i | token_{i-1}, ..., token_1)). Analyzing the sequence of these log probabilities, rather than the probabilities themselves, provides numerical stability and facilitates differentiation for subsequent analysis. Higher log probability values indicate greater model certainty in its predictions, while lower values suggest increased uncertainty or a less predictable continuation of the text. This sequence therefore serves as a quantifiable measure of the model’s internal confidence throughout text generation.

Local volatility, in the context of AI text detection, refers to the degree of variation observed in the log probabilities assigned to sequential tokens by a language model. Log Probability reflects the model’s confidence in its predictions; a higher log probability indicates greater confidence. Fluctuations in these log probabilities – the local volatility – serve as a differentiating factor between human and AI-generated text. Human writing typically exhibits a more natural and varied range of log probabilities due to stylistic choices and nuanced expression, resulting in higher local volatility. Conversely, AI models, particularly those optimized for coherence and predictability, tend to produce text with comparatively reduced variability in log probabilities, and therefore, lower local volatility.

Analysis of the derivative of log probability – the rate of change in a language model’s confidence in its predictions – reveals a statistically significant distinction between AI-generated and human-written text. Quantitative evaluation on the EvoBench and MAGE datasets demonstrates that AI-generated text exhibits a 32% and 24% reduction, respectively, in this derivative metric compared to human text. This indicates that language models tend to maintain a more consistent, and therefore less variable, prediction confidence throughout a sequence than human writers, resulting in a lower rate of change in log probability. This difference in \frac{d}{dx} log(P(x)) provides a measurable characteristic for differentiating between the two text sources.

Analysis of <span class="katex-eq" data-katex-display="false">\log</span> probability features reveals that AI-generated text exhibits greater temporal volatility and change rates compared to human text on both EvoBench and MAGE, with a widening divergence-particularly in the latter half-indicated by the shaded regions and Δ values representing the AI-Human difference.
Analysis of \log probability features reveals that AI-generated text exhibits greater temporal volatility and change rates compared to human text on both EvoBench and MAGE, with a widening divergence-particularly in the latter half-indicated by the shaded regions and Δ values representing the AI-Human difference.

Late-Stage Volatility Decay: A Predictable Pattern

Late-stage volatility decay in AI-generated text refers to the observed reduction in the variability of predicted token probabilities as the sequence length increases. Specifically, research indicates that while initial tokens in a generated sequence exhibit a broader range of possible next-token predictions, the probability distribution increasingly concentrates on a smaller set of likely tokens further into the sequence. This phenomenon is characterized by a decrease in entropy and suggests that, as AI models generate longer texts, they tend to become more deterministic in their predictions, favoring high-probability continuations over more diverse or creative options. This contrasts with human writing, where variability in word choice and sentence structure tends to remain more consistent throughout a text.

Quantitative analysis reveals a consistent decrease in variability within AI-generated text sequences. Specifically, measurements of Derivative Dispersion and Local Volatility demonstrate a 31% reduction in local standard deviation when evaluating text generated on the EvoBench dataset, and a 25% reduction on the MAGE dataset. These features effectively capture the narrowing range of probable tokens as the sequence length increases, indicating a decreasing capacity for diverse expression in later portions of AI-authored content. This reduction in local standard deviation serves as a key metric for characterizing Late-Stage Volatility Decay.

The Temporal Stability Detection method leverages quantified features of late-stage volatility decay – specifically Derivative Dispersion and Local Volatility – to differentiate between human and AI-generated text. Analysis across 13 frontier language models demonstrates a consistent decay ratio of 1.8 to 3.0, indicating a significantly more pronounced reduction in token probability variability over sequence length in machine-authored text compared to human writing. This ratio represents the magnitude of volatility decay observed, effectively serving as a distinguishing characteristic for authorship identification; higher ratios consistently correlate with AI-generated content, while human text exhibits a lower degree of this temporal stability decline.

Analysis of token-level metrics on the MAGE benchmark reveals increasing discrepancies between human and AI performance in derivative and volatility-especially for log probability and sampling discrepancy-during the latter half of sequences.
Analysis of token-level metrics on the MAGE benchmark reveals increasing discrepancies between human and AI performance in derivative and volatility-especially for log probability and sampling discrepancy-during the latter half of sequences.

Validation and Practicality: Ensuring Robustness

Rigorous evaluation of the proposed methodology centered on established benchmark datasets, specifically MAGE and EvoBench, to comprehensively assess its performance characteristics. These datasets facilitated testing across a wide spectrum of text generators, ranging from established models to those continually evolving with advancements in large language models (LLMs). This diverse evaluation strategy ensured the method’s ability to maintain consistent accuracy not only with current LLMs but also as new and more sophisticated generators emerge, thereby demonstrating its adaptability and long-term viability in a rapidly changing technological landscape. The selection of these benchmarks provided a standardized and challenging environment for quantifying the method’s robustness and generalization capabilities.

Evaluations using established benchmark datasets reveal the consistent performance of the proposed method, termed TSD, in discerning machine-generated text. Specifically, TSD achieves an Area Under the Receiver Operating Characteristic curve (AUROC) of 83.36% on the EvoBench dataset, which assesses generalization across evolving large language models. Further demonstrating its reliability, the method also attains a 75.20% AUROC on the MAGE benchmark, designed to test robustness against diverse text generators. These results collectively indicate that TSD not only maintains high accuracy but also exhibits consistent behavior across varied and challenging scenarios, suggesting its potential for reliable deployment in real-world applications.

To address the need for real-time detection capabilities, a computationally efficient method, Fast-DetectGPT, was developed. This approach leverages the principle of Sampling Discrepancy – examining the statistical differences between text generated by language models and authentic human writing – but streamlines the process for significantly faster analysis. By focusing on key statistical divergences rather than exhaustive comparisons, Fast-DetectGPT achieves a practical balance between accuracy and speed, enabling its deployment in applications requiring immediate assessment of text authenticity. This enhanced practicality broadens the scope of the research, moving beyond purely analytical evaluations to offer a tool suitable for dynamic, real-world scenarios where prompt detection is critical.

Task success rate (<span class="katex-eq" data-katex-display="false">TSD</span>) fluctuates with initial position across datasets, with EvoBench exhibiting a preference for starting closer to the goal and MAGE demonstrating greater robustness to initial position.
Task success rate (TSD) fluctuates with initial position across datasets, with EvoBench exhibiting a preference for starting closer to the goal and MAGE demonstrating greater robustness to initial position.

The pursuit of identifying AI-generated text hinges on discerning patterns beyond superficial fluency. This research illuminates a compelling indicator – ‘Late-Stage Volatility Decay’ – a demonstrable reduction in statistical variability as sequences progress. It suggests a fundamental difference in how algorithms and humans construct language, moving beyond simple feature detection. As Bertrand Russell observed, “The point of the experiment is to distinguish between what is known and what is unknown.” This work doesn’t simply detect AI text; it defines a boundary, a measurable characteristic, revealing the inherent predictability of algorithmic generation-a predictability not mirrored in natural human expression. The elegance lies in identifying this consistent decay, a mathematically demonstrable signal within the generated text.

Beyond the Horizon

The observation of ‘Late-Stage Volatility Decay’ in autoregressive models is not merely a detection heuristic; it is a symptom. A symptom of the fundamental constraints imposed upon probabilistic sequence generation. Current work provides a compelling signal, but the underlying mathematical reason why this decay consistently manifests remains largely unexplored. A proof of convergence towards predictable token distributions at sequence termini would elevate this from empirical observation to demonstrable principle. The field should not settle for increasingly complex surrogate models trained to recognize this decay, but rather seek the axioms that necessitate it.

Furthermore, the limitations of this approach are clear. Any attempt to ‘vaccinate’ language models against this signal – by, for instance, introducing artificial stochasticity at late stages – will undoubtedly emerge. This creates an arms race, and the history of machine learning is littered with such futile endeavors. The more fruitful path lies in acknowledging that perfect mimicry of human writing is not the goal, and may be fundamentally unattainable. Instead, research should focus on identifying and characterizing the unique mathematical signatures of all artificial text generation processes, irrespective of their superficial resemblance to natural language.

Ultimately, the pursuit of ‘AI detection’ may be a misdirection. A more elegant solution resides in building generative models capable of certifiable properties – models where the statistical characteristics of the generated text are not merely observed, but mathematically guaranteed. To strive for a provably correct generator is a far more ambitious – and intellectually satisfying – undertaking than perfecting the art of detection.


Original article: https://arxiv.org/pdf/2601.04833.pdf

Contact the author: https://www.linkedin.com/in/avetisyan/

See also:

2026-01-09 16:34