Unmasking Deepfakes: A Hidden Code for Image Authenticity
![A method embeds content-related information into images via quantized multi-scale tokens-created with a VQ-VAE and constrained by watermark capacity [latex] |h| \leq |m| [/latex]-enabling recovery of deepfakes even after malicious manipulations like object removal or inpainting, achieved through decoding the watermarked image to extract hidden tokens and generate a deepfake localization map [latex] M_{loc} [/latex].](https://arxiv.org/html/2602.22759v1/2602.22759v1/picture/workflow.png)
Researchers have developed a novel watermarking technique that embeds a multi-scale ‘fingerprint’ within images, enabling both deepfake detection and faithful recovery of original content.
![A method embeds content-related information into images via quantized multi-scale tokens-created with a VQ-VAE and constrained by watermark capacity [latex] |h| \leq |m| [/latex]-enabling recovery of deepfakes even after malicious manipulations like object removal or inpainting, achieved through decoding the watermarked image to extract hidden tokens and generate a deepfake localization map [latex] M_{loc} [/latex].](https://arxiv.org/html/2602.22759v1/2602.22759v1/picture/workflow.png)
Researchers have developed a novel watermarking technique that embeds a multi-scale ‘fingerprint’ within images, enabling both deepfake detection and faithful recovery of original content.

A new network architecture tackles the challenges of removing reflections from single images, offering improved clarity and realism.

New research reveals that stylistic nuances present in text generated by language models are often lost when those models are used to create images, exposing a critical limitation in cross-modal AI.
![The system adapts to visual stimuli by first refining an EEG encoder through contrastive learning-better aligning brain activity with observed imagery-and then leveraging this enhanced signal to train a visual autoregressive transformer to predict subsequent visual scales, a process formalized by the next-scale prediction objective [latex]Equation 7[/latex] and guided by standard cross-entropy loss.](https://arxiv.org/html/2602.22555v1/2602.22555v1/x2.png)
Researchers have developed a new framework that accurately reconstructs visual information directly from brain activity recorded via electroencephalography (EEG).
![A reconstruction network, when burdened by inherent bias, falters in accurately representing signals - particularly those with low amplitude - but achieves natural homogeneity and improved performance when that bias is removed, as demonstrated by its ability to faithfully reconstruct both a signal [latex]\bm{x}[/latex] and a scaled version [latex]\frac{\bm{x}}{10}[/latex].](https://arxiv.org/html/2602.22279v1/2602.22279v1/x3.png)
A new self-supervised learning technique restores information lost when signals exceed measurement limits, opening doors for improved audio and image processing.

A new analysis reveals that diffusion language models struggle with truly parallel decoding due to inherent sequential dependencies learned from typical training data.

New research demonstrates how combining efficient data retrieval with intelligent reranking can dramatically improve the accuracy of product recommendations and information access in online shopping.

A new reinforcement learning framework automatically generates complex hardware designs to rigorously test formal verification tools.

Researchers have developed a framework to enhance the reasoning capabilities of large models, allowing them to tackle complex problems with improved efficiency and accuracy.

A new agentic framework uses artificial intelligence to actively investigate and verify the authenticity of video content, moving beyond passive detection.