Tuna-2: Pixel Embeddings Beat Vision Encoders for Multimodal Understanding and Generation

1Meta AI 2The University of Hong Kong 3University of Waterloo

* Joint first authors, listed alphabetically by last name

arXiv Paper Code Tuna CVPR'26 Highlight
Tuna-2 Architecture
Figure 1. Evolution of the Tuna-2 architecture and multimodal performance comparison. We simplify Tuna (Liu et al., 2025) by progressively stripping away its visual encoding components. By removing the VAE, we first derive Tuna-R, a pixel-space UMM that relies solely on a representation encoder. Tuna-2 further streamlines the design by bypassing the representation encoder entirely, utilizing direct patch embedding layers for raw image inputs. Tuna-2 using pixel embeddings outperforms both Tuna-R and Tuna across a diverse suite of multimodal benchmarks.
Tuna-2 generates diverse, high-quality images
Figure 2. While being completely encoder-free, Tuna-2 is capable of performing high-fidelity text-to-image generation and image editing.
01

Abstract

Unified multimodal models typically rely on pretrained vision encoders and use separate visual representations for understanding and generation, creating misalignment between the two tasks and preventing fully end-to-end optimization from raw pixels. We introduce Tuna-2, a native unified multimodal model that performs visual understanding and generation directly based on pixel embeddings. Tuna-2 drastically simplifies the model architecture by employing simple patch embedding layers to encode visual input, completely discarding the modular vision encoder designs such as the VAE or the representation encoder. Experiments show that Tuna-2 achieves state-of-the-art performance in multimodal benchmarks, demonstrating that unified pixel-space modelling can fully compete with latent-space approaches for high-quality image generation. Moreover, while the encoder-based variant converges faster in early pretraining, Tuna-2's encoder-free design achieves stronger multimodal understanding at scale, particularly on tasks requiring fine-grained visual perception. These results show that pretrained vision encoders are not necessary for multimodal modelling, and end-to-end pixel-space learning offers a scalable path toward stronger visual representations for both generation and perception.

02

Key Contributions

01

Encoder-Free UMMs

We propose \model, a native unified multimodal model that supports multimodal understanding and generation with encoder-free designs, achieving state-of-the-art performance across a wide range of understanding and generation benchmarks.

02

Encoder-Free Surpasses Encoder-Based

With sufficient end-to-end vision pretraining, the encoder-free Tuna-2 consistently outperforms the encoder-based Tuna-R on multimodal understanding, especially on fine-grained, perception-heavy benchmarks—demonstrating that large-scale pixel-level training can fully replace pretrained vision encoders.

03

Comprehensive Analysis

Through controlled comparison under the same unified framework, we reveal that scaling up vision pretraining is the key to closing and surpassing the gap between encoder-free and encoder-based designs, offering actionable insights for future native unified multimodal models.

03

Method

Pixel-Space Flow Matching

Tuna-2 discards the VAE module and operates the vision-language backbone and flow matching head entirely in pixel space. We adopt the x-prediction and v-loss paradigm for pixel-space flow matching, employing rectified flow and its linear schedule to construct noisy samples directly in pixel space.

Masking-Based Feature Learning

We introduce a masking-based visual feature learning scheme that randomly selects a subset of image patches and replaces them with a learnable mask token. This creates a harder denoising problem for generation and forces the model to perform multimodal reasoning under partial visual observation for understanding.

Two-Stage Training

Tuna-2 is trained using a two-stage, end-to-end strategy: (1) full model pretraining on image captioning and image generation data, and (2) supervised fine-tuning (SFT) for high-quality image generation and instruction following.

Encoder-Free Architecture

Tuna-2 uses simple patch embedding layers to encode input images into vision tokens, offloading the vision-language modelling task purely into the LMM decoder. This eliminates the inductive bias in pretrained representation encoders and simplifies the architecture to a single unified transformer.

04

Benchmark Results

Under a controlled, fair comparison with the same training recipe, the encoder-free Tuna-2 outperforms the encoder-based Tuna-R on the majority of multimodal understanding benchmarks after sufficient vision pretraining, particularly on pixel-centric tasks that demand fine-grained visual perception.

Table 1: Benchmark comparisons
Table 1. Comparisons between Tuna-2 and baseline models on multimodal understanding benchmarks. Results with model size greater than 13B are grayed. Bold: best results among all UMMs. Underline: second-best among all UMMs.
05

Attention Visualization

Visualization of attention maps across different models reveals that Tuna-2 variants attend more precisely to the relevant visual regions, demonstrating stronger fine-grained visual perception capabilities.

Attention visualization
Figure 3. Attention map visualization for Tuna-R, Tuna-2 and other baseline models, including LLaVA-OneVision-1.5, Qwen2.5-VL, Penguin-VL, and Tuna. Red area denotes high attention scores and blue area denotes low attention scores. Tuna-2 variants demonstrate more focused and accurate attention patterns on the queried visual elements.
06

Citation

BibTeX
@article{tuna2,
  title={Tuna-2: Pixel Embeddings Beat Vision Encoders
         for Multimodal Understanding and Generation},
  author={Liu, Zhiheng and Ren, Weiming and Huang, Xiaoke
          and Chen, Shoufa and Li, Tianhong and Chen, Mengzhao
          and Ji, Yatai and He, Sen and Schult, Jonas
          and Zeng, Belinda and Xiang, Tao and Chen, Wenhu
          and Luo, Ping and Zettlemoyer, Luke and Cong, Yuren},
  journal={arXiv preprint arXiv:2604.24763},
  year={2026}
}