NeoBabel Icon NeoBabel: A Multilingual Open Tower for Visual Generation

Mohammad Mahdi Derakhshani2, Dheeraj Varghese2, Marzieh Fadaee1,†, Cees G. M. Snoek2,†
1Cohere Labs, 2University of Amsterdam
Principal senior advisors
Plot showing NeoBabel's performance compared to other models.

NeoBabel establishes a new Pareto frontier in multilingual image generation performance, efficiency, and inclusivity. It matches state-of-the-art models on English benchmarks while being 2-4x smaller, and significantly outperforms them on multilingual tasks.

Abstract

Text-to-image generation advancements have been predominantly English-centric, creating barriers for non-English speakers and perpetuating digital inequities. While existing systems rely on translation pipelines, these introduce semantic drift, computational overhead, and cultural misalignment. We introduce NEOBABEL, a novel multilingual image generation framework that sets a new Pareto frontier in performance, efficiency and inclusivity, supporting six languages: English, Chinese, Dutch, French, Hindi, and Persian. The model is trained using a combination of large-scale multilingual pretraining and high-resolution instruction tuning. To evaluate its capabilities, we expand two English-only benchmarks to multilingual equivalents: m-GenEval and m-DPG. NEOBABEL achieves state-of-the-art multilingual performance while retaining strong English capability. Notably, it performs on par with leading models on English tasks while outperforming them by +0.11 and +0.09 on multilingual benchmarks. We release an open toolkit, including all code, model checkpoints, a curated dataset of 124M multilingual text-image pairs, and standardized multilingual evaluation protocols, to advance inclusive AI research.

NeoBabel Architecture

NeoBabel architecture diagram

Overview of the NeoBabel architecture.

NeoBabel builds upon a pretrained multilingual LLM (Gemma-2) and introduces two key modifications for visual generation:

  • Unified Multimodal Embedding: We extend the LLM's vocabulary with 8,192 new embeddings for discrete image tokens. This allows the model to process both text and image tokens natively within a shared semantic space, eliminating the need for separate encoders.
  • Modality-Aware Attention: A hybrid attention mechanism is used. Text tokens use causal attention to preserve autoregressive language capabilities, while image tokens use full bidirectional attention for rich, high-fidelity image synthesis.
  • Unified Task Representation: All tasks, such as text-to-image generation and inpainting, are treated as a single autoregressive sequence prediction problem, simplifying the training pipeline.

NeoBabel Multilingual Datasets

A primary challenge in multilingual generation is the scarcity of high-quality, culturally annotated visual-linguistic data. To address this, we developed a comprehensive data curation pipeline to expand existing English-only datasets into six target languages: English, Chinese, Dutch, French, Hindi, and Persian. Our process involves:

  1. Recaptioning: Generating detailed, high-quality English captions for images using a powerful vision-language model (InternVL).
  2. Quality Filtering: Applying a multi-step filtering process to ensure captions are of appropriate length, are in the correct language, align with the visual content, and are free of toxic or NSFW material.
  3. Translation: Translating the high-quality English captions into the five other target languages using state-of-the-art translation models (NLLB and Gemini).

This pipeline expanded our total dataset from 39 million to 124 million multilingual text-image pairs.

Original DatasetImage SourceOriginal SizeExpansion MethodNew Size (6 Langs)
ImageNet 1KWeb1MTranslation6M
CC12MWeb12MRecaptioning12M (Eng only)
SA-1BPhotography10MRecaptioning10M (Eng only)
LAION-AestheticWeb12MRecaption + Translation72M
JourneyDBSynthetic4MRecaption + Translation24M
BLIP3-o InstructWeb + Synthetic60KTranslation360K
Total-39M-124M

Expansion of datasets for multilingual training.

NeoBabel Training Stages

Graph showing performance improvement across training stages.

Performance on m-GenEval and m-DPG improves steadily across pretraining and instruction tuning stages.

NeoBabel is trained using a staged learning framework that progressively builds its capabilities:

Progressive Pretraining (at 256x256):

  • Stage 1 - Pixel Dependency Learning: Training on ImageNet with class labels to learn foundational visual representations.
  • Stage 2 - Scaling Alignment: Fine-tuning on 94M image-caption pairs to ground the model in natural language.
  • Stage 3 - Refined Multilingual Pretraining: Final pretraining on 96M high-quality pairs to improve aesthetic generation.

Progressive Instruction Tuning (at 512x512):

  • Stage 1 - Initial Alignment: Training on a mix of aesthetic and instruction-following data to build high-resolution capabilities.
  • Stage 2 - Instruction Refinement: Shifting the data mix to emphasize complex, instruction-rich samples to refine the model's ability to follow commands.

Multilingual Evaluation

Existing image generation benchmarks are overwhelmingly English-centric. To properly assess NeoBabel, we created a comprehensive multilingual evaluation suite.

  • Multilingual Benchmarks: We extended two popular English-only benchmarks by translating their prompts into our five other target languages, creating m-GenEval (for compositional reasoning) and m-DPG (for general-purpose generation).
  • New Evaluation Metrics: We introduced two novel metrics to measure multilingual consistency:
    • Cross-Lingual Consistency (CLC): Measures how visually similar the generated images are for the same prompt across different languages. A high CLC score indicates the model has a consistent internal representation of concepts, regardless of language.
    • Code-Switching Similarity (CSS): Assesses the model's robustness to prompts that mix multiple languages within a single sentence, a common real-world scenario.

This new suite allows for a more rigorous and realistic evaluation of a model's true multilingual capabilities.

Qualitative Results

Multilingual Generation

NeoBabel produces semantically accurate and visually cohesive outputs for the same concept across all six supported languages. Below are examples where the English prompt was translated, and one image was generated for each language.

Qualitative evaluation results for multilingual generation.

Multilingual Inpainting & Extrapolation

The model can edit and extend images using prompts in various languages without additional fine-tuning.

Inpainting a chair to look like a strawberry in Chinese.

Inpainting across languages.

Extrapolating a futuristic car in a city of mirrors using English and Hindi prompts.

Extrapolation across languages.

Cross-Lingual (Code-Switched) Prompt Generation

A challenging test where a single prompt contains phrases from multiple languages. NeoBabel successfully integrates these multilingual instructions into a single, coherent image.

Examples of images generated from code-switched prompts mixing three languages.

Generated from prompts mixing (Top) English, Dutch, French (Bottom) Hindi, Persian, Chinese.

Quantitative Results

NeoBabel achieves state-of-the-art results on both English and multilingual benchmarks, despite being 2-4x smaller than competing models.

Multilingual m-GenEval Benchmark

Model Performance Heatmap on m-GenEval

Figure 3 from the paper: NeoBabel matches the SOTA English score and outperforms on all multilingual cases.

Multilingual m-DPG Benchmark

MethodParams.EnglishChineseDutchFrenchHindiPersianOverall
Show-o1.3B0.670.100.220.320.040.040.23
BLIP3-o4B0.790.600.580.590.470.490.58
Janus Pro7B0.840.500.610.680.120.120.47
BLIP3-o8B0.800.560.590.610.500.530.59
NeoBabel (Ours)2B0.750.700.690.700.630.650.68

Table 3 from the paper: NeoBabel significantly outperforms all baselines across non-English languages.

BibTeX

@misc{derakhshani2025neobabelmultilingualopentower,
      title={NeoBabel: A Multilingual Open Tower for Visual Generation}, 
      author={Mohammad Mahdi Derakhshani and Dheeraj Varghese and Marzieh Fadaee and Cees G. M. Snoek},
      year={2025},
}