Skip to content
/ emma Public

Official repo for paper "EMMA: Efficient Multimodal Understanding, Generation, and Editing with a Unified Architecture."

Notifications You must be signed in to change notification settings

umm-emma/emma

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

21 Commits
Β 
Β 
Β 
Β 

Repository files navigation

EMMA: Efficient Multimodal Understanding, Generation, and Editing with a Unified Architecture

Paper PDF

πŸ“° News

  • [2025.12.04] πŸ”₯ We have released EMMA: Efficient Multimodal Understanding, Generation, and Editing with a Unified Architecture. Check out the [ Paper ].

fig1

πŸ† Contributions

We propose EMMA, an efficient and unified architecture for multimodal understanding, generation and editing. Specifically, EMMA primarily consists of 1) An efficient autoencoder with a 32x compression ratio, which significantly reduces the number of tokens required for generation.This also ensures the training balance between understanding and generation tasks by applying the same compression ratio to images. 2) Channel-wise concatenation instead of token-wise concatenation among visual understanding and generation tokens, which further reduces the visual tokens in unified architectures. 3) A shared-and-decoupled network that enables mutual improvements across tasks while meeting the task-specific modeling requirements 4) A mixture-of-experts mechanism adopted for visual understanding encoder, which substantially improves perceptual capabilities with a few parameters increase. Extensive experiments have shown that EMMA-4B can significantly outperform state-of-the-art unified multimodal approaches (e.g., BAGEL-7B) in both efficiency and performance, while also achieving competitive results compared to recent multimodal understanding and generation experts (e.g., Qwen3-VL and Qwen-Image). We believe that EMMA lays a solid foundation for the future development of unified multimodal architectures.

✨ Comparison Result

  • Comparison between different unified multimodal approaches across multimodal understanding, text-to-image generation, and image editing benchmarks. fig1
  • Comparisons with state-of-the-arts on multimodal understanding benchmarks. fig1

✨ Image Generation

fig1

✨ Image Editing

fig1

πŸ“• BibTeX

@misc{he2025emmaefficientmultimodalunderstanding,
      title={EMMA: Efficient Multimodal Understanding, Generation, and Editing with a Unified Architecture}, 
      author={Xin He and Longhui Wei and Jianbo Ouyang and Minghui Liao and Lingxi Xie and Qi Tian},
      year={2025},
      eprint={2512.04810},
      archivePrefix={arXiv},
      primaryClass={cs.CV},
      url={https://arxiv.org/abs/2512.04810}, 
}

We are looking for interns to join our team. If you are interested, please feel free to concat us at [email protected] or [email protected]

About

Official repo for paper "EMMA: Efficient Multimodal Understanding, Generation, and Editing with a Unified Architecture."

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages