Papers
arxiv:2411.18363

ChatRex: Taming Multimodal LLM for Joint Perception and Understanding

Published on Nov 27
· Submitted by Mountchicken on Nov 28
Authors:
,
,
,
,
,

Abstract

Perception and understanding are two pillars of computer vision. While multimodal large language models (MLLM) have demonstrated remarkable visual understanding capabilities, they arguably lack accurate perception abilities, e.g. the stage-of-the-art model Qwen2-VL only achieves a 43.9 recall rate on the COCO dataset, limiting many tasks requiring the combination of perception and understanding. In this work, we aim to bridge this perception gap from both model designing and data development perspectives. We first introduce ChatRex, an MLLM with a decoupled perception design. Instead of having the LLM directly predict box coordinates, we feed the output boxes from a universal proposal network into the LLM, allowing it to output the corresponding box indices to represent its detection results, turning the regression task into a retrieval-based task that LLM handles more proficiently. From the data perspective, we build a fully automated data engine and construct the Rexverse-2M dataset which possesses multiple granularities to support the joint training of perception and understanding. After standard two-stage training, ChatRex demonstrates strong perception capabilities while preserving multimodal understanding performance. The combination of these two capabilities simultaneously unlocks many attractive applications, demonstrating the complementary roles of both perception and understanding in MLLM. Code is available at https://github.com/IDEA-Research/ChatRex.

Community

ChatRex is a multimodal large language model (MLLM) with strong perception capabilities, including detection, grounding, referring, grounded conversation, and more.

Here are some demo results on tasks like:

Common Object Detection:

44091732772699_.pic.jpg

Long-tailed Object Detection:

44101732772711_.pic.jpg

Referring Object Detection

44111732772725_.pic.jpg

Region QA

44121732772741_.pic.jpg

Grounded Image Captioning

44141732772785_.pic.jpg

Sign up or log in to comment

Models citing this paper 1

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2411.18363 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2411.18363 in a Space README.md to link it from this page.

Collections including this paper 1