Skip to content

VisionLLM: Large Language Model is also an Open-Ended Decoder for Vision-Centric Tasks

Notifications You must be signed in to change notification settings

OpenGVLab/VisionLLM

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

9 Commits
Β 
Β 

Repository files navigation

VisionLLM [Paper]

Large language models (LLMs) have notably accelerated progress towards artificial general intelligence (AGI), with their impressive zero-shot capacity for user-tailored tasks, endowing them with immense potential across a range of applications. However, in the field of computer vision, despite the availability of numerous powerful vision foundation models (VFMs), they are still restricted to tasks in a pre-defined form, struggling to match the open-ended task capabilities of LLMs. In this work, we present an LLM-based framework for vision-centric tasks, termed VisionLLM. This framework provides a unified perspective for vision and language tasks by treating images as a foreign language and aligning vision-centric tasks with language tasks that can be flexibly defined and managed using language instructions. An LLM-based decoder can then make appropriate predictions based on these instructions for open-ended tasks. Extensive experiments show that the proposed VisionLLM can achieve different levels of task customization through language instructions, from fine-grained object-level to coarse-grained task-level customization, all with good results. It's noteworthy that, with a generalist LLM-based framework, our model can achieve over 60% mAP on COCO, on par with detection-specific models. We hope this model can set a new baseline for generalist vision and language models.

πŸ€–πŸ’¬ Online Demo

[TODO] VisionLLM will be integrated into InternGPT.

InternGPT is online (see https://igpt.opengvlab.com). Let's try it!

πŸ—“οΈ Schedule

  • Integrate into InternGPT
  • Release code and models

🏠 Overview

image

🎁 Major Features

image

🎫 License

This project is released under the Apache 2.0 license.

πŸ–ŠοΈ Citation

If you find this project useful in your research, please consider cite:

@misc{2023visionllm,
    title={VisionLLM: Large Language Model is also an Open-Ended Decoder for Vision-Centric Tasks},
    author={Wenhai Wang and Zhe Chen and Xiaokang Chen and Jiannan Wu and Xizhou Zhu and Gang Zeng and Ping Luo and Tong Lu and Jie Zhou and Yu Qiao and Jifeng Dai},
    howpublished = {\url{https://arxiv.org/abs/2305.11175)},
    year={2023}
}

About

VisionLLM: Large Language Model is also an Open-Ended Decoder for Vision-Centric Tasks

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published