Skip to content
@OpenGVLab

OpenGVLab

General Vision Team of Shanghai AI Laboratory

Static Badge Twitter

Welcome to OpenGVLab! 👋

We are a research group from Shanghai AI Lab focused on Vision-Centric AI research. The GV in our name, OpenGVLab, means general vision, a general understanding of vision, so little effort is needed to adapt to new vision-based tasks.

We develop model architecture and release pre-trained foundation models to the community to motivate further research in this area. We have made promising progress in general vision AI, with 109 SOTA🚀. In 2022, our open-sourced foundation model 65.5 mAP on the COCO object detection benchmark, 91.1% Top1 accuracy in Kinetics 400, achieved landmarks for AI vision👀 tasks for image🖼️ and video📹 understanding.

Based on solid vision foundations, we have expanded to Multi-Modality models and Generative AI(partner with Vchitect). We aim to empower individuals and businesses by offering a higher starting point for developing vision-based AI products and lessening the burden of building an AI model from scratch.

Branches: Alpha (explore lattest advances in vision+language research) and uni-medical (focus on medical AI)

Follow us:    Twitter X logo Twitter   🤗Hugging Face    Medium logo Medium    WeChat logo WeChat    zhihu logo Zhihu

Pinned Loading

  1. InternVL InternVL Public

    [CVPR 2024 Oral] InternVL Family: A Pioneering Open-Source Alternative to GPT-4o. 接近GPT-4o表现的可商用开源多模态对话模型

    Python 4.4k 338

  2. InternVideo InternVideo Public

    [ECCV2024] Video Foundation Models & Data for Multimodal Understanding

    Python 1.2k 75

  3. Ask-Anything Ask-Anything Public

    [CVPR2024 Highlight][VideoChatGPT] ChatGPT with video understanding! And many more supported LMs such as miniGPT4, StableLM, and MOSS.

    Python 2.9k 236

  4. VideoMamba VideoMamba Public

    [ECCV2024] VideoMamba: State Space Model for Efficient Video Understanding

    Python 714 54

  5. OmniQuant OmniQuant Public

    [ICLR2024 spotlight] OmniQuant is a simple and powerful quantization technique for LLMs.

    Python 629 49

  6. LLaMA-Adapter LLaMA-Adapter Public

    [ICLR 2024] Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parameters

    Python 5.6k 366

Repositories

Showing 10 of 66 repositories
  • InternVL Public

    [CVPR 2024 Oral] InternVL Family: A Pioneering Open-Source Alternative to GPT-4o. 接近GPT-4o表现的可商用开源多模态对话模型

    OpenGVLab/InternVL’s past year of commit activity
    Python 4,394 MIT 338 176 1 Updated Jul 25, 2024
  • EfficientQAT Public

    EfficientQAT: Efficient Quantization-Aware Training for Large Language Models

    OpenGVLab/EfficientQAT’s past year of commit activity
    Python 70 4 0 1 Updated Jul 25, 2024
  • Diffree Public

    Diffree: Text-Guided Shape Free Object Inpainting with Diffusion Model

    OpenGVLab/Diffree’s past year of commit activity
    Python 5 0 0 0 Updated Jul 25, 2024
  • OmniQuant Public

    [ICLR2024 spotlight] OmniQuant is a simple and powerful quantization technique for LLMs.

    OpenGVLab/OmniQuant’s past year of commit activity
    Python 629 MIT 49 25 1 Updated Jul 24, 2024
  • InternVideo Public

    [ECCV2024] Video Foundation Models & Data for Multimodal Understanding

    OpenGVLab/InternVideo’s past year of commit activity
    Python 1,151 Apache-2.0 75 63 3 Updated Jul 23, 2024
  • ChartAst Public

    ChartAssistant is a chart-based vision-language model for universal chart comprehension and reasoning.

    OpenGVLab/ChartAst’s past year of commit activity
    Python 68 5 11 0 Updated Jul 22, 2024
  • .github Public
    OpenGVLab/.github’s past year of commit activity
    0 1 0 0 Updated Jul 22, 2024
  • EgoExoLearn Public

    [CVPR 2024] Data and benchmark code for the EgoExoLearn dataset

    OpenGVLab/EgoExoLearn’s past year of commit activity
    Python 40 MIT 0 0 0 Updated Jul 19, 2024
  • MMT-Bench Public

    ICML'2024 | MMT-Bench: A Comprehensive Multimodal Benchmark for Evaluating Large Vision-Language Models Towards Multitask AGI

    OpenGVLab/MMT-Bench’s past year of commit activity
    Python 75 2 5 0 Updated Jul 18, 2024
  • ControlLLM Public

    ControlLLM: Augment Language Models with Tools by Searching on Graphs

    OpenGVLab/ControlLLM’s past year of commit activity
    Python 177 9 5 0 Updated Jul 15, 2024