The AIxiv column is a column where academic and technical content is published on this site. In the past few years, the AIxiv column of this site has received more than 2,000 reports, covering top laboratories from major universities and companies around the world, effectively promoting academic exchanges and dissemination. If you have excellent work that you want to share, please feel free to contribute or contact us for reporting. Submission email: liyazhou@jiqizhixin.com; zhaoyunfeng@jiqizhixin.com.
In recent years, the application of multimodal large language models (MLLM) in various fields has achieved remarkable success. However, as the underlying model for many downstream tasks, current MLLMs consist of the well-known Transformer network, which has less efficient quadratic computational complexity. In order to improve the efficiency of such basic models, a large number of experiments show that: (1) Cobra has extremely competitive performance with the current state-of-the-art methods with high computational efficiency (e.g., LLaVA-Phi, TinyLLaVA and MobileVLM v2), and due to Cobra linear sequence modeling, which is faster. (2) Interestingly, the results of the closed-set challenging prediction benchmark show that Cobra performs well in overcoming visual illusions and spatial relationship judgments. (3) It is worth noting that Cobra achieves comparable performance to LLaVA even when the number of parameters is only about 43% of LLaVA. Large language models (LLMs) are limited to interacting only through language, limiting their adaptability to handle more diverse tasks. Multimodal understanding is critical to enhance a model’s ability to effectively address real-world challenges. Therefore, researchers are actively working to extend large language models to incorporate multimodal information processing capabilities. Visual-Language Models (VLMs) such as GPT-4, LLaMA-Adapter, and LLaVA have been developed to enhance the visual understanding capabilities of LLMs. However, previous research mainly tried to obtain efficient VLMs in a similar way, that is, reducing the parameters of the basic language model or the number of visual tokens while keeping the attention-based Transformer structure unchanged. This paper proposes a different perspective: directly using the state space model (SSM) as the backbone network, an MLLM with linear computational complexity is obtained. Additionally, this paper explores and studies various modal fusion schemes to create an effective multi-modal Mamba. Specifically, this paper adopts the Mamba language model as the base model of VLM, which has shown performance that can compete with the Transformer language model, but with higher inference efficiency. Tests show that Cobra's inference performance is 3x to 4x faster than MobileVLM v2 3B and TinyLLaVA 3B of the same parameter magnitude. Even when compared to the LLaVA v1.5 model (7B parameters), which has a much higher number of parameters, Cobra still achieves matching performance on several benchmarks with about 43% the number of parameters.和 The main contributions of DEMO
this article of Cobra and LLAVA V1.5 7B are as follows:
- investigated the existing multimodilica large -scale Language models (MLLMs) often rely on Transformer networks, which exhibit quadratic computational complexity. To address this inefficiency, this paper introduces Cobra, a novel MLLM with linear computational complexity.
- Dives into various modal fusion schemes to optimize the integration of visual and linguistic information in the Mamba language model. Through experiments, this paper explores the effectiveness of different fusion strategies and determines the method that produces the most effective multimodal representation.
- Extensive experiments were conducted to evaluate the performance of Cobra with parallel studies aimed at improving the computational efficiency of underlying MLLM. Notably, Cobra achieves comparable performance to LLaVA even with fewer parameters, highlighting its efficiency.
- Original link: https://arxiv.org/pdf/2403.14520v2.pdf
- Project link: https://sites.google.com/view/cobravlm/
- Paper title: Cobra: Extending Mamba to Multi-Modal Large Language Model for Efficient Inference
Cobra uses a classic visual encoder to connect two models The VLM structure consists of a stateful projector and the LLM language backbone. The backbone part of LLM uses the 2.8B parameter pre-trained Mamba language model, which was pre-trained on the SlimPajama data set with 600B tokens and fine-tuned with the instructions of the conversation data.网络 Cobra network structure diagram
Different from LLAVA, etc., COBRA uses visual representation of Dinov2 and SIGLIP fusion. By stitching the output of the two visual coders together Feeding into the projector, the model can better capture the high-level semantic features brought by SigLIP and the low-level fine-grained image features extracted by DINOv2. Training scheme
Recent research shows that for existing training paradigms based on LLaVA (i.e., only training the pre-alignment stage of the projection layer and the fine-tuning stage of the LLM backbone once each), pre-alignment stages may be unnecessary and the fine-tuned model may still be underfitted. Therefore, Cobra abandons the pre-alignment stage and directly fine-tunes the entire LLM language backbone and projectors. This fine-tuning process was performed for two epochs with random sampling on a combined dataset consisting of: Hybrid dataset used in LLaVA v1.5, which contains a total of 655K visual multi-turn conversations, including Academic VQA samples, as well as visual instruction tuning data in LLaVA-Instruct and plain text instruction tuning data in ShareGPT. LVIS-Instruct-4V, which contains 220K images with visual alignment and context-aware instructions generated by GPT-4V.
-
LRV-Instruct, a dataset containing 400K visual instructions covering 16 visual language tasks, aimed at mitigating hallucination phenomena.
- The entire data set contains approximately 1.2 million images and corresponding multiple rounds of conversation data, as well as plain text conversation data.
Experiment
In the experimental part, this paper compares the proposed Cobra model and the open source SOTA VLM model on the basic benchmark, and compares it with the same The magnitude is based on the answering speed of the VLM model based on the Transformer architecture. At the same time, the generating speed and performance comparison of the graph at the same time, the COBRA is also the four open VQA tasks of VQA-V2, GQA, Vizwiz, TextVQA, and VSR, POPE two For a closed set prediction task, scores were compared on a total of 6 benchmarks. The comparison of the map on the Benchmark and other open source models Qualitative test
In addition, Cobra also gives two VQA examples to qualitatively illustrate the Cobra in the object of the object. Superiority in the ability to recognize spatial relationships and reduce model illusion.和 Figure COBRA and other baseline models in the judgment of object spatial relations 和 Figure Cobra and other baseline models in the example of visual illusion
In examples, Llava V1.5 and Mobilevlm are given an error answer, while COBRA does An accurate description was given, especially in the second instance, Cobra accurately identified that the picture came from the robot's simulation environment.
This article conducts ablation research on the solution adopted by Cobra from the two dimensions of performance and generation speed. The experimental plan conducts ablation experiments on the projector, visual encoder, and LLM language backbone respectively. The performance comparison of the performance of the diagram ablation experiment shows that the ablation experiments of the project part of the projector show that the effect of the MLP projector adopted in this article is significantly better than dedicated to reducing the number of visual Token to The LDP module improves the computing speed. At the same time, because Cobra's sequence processing speed and computational complexity are better than Transformer, the LDP module has no obvious advantage in generation speed. Therefore, the Mamba class model is used to reduce the number of visual tokens by sacrificing accuracy. The sampler may not be necessary.和 Figure COBRA and other models in the range of generating speed comparison
The ablation results of the visual encoder part show that the fusion of Dinov2 features effectively improves the performance of COBRA. In the language backbone experiment, the Mamba language model without instruction fine-tuning was completely unable to give reasonable answers in the open question and answer test, while the fine-tuned Mamba language model can achieve considerable performance on various tasks.
Conclusion
This paper proposes Cobra, which solves the efficiency bottleneck of existing multi-modal large-scale language models that rely on Transformer networks with quadratic computational complexity. This paper explores the combination of language models with linear computational complexity and multimodal input. In terms of fusing visual and language information, this paper successfully optimizes the internal information integration of the Mamba language model and achieves more effective multi-modal representation through in-depth research on different modal fusion schemes. Experiments show that Cobra not only significantly improves computational efficiency, but is also comparable in performance to advanced models such as LLaVA, especially in overcoming visual illusions and spatial relationship judgments. It even significantly reduces the number of parameters. This opens up new possibilities for future deployment of high-performance AI models in environments that require high-frequency processing of visual information, such as vision-based robot feedback control.
The above is the detailed content of The first Mamba-based MLLM is here! Model weights, training code, etc. have all been open source. For more information, please follow other related articles on the PHP Chinese website!