LVIDA: Layer-wise Vision Injection with Disentangled Attention for Efficient LVLMs

Xuange Zhang1† Dengjie Li2† Bo Liu1 Zenghao Bao2 Yao Zhou2
Baisong Yang2 Zhongying Liu2 Yujie Zhong2* Tongtong Yuan1*

1Beijing University of Technology, CN   2Meituan Inc., CN

Paper Icon
Paper
Code Icon
Code

clean-usnob
Comprehensive Comparison of LVIDA against Baseline LVLMs Across Different LLM Decoders.
Left: FLOPs comparison across different models, showing 90% computation reduction for models under 3B and 88% for 7B.
Right: Performance comparison across 7 benchmarks, demonstrating capability of LVIDA preservation compared to original models.

Abstract
Benefiting from recent advancements in large language models and modality alignment techniques, existing Large Vision-Language Models~(LVLMs) have achieved prominent performance across a wide range of scenarios. However, the excessive computational complexity limits the widespread use of these models in practical applications. We argue that one main bottleneck in computational complexity is caused by the involvement of redundant vision sequences in model computation. This is inspired by a reassessment of the efficiency of vision and language information transmission in the language decoder of LVLMs. Then, we propose a novel vision-language interaction mechanism called Layer-wise Vision Injection with Disentangled Attention (LVIDA). In LVIDA, only the language sequence undergoes full forward propagation, while the vision sequence interacts with the language at specific stages within each language decoder layer. It is striking that our approach significantly reduces computational complexity with minimal performance loss. Specifically, LVIDA achieves approximately a 10× reduction in the computational cost of the language decoder across multiple LVLM models while maintaining comparable performance

Framework
Model Structure
Comparison of Vanilla Model and HiMix Architectures. Left: Overall structure of traditional Vanilla. Middle: Overall structure of HiMix. Right: Details of Mixture Attention. Hierarchical Vision Injection for Mixture Attention (HiMix) is a method designed to reduce computational overhead while maintaining LVLM performance. After fusing the vision and language features through Mixture Attention, the vision sequence no longer participates in the forward propagation process within the language decoder, thereby substantially decreasing the overall computational load.

Results
Performance and Computational Efficiency Comparison
Comprehensive Comparison of LVIDA and Baseline Models. The V-L input ratio in these LVLMs is 728:64. The table reports computational efficiency (measured by FLOPs, Time to First Token, Peak Memory consumption) and performance, highlighting the efficiency of LVIDA with comparable results.
Visualisation
Visualisation results 1 Visualisation results 2

Citation
If you find our work useful in your research or applications, please cite our paper: