├── .gitignore ├── images └── motivation.png ├── LICENSE ├── audio-transformer.md ├── vision-transformer.md ├── audio-llm.md ├── image-llm.md └── README.md /.gitignore: -------------------------------------------------------------------------------- 1 | .DS_Store 2 | .vscode 3 | *.py 4 | *.csv 5 | 6 | *template* -------------------------------------------------------------------------------- /images/motivation.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/cokeshao/Awesome-Multimodal-Token-Compression/HEAD/images/motivation.png -------------------------------------------------------------------------------- /LICENSE: -------------------------------------------------------------------------------- 1 | MIT License 2 | 3 | Copyright (c) 2025 cokeshao 4 | 5 | Permission is hereby granted, free of charge, to any person obtaining a copy 6 | of this software and associated documentation files (the "Software"), to deal 7 | in the Software without restriction, including without limitation the rights 8 | to use, copy, modify, merge, publish, distribute, sublicense, and/or sell 9 | copies of the Software, and to permit persons to whom the Software is 10 | furnished to do so, subject to the following conditions: 11 | 12 | The above copyright notice and this permission notice shall be included in all 13 | copies or substantial portions of the Software. 14 | 15 | THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 16 | IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 17 | FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE 18 | AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER 19 | LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, 20 | OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE 21 | SOFTWARE. 22 | -------------------------------------------------------------------------------- /audio-transformer.md: -------------------------------------------------------------------------------- 1 | 2 |
3 | 2025 AST 4 | 5 | | **Title & Authors** | **Areas** | **Tags** | **Links** | 6 | | --- | --- | --- | :---: | 7 | | [![Publish](https://img.shields.io/badge/ICML-2025-blue)]() [![Star](https://img.shields.io/github/stars/yangdongchao/ALMTokenizer.svg?style=social&label=Star)](https://github.com/yangdongchao/ALMTokenizer)
[ALMTokenizer: A Low-bitrate and Semantic-rich Audio Codec Tokenizer for Audio Language Modeling](https://arxiv.org/abs/2504.10344)
Dongchao Yang, Songxiang Liu, Haohan Guo, Jiankun Zhao, Yuanyuan Wang, Helin Wang, Zeqian Ju, Xubo Liu, Xueyuan Chen, Xu Tan, Xixin Wu, Helen Meng | [![Area](https://img.shields.io/badge/Audio--Transformer-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2504.10344)
[GitHub](https://github.com/yangdongchao/ALMTokenizer)
| 8 | | [![Publish](https://img.shields.io/badge/ECAI-2025-blue)]() [![Star](https://img.shields.io/github/stars/andylee-24/token-pruning-audio-transformer.svg?style=social&label=Star)](https://github.com/andylee-24/token-pruning-audio-transformer)
[Token Pruning in Audio-Transformers: Optimizing Performance and Decoding Patch Importance](https://arxiv.org/abs/2504.01690)
Taehan Lee, Hyukjun Lee | [![Area](https://img.shields.io/badge/Audio--Transformer-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2504.01690)
[GitHub](https://github.com/andylee-24/token-pruning-audio-transformer)
[Model](https://drive.google.com/drive/folders/1cBDXh98m2qDlYLLX3q6xB-gtU1uUtxhK)
| 9 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.01-red)]() [![Star](https://img.shields.io/github/stars/VITA-MLLM/LUCY.svg?style=social&label=Star)](https://github.com/VITA-MLLM/LUCY)
[LUCY: Linguistic Understanding and Control Yielding Early Stage of Her](https://arxiv.org/abs/2501.16327)
Heting Gao, Hang Shao, Xiong Wang, Chaofan Qiu, Yunhang Shen, Siqi Cai, Yuchen Shi, Zihan Xu, Zuwei Long, Yike Zhang, Shaoqi Dong, Chaoyou Fu, Ke Li, Long Ma, Xing Sun | [![Area](https://img.shields.io/badge/Audio--Transformer-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2501.16327)
[GitHub](https://github.com/VITA-MLLM/LUCY)
[Model](https://huggingface.co/VITA-MLLM)
| 10 |
11 | 12 |
13 | 2024 AST 14 | 15 | | **Title & Authors** | **Areas** | **Tags** | **Links** | 16 | | --- | --- | --- | :---: | 17 | | [![Publish](https://img.shields.io/badge/Interspeech-2024-blue)]() [![Star](https://img.shields.io/github/stars/swarupbehera/FastAST.svg?style=social&label=Star)](https://github.com/swarupbehera/FastAST)
[FastAST: Accelerating Audio Spectrogram Transformer via Token Merging and Cross-Model Knowledge Distillation](https://arxiv.org/abs/2406.07676)
Swarup Ranjan Behera, Abhishek Dhiman, Karthik Gowda, Aalekhya Satya Narayani | [![Area](https://img.shields.io/badge/Audio--Transformer-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2406.07676)
[GitHub](https://github.com/swarupbehera/FastAST)
| 18 |
19 | 20 |
21 | 2023 AST 22 | 23 | | **Title & Authors** | **Areas** | **Tags** | **Links** | 24 | | --- | --- | --- | :---: | 25 | | [![Publish](https://img.shields.io/badge/Interspeech-2023-blue)]()
[Accelerating Transducers through Adjacent Token Merging](https://arxiv.org/abs/2306.16009)
Yuang Li, Yu Wu, Jinyu Li, Shujie Liu | [![Area](https://img.shields.io/badge/Audio--Transformer-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2306.16009)
| 26 |
27 | 28 |
29 | 2022 AST 30 | 31 | | **Title & Authors** | **Areas** | **Tags** | **Links** | 32 | | --- | --- | --- | :---: | 33 | | [![Publish](https://img.shields.io/badge/ICASSP-2022-blue)]() [![Star](https://img.shields.io/github/stars/RetroCirce/HTS-Audio-Transformer.svg?style=social&label=Star)](https://github.com/RetroCirce/HTS-Audio-Transformer)
[HTS-AT: A Hierarchical Token-Semantic Audio-Transformer for Sound Classification and Detection](https://arxiv.org/abs/2202.00874)
Ke Chen, Xingjian Du, Bilei Zhu, Zejun Ma, Taylor Berg-Kirkpatrick, Shlomo Dubnov | [![Area](https://img.shields.io/badge/Audio--Transformer-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2202.00874)
[GitHub](https://github.com/RetroCirce/HTS-Audio-Transformer)
| 34 |
35 | -------------------------------------------------------------------------------- /vision-transformer.md: -------------------------------------------------------------------------------- 1 | 2 |
3 | 2025 ViT 4 | 5 | | **Title & Authors** | **Areas** | **Tags** | **Links** | 6 | | --- | --- | --- | :---: | 7 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.08-red)]()
[TCSAFormer: Efficient Vision Transformer with Token Compression and Sparse Attention for Medical Image Segmentation](https://arxiv.org/abs/2508.04058)
Zunhui Xia, Hongxing Li, Libin Lan | [![Area](https://img.shields.io/badge/Vision--Transformer-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]() [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2508.04058)
| 8 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.08-red)]()
[Neutralizing Token Aggregation via Information Augmentation for Efficient Test-Time Adaptation](https://arxiv.org/abs/2508.03388)
Yizhe Xiong, Zihan Zhou, Yiwen Liang, Hui Chen, Zijia Lin, Tianxiang Hao, Fan Zhang, Jungong Han, Guiguang Ding | [![Area](https://img.shields.io/badge/Vision--Transformer-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2508.03388)
| 9 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]() [![Star](https://img.shields.io/github/stars/mlvlab/Representation-Shift.svg?style=social&label=Star)](https://github.com/mlvlab/Representation-Shift)
[Representation Shift: Unifying Token Compression with FlashAttention](https://arxiv.org/abs/2508.00367)
Joonmyung Choi, Sanghyeok Lee, Byungoh Ko, Eunseo Kim, Jihyung Kil, Hyunwoo J. Kim | [![Area](https://img.shields.io/badge/Vision--Transformer-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2508.00367)
[GitHub](https://github.com/mlvlab/Representation-Shift)
| 10 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.07-red)]()
[ToFe: Lagged Token Freezing and Reusing for Efficient Vision Transformer Inference](https://arxiv.org/abs/2507.16260)
Haoyue Zhang, Jie Zhang, Song Guo | [![Area](https://img.shields.io/badge/Vision--Transformer-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2507.16260)
| 11 |
12 | 13 |
14 | 2024 ViT 15 | 16 | | **Title & Authors** | **Areas** | **Tags** | **Links** | 17 | | --- | --- | --- | :---: | 18 | | [![Arxiv](https://img.shields.io/badge/arXiv-2024\.05-red)]() [![Star](https://img.shields.io/github/stars/yaolinli/DeCo.svg?style=social&label=Star)](https://github.com/yaolinli/DeCo)
[DeCo: Decoupling Token Compression from Semantic Abstraction in Multimodal Large Language Models](https://arxiv.org/abs/2405.20985)
Linli Yao, Lei Li, Shuhuai Ren, Lean Wang, Yuanxin Liu, Xu Sun, Lu Hou | [![Area](https://img.shields.io/badge/Vision--Transformer-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2405.20985)
[GitHub](https://github.com/yaolinli/DeCo)
| 19 | | [![Publish](https://img.shields.io/badge/CVPR-2024-blue)]() [![Star](https://img.shields.io/github/stars/double125/MADTP-plus.svg?style=social&label=Star)](https://github.com/double125/MADTP-plus)
[MADTP: Multimodal Alignment-Guided Dynamic Token Pruning for Accelerating Vision-Language Transformer](https://arxiv.org/abs/2403.02991)
Jianjian Cao, Peng Ye, Shengze Li, Chong Yu, Yansong Tang, Jiwen Lu, Tao Chen | [![Area](https://img.shields.io/badge/Vision--Transformer-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2403.02991)
[GitHub](https://github.com/double125/MADTP-plus)
| 20 |
21 | 22 |
23 | 2023 ViT 24 | 25 | | **Title & Authors** | **Areas** | **Tags** | **Links** | 26 | | --- | --- | --- | :---: | 27 | | [![Publish](https://img.shields.io/badge/ACL-2023-blue)]() [![Star](https://img.shields.io/github/stars/csarron/PuMer.svg?style=social&label=Star)](https://github.com/csarron/PuMer)
[PuMer: Pruning and Merging Tokens for Efficient Vision Language Models](https://arxiv.org/abs/2305.17530)
Qingqing Cao, Bhargavi Paranjape, Hannaneh Hajishirzi | [![Area](https://img.shields.io/badge/Vision--Transformer-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2305.17530)
[GitHub](https://github.com/csarron/PuMer)
| 28 |
29 | 30 |
31 | 2022 ViT 32 | 33 | | **Title & Authors** | **Areas** | **Tags** | **Links** | 34 | | --- | --- | --- | :---: | 35 | | [![Publish](https://img.shields.io/badge/ICLR_Oral-2023-blue)]() [![Star](https://img.shields.io/github/stars/facebookresearch/ToMe.svg?style=social&label=Star)](https://github.com/facebookresearch/ToMe)
[Token Merging: Your ViT But Faster](https://arxiv.org/abs/2210.09461)
Daniel Bolya, Cheng-Yang Fu, Xiaoliang Dai, Peizhao Zhang, Christoph Feichtenhofer, Judy Hoffman | [![Area](https://img.shields.io/badge/Vision--Transformer-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() [![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2210.09461)
[GitHub](https://github.com/facebookresearch/ToMe)
| 36 |
37 | 38 |
39 | 2021 ViT 40 | 41 | | **Title & Authors** | **Areas** | **Tags** | **Links** | 42 | | --- | --- | --- | :---: | 43 | | [![Publish](https://img.shields.io/badge/ICML-2021-blue)]()
[Perceiver: General Perception with Iterative Attention](https://arxiv.org/abs/2103.03206)
Andrew Jaegle, Felix Gimeno, Andrew Brock, Andrew Zisserman, Oriol Vinyals, Joao Carreira | [![Area](https://img.shields.io/badge/Vision--Transformer-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2103.03206)
| 44 |
45 | -------------------------------------------------------------------------------- /audio-llm.md: -------------------------------------------------------------------------------- 1 | 2 |
3 | 2025 Audio 4 | 5 | | **Title & Authors** | **Areas** | **Tags** | **Links** | 6 | | --- | --- | --- | :---: | 7 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.12-red)]()
[EchoingPixels: Cross-Modal Adaptive Token Reduction for Efficient Audio-Visual LLMs](https://arxiv.org/abs/2512.10324)
Chao Gong, Depeng Wang, Zhipeng Wei, Ya Guo, Huijia Zhu, Jingjing Chen | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2512.10324)
| 8 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.07-red)]() [![Star](https://img.shields.io/github/stars/cokeshao/Awesome-Multimodal-Token-Compression.svg?style=social&label=Star)](https://github.com/cokeshao/Awesome-Multimodal-Token-Compression)
[When Tokens Talk Too Much: A Survey of Multimodal Long-Context Token Compression across Images, Videos, and Audios](https://arxiv.org/abs/2507.20198)
Kele Shao, Keda Tao, Kejia Zhang, Sicheng Feng, Mu Cai, Yuzhang Shang, Haoxuan You, Can Qin, Yang Sui, Huan Wang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() [![Area](https://img.shields.io/badge/Survey-purple)]() | | [Paper](https://arxiv.org/abs/2507.20198)
[GitHub](https://github.com/cokeshao/Awesome-Multimodal-Token-Compression)
| 9 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.05-red)]() [![Star](https://img.shields.io/github/stars/ZLKong/Awesome-Collection-Token-Reduction.svg?style=social&label=Star)](https://github.com/ZLKong/Awesome-Collection-Token-Reduction)
[Token Reduction Should Go Beyond Efficiency in Generative Models -- From Vision, Language to Multimodality](https://arxiv.org/abs/2505.18227)
Zhenglun Kong, Yize Li, Fanhu Zeng, Lei Xin, Shvat Messica, Xue Lin, Pu Zhao, Manolis Kellis, Hao Tang, Marinka Zitnik | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() [![Area](https://img.shields.io/badge/Position--Paper-purple)]() | | [Paper](https://arxiv.org/abs/2505.18227)
[GitHub](https://github.com/ZLKong/Awesome-Collection-Token-Reduction)
| 10 | | [![Publish](https://img.shields.io/badge/ICML-2025-blue)]() [![Star](https://img.shields.io/github/stars/yangdongchao/ALMTokenizer.svg?style=social&label=Star)](https://github.com/yangdongchao/ALMTokenizer)
[ALMTokenizer: A Low-bitrate and Semantic-rich Audio Codec Tokenizer for Audio Language Modeling](https://arxiv.org/abs/2504.10344)
Dongchao Yang, Songxiang Liu, Haohan Guo, Jiankun Zhao, Yuanyuan Wang, Helin Wang, Zeqian Ju, Xubo Liu, Xueyuan Chen, Xu Tan, Xixin Wu, Helen Meng | [![Area](https://img.shields.io/badge/Audio--Transformer-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2504.10344)
[GitHub](https://github.com/yangdongchao/ALMTokenizer)
| 11 | | [![Publish](https://img.shields.io/badge/ECAI-2025-blue)]() [![Star](https://img.shields.io/github/stars/andylee-24/token-pruning-audio-transformer.svg?style=social&label=Star)](https://github.com/andylee-24/token-pruning-audio-transformer)
[Token Pruning in Audio-Transformers: Optimizing Performance and Decoding Patch Importance](https://arxiv.org/abs/2504.01690)
Taehan Lee, Hyukjun Lee | [![Area](https://img.shields.io/badge/Audio--Transformer-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2504.01690)
[GitHub](https://github.com/andylee-24/token-pruning-audio-transformer)
[Model](https://drive.google.com/drive/folders/1cBDXh98m2qDlYLLX3q6xB-gtU1uUtxhK)
| 12 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.03-red)]() [![Star](https://img.shields.io/github/stars/QwenLM/Qwen2.5-Omni.svg?style=social&label=Star)](https://github.com/QwenLM/Qwen2.5-Omni)
[Qwen2.5-Omni Technical Report](https://arxiv.org/abs/2503.20215)
Qwen Team | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2503.20215)
[GitHub](https://github.com/QwenLM/Qwen2.5-Omni)
[Model](https://huggingface.co/collections/Qwen/qwen25-omni-67de1e5f0f9464dc6314b36e)
| 13 | | [![Publish](https://img.shields.io/badge/ACL_Findings-2025-blue)]() [![Star](https://img.shields.io/github/stars/JeongHun0716/MMS-LLaMA.svg?style=social&label=Star)](https://github.com/JeongHun0716/MMS-LLaMA)
[MMS-LLaMA: Efficient LLM-based Audio-Visual Speech Recognition with Minimal Multimodal Speech Tokens](https://arxiv.org/abs/2503.11315)
Jeong Hun Yeo, Hyeongseop Rha, Se Jin Park, Yong Man Ro | [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2503.11315)
[GitHub](https://github.com/JeongHun0716/MMS-LLaMA)
| 14 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.03-red)]()
[Adaptive Audio-Visual Speech Recognition via Matryoshka-Based Multimodal LLMs](https://arxiv.org/abs/2503.06362)
Umberto Cappellazzo, Minsu Kim, Stavros Petridis | [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2503.06362)
| 15 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.02-red)]() [![Star](https://img.shields.io/github/stars/baichuan-inc/Baichuan-Audio.svg?style=social&label=Star)](https://github.com/baichuan-inc/Baichuan-Audio)
[Baichuan-Audio: A Unified Framework for End-to-End Speech Interaction](https://arxiv.org/abs/2502.17239)
Tianpeng Li, Jun Liu, Tao Zhang, Yuanbo Fang, Da Pan, Mingrui Wang, Zheng Liang, Zehuan Li, Mingan Lin, Guosheng Dong, Jianhua Xu, Haoze Sun, Zenan Zhou, Weipeng Chen | [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2502.17239)
[GitHub](https://github.com/baichuan-inc/Baichuan-Audio)
[Model](https://huggingface.co/baichuan-inc/Baichuan-Audio-Instruct)
| 16 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.01-red)]() [![Star](https://img.shields.io/github/stars/VITA-MLLM/LUCY.svg?style=social&label=Star)](https://github.com/VITA-MLLM/LUCY)
[LUCY: Linguistic Understanding and Control Yielding Early Stage of Her](https://arxiv.org/abs/2501.16327)
Heting Gao, Hang Shao, Xiong Wang, Chaofan Qiu, Yunhang Shen, Siqi Cai, Yuchen Shi, Zihan Xu, Zuwei Long, Yike Zhang, Shaoqi Dong, Chaoyou Fu, Ke Li, Long Ma, Xing Sun | [![Area](https://img.shields.io/badge/Audio--Transformer-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2501.16327)
[GitHub](https://github.com/VITA-MLLM/LUCY)
[Model](https://huggingface.co/VITA-MLLM)
| 17 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.01-red)]() [![Star](https://img.shields.io/github/stars/ASLP-lab/OSUM.svg?style=social&label=Star)](https://github.com/ASLP-lab/OSUM)
[OSUM: Advancing Open Speech Understanding Models with Limited Resources in Academia](https://arxiv.org/abs/2501.13306)
ASLP@NPU | [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2501.13306)
[GitHub](https://github.com/ASLP-lab/OSUM)
[Model](https://huggingface.co/ASLP-lab/OSUM)
| 18 |
19 | 20 |
21 | 2024 Audio 22 | 23 | | **Title & Authors** | **Areas** | **Tags** | **Links** | 24 | | --- | --- | --- | :---: | 25 | | [![Arxiv](https://img.shields.io/badge/arXiv-2024\.12-red)]() [![Star](https://img.shields.io/github/stars/scb-10x/typhoon2-audio.svg?style=social&label=Star)](https://github.com/scb-10x/typhoon2-audio)
[Typhoon 2: A Family of Open Text and Multimodal Thai Large Language Models](https://arxiv.org/abs/2412.13702)
Kunat Pipatanakul, Potsawee Manakul, Natapong Nitarach, Warit Sirichotedumrong, Surapon Nonesung, Teetouch Jaknamon, Parinthapat Pengpun, Pittawat Taveekitworachai, Adisai Na-Thalang, Sittipong Sripaisarnmongkol, Krisanapong Jirayoot, Kasima Tharnpipitchai | [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2412.13702)
[GitHub](https://github.com/scb-10x/typhoon2-audio)
[Model](https://huggingface.co/scb10x/llama3.1-typhoon2-audio-8b-instruct)
| 26 | | [![Publish](https://img.shields.io/badge/ICME-2025-blue)]()
[SpeechPrune: Context-aware Token Pruning for Speech Information Retrieval](https://arxiv.org/abs/2412.12009)
Yueqian Lin, Yuzhe Fu, Jingyang Zhang, Yudong Liu, Jianyi Zhang, Jingwei Sun, Hai "Helen" Li, Yiran Chen | [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2412.12009)
| 27 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]() [![Star](https://img.shields.io/github/stars/dvlab-research/Lyra.svg?style=social&label=Star)](https://github.com/dvlab-research/Lyra)
[Lyra: An Efficient and Speech-Centric Framework for Omni-Cognition](https://arxiv.org/abs/2412.09501)
Zhisheng Zhong, Chengyao Wang, Yuqi Liu, Senqiao Yang, Longxiang Tang, Yuechen Zhang, Jingyao Li, Tianyuan Qu, Yanwei Li, Yukang Chen, Shaozuo Yu, Sitong Wu, Eric Lo, Shu Liu, Jiaya Jia | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2412.09501)
[GitHub](https://github.com/dvlab-research/Lyra)
[Model](https://huggingface.co/collections/zszhong/lyra-model-674ea5bb3b39ff8f15de75fc)
[Dataset](https://huggingface.co/collections/zszhong/lyra-data-675d80fbab80334eb52cdd82)
| 28 | | [![Publish](https://img.shields.io/badge/ICASSP-2025-blue)]() [![Star](https://img.shields.io/github/stars/umbertocappellazzo/Llama-AVSR.svg?style=social&label=Star)](https://github.com/umbertocappellazzo/Llama-AVSR)
[Large Language Models are Strong Audio-Visual Speech Recognition Learners](https://arxiv.org/abs/2409.12319)
Umberto Cappellazzo, Minsu Kim, Honglie Chen, Pingchuan Ma, Stavros Petridis, Daniele Falavigna, Alessio Brutti, Maja Pantic | [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2409.12319)
[GitHub](https://github.com/umbertocappellazzo/Llama-AVSR)
| 29 | | [![Publish](https://img.shields.io/badge/ICLR-2025-blue)]() [![Star](https://img.shields.io/github/stars/ictnlp/LLaMA-Omni.svg?style=social&label=Star)](https://github.com/ictnlp/LLaMA-Omni)
[LLaMA-Omni: Seamless Speech Interaction with Large Language Models](https://arxiv.org/abs/2409.06666)
Qingkai Fang, Shoutao Guo, Yan Zhou, Zhengrui Ma, Shaolei Zhang, Yang Feng | [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2409.06666)
[GitHub](https://github.com/ictnlp/LLaMA-Omni)
[Model](https://huggingface.co/ICTNLP/Llama-3.1-8B-Omni)
[Dataset](https://huggingface.co/datasets/ICTNLP/InstructS2S-200K)
| 30 | | [![Arxiv](https://img.shields.io/badge/arXiv-2024\.07-red)]() [![Star](https://img.shields.io/github/stars/QwenLM/Qwen2-Audio.svg?style=social&label=Star)](https://github.com/QwenLM/Qwen2-Audio)
[Qwen2-Audio Technical Report](https://arxiv.org/abs/2407.10759)
Yunfei Chu, Jin Xu, Qian Yang, Haojie Wei, Xipin Wei, Zhifang Guo, Yichong Leng, Yuanjun Lv, Jinzheng He, Junyang Lin, Chang Zhou, Jingren Zhou | [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2407.10759)
[GitHub](https://github.com/QwenLM/Qwen2-Audio)
[Model](https://huggingface.co/collections/Qwen/qwen2-audio-66b628d694096020e0c52ff6)
| 31 | | [![Publish](https://img.shields.io/badge/ICML-2024-blue)]() [![Star](https://img.shields.io/github/stars/bytedance/SALMONN.svg?style=social&label=Star)](https://github.com/bytedance/SALMONN/tree/videosalmonn)
[video-SALMONN: Speech-Enhanced Audio-Visual Large Language Models](https://arxiv.org/abs/2406.15704)
Guangzhi Sun, Wenyi Yu, Changli Tang, Xianzhao Chen, Tian Tan, Wei Li, Lu Lu, Zejun Ma, Yuxuan Wang, Chao Zhang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2406.15704)
[GitHub](https://github.com/bytedance/SALMONN/tree/videosalmonn)
[Model](https://huggingface.co/tsinghua-ee/Video-SALMONN/tree/main)
| 32 | | [![Publish](https://img.shields.io/badge/Interspeech-2024-blue)]()
[Enhancing Automated Audio Captioning via Large Language Models with Optimized Audio Encoding](https://arxiv.org/abs/2406.13275)
Jizhong Liu, Gang Li, Junbo Zhang, Heinrich Dinkel, Yongqing Wang, Zhiyong Yan, Yujun Wang, Bin Wang | [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2406.13275)
| 33 | | [![Publish](https://img.shields.io/badge/Interspeech-2024-blue)]() [![Star](https://img.shields.io/github/stars/swarupbehera/FastAST.svg?style=social&label=Star)](https://github.com/swarupbehera/FastAST)
[FastAST: Accelerating Audio Spectrogram Transformer via Token Merging and Cross-Model Knowledge Distillation](https://arxiv.org/abs/2406.07676)
Swarup Ranjan Behera, Abhishek Dhiman, Karthik Gowda, Aalekhya Satya Narayani | [![Area](https://img.shields.io/badge/Audio--Transformer-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2406.07676)
[GitHub](https://github.com/swarupbehera/FastAST)
| 34 | | [![Arxiv](https://img.shields.io/badge/arXiv-2024\.06-red)]() [![Star](https://img.shields.io/github/stars/DAMO-NLP-SG/VideoLLaMA2.svg?style=social&label=Star)](https://github.com/DAMO-NLP-SG/VideoLLaMA2)
[VideoLLaMA 2: Advancing Spatial-Temporal Modeling and Audio Understanding in Video-LLMs](https://arxiv.org/abs/2406.07476)
Zesen Cheng, Sicong Leng, Hang Zhang, Yifei Xin, Xin Li, Guanzheng Chen, Yongxin Zhu, Wenqi Zhang, Ziyang Luo, Deli Zhao, Lidong Bing | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2406.07476)
[GitHub](https://github.com/DAMO-NLP-SG/VideoLLaMA2)
[Model](https://huggingface.co/collections/DAMO-NLP-SG/videollama2-6669b6b6f0493188305c87ed)
| 35 | | [![Publish](https://img.shields.io/badge/Interspeech-2024-blue)]()
[Improving Audio Codec-based Zero-Shot Text-to-Speech Synthesis with Multi-Modal Context and Large Language Model](https://arxiv.org/abs/2406.03706)
Jinlong Xue, Yayue Deng, Yicheng Han, Yingming Gao, Ya Li | [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2406.03706)
| 36 | | [![Arxiv](https://img.shields.io/badge/arXiv-2024\.05-red)]()
[SpeechVerse: A Large-scale Generalizable Audio Language Model](https://arxiv.org/abs/2405.08295)
AWS AI Team | [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2405.08295)
| 37 | | [![Arxiv](https://img.shields.io/badge/arXiv-2024\.02-red)]()
[An Embarrassingly Simple Approach for LLM with Strong ASR Capacity](https://arxiv.org/abs/2402.08846)
Ziyang Ma, Guanrou Yang, Yifan Yang, Zhifu Gao, Jiaming Wang, Zhihao Du, Fan Yu, Qian Chen, Siqi Zheng, Shiliang Zhang, Xie Chen | [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2402.08846)
| 38 |
39 | 40 |
41 | 2023 Audio 42 | 43 | | **Title & Authors** | **Areas** | **Tags** | **Links** | 44 | | --- | --- | --- | :---: | 45 | | [![Publish](https://img.shields.io/badge/ICLR-2024-blue)]() [![Star](https://img.shields.io/github/stars/Render-AI/salmonn.svg?style=social&label=Star)](https://github.com/Render-AI/salmonn)
[SALMONN: Towards Generic Hearing Abilities for Large Language Models](https://arxiv.org/abs/2310.13289)
Changli Tang, Wenyi Yu, Guangzhi Sun, Xianzhao Chen, Tian Tan, Wei Li, Lu Lu, Zejun Ma, Chao Zhang | [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2310.13289)
[GitHub](https://github.com/Render-AI/salmonn)
[Model](https://huggingface.co/tsinghua-ee/SALMONN)
| 46 | | [![Publish](https://img.shields.io/badge/ICASSP-2024-blue)]()
[Connecting Speech Encoder and Large Language Model for ASR](https://arxiv.org/abs/2309.13963)
Wenyi Yu, Changli Tang, Guangzhi Sun, Xianzhao Chen, Tian Tan, Wei Li, Lu Lu, Zejun Ma, Chao Zhang | [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2309.13963)
| 47 | | [![Publish](https://img.shields.io/badge/ICASSP-2024-blue)]()
[Prompting Large Language Models with Speech Recognition Abilities](https://arxiv.org/abs/2307.11795)
Yassir Fathullah, Chunyang Wu, Egor Lakomkin, Junteng Jia, Yuan Shangguan, Ke Li, Jinxi Guo, Wenhan Xiong, Jay Mahadeokar, Ozlem Kalinli, Christian Fuegen, Mike Seltzer | [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2307.11795)
| 48 | | [![Publish](https://img.shields.io/badge/Interspeech-2023-blue)]()
[Accelerating Transducers through Adjacent Token Merging](https://arxiv.org/abs/2306.16009)
Yuang Li, Yu Wu, Jinyu Li, Shujie Liu | [![Area](https://img.shields.io/badge/Audio--Transformer-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2306.16009)
| 49 | | [![Publish](https://img.shields.io/badge/EMNLP-2023-blue)]() [![Star](https://img.shields.io/github/stars/DAMO-NLP-SG/Video-LLaMA.svg?style=social&label=Star)](https://github.com/DAMO-NLP-SG/Video-LLaMA)
[Video-LLaMA: An Instruction-tuned Audio-Visual Language Model for Video Understanding](https://arxiv.org/abs/2306.02858)
Hang Zhang, Xin Li, Lidong Bing | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2306.02858)
[GitHub](https://github.com/DAMO-NLP-SG/Video-LLaMA)
[Model](https://huggingface.co/DAMO-NLP-SG/Video-LLaMA-Series)
| 50 |
51 | 52 |
53 | 2022 Audio 54 | 55 | | **Title & Authors** | **Areas** | **Tags** | **Links** | 56 | | --- | --- | --- | :---: | 57 | | [![Publish](https://img.shields.io/badge/ICASSP-2022-blue)]() [![Star](https://img.shields.io/github/stars/RetroCirce/HTS-Audio-Transformer.svg?style=social&label=Star)](https://github.com/RetroCirce/HTS-Audio-Transformer)
[HTS-AT: A Hierarchical Token-Semantic Audio-Transformer for Sound Classification and Detection](https://arxiv.org/abs/2202.00874)
Ke Chen, Xingjian Du, Bilei Zhu, Zejun Ma, Taylor Berg-Kirkpatrick, Shlomo Dubnov | [![Area](https://img.shields.io/badge/Audio--Transformer-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2202.00874)
[GitHub](https://github.com/RetroCirce/HTS-Audio-Transformer)
| 58 |
59 | -------------------------------------------------------------------------------- /image-llm.md: -------------------------------------------------------------------------------- 1 | 2 |
3 | 2025 Image 4 | 5 | | **Title & Authors** | **Areas** | **Tags** | **Links** | 6 | | --- | --- | --- | :---: | 7 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.10-red)]() [![Star](https://img.shields.io/github/stars/deepseek-ai/DeepSeek-OCR.svg?style=social&label=Star)](https://github.com/deepseek-ai/DeepSeek-OCR)
[DeepSeek-OCR: Contexts Optical Compression](https://arxiv.org/abs/2510.18234)
Haoran Wei, Yaofeng Sun, Yukun Li | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2510.18234)
[GitHub](https://github.com/deepseek-ai/DeepSeek-OCR)
[Model](https://huggingface.co/deepseek-ai/DeepSeek-OCR)
| 8 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.10-red)]() [![Star](https://img.shields.io/github/stars/JulietChoo/VisionSelector.svg?style=social&label=Star)](https://github.com/JulietChoo/VisionSelector)
[VisionSelector: End-to-End Learnable Visual Token Compression for Efficient Multimodal LLMs](https://arxiv.org/abs/2510.16598)
Jiaying Zhu, Yurui Zhu, Xin Lu, Wenrui Yan, Dong Li, Kunlin Liu, Xueyang Fu, Zheng-Jun Zha | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2510.16598)
[GitHub](https://github.com/JulietChoo/VisionSelector)
[Model](https://huggingface.co/JulietChoo/VisionSelector-Qwen2.5-VL-7B)
| 9 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.10-red)]() [![Star](https://img.shields.io/github/stars/Chenfei-Liao/VTC-Bench.svg?style=social&label=Star)](https://github.com/Chenfei-Liao/VTC-Bench)
[Are We Using the Right Benchmark: An Evaluation Framework for Visual Token Compression Methods](https://arxiv.org/abs/2510.07143)
Chenfei Liao, Wensong Wang, Zichen Wen, Xu Zheng, Yiyu Wang, Haocong He, Yuanhuiyi Lyu, Lutao Jiang, Xin Zou, Yuqian Fu, Bin Ren, Linfeng Zhang, Xuming Hu | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Benchmark-purple)]() | | [Paper](https://arxiv.org/abs/2510.07143)
[GitHub](https://github.com/Chenfei-Liao/VTC-Bench)
| 10 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.09-red)]()
[Training-Free Token Pruning via Zeroth-Order Gradient Estimation in Vision-Language Models](https://arxiv.org/abs/2509.24837)
Youngeun Kim, Youjia Zhang, Huiling Liu, Aecheon Jung, Sunwoo Lee, Sungeun Hong | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2509.24837)
| 11 | | [![Publish](https://img.shields.io/badge/NeurIPS-2025-blue)]() [![Star](https://img.shields.io/github/stars/AutoLab-SAI-SJTU/AutoPrune.svg?style=social&label=Star)](https://github.com/AutoLab-SAI-SJTU/AutoPrune)
[AutoPrune: Each Complexity Deserves a Pruning Policy](https://arxiv.org/abs/2509.23931)
Hanshi Wang, Yuhao Xu, Zekun Xu, Jin Gao, Yufan Liu, Weiming Hu, Ke Wang, Zhipeng Zhang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2509.23931)
[GitHub](https://github.com/AutoLab-SAI-SJTU/AutoPrune)
| 12 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.09-red)]()
[HIVTP: A Training-Free Method to Improve VLMs Efficiency via Hierarchical Visual Token Pruning Using Middle-Layer-Based Importance Score](https://arxiv.org/abs/2509.23663)
Jingqi Xu, Jingxi Lu, Chenghao Li, Sreetama Sarkar, Peter A. Beerel | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2509.23663)
| 13 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.09-red)]()
[Pyramid Token Pruning for High-Resolution Large Vision-Language Models via Region, Token, and Instruction-Guided Importance](https://arxiv.org/abs/2509.15704)
Yuxuan Liang, Xu Li, Xiaolei Chen, Yi Zheng, Haotian Chen, Bin Li, Xiangyang Xue | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Query--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2509.15704)
| 14 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.09-red)]()
[EfficientUICoder: Efficient MLLM-based UI Code Generation via Input and Output Token Compression](https://arxiv.org/abs/2509.12159)
Jingyu Xiao, Zhongyi Zhang, Yuxuan Wan, Yintong Huo, Yang Liu, Michael R.Lyu | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/GUI--Agent-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2509.12159)
| 15 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.09-red)]()
[Adaptive Token Merging for Efficient Transformer Semantic Communication at the Edge](https://arxiv.org/abs/2509.09955)
Omar Erak, Omar Alhussein, Hatem Abou-Zeid, Mehdi Bennis, Sami Muhaidat | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2509.09955)
| 16 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.08-red)]() [![Star](https://img.shields.io/github/stars/OpenGVLab/InternVL.svg?style=social&label=Star)](https://github.com/OpenGVLab/InternVL)
[InternVL3.5: Advancing Open-Source Multimodal Models in Versatility, Reasoning, and Efficiency](https://arxiv.org/abs/2508.18265)
InternVL Team | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2508.18265)
[GitHub](https://github.com/OpenGVLab/InternVL)
[Model](https://huggingface.co/collections/OpenGVLab/internvl35-68ac87bd52ebe953485927fb)
| 17 | | [![Publish](https://img.shields.io/badge/ACM_MM-2025-blue)]()
[VISA: Group-wise Visual Token Selection and Aggregation via Graph Summarization for Efficient MLLMs Inference](https://arxiv.org/abs/2508.17857)
Pengfei Jiang, Hanjun Li, Linglan Zhao, Fei Chao, Ke Yan, Shouhong Ding, Rongrong Ji | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2508.17857)
| 18 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.08-red)]()
[Revisiting MLLM Token Technology through the Lens of Classical Visual Coding](https://arxiv.org/abs/2508.13460)
Jinming Liu, Junyan Lin, Yuntao Wei, Kele Shao, Keda Tao, Jianguo Huang, Xudong Yang, Zhibo Chen, Huan Wang, Xin Jin | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Visual--Coding-purple)]() | | [Paper](https://arxiv.org/abs/2508.13460)
| 19 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.08-red)]()
[EVTP-IVS: Effective Visual Token Pruning For Unifying Instruction Visual Segmentation In Multi-Modal Large Language Models](https://arxiv.org/abs/2508.11886)
Wenhui Zhu, Xiwen Chen, Zhipeng Wang, Shao Tang, Sayan Ghosh, Xuanzhao Dong, Rajat Koner, Yalin Wang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2508.11886)
| 20 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.08-red)]()
[CATP: Contextually Adaptive Token Pruning for Efficient and Enhanced Multimodal In-Context Learning](https://arxiv.org/abs/2508.07871)
Yanshu Li, Jianjiang Yang, Zhennan Shen, Ligong Han, Haoyan Xu, Ruixiang Tang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2508.07871)
| 21 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.08-red)]()
[AdaptInfer: Adaptive Token Pruning for Vision-Language Model Inference with Dynamical Text Guidance](https://arxiv.org/abs/2508.06084)
Weichen Zhang, Zhui Zhu, Ningbo Li, Kebin Liu, Yunhao Liu | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2508.06084)
| 22 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.08-red)]()
[Fourier-VLM: Compressing Vision Tokens in the Frequency Domain for Large Vision-Language Models](https://arxiv.org/abs/2508.06038)
Huanyu Wang, Jushi Kai, Haoli Bai, Lu Hou, Bo Jiang, Ziwei He, Zhouhan Lin | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2508.06038)
| 23 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]() [![Star](https://img.shields.io/github/stars/sihany077/VFlowOpt.svg?style=social&label=Star)](https://github.com/sihany077/VFlowOpt)
[VFlowOpt: A Token Pruning Framework for LMMs with Visual Information Flow-Guided Optimization](https://arxiv.org/abs/2508.05211)
Sihan Yang, Runsen Xu, Chenhang Cui, Tai Wang, Dahua Lin, Jiangmiao Pang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2508.05211)
[GitHub](https://github.com/sihany077/VFlowOpt)
| 24 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.08-red)]() [![Star](https://img.shields.io/github/stars/HVision-NKU/GlimpsePrune.svg?style=social&label=Star)](https://github.com/HVision-NKU/GlimpsePrune)
[A Glimpse to Compress: Dynamic Visual Token Pruning for Large Vision-Language Models](https://arxiv.org/abs/2508.01548)
Quan-Sheng Zeng, Yunheng Li, Qilong Wang, Peng-Tao Jiang, Zuxuan Wu, Ming-Ming Cheng, Qibin Hou | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2508.01548)
[GitHub](https://github.com/HVision-NKU/GlimpsePrune)
[Model](https://huggingface.co/collections/ashun989/glimpseprune-688d8826ef5bd09db6af145e)
| 25 | | [![Publish](https://img.shields.io/badge/ACM_MM-2025-blue)]()
[Mitigating Information Loss under High Pruning Rates for Efficient Large Vision Language Models](https://arxiv.org/abs/2508.01236)
Mingyu Fu, Wei Suo, Ji Ma, Lin Yuanbo Wu, Peng Wang, Yanning Zhang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2508.01236)
| 26 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.08-red)]()
[HiPrune: Training-Free Visual Token Pruning via Hierarchical Attention in Vision-Language Models](https://arxiv.org/abs/2508.00553)
Jizhihui Liu, Feiyi Du, Guangdao Zhu, Niu Lian, Jun Li, Bin Chen | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2508.00553)
| 27 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.07-red)]()
[FastDriveVLA: Efficient End-to-End Driving via Plug-and-Play Reconstruction-based Token Pruning](https://arxiv.org/abs/2507.23318)
Jiajun Cao, Qizhe Zhang, Peidong Jia, Xuhui Zhao, Bo Lan, Xiaoan Zhang, Xiaobao Wei, Sixiang Chen, Zhuo Li, Yang Wang, Liyun Li, Xianming Liu, Ming Lu, Shanghang Zhang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/VLA-purple)]() | [![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2507.23318)
| 28 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]() [![Star](https://img.shields.io/github/stars/YuchenLiu98/METEOR.svg?style=social&label=Star)](https://github.com/YuchenLiu98/METEOR)
[METEOR: Multi-Encoder Collaborative Token Pruning for Efficient Vision Language Models](https://arxiv.org/abs/2507.20842)
Yuchen Liu, Yaoming Wang, Bowen Shi, Xiaopeng Zhang, Wenrui Dai, Chenglin Li, Hongkai Xiong, Qi Tian | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2507.20842)
[GitHub](https://github.com/YuchenLiu98/METEOR)
| 29 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.07-red)]() [![Star](https://img.shields.io/github/stars/liaolea/TransPrune.svg?style=social&label=Star)](https://github.com/liaolea/TransPrune)
[TransPrune: Token Transition Pruning for Efficient Large Vision-Language Model](https://arxiv.org/abs/2507.20630)
Ao Li, Yuxiang Duan, Jinghui Zhang, Congbo Ma, Yutong Xie, Gustavo Carneiro, Mohammad Yaqub, Hu Wang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2507.20630)
[GitHub](https://github.com/liaolea/TransPrune)
| 30 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.07-red)]() [![Star](https://img.shields.io/github/stars/cokeshao/Awesome-Multimodal-Token-Compression.svg?style=social&label=Star)](https://github.com/cokeshao/Awesome-Multimodal-Token-Compression)
[When Tokens Talk Too Much: A Survey of Multimodal Long-Context Token Compression across Images, Videos, and Audios](https://arxiv.org/abs/2507.20198)
Kele Shao, Keda Tao, Kejia Zhang, Sicheng Feng, Mu Cai, Yuzhang Shang, Haoxuan You, Can Qin, Yang Sui, Huan Wang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() [![Area](https://img.shields.io/badge/Survey-purple)]() | | [Paper](https://arxiv.org/abs/2507.20198)
[GitHub](https://github.com/cokeshao/Awesome-Multimodal-Token-Compression)
| 31 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.07-red)]()
[Efficient Whole Slide Pathology VQA via Token Compression](https://arxiv.org/abs/2507.14497)
Weimin Lyu, Qingqiao Hu, Kehan Qi, Zhan Shi, Wentao Huang, Saumya Gupta, Chao Chen | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2507.14497)
| 32 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.07-red)]()
[Training-free Token Reduction for Vision Mamba](https://arxiv.org/abs/2507.14042)
Qiankun Ma, Ziyao Zhang, Chi Su, Jie Chen, Zhen Song, Hairong Zheng, Wen Gao | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2507.14042)
| 33 | | [![Publish](https://img.shields.io/badge/NeurIPS-2025-blue)]() [![Star](https://img.shields.io/github/stars/dvlab-research/VisionThink.svg?style=social&label=Star)](https://github.com/dvlab-research/VisionThink)
[VisionThink: Smart and Efficient Vision Language Model via Reinforcement Learning](https://arxiv.org/abs/2507.13348)
Senqiao Yang, Junyi Li, Xin Lai, Bei Yu, Hengshuang Zhao, Jiaya Jia | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2507.13348)
[GitHub](https://github.com/dvlab-research/VisionThink)
[Model](https://huggingface.co/collections/Senqiao/visionthink-6878d839fae02a079c9c7bfe)
[Dataset](https://huggingface.co/collections/Senqiao/visionthink-6878d839fae02a079c9c7bfe)
| 34 | | [![Publish](https://img.shields.io/badge/EMNLP_Findings-2024-blue)]()
[LaCo: Efficient Layer-wise Compression of Visual Tokens for Multimodal Large Language Models](https://arxiv.org/abs/2507.02279)
Juntao Liu, Liqiang Niu, Wenchao Chen, Jie Zhou, Fandong Meng | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2507.02279)
| 35 | | [![Publish](https://img.shields.io/badge/IROS-2025-blue)]()
[ToSA: Token Merging with Spatial Awareness](https://arxiv.org/abs/2506.20066)
Hsiang-Wei Huang, Wenhao Chai, Kuang-Ming Chen, Cheng-Yen Yang, Jenq-Neng Hwang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2506.20066)
| 36 | | [![Publish](https://img.shields.io/badge/NeurIPS-2025-blue)]() [![Star](https://img.shields.io/github/stars/Theia-4869/CDPruner.svg?style=social&label=Star)](https://github.com/Theia-4869/CDPruner)
[Beyond Attention or Similarity: Maximizing Conditional Diversity for Token Pruning in MLLMs](https://arxiv.org/abs/2506.10967)
Qizhe Zhang, Mengzhen Liu, Lichen Li, Ming Lu, Yuan Zhang, Junwen Pan, Qi She, Shanghang Zhang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2506.10967)
[GitHub](https://github.com/Theia-4869/CDPruner)
| 37 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.06-red)]()
[Generic Token Compression in Multimodal Large Language Models from an Explainability Perspective](https://arxiv.org/abs/2506.01097)
Lei Lei, Jie Gu, Xiaokang Ma, Chu Tang, Jingmin Chen, Tong Xu | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2506.01097)
| 38 | | [![Publish](https://img.shields.io/badge/ACL-2025-blue)]() [![Star](https://img.shields.io/github/stars/EffiVLM-Bench/EffiVLM-Bench.svg?style=social&label=Star)](https://github.com/EffiVLM-Bench/EffiVLM-Bench)
[EffiVLM-Bench: A Comprehensive Benchmark for Evaluating Training-Free Acceleration in Large Visual-Languge Models](https://arxiv.org/abs/2506.00479)
Zekun Wang, Minghua Ma, Zexin Wang, Rongchuan Mu, Liping Shan, Ming Liu, Bing Qin | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Benchmark-purple)]() | | [Paper](https://arxiv.org/abs/2506.00479)
[GitHub](https://github.com/EffiVLM-Bench/EffiVLM-Bench)
| 39 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.05-red)]() [![Star](https://img.shields.io/github/stars/Tencent/SelfEvolvingAgent.svg?style=social&label=Star)](https://github.com/Tencent/SelfEvolvingAgent/tree/main/VScan)
[VScan: Rethinking Visual Token Reduction for Efficient Large Vision-Language Models](https://arxiv.org/abs/2505.22654)
Ce Zhang, Kaixin Ma, Tianqing Fang, Wenhao Yu, Hongming Zhang, Zhisong Zhang, Yaqi Xie, Katia Sycara, Haitao Mi, Dong Yu | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2505.22654)
[GitHub](https://github.com/Tencent/SelfEvolvingAgent/tree/main/VScan)
| 40 | | [![Publish](https://img.shields.io/badge/NeurIPS-2025-blue)]()
[Balanced Token Pruning: Accelerating Vision Language Models Beyond Local Optimization](https://arxiv.org/abs/2505.22038)
Kaiyuan Li, Xiaoyue Chen, Chen Gao, Yong Li, Xinlei Chen | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2505.22038)
| 41 | | [![Publish](https://img.shields.io/badge/ICML-2025-blue)]() [![Star](https://img.shields.io/github/stars/wangqinsi1/2025-ICML-CoreMatching.svg?style=social&label=Star)](https://github.com/wangqinsi1/2025-ICML-CoreMatching)
[CoreMatching: A Co-adaptive Sparse Inference Framework with Token and Neuron Pruning for Comprehensive Acceleration of Vision-Language Models](https://arxiv.org/abs/2505.19235)
Qinsi Wang, Hancheng Ye, Ming-Yu Chung, Yudong Liu, Yueqian Lin, Martin Kuo, Mingyuan Ma, Jianyi Zhang, Yiran Chen | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2505.19235)
[GitHub](https://github.com/wangqinsi1/2025-ICML-CoreMatching)
| 42 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.05-red)]() [![Star](https://img.shields.io/github/stars/xuyang-liu16/Awesome-Token-level-Model-Compression.svg?style=social&label=Star)](https://github.com/xuyang-liu16/Awesome-Token-level-Model-Compression)
[Shifting AI Efficiency From Model-Centric to Data-Centric Compression](https://arxiv.org/abs/2505.19147)
Xuyang Liu, Zichen Wen, Shaobo Wang, Junjie Chen, Zhishan Tao, Yubo Wang, Xiangqi Jin, Chang Zou, Yiyu Wang, Chenfei Liao, Xu Zheng, Honggang Chen, Weijia Li, Xuming Hu, Conghui He, Linfeng Zhang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Position--Paper-purple)]() | | [Paper](https://arxiv.org/abs/2505.19147)
[GitHub](https://github.com/xuyang-liu16/Awesome-Token-level-Model-Compression)
| 43 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.05-red)]() [![Star](https://img.shields.io/github/stars/ZLKong/Awesome-Collection-Token-Reduction.svg?style=social&label=Star)](https://github.com/ZLKong/Awesome-Collection-Token-Reduction)
[Token Reduction Should Go Beyond Efficiency in Generative Models -- From Vision, Language to Multimodality](https://arxiv.org/abs/2505.18227)
Zhenglun Kong, Yize Li, Fanhu Zeng, Lei Xin, Shvat Messica, Xue Lin, Pu Zhao, Manolis Kellis, Hao Tang, Marinka Zitnik | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() [![Area](https://img.shields.io/badge/Position--Paper-purple)]() | | [Paper](https://arxiv.org/abs/2505.18227)
[GitHub](https://github.com/ZLKong/Awesome-Collection-Token-Reduction)
| 44 | | [![Publish](https://img.shields.io/badge/NeurIPS-2025-blue)]()
[Why 1 + 1 < 1 in Visual Token Pruning: Beyond Naive Integration via Multi-Objective Balanced Covering](https://arxiv.org/abs/2505.10118)
Yangfu Li, Hongjian Zhan, Tianyi Chen, Qi Liu, Yue Lu | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2505.10118)
| 45 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.05-red)]() [![Star](https://img.shields.io/github/stars/ByteDance-Seed/Seed1.5-VL.svg?style=social&label=Star)](https://github.com/ByteDance-Seed/Seed1.5-VL)
[Seed1.5-VL Technical Report](https://arxiv.org/abs/2505.07062)
Seed Team | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2505.07062)
[GitHub](https://github.com/ByteDance-Seed/Seed1.5-VL)
| 46 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.04-red)]() [![Star](https://img.shields.io/github/stars/MikeWangWZHL/dymu.svg?style=social&label=Star)](https://github.com/MikeWangWZHL/dymu)
[DyMU: Dynamic Merging and Virtual Unmerging for Efficient VLMs](https://arxiv.org/abs/2504.17040)
Zhenhailong Wang, Senthil Purushwalkam, Caiming Xiong, Silvio Savarese, Heng Ji, Ran Xu | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2504.17040)
[GitHub](https://github.com/MikeWangWZHL/dymu)
| 47 | | [![Publish](https://img.shields.io/badge/CVPR-2025-blue)]() [![Star](https://img.shields.io/github/stars/orailix/PACT.svg?style=social&label=Star)](https://github.com/orailix/PACT)
[PACT: Pruning and Clustering-Based Token Reduction for Faster Visual Language Models](https://arxiv.org/abs/2504.08966)
Mohamed Dhouib, Davide Buscaldi, Sonia Vanier, Aymen Shabou | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2504.08966)
[GitHub](https://github.com/orailix/PACT)
| 48 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.04-red)]()
[QG-VTC: Question-Guided Visual Token Compression in MLLMs for Efficient VQA](https://arxiv.org/abs/2504.00654)
Shuai Li, Jian Xu, Xiao-Hui Li, Chao Deng, Lin-Lin Huang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2504.00654)
| 49 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]() [![Star](https://img.shields.io/github/stars/zwl666666/Skip-Vision.svg?style=social&label=Star)](https://github.com/zwl666666/Skip-Vision)
[Skip-Vision: Efficient and Scalable Acceleration of Vision-Language Models via Adaptive Token Skipping](https://arxiv.org/abs/2503.21817)
Weili Zeng, Ziyuan Huang, Kaixiang Ji, Yichao Yan | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2503.21817)
[GitHub](https://github.com/zwl666666/Skip-Vision)
| 50 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.03-red)]() [![Star](https://img.shields.io/github/stars/ludc506/InternVL-X.svg?style=social&label=Star)](https://github.com/ludc506/InternVL-X)
[InternVL-X: Advancing and Accelerating InternVL Series with Efficient Visual Token Compression](https://arxiv.org/abs/2503.21307)
Dongchen Lu, Yuyao Sun, Zilu Zhang, Leping Huang, Jianliang Zeng, Mao Shu, Huo Cao | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]() [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2503.21307)
[GitHub](https://github.com/ludc506/InternVL-X)
[Model](https://huggingface.co/LLCC506/InternVL-X-8B-HD)
| 51 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.03-red)]() [![Star](https://img.shields.io/github/stars/QwenLM/Qwen2.5-Omni.svg?style=social&label=Star)](https://github.com/QwenLM/Qwen2.5-Omni)
[Qwen2.5-Omni Technical Report](https://arxiv.org/abs/2503.20215)
Qwen Team | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2503.20215)
[GitHub](https://github.com/QwenLM/Qwen2.5-Omni)
[Model](https://huggingface.co/collections/Qwen/qwen25-omni-67de1e5f0f9464dc6314b36e)
| 52 | | [![Publish](https://img.shields.io/badge/CVPR-2025-blue)]()
[TopV: Compatible Token Pruning with Inference Time Optimization for Fast and Low-Memory Multimodal Vision Language Model](https://arxiv.org/abs/2503.18278)
Cheng Yang, Yang Sui, Jinqi Xiao, Lingyi Huang, Yu Gong, Chendi Li, Jinghua Yan, Yu Bai, Ponnuswamy Sadayappan, Xia Hu, Bo Yuan | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2503.18278)
| 53 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]()
[Growing a Twig to Accelerate Large Vision-Language Models](https://arxiv.org/abs/2503.14075)
Zhenwei Shao, Mingyang Wang, Zhou Yu, Wenwen Pan, Yan Yang, Tao Wei, Hongyuan Zhang, Ning Mao, Wei Chen, Jun Yu | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2503.14075)
| 54 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.03-red)]() [![Star](https://img.shields.io/github/stars/ShawnTan86/TokenCarve.svg?style=social&label=Star)](https://github.com/ShawnTan86/TokenCarve)
[TokenCarve: Information-Preserving Visual Token Compression in Multimodal Large Language Models](https://arxiv.org/abs/2503.10501)
Xudong Tan, Peng Ye, Chongjun Tu, Jianjian Cao, Yaoxin Yang, Lin Zhang, Dongzhan Zhou, Tao Chen | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2503.10501)
[GitHub](https://github.com/ShawnTan86/TokenCarve)
| 55 | | [![Publish](https://img.shields.io/badge/CVPR-2025-blue)]() [![Star](https://img.shields.io/github/stars/vbdi/divprune.svg?style=social&label=Star)](https://github.com/vbdi/divprune)
[DivPrune: Diversity-based Visual Token Pruning for Large Multimodal Models](https://arxiv.org/abs/2503.02175)
Saeed Ranjbar Alvar, Gursimran Singh, Mohammad Akbari, Yong Zhang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2503.02175)
[GitHub](https://github.com/vbdi/divprune)
| 56 | | [![Publish](https://img.shields.io/badge/NAACL-2025-blue)]() [![Star](https://img.shields.io/github/stars/AIoT-MLSys-Lab/MEDA.svg?style=social&label=Star)](https://github.com/AIoT-MLSys-Lab/MEDA)
[MEDA: Dynamic KV Cache Allocation for Efficient Multimodal Long-Context Inference](https://arxiv.org/abs/2502.17599)
Zhongwei Wan, Hui Shen, Xin Wang, Che Liu, Zheda Mai, Mi Zhang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2502.17599)
[GitHub](https://github.com/AIoT-MLSys-Lab/MEDA)
| 57 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.02-red)]() [![Star](https://img.shields.io/github/stars/ZichenWen1/DART.svg?style=social&label=Star)](https://github.com/ZichenWen1/DART)
[Stop Looking for Important Tokens in Multimodal Language Models: Duplication Matters More](https://arxiv.org/abs/2502.11494)
Zichen Wen, Yifeng Gao, Shaobo Wang, Junyuan Zhang, Qintong Zhang, Weijia Li, Conghui He, Linfeng Zhang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() [![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2502.11494)
[GitHub](https://github.com/ZichenWen1/DART)
| 58 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.01-red)]()
[AdaFV: Rethinking of Visual-Language alignment for VLM acceleration](https://arxiv.org/abs/2501.09532)
Jiayi Han, Liang Du, Yiwen Wu, Xiangguo Zhou, Hongwei Du, Weibo Zheng | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2501.09532)
| 59 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.01-red)]() [![Star](https://img.shields.io/github/stars/xuyang-liu16/GlobalCom2.svg?style=social&label=Star)](https://github.com/xuyang-liu16/GlobalCom2)
[Global Compression Commander: Plug-and-Play Inference Acceleration for High-Resolution Large Vision-Language Models](https://arxiv.org/abs/2501.05179)
Xuyang Liu, Ziming Wang, Yuhang Han, Yingyao Wang, Jiale Yuan, Jun Song, Bo Zheng, Linfeng Zhang, Siteng Huang, Honggang Chen | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2501.05179)
[GitHub](https://github.com/xuyang-liu16/GlobalCom2)
| 60 | | [![Publish](https://img.shields.io/badge/ICLR-2025-blue)]() [![Star](https://img.shields.io/github/stars/ictnlp/LLaVA-Mini.svg?style=social&label=Star)](https://github.com/ictnlp/LLaVA-Mini)
[LLaVA-Mini: Efficient Image and Video Large Multimodal Models with One Vision Token](https://arxiv.org/abs/2501.03895)
Shaolei Zhang, Qingkai Fang, Zhe Yang, Yang Feng | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2501.03895)
[GitHub](https://github.com/ictnlp/LLaVA-Mini)
[Model](https://huggingface.co/ICTNLP/llava-mini-llama-3.1-8b)
| 61 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]() [![Star](https://img.shields.io/github/stars/anakin-skywalker-Joseph/Folder.svg?style=social&label=Star)](https://github.com/anakin-skywalker-Joseph/Folder)
[FOLDER: Accelerating Multi-modal Large Language Models with Enhanced Performance](https://arxiv.org/abs/2501.02430)
Haicheng Wang, Zhemeng Yu, Gabriele Spadaro, Chen Ju, Victor Quétu, Shuai Xiao, Enzo Tartaglione | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2501.02430)
[GitHub](https://github.com/anakin-skywalker-Joseph/Folder)
| 62 | | [![Publish](https://img.shields.io/badge/AAAI-2025-blue)]() [![Star](https://img.shields.io/github/stars/jytmelon/G-Prune.svg?style=social&label=Star)](https://github.com/jytmelon/G-Prune)
[What Kind of Visual Tokens Do We Need? Training-free Visual Token Pruning for Multi-modal Large Language Models from the Perspective of Graph](https://arxiv.org/abs/2501.02268)
Yutao Jiang, Qiong Wu, Wenhao Lin, Wei Yu, Yiyi Zhou | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2501.02268)
[GitHub](https://github.com/jytmelon/G-Prune)
| 63 |
64 | 65 |
66 | 2024 Image 67 | 68 | | **Title & Authors** | **Areas** | **Tags** | **Links** | 69 | | --- | --- | --- | :---: | 70 | | [![Publish](https://img.shields.io/badge/AAAI-2025-blue)]()
[ST3: Accelerating Multimodal Large Language Model by Spatial-Temporal Visual Token Trimming](https://arxiv.org/abs/2412.20105)
Jiedong Zhuang, Lu Lu, Ming Dai, Rui Hu, Jian Chen, Qiang Liu, Haoji Hu | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2412.20105)
| 71 | | [![Publish](https://img.shields.io/badge/CVPR-2025-blue)]() [![Star](https://img.shields.io/github/stars/OpenGVLab/PVC.svg?style=social&label=Star)](https://github.com/OpenGVLab/PVC)
[PVC: Progressive Visual Token Compression for Unified Image and Video Processing in Large Vision-Language Models](https://arxiv.org/abs/2412.09613)
Chenyu Yang, Xuan Dong, Xizhou Zhu, Weijie Su, Jiahao Wang, Hao Tian, Zhe Chen, Wenhai Wang, Lewei Lu, Jifeng Dai | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2412.09613)
[GitHub](https://github.com/OpenGVLab/PVC)
[Model](https://huggingface.co/OpenGVLab/PVC-InternVL2-8B)
| 72 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]() [![Star](https://img.shields.io/github/stars/dvlab-research/Lyra.svg?style=social&label=Star)](https://github.com/dvlab-research/Lyra)
[Lyra: An Efficient and Speech-Centric Framework for Omni-Cognition](https://arxiv.org/abs/2412.09501)
Zhisheng Zhong, Chengyao Wang, Yuqi Liu, Senqiao Yang, Longxiang Tang, Yuechen Zhang, Jingyao Li, Tianyuan Qu, Yanwei Li, Yukang Chen, Shaozuo Yu, Sitong Wu, Eric Lo, Shu Liu, Jiaya Jia | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2412.09501)
[GitHub](https://github.com/dvlab-research/Lyra)
[Model](https://huggingface.co/collections/zszhong/lyra-model-674ea5bb3b39ff8f15de75fc)
[Dataset](https://huggingface.co/collections/zszhong/lyra-data-675d80fbab80334eb52cdd82)
| 73 | | [![Arxiv](https://img.shields.io/badge/arXiv-2024\.12-red)]() [![Star](https://img.shields.io/github/stars/hulianyuyy/iLLaVA.svg?style=social&label=Star)](https://github.com/hulianyuyy/iLLaVA)
[iLLaVA: An Image is Worth Fewer Than 1/3 Input Tokens in Large Multimodal Models](https://arxiv.org/abs/2412.06263)
Lianyu Hu, Fanhua Shang, Liang Wan, Wei Feng | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2412.06263)
[GitHub](https://github.com/hulianyuyy/iLLaVA)
| 74 | | [![Publish](https://img.shields.io/badge/CVPR-2025-blue)]() [![Star](https://img.shields.io/github/stars/dvlab-research/VisionZip.svg?style=social&label=Star)](https://github.com/dvlab-research/VisionZip)
[VisionZip: Longer is Better but Not Necessary in Vision Language Models](https://arxiv.org/abs/2412.04467)
Senqiao Yang, Yukang Chen, Zhuotao Tian, Chengyao Wang, Jingyao Li, Bei Yu, Jiaya Jia | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2412.04467)
[GitHub](https://github.com/dvlab-research/VisionZip)
| 75 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]() [![Star](https://img.shields.io/github/stars/Theia-4869/VisPruner.svg?style=social&label=Star)](https://github.com/Theia-4869/VisPruner)
[Beyond Text-Visual Attention: Exploiting Visual Cues for Effective Token Pruning in VLMs](https://arxiv.org/abs/2412.01818)
Qizhe Zhang, Aosong Cheng, Ming Lu, Renrui Zhang, Zhiyong Zhuo, Jiajun Cao, Shaobo Guo, Qi She, Shanghang Zhang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2412.01818)
[GitHub](https://github.com/Theia-4869/VisPruner)
| 76 | | [![Publish](https://img.shields.io/badge/CVPR-2025-blue)]()
[Accelerating Multimodal Large Language Models by Searching Optimal Vision Token Reduction](https://arxiv.org/abs/2412.00556)
Shiyu Zhao, Zhenting Wang, Felix Juefei-Xu, Xide Xia, Miao Liu, Xiaofang Wang, Mingfu Liang, Ning Zhang, Dimitris N. Metaxas, Licheng Yu | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2412.00556)
| 77 | | [![Publish](https://img.shields.io/badge/CVPR-2025-blue)]()
[ATP-LLaVA: Adaptive Token Pruning for Large Vision Language Models](https://arxiv.org/abs/2412.00447)
Xubing Ye, Yukang Gan, Yixiao Ge, Xiao-Ping Zhang, Yansong Tang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2412.00447)
| 78 | | [![Arxiv](https://img.shields.io/badge/arXiv-2024\.11-red)]()
[Efficient Multi-modal Large Language Models via Visual Token Grouping](https://arxiv.org/abs/2411.17773)
Minbin Huang, Runhui Huang, Han Shi, Yimeng Chen, Chuanyang Zheng, Xiangguo Sun, Xin Jiang, Zhenguo Li, Hong Cheng | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2411.17773)
| 79 | | [![Arxiv](https://img.shields.io/badge/arXiv-2024\.11-red)]() [![Star](https://img.shields.io/github/stars/kawhiiiileo/FiCoCo.svg?style=social&label=Star)](https://github.com/kawhiiiileo/FiCoCo)
[Filter, Correlate, Compress: Training-Free Token Reduction for MLLM Acceleration](https://arxiv.org/abs/2411.17686)
Yuhang Han, Xuyang Liu, Zihan Zhang, Pengxiang Ding, Donglin Wang, Honggang Chen, Qingsen Yan, Siteng Huang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2411.17686)
[GitHub](https://github.com/kawhiiiileo/FiCoCo)
| 80 | | [![Arxiv](https://img.shields.io/badge/arXiv-2024\.11-red)]()
[FocusLLaVA: A Coarse-to-Fine Approach for Efficient and Effective Visual Token Compression](https://arxiv.org/abs/2411.14228)
Yuke Zhu, Chi Xie, Shuang Liang, Bo Zheng, Sheng Guo | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2411.14228)
| 81 | | [![Publish](https://img.shields.io/badge/CVPR_Highlight-2025-blue)]()
[AdaCM2: Adaptive Cross‑Modality Memory Reduction](https://arxiv.org/abs/2411.12593)
Yuanbin Man, Ying Huang, Chengming Zhang, Bingzhe Li, Wei Niu, Miao Yin | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2411.12593)
| 82 | | [![Arxiv](https://img.shields.io/badge/arXiv-2024\.11-red)]() [![Star](https://img.shields.io/github/stars/liuting20/MustDrop.svg?style=social&label=Star)](https://github.com/liuting20/MustDrop)
[Multi-Stage Vision Token Dropping: Towards Efficient Multimodal Large Language Model](https://arxiv.org/abs/2411.10803)
Ting Liu, Liangtao Shi, Richang Hong, Yue Hu, Quanjun Yin, Linfeng Zhang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2411.10803)
[GitHub](https://github.com/liuting20/MustDrop)
| 83 | | [![Publish](https://img.shields.io/badge/CVPR-2025-blue)]() [![Star](https://img.shields.io/github/stars/Cooperx521/PyramidDrop.svg?style=social&label=Star)](https://github.com/Cooperx521/PyramidDrop)
[PyramidDrop: Accelerating Your Large Vision-Language Models via Pyramid Visual Redundancy Reduction](https://arxiv.org/abs/2410.17247)
Long Xing, Qidong Huang, Xiaoyi Dong, Jiajie Lu, Pan Zhang, Yuhang Zang, Yuhang Cao, Conghui He, Jiaqi Wang, Feng Wu, Dahua Lin | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() [![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2410.17247)
[GitHub](https://github.com/Cooperx521/PyramidDrop)
| 84 | | [![Arxiv](https://img.shields.io/badge/arXiv-2024\.10-red)]()
[Efficient Vision-Language Models by Summarizing Visual Tokens into Compact Registers](https://arxiv.org/abs/2410.14072)
Yuxin Wen, Qingqing Cao, Qichen Fu, Sachin Mehta, Mahyar Najibi | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2410.14072)
| 85 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]()
[ZipVL: Efficient Large Vision-Language Models with Dynamic Token Sparsification](https://arxiv.org/abs/2410.08584)
Yefei He, Feng Chen, Jing Liu, Wenqi Shao, Hong Zhou, Kaipeng Zhang, Bohan Zhuang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2410.08584)
| 86 | | [![Publish](https://img.shields.io/badge/ICML-2025-blue)]() [![Star](https://img.shields.io/github/stars/Gumpest/SparseVLMs.svg?style=social&label=Star)](https://github.com/Gumpest/SparseVLMs)
[SparseVLM: Visual Token Sparsification for Efficient Vision-Language Model Inference](https://arxiv.org/abs/2410.04417)
Yuan Zhang, Chun-Kai Fan, Junpeng Ma, Wenzhao Zheng, Tao Huang, Kuan Cheng, Denis Gudovskiy, Tomoyuki Okuno, Yohei Nakata, Kurt Keutzer, Shanghang Zhang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2410.04417)
[GitHub](https://github.com/Gumpest/SparseVLMs)
| 87 | | [![Publish](https://img.shields.io/badge/ICLR-2025-blue)]() [![Star](https://img.shields.io/github/stars/rese1f/aurora.svg?style=social&label=Star)](https://github.com/rese1f/aurora)
[AuroraCap: Efficient, Performant Video Detailed Captioning and a New Benchmark](https://arxiv.org/abs/2410.03051)
Wenhao Chai, Enxin Song, Yilun Du, Chenlin Meng, Vashisht Madhavan, Omer Bar-Tal, Jenq-Neng Hwang, Saining Xie, Christopher D. Manning | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Benchmark-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2410.03051)
[GitHub](https://github.com/rese1f/aurora)
[Model](https://huggingface.co/collections/wchai/auroracap-66d117ffe13bedda96702013)
[Dataset](https://huggingface.co/datasets/wchai/Video-Detailed-Caption)
| 88 | | [![Arxiv](https://img.shields.io/badge/arXiv-2024\.10-red)]() [![Star](https://img.shields.io/github/stars/LLaVA-VL/LLaVA-NeXT.svg?style=social&label=Star)](https://github.com/LLaVA-VL/LLaVA-NeXT)
[Video Instruction Tuning with Synthetic Data](https://arxiv.org/abs/2410.02713)
Yuanhan Zhang, Jinming Wu, Wei Li, Bo Li, Zejun Ma, Ziwei Liu, Chunyuan Li | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2410.02713)
[GitHub](https://github.com/LLaVA-VL/LLaVA-NeXT)
[Model](https://huggingface.co/collections/lmms-lab/llava-video-661e86f5e8dabc3ff793c944)
| 89 | | [![Arxiv](https://img.shields.io/badge/arXiv-2024\.09-red)]() [![Star](https://img.shields.io/github/stars/NVIDIA/Megatron-LM.svg?style=social&label=Star)](https://github.com/NVIDIA/Megatron-LM/tree/NVLM-1.0/examples/multimodal/nvlm)
[NVLM: Open Frontier-Class Multimodal LLMs](https://arxiv.org/abs/2409.11402)
Wenliang Dai, Nayeon Lee, Boxin Wang, Zhuolin Yang, Zihan Liu, Jon Barker, Tuomas Rintamaki, Mohammad Shoeybi, Bryan Catanzaro, Wei Ping | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2409.11402)
[GitHub](https://github.com/NVIDIA/Megatron-LM/tree/NVLM-1.0/examples/multimodal/nvlm)
[Model](https://huggingface.co/collections/nvidia/nvlm-10-66e9f407c764a0ee6e37b7f4)
| 90 | | [![Publish](https://img.shields.io/badge/COLING-2025-blue)]() [![Star](https://img.shields.io/github/stars/FreedomIntelligence/TRIM.svg?style=social&label=Star)](https://github.com/FreedomIntelligence/TRIM)
[Less is More: A Simple yet Effective Token Reduction Method for Efficient Multi-modal LLMs](https://arxiv.org/abs/2409.10994)
Dingjie Song, Wenjun Wang, Shunian Chen, Xidong Wang, Michael Guan, Benyou Wang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2409.10994)
[GitHub](https://github.com/FreedomIntelligence/TRIM)
| 91 | | [![Publish](https://img.shields.io/badge/AAAI-2025-blue)]() [![Star](https://img.shields.io/github/stars/ywh187/FitPrune.svg?style=social&label=Star)](https://github.com/ywh187/FitPrune)
[Fit and Prune: Fast and Training-free Visual Token Pruning for Multi-modal Large Language Models](https://arxiv.org/abs/2409.10197)
Weihao Ye, Qiong Wu, Wenhao Lin, Yiyi Zhou | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2409.10197)
[GitHub](https://github.com/ywh187/FitPrune)
| 92 | | [![Publish](https://img.shields.io/badge/AAAI-2025-blue)]() [![Star](https://img.shields.io/github/stars/hasanar1f/HiRED.svg?style=social&label=Star)](https://github.com/hasanar1f/HiRED)
[HiRED: Attention-Guided Token Dropping for Efficient Inference of High-Resolution Vision-Language Models](https://arxiv.org/abs/2408.10945)
Kazi Hasan Ibn Arif, JinYi Yoon, Dimitrios S. Nikolopoulos, Hans Vandierendonck, Deepu John, Bo Ji | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2408.10945)
[GitHub](https://github.com/hasanar1f/HiRED)
| 93 | | [![Publish](https://img.shields.io/badge/Trans._Mach._Learn._Res.-2025-blue)]() [![Star](https://img.shields.io/github/stars/LLaVA-VL/LLaVA-NeXT.svg?style=social&label=Star)](https://github.com/LLaVA-VL/LLaVA-NeXT)
[LLaVA-OneVision: Easy Visual Task Transfer](https://arxiv.org/abs/2408.03326)
Bo Li, Yuanhan Zhang, Dong Guo, Renrui Zhang, Feng Li, Hao Zhang, Kaichen Zhang, Peiyuan Zhang, Yanwei Li, Ziwei Liu, Chunyuan Li | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2408.03326)
[GitHub](https://github.com/LLaVA-VL/LLaVA-NeXT)
[Model](https://huggingface.co/collections/lmms-lab/llava-onevision-66a259c3526e15166d6bba37)
| 94 | | [![Arxiv](https://img.shields.io/badge/arXiv-2024\.07-red)]() [![Star](https://img.shields.io/github/stars/JiuTian-VL/TokenCorrCompressor.svg?style=social&label=Star)](https://github.com/JiuTian-VL/TokenCorrCompressor)
[Token-level Correlation-guided Compression for Efficient Multimodal Document Understanding](https://arxiv.org/abs/2407.14439)
Renshan Zhang, Yibo Lyu, Rui Shao, Gongwei Chen, Weili Guan, Liqiang Nie | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() [![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2407.14439)
[GitHub](https://github.com/JiuTian-VL/TokenCorrCompressor)
| 95 | | [![Publish](https://img.shields.io/badge/IJCV-2025-blue)]() [![Star](https://img.shields.io/github/stars/CircleRadon/TokenPacker.svg?style=social&label=Star)](https://github.com/CircleRadon/TokenPacker)
[TokenPacker: Efficient Visual Projector for Multimodal LLM](https://arxiv.org/abs/2407.02392)
Wentong Li, Yuqian Yuan, Jian Liu, Dongqi Tang, Song Wang, Jie Qin, Jianke Zhu, Lei Zhang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2407.02392)
[GitHub](https://github.com/CircleRadon/TokenPacker)
| 96 | | [![Publish](https://img.shields.io/badge/EMNLP_Findings-2024-blue)]() [![Star](https://img.shields.io/github/stars/SUSTechBruce/LOOK-M.svg?style=social&label=Star)](https://github.com/SUSTechBruce/LOOK-M)
[LOOK-M: Look-Once Optimization in KV Cache for Efficient Multimodal Long-Context Inference](https://arxiv.org/abs/2406.18139)
Zhongwei Wan, Ziang Wu, Che Liu, Jinfa Huang, Zhihong Zhu, Peng Jin, Longyue Wang, Li Yuan | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2406.18139)
[GitHub](https://github.com/SUSTechBruce/LOOK-M)
| 97 | | [![Publish](https://img.shields.io/badge/Trans._Mach._Learn._Res.-2025-blue)]() [![Star](https://img.shields.io/github/stars/EvolvingLMMs-Lab/LongVA.svg?style=social&label=Star)](https://github.com/EvolvingLMMs-Lab/LongVA)
[Long Context Transfer from Language to Vision](https://arxiv.org/abs/2406.16852)
Peiyuan Zhang, Kaichen Zhang, Bo Li, Guangtao Zeng, Jingkang Yang, Yuanhan Zhang, Ziyue Wang, Haoran Tan, Chunyuan Li, Ziwei Liu | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2406.16852)
[GitHub](https://github.com/EvolvingLMMs-Lab/LongVA)
| 98 | | [![Publish](https://img.shields.io/badge/ICML-2024-blue)]() [![Star](https://img.shields.io/github/stars/bytedance/SALMONN.svg?style=social&label=Star)](https://github.com/bytedance/SALMONN/tree/videosalmonn)
[video-SALMONN: Speech-Enhanced Audio-Visual Large Language Models](https://arxiv.org/abs/2406.15704)
Guangzhi Sun, Wenyi Yu, Changli Tang, Xianzhao Chen, Tian Tan, Wei Li, Lu Lu, Zejun Ma, Yuxuan Wang, Chao Zhang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2406.15704)
[GitHub](https://github.com/bytedance/SALMONN/tree/videosalmonn)
[Model](https://huggingface.co/tsinghua-ee/Video-SALMONN/tree/main)
| 99 | | [![Publish](https://img.shields.io/badge/CVPR-2025-blue)]() [![Star](https://img.shields.io/github/stars/Yxxxb/VoCo-LLaMA.svg?style=social&label=Star)](https://github.com/Yxxxb/VoCo-LLaMA)
[VoCo-LLaMA: Towards Vision Compression with Large Language Models](https://arxiv.org/abs/2406.12275)
Xubing Ye, Yukang Gan, Xiaoke Huang, Yixiao Ge, Yansong Tang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2406.12275)
[GitHub](https://github.com/Yxxxb/VoCo-LLaMA)
| 100 | | [![Publish](https://img.shields.io/badge/ICLR-2025-blue)]() [![Star](https://img.shields.io/github/stars/mu-cai/matryoshka-mm.svg?style=social&label=Star)](https://github.com/mu-cai/matryoshka-mm)
[Matryoshka Multimodal Models](https://arxiv.org/abs/2405.17430)
Mu Cai, Jianwei Yang, Jianfeng Gao, Yong Jae Lee | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2405.17430)
[GitHub](https://github.com/mu-cai/matryoshka-mm)
| 101 | | [![Publish](https://img.shields.io/badge/AAAI_Oral-2025-blue)]() [![Star](https://img.shields.io/github/stars/lzhxmu/VTW.svg?style=social&label=Star)](https://github.com/lzhxmu/VTW)
[Boosting multimodal large language models with visual tokens withdrawal for rapid inference.](https://arxiv.org/abs/2405.05803)
Zhihang Lin, Mingbao Lin, Luxi Lin, Rongrong Ji | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2405.05803)
[GitHub](https://github.com/lzhxmu/VTW)
| 102 | | [![Arxiv](https://img.shields.io/badge/arXiv-2024\.04-red)]() [![Star](https://img.shields.io/github/stars/OpenGVLab/InternVL.svg?style=social&label=Star)](https://github.com/OpenGVLab/InternVL)
[How Far Are We to GPT-4V? Closing the Gap to Commercial Multimodal Models with Open-Source Suites](https://arxiv.org/abs/2404.16821)
InternVL Team | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2404.16821)
[GitHub](https://github.com/OpenGVLab/InternVL)
[Model](https://huggingface.co/collections/OpenGVLab/internvl15-6675ae031d45e5a07007f260)
| 103 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]() [![Star](https://img.shields.io/github/stars/42Shawn/LLaVA-PruMerge.svg?style=social&label=Star)](https://github.com/42Shawn/LLaVA-PruMerge)
[LLaVA-PruMerge: Adaptive Token Reduction for Efficient Large Multimodal Models](https://arxiv.org/abs/2403.15388)
Yuzhang Shang, Mu Cai, Bingxin Xu, Yong Jae Lee, Yan Yan | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2403.15388)
[GitHub](https://github.com/42Shawn/LLaVA-PruMerge)
| 104 | | [![Publish](https://img.shields.io/badge/ECCV_Oral-2024-blue)]() [![Star](https://img.shields.io/github/stars/pkunlp-icler/FastV.svg?style=social&label=Star)](https://github.com/pkunlp-icler/FastV)
[An Image is Worth 1/2 Tokens After Layer 2: Plug-and-Play Inference Acceleration for Large Vision-Language Models](https://arxiv.org/abs/2403.06764)
Liang Chen, Haozhe Zhao, Tianyu Liu, Shuai Bai, Junyang Lin, Chang Zhou, Baobao Chang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2403.06764)
[GitHub](https://github.com/pkunlp-icler/FastV)
| 105 | | [![Arxiv](https://img.shields.io/badge/arXiv-2024\.02-red)]() [![Star](https://img.shields.io/github/stars/Meituan-AutoML/MobileVLM.svg?style=social&label=Star)](https://github.com/Meituan-AutoML/MobileVLM)
[MobileVLM V2: Faster and Stronger Baseline for Vision Language Model](https://arxiv.org/abs/2402.03766)
Xiangxiang Chu, Limeng Qiao, Xinyu Zhang, Shuang Xu, Fei Wei, Yang Yang, Xiaofei Sun, Yiming Hu, Xinyang Lin, Bo Zhang, Chunhua Shen | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2402.03766)
[GitHub](https://github.com/Meituan-AutoML/MobileVLM)
[Model](https://huggingface.co/mtgv/models)
| 106 |
107 | 108 |
109 | 2023 Image 110 | 111 | | **Title & Authors** | **Areas** | **Tags** | **Links** | 112 | | --- | --- | --- | :---: | 113 | | [![Arxiv](https://img.shields.io/badge/arXiv-2023\.12-red)]() [![Star](https://img.shields.io/github/stars/Meituan-AutoML/MobileVLM.svg?style=social&label=Star)](https://github.com/Meituan-AutoML/MobileVLM)
[MobileVLM : A Fast, Strong and Open Vision Language Assistant for Mobile Devices](https://arxiv.org/abs/2312.16886)
Xiangxiang Chu, Limeng Qiao, Xinyang Lin, Shuang Xu, Yang Yang, Yiming Hu, Fei Wei, Xinyu Zhang, Bo Zhang, Xiaolin Wei, Chunhua Shen | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2312.16886)
[GitHub](https://github.com/Meituan-AutoML/MobileVLM)
[Model](https://huggingface.co/mtgv/models)
| 114 | | [![Publish](https://img.shields.io/badge/CVPR-2024-blue)]() [![Star](https://img.shields.io/github/stars/khanrc/honeybee.svg?style=social&label=Star)](https://github.com/khanrc/honeybee?tab=readme-ov-file)
[Honeybee: Locality-enhanced Projector for Multimodal LLM](https://arxiv.org/abs/2312.06742)
Junbum Cha, Wooyoung Kang, Jonghwan Mun, Byungseok Roh | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2312.06742)
[GitHub](https://github.com/khanrc/honeybee?tab=readme-ov-file)
| 115 | | [![Publish](https://img.shields.io/badge/ECCV-2024-blue)]() [![Star](https://img.shields.io/github/stars/dvlab-research/LLaMA-VID.svg?style=social&label=Star)](https://github.com/dvlab-research/LLaMA-VID)
[LLaMA-VID: An Image is Worth 2 Tokens in Large Language Models](https://arxiv.org/abs/2311.17043)
Yanwei Li, Chengyao Wang, Jiaya Jia | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]() [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2311.17043)
[GitHub](https://github.com/dvlab-research/LLaMA-VID)
[Model](https://huggingface.co/collections/YanweiLi/llama-vid-656741a92f3ec92d7e484dea)
[Dataset](https://huggingface.co/datasets/YanweiLi/LLaMA-VID-Data/tree/main)
| 116 | | [![Publish](https://img.shields.io/badge/CVPR_Highlight-2024-blue)]() [![Star](https://img.shields.io/github/stars/PKU-YuanGroup/Chat-UniVi.svg?style=social&label=Star)](https://github.com/PKU-YuanGroup/Chat-UniVi)
[Chat-UniVi: Unified Visual Representation Empowers Large Language Models with Image and Video Understanding](https://arxiv.org/abs/2311.08046)
Peng Jin, Ryuichi Takanobu, Wancai Zhang, Xiaochun Cao, Li Yuan | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2311.08046)
[GitHub](https://github.com/PKU-YuanGroup/Chat-UniVi)
[Model](https://huggingface.co/collections/Chat-UniVi/chat-univi-66f4265ee4c51e5acf255f2e)
[Dataset](https://github.com/PKU-YuanGroup/Chat-UniVi/blob/main/DATA.md)
| 117 | | [![Arxiv](https://img.shields.io/badge/arXiv-2023\.08-red)]() [![Star](https://img.shields.io/github/stars/QwenLM/Qwen-VL.svg?style=social&label=Star)](https://github.com/QwenLM/Qwen-VL)
[Qwen-VL: A Versatile Vision-Language Model for Understanding, Localization, Text Reading, and Beyond](https://arxiv.org/abs/2308.12966)
Jinze Bai, Shuai Bai, Shusheng Yang, Shijie Wang, Sinan Tan, Peng Wang, Junyang Lin, Chang Zhou, Jingren Zhou | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2308.12966)
[GitHub](https://github.com/QwenLM/Qwen-VL)
[Model](https://huggingface.co/Qwen/Qwen-VL)
| 118 | | [![Publish](https://img.shields.io/badge/EMNLP-2023-blue)]() [![Star](https://img.shields.io/github/stars/DAMO-NLP-SG/Video-LLaMA.svg?style=social&label=Star)](https://github.com/DAMO-NLP-SG/Video-LLaMA)
[Video-LLaMA: An Instruction-tuned Audio-Visual Language Model for Video Understanding](https://arxiv.org/abs/2306.02858)
Hang Zhang, Xin Li, Lidong Bing | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2306.02858)
[GitHub](https://github.com/DAMO-NLP-SG/Video-LLaMA)
[Model](https://huggingface.co/DAMO-NLP-SG/Video-LLaMA-Series)
| 119 | | [![Publish](https://img.shields.io/badge/NeurIPS-2023-blue)]() [![Star](https://img.shields.io/github/stars/salesforce/LAVIS.svg?style=social&label=Star)](https://github.com/salesforce/LAVIS/tree/main/projects/instructblip)
[InstructBLIP: Towards General-purpose Vision-Language Models with Instruction Tuning](https://arxiv.org/abs/2305.06500)
Wenliang Dai, Junnan Li, Dongxu Li, Anthony Meng Huat Tiong, Junqi Zhao, Weisheng Wang, Boyang Li, Pascale Fung, Steven Hoi | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2305.06500)
[GitHub](https://github.com/salesforce/LAVIS/tree/main/projects/instructblip)
| 120 | | [![Arxiv](https://img.shields.io/badge/arXiv-2023\.04-red)]() [![Star](https://img.shields.io/github/stars/X-PLUG/mPLUG-Owl.svg?style=social&label=Star)](https://github.com/X-PLUG/mPLUG-Owl)
[mPLUG-Owl: Modularization Empowers Large Language Models with Multimodality](https://arxiv.org/abs/2304.14178)
DAMO Team | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2304.14178)
[GitHub](https://github.com/X-PLUG/mPLUG-Owl)
| 121 | | [![Publish](https://img.shields.io/badge/ICLR-2024-blue)]() [![Star](https://img.shields.io/github/stars/Vision-CAIR/MiniGPT-4.svg?style=social&label=Star)](https://github.com/Vision-CAIR/MiniGPT-4)
[Minigpt-4: Enhancing vision-language understanding with advanced large language models.](https://arxiv.org/abs/2304.10592)
Deyao Zhu, Jun Chen, Xiaoqian Shen, Xiang Li, Mohamed Elhoseiny | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2304.10592)
[GitHub](https://github.com/Vision-CAIR/MiniGPT-4)
[Model](https://huggingface.co/Vision-CAIR/MiniGPT-4)
[Dataset](https://github.com/Vision-CAIR/MiniGPT-4/tree/main/dataset)
| 122 | | [![Publish](https://img.shields.io/badge/ICML-2023-blue)]() [![Star](https://img.shields.io/github/stars/salesforce/LAVIS.svg?style=social&label=Star)](https://github.com/salesforce/LAVIS/tree/main/projects/blip2)
[BLIP-2: Bootstrapping Language-Image Pre-training with Frozen Image Encoders and Large Language Models](https://arxiv.org/abs/2301.12597)
Junnan Li, Dongxu Li, Silvio Savarese, Steven Hoi | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2301.12597)
[GitHub](https://github.com/salesforce/LAVIS/tree/main/projects/blip2)
| 123 |
124 | 125 |
126 | 2022 Image 127 | 128 | | **Title & Authors** | **Areas** | **Tags** | **Links** | 129 | | --- | --- | --- | :---: | 130 | | [![Publish](https://img.shields.io/badge/NeurIPS-2022-blue)]()
[Flamingo: a Visual Language Model for Few-Shot Learning](https://arxiv.org/abs/2204.14198)
DeepMind Team | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2204.14198)
| 131 |
132 | -------------------------------------------------------------------------------- /README.md: -------------------------------------------------------------------------------- 1 |
2 | 3 | # Awesome Multimodal Token Compression 4 | 5 | [![License: MIT](https://img.shields.io/badge/License-MIT-yellow.svg)](https://opensource.org/licenses/MIT) 6 | [![PRs Welcome](https://img.shields.io/badge/PRs-welcome-brightgreen.svg?style=flat-square)](http://makeapullrequest.com) 7 | [![arXiv](https://img.shields.io/badge/arXiv-2507\.20198-red.svg)](https://arxiv.org/abs/2507.20198) 8 | [![Last Commit](https://img.shields.io/github/last-commit/cokeshao/Awesome-Multimodal-Token-Compression.svg?style=flat&color=orange)](https://github.com/cokeshao/Awesome-Multimodal-Token-Compression) 9 | 10 | [[arXiv]](https://arxiv.org/abs/2507.20198) [[HuggingFace]](https://huggingface.co/papers/2507.20198) [[Database]](https://oasis-paddleboat-fc1.notion.site/when-tokens-talk-too-much-database) 11 | 12 |
13 | 14 | > **When Tokens Talk Too Much: A Survey of Multimodal Long-Context Token Compression across Images, Videos, and Audios** [[arXiv]](https://arxiv.org/pdf/2507.20198) 15 | > [Kele Shao](https://cokeshao.github.io/)\*,1,2, [Keda Tao](https://kd-tao.github.io/)\*,1,2, [Kejia Zhang](https://kejiazhang-robust.github.io/)3, [Sicheng Feng](https://fscdc.github.io/)2,4, [Mu Cai](https://pages.cs.wisc.edu/~mucai/)5, [Yuzhang Shang](https://42shawn.github.io/)6, [Haoxuan You](https://hxyou.github.io/)7, [Can Qin](https://canqin.tech/)8, [Yang Sui](https://eclipsess.github.io/yangsui.github.io/)9, [Huan Wang](https://huanwang.tech/)†,2 16 | > 17 | > 1Zhejiang University, 2Westlake University, 3Xiamen University, 4National University of Singapore, 5University of Wisconsin-Madison, 6University of Central Florida, 7Columbia University, 8Salesforce AI Research, 9Rice University 18 | > 19 | > \* Equal Contribution. † Corresponding Author (wanghuan@westlake.edu.cn). 20 | 21 | --- 22 | 23 | > [!IMPORTANT] 24 | > We welcome your help in improving the repository and paper. Please feel free to submit a [pull request](https://github.com/cokeshao/Awesome-Multimodal-Token-Compression/pulls) or [contact us](#️-contact) to: 25 | > 26 | > - Add a relevant paper not yet included. 27 | > 28 | > - Suggest a more suitable category. 29 | > 30 | > - Update the information. 31 | > 32 | > - Ask for clarification about any content. 33 | 34 | --- 35 | 36 | ## 🔥 News 37 | 38 | - **[2025.10.11]** Papers accepted by **NeurIPS'25** about MLLM token compression have been updated [here](#published-in-recent-conferencejournal). Congratulations! 🎉🎉🎉 39 | - **[2025.08.14]** ❗ Added [Recent Papers](#recent-papers-last-6-months), [Papers Published in Recent Conference/Journal](#published-in-recent-conferencejournal), and a [database](https://oasis-paddleboat-fc1.notion.site/when-tokens-talk-too-much-database) for quick-search. 40 | - **[2025.07.29]** The v1 survey is now published! We've also initialized the repository. 41 | 42 | ## 🎯 Motivation 43 |
44 | Awesome Token Compression 45 |
46 | 47 | > **Motivation:** **Up:** Image, video, and audio data types can scale in their representation dimensions, leading to a corresponding increase in the number of tokens. **Down:** Top-performing MLLMs cannot address real-world demands, as the number of tokens for multimodal information, especially video, vastly exceeds that of text. Therefore, token compression is crucial to address this limitation. 48 | 49 | ## 📌 Citation 50 | 51 | If you find our paper or this resource helpful, please consider cite: 52 | 53 | ```bibtex 54 | @article{shao2025tokens, 55 | title={When Tokens Talk Too Much: A Survey of Multimodal Long-Context Token Compression across Images, Videos, and Audios}, 56 | author={Shao, Kele and Tao, Keda and Zhang, Kejia and Feng, Sicheng and Cai, Mu and Shang, Yuzhang and You, Haoxuan and Qin, Can and Sui, Yang and Wang, Huan}, 57 | journal={arXiv preprint arXiv:2507.20198}, 58 | year={2025} 59 | } 60 | ``` 61 | 62 | ## 📚 Contents 63 | 64 | - [Awesome Token Compression](#awesome-multimodal-token-compression) 65 | - [Image LLM](https://github.com/cokeshao/Awesome-Multimodal-Token-Compression/tree/main/image-llm.md) 66 | - [Video LLM](https://github.com/cokeshao/Awesome-Multimodal-Token-Compression/tree/main/video-llm.md) 67 | - [Audio LLM](https://github.com/cokeshao/Awesome-Multimodal-Token-Compression/tree/main/audio-llm.md) 68 | - [Vision Transformer](https://github.com/cokeshao/Awesome-Multimodal-Token-Compression/tree/main/vision-transformer.md) 69 | - [Audio Transformer](https://github.com/cokeshao/Awesome-Multimodal-Token-Compression/tree/main/audio-transformer.md) 70 | 71 | **Please check out all the papers by selecting the sub-area you're interested in. On this main page, only papers released in the past 6 months are shown.** 72 | 73 | --- 74 | 75 | ### Badge Colors 76 | - ![arXiv Badge](https://img.shields.io/badge/arXiv-red) `red` for arXiv papers 77 | - ![PDF Badge](https://img.shields.io/badge/PDF-blue) `blue` for conference/journal papers 78 | - ![GitHub Badge](https://img.shields.io/badge/GitHub-white) `white` for GitHub repositories 79 | - ![Research Areas Badge](https://img.shields.io/badge/Areas-purple) `purple` for research areas 80 | - ![Categories Badge](https://img.shields.io/badge/Categories-green) `green` for categories 81 | - ![Cost Badge](https://img.shields.io/badge/Cost-yellow) `yellow` for training cost 82 | 83 | ### Recent Papers (Last 6 Months) 84 | 85 | 86 |
87 | Image 88 | 89 | | **Title & Authors** | **Areas** | **Tags** | **Links** | 90 | | --- | --- | --- | :---: | 91 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.10-red)]() [![Star](https://img.shields.io/github/stars/deepseek-ai/DeepSeek-OCR.svg?style=social&label=Star)](https://github.com/deepseek-ai/DeepSeek-OCR)
[DeepSeek-OCR: Contexts Optical Compression](https://arxiv.org/abs/2510.18234)
Haoran Wei, Yaofeng Sun, Yukun Li | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2510.18234)
[GitHub](https://github.com/deepseek-ai/DeepSeek-OCR)
[Model](https://huggingface.co/deepseek-ai/DeepSeek-OCR)
| 92 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.10-red)]() [![Star](https://img.shields.io/github/stars/JulietChoo/VisionSelector.svg?style=social&label=Star)](https://github.com/JulietChoo/VisionSelector)
[VisionSelector: End-to-End Learnable Visual Token Compression for Efficient Multimodal LLMs](https://arxiv.org/abs/2510.16598)
Jiaying Zhu, Yurui Zhu, Xin Lu, Wenrui Yan, Dong Li, Kunlin Liu, Xueyang Fu, Zheng-Jun Zha | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2510.16598)
[GitHub](https://github.com/JulietChoo/VisionSelector)
[Model](https://huggingface.co/JulietChoo/VisionSelector-Qwen2.5-VL-7B)
| 93 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.10-red)]() [![Star](https://img.shields.io/github/stars/Chenfei-Liao/VTC-Bench.svg?style=social&label=Star)](https://github.com/Chenfei-Liao/VTC-Bench)
[Are We Using the Right Benchmark: An Evaluation Framework for Visual Token Compression Methods](https://arxiv.org/abs/2510.07143)
Chenfei Liao, Wensong Wang, Zichen Wen, Xu Zheng, Yiyu Wang, Haocong He, Yuanhuiyi Lyu, Lutao Jiang, Xin Zou, Yuqian Fu, Bin Ren, Linfeng Zhang, Xuming Hu | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Benchmark-purple)]() | | [Paper](https://arxiv.org/abs/2510.07143)
[GitHub](https://github.com/Chenfei-Liao/VTC-Bench)
| 94 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.09-red)]()
[Training-Free Token Pruning via Zeroth-Order Gradient Estimation in Vision-Language Models](https://arxiv.org/abs/2509.24837)
Youngeun Kim, Youjia Zhang, Huiling Liu, Aecheon Jung, Sunwoo Lee, Sungeun Hong | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2509.24837)
| 95 | | [![Publish](https://img.shields.io/badge/NeurIPS-2025-blue)]() [![Star](https://img.shields.io/github/stars/AutoLab-SAI-SJTU/AutoPrune.svg?style=social&label=Star)](https://github.com/AutoLab-SAI-SJTU/AutoPrune)
[AutoPrune: Each Complexity Deserves a Pruning Policy](https://arxiv.org/abs/2509.23931)
Hanshi Wang, Yuhao Xu, Zekun Xu, Jin Gao, Yufan Liu, Weiming Hu, Ke Wang, Zhipeng Zhang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2509.23931)
[GitHub](https://github.com/AutoLab-SAI-SJTU/AutoPrune)
| 96 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.09-red)]()
[HIVTP: A Training-Free Method to Improve VLMs Efficiency via Hierarchical Visual Token Pruning Using Middle-Layer-Based Importance Score](https://arxiv.org/abs/2509.23663)
Jingqi Xu, Jingxi Lu, Chenghao Li, Sreetama Sarkar, Peter A. Beerel | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2509.23663)
| 97 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.09-red)]()
[Pyramid Token Pruning for High-Resolution Large Vision-Language Models via Region, Token, and Instruction-Guided Importance](https://arxiv.org/abs/2509.15704)
Yuxuan Liang, Xu Li, Xiaolei Chen, Yi Zheng, Haotian Chen, Bin Li, Xiangyang Xue | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Query--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2509.15704)
| 98 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.09-red)]()
[EfficientUICoder: Efficient MLLM-based UI Code Generation via Input and Output Token Compression](https://arxiv.org/abs/2509.12159)
Jingyu Xiao, Zhongyi Zhang, Yuxuan Wan, Yintong Huo, Yang Liu, Michael R.Lyu | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/GUI--Agent-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2509.12159)
| 99 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.09-red)]()
[Adaptive Token Merging for Efficient Transformer Semantic Communication at the Edge](https://arxiv.org/abs/2509.09955)
Omar Erak, Omar Alhussein, Hatem Abou-Zeid, Mehdi Bennis, Sami Muhaidat | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2509.09955)
| 100 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.08-red)]() [![Star](https://img.shields.io/github/stars/OpenGVLab/InternVL.svg?style=social&label=Star)](https://github.com/OpenGVLab/InternVL)
[InternVL3.5: Advancing Open-Source Multimodal Models in Versatility, Reasoning, and Efficiency](https://arxiv.org/abs/2508.18265)
InternVL Team | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2508.18265)
[GitHub](https://github.com/OpenGVLab/InternVL)
[Model](https://huggingface.co/collections/OpenGVLab/internvl35-68ac87bd52ebe953485927fb)
| 101 | | [![Publish](https://img.shields.io/badge/ACM_MM-2025-blue)]()
[VISA: Group-wise Visual Token Selection and Aggregation via Graph Summarization for Efficient MLLMs Inference](https://arxiv.org/abs/2508.17857)
Pengfei Jiang, Hanjun Li, Linglan Zhao, Fei Chao, Ke Yan, Shouhong Ding, Rongrong Ji | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2508.17857)
| 102 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.08-red)]()
[Revisiting MLLM Token Technology through the Lens of Classical Visual Coding](https://arxiv.org/abs/2508.13460)
Jinming Liu, Junyan Lin, Yuntao Wei, Kele Shao, Keda Tao, Jianguo Huang, Xudong Yang, Zhibo Chen, Huan Wang, Xin Jin | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Visual--Coding-purple)]() | | [Paper](https://arxiv.org/abs/2508.13460)
| 103 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.08-red)]()
[EVTP-IVS: Effective Visual Token Pruning For Unifying Instruction Visual Segmentation In Multi-Modal Large Language Models](https://arxiv.org/abs/2508.11886)
Wenhui Zhu, Xiwen Chen, Zhipeng Wang, Shao Tang, Sayan Ghosh, Xuanzhao Dong, Rajat Koner, Yalin Wang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2508.11886)
| 104 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.08-red)]()
[CATP: Contextually Adaptive Token Pruning for Efficient and Enhanced Multimodal In-Context Learning](https://arxiv.org/abs/2508.07871)
Yanshu Li, Jianjiang Yang, Zhennan Shen, Ligong Han, Haoyan Xu, Ruixiang Tang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2508.07871)
| 105 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.08-red)]()
[AdaptInfer: Adaptive Token Pruning for Vision-Language Model Inference with Dynamical Text Guidance](https://arxiv.org/abs/2508.06084)
Weichen Zhang, Zhui Zhu, Ningbo Li, Kebin Liu, Yunhao Liu | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2508.06084)
| 106 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.08-red)]()
[Fourier-VLM: Compressing Vision Tokens in the Frequency Domain for Large Vision-Language Models](https://arxiv.org/abs/2508.06038)
Huanyu Wang, Jushi Kai, Haoli Bai, Lu Hou, Bo Jiang, Ziwei He, Zhouhan Lin | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2508.06038)
| 107 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]() [![Star](https://img.shields.io/github/stars/sihany077/VFlowOpt.svg?style=social&label=Star)](https://github.com/sihany077/VFlowOpt)
[VFlowOpt: A Token Pruning Framework for LMMs with Visual Information Flow-Guided Optimization](https://arxiv.org/abs/2508.05211)
Sihan Yang, Runsen Xu, Chenhang Cui, Tai Wang, Dahua Lin, Jiangmiao Pang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2508.05211)
[GitHub](https://github.com/sihany077/VFlowOpt)
| 108 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.08-red)]() [![Star](https://img.shields.io/github/stars/HVision-NKU/GlimpsePrune.svg?style=social&label=Star)](https://github.com/HVision-NKU/GlimpsePrune)
[A Glimpse to Compress: Dynamic Visual Token Pruning for Large Vision-Language Models](https://arxiv.org/abs/2508.01548)
Quan-Sheng Zeng, Yunheng Li, Qilong Wang, Peng-Tao Jiang, Zuxuan Wu, Ming-Ming Cheng, Qibin Hou | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2508.01548)
[GitHub](https://github.com/HVision-NKU/GlimpsePrune)
[Model](https://huggingface.co/collections/ashun989/glimpseprune-688d8826ef5bd09db6af145e)
| 109 | | [![Publish](https://img.shields.io/badge/ACM_MM-2025-blue)]()
[Mitigating Information Loss under High Pruning Rates for Efficient Large Vision Language Models](https://arxiv.org/abs/2508.01236)
Mingyu Fu, Wei Suo, Ji Ma, Lin Yuanbo Wu, Peng Wang, Yanning Zhang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2508.01236)
| 110 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.08-red)]()
[HiPrune: Training-Free Visual Token Pruning via Hierarchical Attention in Vision-Language Models](https://arxiv.org/abs/2508.00553)
Jizhihui Liu, Feiyi Du, Guangdao Zhu, Niu Lian, Jun Li, Bin Chen | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2508.00553)
| 111 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.07-red)]()
[FastDriveVLA: Efficient End-to-End Driving via Plug-and-Play Reconstruction-based Token Pruning](https://arxiv.org/abs/2507.23318)
Jiajun Cao, Qizhe Zhang, Peidong Jia, Xuhui Zhao, Bo Lan, Xiaoan Zhang, Xiaobao Wei, Sixiang Chen, Zhuo Li, Yang Wang, Liyun Li, Xianming Liu, Ming Lu, Shanghang Zhang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/VLA-purple)]() | [![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2507.23318)
| 112 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]() [![Star](https://img.shields.io/github/stars/YuchenLiu98/METEOR.svg?style=social&label=Star)](https://github.com/YuchenLiu98/METEOR)
[METEOR: Multi-Encoder Collaborative Token Pruning for Efficient Vision Language Models](https://arxiv.org/abs/2507.20842)
Yuchen Liu, Yaoming Wang, Bowen Shi, Xiaopeng Zhang, Wenrui Dai, Chenglin Li, Hongkai Xiong, Qi Tian | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2507.20842)
[GitHub](https://github.com/YuchenLiu98/METEOR)
| 113 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.07-red)]() [![Star](https://img.shields.io/github/stars/liaolea/TransPrune.svg?style=social&label=Star)](https://github.com/liaolea/TransPrune)
[TransPrune: Token Transition Pruning for Efficient Large Vision-Language Model](https://arxiv.org/abs/2507.20630)
Ao Li, Yuxiang Duan, Jinghui Zhang, Congbo Ma, Yutong Xie, Gustavo Carneiro, Mohammad Yaqub, Hu Wang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2507.20630)
[GitHub](https://github.com/liaolea/TransPrune)
| 114 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.07-red)]() [![Star](https://img.shields.io/github/stars/cokeshao/Awesome-Multimodal-Token-Compression.svg?style=social&label=Star)](https://github.com/cokeshao/Awesome-Multimodal-Token-Compression)
[When Tokens Talk Too Much: A Survey of Multimodal Long-Context Token Compression across Images, Videos, and Audios](https://arxiv.org/abs/2507.20198)
Kele Shao, Keda Tao, Kejia Zhang, Sicheng Feng, Mu Cai, Yuzhang Shang, Haoxuan You, Can Qin, Yang Sui, Huan Wang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() [![Area](https://img.shields.io/badge/Survey-purple)]() | | [Paper](https://arxiv.org/abs/2507.20198)
[GitHub](https://github.com/cokeshao/Awesome-Multimodal-Token-Compression)
| 115 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.07-red)]()
[Efficient Whole Slide Pathology VQA via Token Compression](https://arxiv.org/abs/2507.14497)
Weimin Lyu, Qingqiao Hu, Kehan Qi, Zhan Shi, Wentao Huang, Saumya Gupta, Chao Chen | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2507.14497)
| 116 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.07-red)]()
[Training-free Token Reduction for Vision Mamba](https://arxiv.org/abs/2507.14042)
Qiankun Ma, Ziyao Zhang, Chi Su, Jie Chen, Zhen Song, Hairong Zheng, Wen Gao | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2507.14042)
| 117 | | [![Publish](https://img.shields.io/badge/NeurIPS-2025-blue)]() [![Star](https://img.shields.io/github/stars/dvlab-research/VisionThink.svg?style=social&label=Star)](https://github.com/dvlab-research/VisionThink)
[VisionThink: Smart and Efficient Vision Language Model via Reinforcement Learning](https://arxiv.org/abs/2507.13348)
Senqiao Yang, Junyi Li, Xin Lai, Bei Yu, Hengshuang Zhao, Jiaya Jia | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2507.13348)
[GitHub](https://github.com/dvlab-research/VisionThink)
[Model](https://huggingface.co/collections/Senqiao/visionthink-6878d839fae02a079c9c7bfe)
[Dataset](https://huggingface.co/collections/Senqiao/visionthink-6878d839fae02a079c9c7bfe)
| 118 | | [![Publish](https://img.shields.io/badge/EMNLP_Findings-2024-blue)]()
[LaCo: Efficient Layer-wise Compression of Visual Tokens for Multimodal Large Language Models](https://arxiv.org/abs/2507.02279)
Juntao Liu, Liqiang Niu, Wenchao Chen, Jie Zhou, Fandong Meng | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2507.02279)
| 119 | | [![Publish](https://img.shields.io/badge/IROS-2025-blue)]()
[ToSA: Token Merging with Spatial Awareness](https://arxiv.org/abs/2506.20066)
Hsiang-Wei Huang, Wenhao Chai, Kuang-Ming Chen, Cheng-Yen Yang, Jenq-Neng Hwang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2506.20066)
| 120 | | [![Publish](https://img.shields.io/badge/NeurIPS-2025-blue)]() [![Star](https://img.shields.io/github/stars/Theia-4869/CDPruner.svg?style=social&label=Star)](https://github.com/Theia-4869/CDPruner)
[Beyond Attention or Similarity: Maximizing Conditional Diversity for Token Pruning in MLLMs](https://arxiv.org/abs/2506.10967)
Qizhe Zhang, Mengzhen Liu, Lichen Li, Ming Lu, Yuan Zhang, Junwen Pan, Qi She, Shanghang Zhang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2506.10967)
[GitHub](https://github.com/Theia-4869/CDPruner)
| 121 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.06-red)]()
[Generic Token Compression in Multimodal Large Language Models from an Explainability Perspective](https://arxiv.org/abs/2506.01097)
Lei Lei, Jie Gu, Xiaokang Ma, Chu Tang, Jingmin Chen, Tong Xu | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2506.01097)
| 122 | | [![Publish](https://img.shields.io/badge/ACL-2025-blue)]() [![Star](https://img.shields.io/github/stars/EffiVLM-Bench/EffiVLM-Bench.svg?style=social&label=Star)](https://github.com/EffiVLM-Bench/EffiVLM-Bench)
[EffiVLM-Bench: A Comprehensive Benchmark for Evaluating Training-Free Acceleration in Large Visual-Languge Models](https://arxiv.org/abs/2506.00479)
Zekun Wang, Minghua Ma, Zexin Wang, Rongchuan Mu, Liping Shan, Ming Liu, Bing Qin | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Benchmark-purple)]() | | [Paper](https://arxiv.org/abs/2506.00479)
[GitHub](https://github.com/EffiVLM-Bench/EffiVLM-Bench)
| 123 |
124 | 125 |
126 | Video 127 | 128 | | **Title & Authors** | **Areas** | **Tags** | **Links** | 129 | | --- | --- | --- | :---: | 130 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.12-red)]()
[EchoingPixels: Cross-Modal Adaptive Token Reduction for Efficient Audio-Visual LLMs](https://arxiv.org/abs/2512.10324)
Chao Gong, Depeng Wang, Zhipeng Wei, Ya Guo, Huijia Zhu, Jingjing Chen | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2512.10324)
| 131 | | [![Publish](https://img.shields.io/badge/NeurIPS-2025-blue)]()
[Recurrent Attention-based Token Selection for Efficient Streaming Video-LLMs](https://arxiv.org/abs/2510.17364)
Vaggelis Dorovatas, Soroush Seifi, Gunshi Gupta, Rahaf Aljundi | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2510.17364)
| 132 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.10-red)]() [![Star](https://img.shields.io/github/stars/JulietChoo/VisionSelector.svg?style=social&label=Star)](https://github.com/JulietChoo/VisionSelector)
[VisionSelector: End-to-End Learnable Visual Token Compression for Efficient Multimodal LLMs](https://arxiv.org/abs/2510.16598)
Jiaying Zhu, Yurui Zhu, Xin Lu, Wenrui Yan, Dong Li, Kunlin Liu, Xueyang Fu, Zheng-Jun Zha | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2510.16598)
[GitHub](https://github.com/JulietChoo/VisionSelector)
[Model](https://huggingface.co/JulietChoo/VisionSelector-Qwen2.5-VL-7B)
| 133 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.09-red)]()
[PSTTS: A Plug-and-Play Token Selector for Efficient Event-based Spatio-temporal Representation Learning](https://arxiv.org/abs/2509.22481)
Xiangmo Zhao, Nan Yang, Yang Wang, Zhanwen Liu | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Event--Camera-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2509.22481)
| 134 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.09-red)]()
[Walk and Read Less: Improving the Efficiency of Vision-and-Language Navigation via Tuning-Free Multimodal Token Pruning](https://arxiv.org/abs/2509.15250)
Wenda Qin, Andrea Burns, Bryan A. Plummer, Margrit Betke | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/VLN-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2509.15250)
| 135 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.09-red)]()
[The Better You Learn, The Smarter You Prune: Towards Efficient Vision-language-action Models via Differentiable Token Pruning](https://arxiv.org/abs/2509.12594)
Titong Jiang, Xuefeng Jiang, Yuan Ma, Xin Wen, Bailin Li, Kun Zhan, Peng Jia, Yahui Liu, Sheng Sun, Xianpeng Lang | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/VLA-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2509.12594)
| 136 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.09-red)]() [![Star](https://img.shields.io/github/stars/Zizzzzzzz/FocusMamba.svg?style=social&label=Star)](https://github.com/Zizzzzzzz/FocusMamba)
[Focus Through Motion: RGB-Event Collaborative Token Sparsification for Efficient Object Detection](https://arxiv.org/abs/2509.03872)
Nan Yang, Yang Wang, Zhanwen Liu, Yuchao Dai, Yang Liu, Xiangmo Zhao | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Event--Camera-purple)]() | [![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2509.03872)
[GitHub](https://github.com/Zizzzzzzz/FocusMamba)
| 137 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.08-red)]() [![Star](https://img.shields.io/github/stars/OpenGVLab/InternVL.svg?style=social&label=Star)](https://github.com/OpenGVLab/InternVL)
[InternVL3.5: Advancing Open-Source Multimodal Models in Versatility, Reasoning, and Efficiency](https://arxiv.org/abs/2508.18265)
InternVL Team | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2508.18265)
[GitHub](https://github.com/OpenGVLab/InternVL)
[Model](https://huggingface.co/collections/OpenGVLab/internvl35-68ac87bd52ebe953485927fb)
| 138 | | [![Publish](https://img.shields.io/badge/ACM_MM-2025-blue)]()
[VISA: Group-wise Visual Token Selection and Aggregation via Graph Summarization for Efficient MLLMs Inference](https://arxiv.org/abs/2508.17857)
Pengfei Jiang, Hanjun Li, Linglan Zhao, Fei Chao, Ke Yan, Shouhong Ding, Rongrong Ji | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2508.17857)
| 139 | | [![Publish](https://img.shields.io/badge/EMNLP--blue)]() [![Star](https://img.shields.io/github/stars/yogesh-iitj/LGTTP.svg?style=social&label=Star)](https://github.com/yogesh-iitj/LGTTP)
[Language-Guided Temporal Token Pruning for Efficient VideoLLM Processing](https://arxiv.org/abs/2508.17686)
Yogesh Kumar | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2508.17686)
[GitHub](https://github.com/yogesh-iitj/LGTTP)
| 140 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.08-red)]() [![Star](https://img.shields.io/github/stars/zju-jiyicheng/SpecVLM.svg?style=social&label=Star)](https://github.com/zju-jiyicheng/SpecVLM)
[SpecVLM: Enhancing Speculative Decoding of Video LLMs via Verifier-Guided Token Pruning](https://arxiv.org/abs/2508.16201)
Yicheng Ji, Jun Zhang, Heming Xia, Jinpeng Chen, Lidan Shou, Gang Chen, Huan Li | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() | [Paper](https://arxiv.org/abs/2508.16201)
[GitHub](https://github.com/zju-jiyicheng/SpecVLM)
| 141 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.08-red)]()
[StreamMem: Query-Agnostic KV Cache Memory for Streaming Video Understanding](https://arxiv.org/abs/2508.15717)
Yanlai Yang, Zhuokai Zhao, Satya Narayan Shukla, Aashu Singh, Shlok Kumar Mishra, Lizhu Zhang, Mengye Ren | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Streaming--Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2508.15717)
| 142 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.08-red)]()
[EVTP-IVS: Effective Visual Token Pruning For Unifying Instruction Visual Segmentation In Multi-Modal Large Language Models](https://arxiv.org/abs/2508.11886)
Wenhui Zhu, Xiwen Chen, Zhipeng Wang, Shao Tang, Sayan Ghosh, Xuanzhao Dong, Rajat Koner, Yalin Wang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2508.11886)
| 143 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]() [![Star](https://img.shields.io/github/stars/sihany077/VFlowOpt.svg?style=social&label=Star)](https://github.com/sihany077/VFlowOpt)
[VFlowOpt: A Token Pruning Framework for LMMs with Visual Information Flow-Guided Optimization](https://arxiv.org/abs/2508.05211)
Sihan Yang, Runsen Xu, Chenhang Cui, Tai Wang, Dahua Lin, Jiangmiao Pang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2508.05211)
[GitHub](https://github.com/sihany077/VFlowOpt)
| 144 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.07-red)]() [![Star](https://img.shields.io/github/stars/cokeshao/Awesome-Multimodal-Token-Compression.svg?style=social&label=Star)](https://github.com/cokeshao/Awesome-Multimodal-Token-Compression)
[When Tokens Talk Too Much: A Survey of Multimodal Long-Context Token Compression across Images, Videos, and Audios](https://arxiv.org/abs/2507.20198)
Kele Shao, Keda Tao, Kejia Zhang, Sicheng Feng, Mu Cai, Yuzhang Shang, Haoxuan You, Can Qin, Yang Sui, Huan Wang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() [![Area](https://img.shields.io/badge/Survey-purple)]() | | [Paper](https://arxiv.org/abs/2507.20198)
[GitHub](https://github.com/cokeshao/Awesome-Multimodal-Token-Compression)
| 145 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.07-red)]()
[EgoPrune: Efficient Token Pruning for Egomotion Video Reasoning in Embodied Agent](https://arxiv.org/abs/2507.15428)
Jiaao Li, Kaiyuan Li, Chen Gao, Yong Li, Xinlei Chen | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2507.15428)
| 146 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]() [![Star](https://img.shields.io/github/stars/HYUNJS/STTM.svg?style=social&label=Star)](https://github.com/HYUNJS/STTM)
[Multi-Granular Spatio-Temporal Token Merging for Training-Free Acceleration of Video-LLMs](https://arxiv.org/abs/2507.07990)
Jeongseok Hyun, Sukjun Hwang, Su Ho Han, Taeoh Kim, Inwoong Lee, Dongyoon Wee, Joon-Young Lee, Seon Joo Kim, Minho Shim | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2507.07990)
[GitHub](https://github.com/HYUNJS/STTM)
| 147 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.07-red)]() [![Star](https://img.shields.io/github/stars/InternRobotics/StreamVLN.svg?style=social&label=Star)](https://github.com/InternRobotics/StreamVLN)
[StreamVLN: Streaming Vision-and-Language Navigation via SlowFast Context Modeling](https://arxiv.org/abs/2507.05240)
Meng Wei, Chenyang Wan, Xiqian Yu, Tai Wang, Yuqiang Yang, Xiaohan Mao, Chenming Zhu, Wenzhe Cai, Hanqing Wang, Yilun Chen, Xihui Liu, Jiangmiao Pang | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/VLN-purple)]() | [![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2507.05240)
[GitHub](https://github.com/InternRobotics/StreamVLN)
[Dataset](https://huggingface.co/datasets/cywan/StreamVLN-Trajectory-Data)
| 148 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]()
[AuroraLong: Bringing RNNs Back to Efficient Open-Ended Video Understanding](https://arxiv.org/abs/2507.02591)
Weili Xu, Enxin Song, Wenhao Chai, Xuexiang Wen, Tian Ye, Gaoang Wang | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2507.02591)
| 149 | | [![Publish](https://img.shields.io/badge/EMNLP_Findings-2024-blue)]()
[LaCo: Efficient Layer-wise Compression of Visual Tokens for Multimodal Large Language Models](https://arxiv.org/abs/2507.02279)
Juntao Liu, Liqiang Niu, Wenchao Chen, Jie Zhou, Fandong Meng | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2507.02279)
| 150 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.06-red)]() [![Star](https://img.shields.io/github/stars/HumanMLLM/LLaVA-Scissor.svg?style=social&label=Star)](https://github.com/HumanMLLM/LLaVA-Scissor)
[LLaVA-Scissor: Token Compression with Semantic Connected Components for Video-LLMs](https://arxiv.org/abs/2506.21862)
Boyuan Sun, Jiaxing Zhao, Xihan Wei, Qibin Hou | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2506.21862)
[GitHub](https://github.com/HumanMLLM/LLaVA-Scissor)
| 151 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.06-red)]() [![Star](https://img.shields.io/github/stars/VectorSpaceLab/Video-XL.svg?style=social&label=Star)](https://github.com/VectorSpaceLab/Video-XL)
[Video-XL-2: Towards Very Long-Video Understanding Through Task-Aware KV Sparsification](https://arxiv.org/abs/2506.19225)
Minghao Qin, Xiangrui Liu, Zhengyang Liang, Yan Shu, Huaying Yuan, Juenjie Zhou, Shitao Xiao, Bo Zhao, Zheng Liu | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]() [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2506.19225)
[GitHub](https://github.com/VectorSpaceLab/Video-XL)
[Model](https://huggingface.co/collections/BAAI/video-xl-683973cd45636acda09a11bd)
| 152 | | [![Publish](https://img.shields.io/badge/NeurIPS-2025-blue)]() [![Star](https://img.shields.io/github/stars/Theia-4869/CDPruner.svg?style=social&label=Star)](https://github.com/Theia-4869/CDPruner)
[Beyond Attention or Similarity: Maximizing Conditional Diversity for Token Pruning in MLLMs](https://arxiv.org/abs/2506.10967)
Qizhe Zhang, Mengzhen Liu, Lichen Li, Ming Lu, Yuan Zhang, Junwen Pan, Qi She, Shanghang Zhang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2506.10967)
[GitHub](https://github.com/Theia-4869/CDPruner)
| 153 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.06-red)]()
[DynTok: Dynamic Compression of Visual Tokens for Efficient and Effective Video Understanding](https://arxiv.org/abs/2506.03990)
Hongzhi Zhang, Jingyuan Zhang, Xingguang Ji, Qi Wang, Fuzheng Zhang | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2506.03990)
| 154 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.06-red)]() [![Star](https://img.shields.io/github/stars/mnyuew/METok.svg?style=social&label=Star)](https://github.com/mnyuew/METok)
[METok: Multi-Stage Event-based Token Compression for Efficient Long Video Understanding](https://arxiv.org/abs/2506.02850)
Mengyue Wang, Shuo Chen, Kristian Kersting, Volker Tresp, Yunpu Ma | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Query--Based-green)]() [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2506.02850)
[GitHub](https://github.com/mnyuew/METok)
| 155 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.06-red)]()
[Generic Token Compression in Multimodal Large Language Models from an Explainability Perspective](https://arxiv.org/abs/2506.01097)
Lei Lei, Jie Gu, Xiaokang Ma, Chu Tang, Jingmin Chen, Tong Xu | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2506.01097)
| 156 | | [![Publish](https://img.shields.io/badge/NeurIPS-2025-blue)]() [![Star](https://img.shields.io/github/stars/yunzhuzhang0918/flexselect.svg?style=social&label=Star)](https://github.com/yunzhuzhang0918/flexselect)
[FlexSelect: Flexible Token Selection for Efficient Long Video Understanding](https://arxiv.org/abs/2506.00993)
Yunzhu Zhang, Yu Lu, Tianyi Wang, Fengyun Rao, Yi Yang, Linchao Zhu | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2506.00993)
[GitHub](https://github.com/yunzhuzhang0918/flexselect)
| 157 | | [![Publish](https://img.shields.io/badge/ACL-2025-blue)]() [![Star](https://img.shields.io/github/stars/EffiVLM-Bench/EffiVLM-Bench.svg?style=social&label=Star)](https://github.com/EffiVLM-Bench/EffiVLM-Bench)
[EffiVLM-Bench: A Comprehensive Benchmark for Evaluating Training-Free Acceleration in Large Visual-Languge Models](https://arxiv.org/abs/2506.00479)
Zekun Wang, Minghua Ma, Zexin Wang, Rongchuan Mu, Liping Shan, Ming Liu, Bing Qin | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Benchmark-purple)]() | | [Paper](https://arxiv.org/abs/2506.00479)
[GitHub](https://github.com/EffiVLM-Bench/EffiVLM-Bench)
| 158 |
159 | 160 |
161 | Audio 162 | 163 | | **Title & Authors** | **Areas** | **Tags** | **Links** | 164 | | --- | --- | --- | :---: | 165 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.12-red)]()
[EchoingPixels: Cross-Modal Adaptive Token Reduction for Efficient Audio-Visual LLMs](https://arxiv.org/abs/2512.10324)
Chao Gong, Depeng Wang, Zhipeng Wei, Ya Guo, Huijia Zhu, Jingjing Chen | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2512.10324)
| 166 | | [![Arxiv](https://img.shields.io/badge/arXiv-2025\.07-red)]() [![Star](https://img.shields.io/github/stars/cokeshao/Awesome-Multimodal-Token-Compression.svg?style=social&label=Star)](https://github.com/cokeshao/Awesome-Multimodal-Token-Compression)
[When Tokens Talk Too Much: A Survey of Multimodal Long-Context Token Compression across Images, Videos, and Audios](https://arxiv.org/abs/2507.20198)
Kele Shao, Keda Tao, Kejia Zhang, Sicheng Feng, Mu Cai, Yuzhang Shang, Haoxuan You, Can Qin, Yang Sui, Huan Wang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() [![Area](https://img.shields.io/badge/Survey-purple)]() | | [Paper](https://arxiv.org/abs/2507.20198)
[GitHub](https://github.com/cokeshao/Awesome-Multimodal-Token-Compression)
| 167 |
168 | 169 | 170 | ### Published in Recent Conference/Journal 171 | 172 | 173 |
174 | NeurIPS 2025 175 | 176 | | **Title & Authors** | **Areas** | **Tags** | **Links** | 177 | | --- | --- | --- | :---: | 178 | | [![Publish](https://img.shields.io/badge/NeurIPS-2025-blue)]()
[Recurrent Attention-based Token Selection for Efficient Streaming Video-LLMs](https://arxiv.org/abs/2510.17364)
Vaggelis Dorovatas, Soroush Seifi, Gunshi Gupta, Rahaf Aljundi | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2510.17364)
| 179 | | [![Publish](https://img.shields.io/badge/NeurIPS-2025-blue)]() [![Star](https://img.shields.io/github/stars/AutoLab-SAI-SJTU/AutoPrune.svg?style=social&label=Star)](https://github.com/AutoLab-SAI-SJTU/AutoPrune)
[AutoPrune: Each Complexity Deserves a Pruning Policy](https://arxiv.org/abs/2509.23931)
Hanshi Wang, Yuhao Xu, Zekun Xu, Jin Gao, Yufan Liu, Weiming Hu, Ke Wang, Zhipeng Zhang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2509.23931)
[GitHub](https://github.com/AutoLab-SAI-SJTU/AutoPrune)
| 180 | | [![Publish](https://img.shields.io/badge/NeurIPS-2025-blue)]() [![Star](https://img.shields.io/github/stars/dvlab-research/VisionThink.svg?style=social&label=Star)](https://github.com/dvlab-research/VisionThink)
[VisionThink: Smart and Efficient Vision Language Model via Reinforcement Learning](https://arxiv.org/abs/2507.13348)
Senqiao Yang, Junyi Li, Xin Lai, Bei Yu, Hengshuang Zhao, Jiaya Jia | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2507.13348)
[GitHub](https://github.com/dvlab-research/VisionThink)
[Model](https://huggingface.co/collections/Senqiao/visionthink-6878d839fae02a079c9c7bfe)
[Dataset](https://huggingface.co/collections/Senqiao/visionthink-6878d839fae02a079c9c7bfe)
| 181 | | [![Publish](https://img.shields.io/badge/NeurIPS-2025-blue)]() [![Star](https://img.shields.io/github/stars/Theia-4869/CDPruner.svg?style=social&label=Star)](https://github.com/Theia-4869/CDPruner)
[Beyond Attention or Similarity: Maximizing Conditional Diversity for Token Pruning in MLLMs](https://arxiv.org/abs/2506.10967)
Qizhe Zhang, Mengzhen Liu, Lichen Li, Ming Lu, Yuan Zhang, Junwen Pan, Qi She, Shanghang Zhang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2506.10967)
[GitHub](https://github.com/Theia-4869/CDPruner)
| 182 | | [![Publish](https://img.shields.io/badge/NeurIPS-2025-blue)]() [![Star](https://img.shields.io/github/stars/yunzhuzhang0918/flexselect.svg?style=social&label=Star)](https://github.com/yunzhuzhang0918/flexselect)
[FlexSelect: Flexible Token Selection for Efficient Long Video Understanding](https://arxiv.org/abs/2506.00993)
Yunzhu Zhang, Yu Lu, Tianyi Wang, Fengyun Rao, Yi Yang, Linchao Zhu | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2506.00993)
[GitHub](https://github.com/yunzhuzhang0918/flexselect)
| 183 | | [![Publish](https://img.shields.io/badge/NeurIPS-2025-blue)]()
[Balanced Token Pruning: Accelerating Vision Language Models Beyond Local Optimization](https://arxiv.org/abs/2505.22038)
Kaiyuan Li, Xiaoyue Chen, Chen Gao, Yong Li, Xinlei Chen | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2505.22038)
| 184 | | [![Publish](https://img.shields.io/badge/NeurIPS-2025-blue)]() [![Star](https://img.shields.io/github/stars/cokeshao/HoliTom.svg?style=social&label=Star)](https://github.com/cokeshao/HoliTom)
[HoliTom: Holistic Token Merging for Fast Video Large Language Models](https://arxiv.org/abs/2505.21334)
Kele Shao, Keda Tao, Can Qin, Haoxuan You, Yang Sui, Huan Wang | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2505.21334)
[GitHub](https://github.com/cokeshao/HoliTom)
| 185 | | [![Publish](https://img.shields.io/badge/NeurIPS-2025-blue)]()
[Why 1 + 1 < 1 in Visual Token Pruning: Beyond Naive Integration via Multi-Objective Balanced Covering](https://arxiv.org/abs/2505.10118)
Yangfu Li, Hongjian Zhan, Tianyi Chen, Qi Liu, Yue Lu | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2505.10118)
| 186 | | [![Publish](https://img.shields.io/badge/NeurIPS-2025-blue)]() [![Star](https://img.shields.io/github/stars/LunarShen/FastVID.svg?style=social&label=Star)](https://github.com/LunarShen/FastVID)
[FastVID: Dynamic Density Pruning for Fast Video Large Language Models](https://arxiv.org/abs/2503.11187)
Leqi Shen, Guoqiang Gong, Tao He, Yifeng Zhang, Pengzhang Liu, Sicheng Zhao, Guiguang Ding | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2503.11187)
[GitHub](https://github.com/LunarShen/FastVID)
| 187 |
188 | 189 |
190 | ICCV 2025 191 | 192 | | **Title & Authors** | **Areas** | **Tags** | **Links** | 193 | | --- | --- | --- | :---: | 194 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]() [![Star](https://img.shields.io/github/stars/sihany077/VFlowOpt.svg?style=social&label=Star)](https://github.com/sihany077/VFlowOpt)
[VFlowOpt: A Token Pruning Framework for LMMs with Visual Information Flow-Guided Optimization](https://arxiv.org/abs/2508.05211)
Sihan Yang, Runsen Xu, Chenhang Cui, Tai Wang, Dahua Lin, Jiangmiao Pang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2508.05211)
[GitHub](https://github.com/sihany077/VFlowOpt)
| 195 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]() [![Star](https://img.shields.io/github/stars/mlvlab/Representation-Shift.svg?style=social&label=Star)](https://github.com/mlvlab/Representation-Shift)
[Representation Shift: Unifying Token Compression with FlashAttention](https://arxiv.org/abs/2508.00367)
Joonmyung Choi, Sanghyeok Lee, Byungoh Ko, Eunseo Kim, Jihyung Kil, Hyunwoo J. Kim | [![Area](https://img.shields.io/badge/Vision--Transformer-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2508.00367)
[GitHub](https://github.com/mlvlab/Representation-Shift)
| 196 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]() [![Star](https://img.shields.io/github/stars/YuchenLiu98/METEOR.svg?style=social&label=Star)](https://github.com/YuchenLiu98/METEOR)
[METEOR: Multi-Encoder Collaborative Token Pruning for Efficient Vision Language Models](https://arxiv.org/abs/2507.20842)
Yuchen Liu, Yaoming Wang, Bowen Shi, Xiaopeng Zhang, Wenrui Dai, Chenglin Li, Hongkai Xiong, Qi Tian | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2507.20842)
[GitHub](https://github.com/YuchenLiu98/METEOR)
| 197 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]() [![Star](https://img.shields.io/github/stars/HYUNJS/STTM.svg?style=social&label=Star)](https://github.com/HYUNJS/STTM)
[Multi-Granular Spatio-Temporal Token Merging for Training-Free Acceleration of Video-LLMs](https://arxiv.org/abs/2507.07990)
Jeongseok Hyun, Sukjun Hwang, Su Ho Han, Taeoh Kim, Inwoong Lee, Dongyoon Wee, Joon-Young Lee, Seon Joo Kim, Minho Shim | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2507.07990)
[GitHub](https://github.com/HYUNJS/STTM)
| 198 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]()
[AuroraLong: Bringing RNNs Back to Efficient Open-Ended Video Understanding](https://arxiv.org/abs/2507.02591)
Weili Xu, Enxin Song, Wenhao Chai, Xuexiang Wen, Tian Ye, Gaoang Wang | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2507.02591)
| 199 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]() [![Star](https://img.shields.io/github/stars/zwl666666/Skip-Vision.svg?style=social&label=Star)](https://github.com/zwl666666/Skip-Vision)
[Skip-Vision: Efficient and Scalable Acceleration of Vision-Language Models via Adaptive Token Skipping](https://arxiv.org/abs/2503.21817)
Weili Zeng, Ziyuan Huang, Kaixiang Ji, Yichao Yan | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2503.21817)
[GitHub](https://github.com/zwl666666/Skip-Vision)
| 200 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]()
[Growing a Twig to Accelerate Large Vision-Language Models](https://arxiv.org/abs/2503.14075)
Zhenwei Shao, Mingyang Wang, Zhou Yu, Wenwen Pan, Yan Yang, Tao Wei, Hongyuan Zhang, Ning Mao, Wei Chen, Jun Yu | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2503.14075)
| 201 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]() [![Star](https://img.shields.io/github/stars/dvlab-research/LSDBench.svg?style=social&label=Star)](https://github.com/dvlab-research/LSDBench)
[Does Your Vision-Language Model Get Lost in the Long Video Sampling Dilemma?](https://arxiv.org/abs/2503.12496)
Tianyuan Qu, Longxiang Tang, Bohao Peng, Senqiao Yang, Bei Yu, Jiaya Jia | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Benchmark-purple)]() | | [Paper](https://arxiv.org/abs/2503.12496)
[GitHub](https://github.com/dvlab-research/LSDBench)
[Dataset](https://huggingface.co/datasets/TainU/LSDBench)
| 202 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]() [![Star](https://img.shields.io/github/stars/anakin-skywalker-Joseph/Folder.svg?style=social&label=Star)](https://github.com/anakin-skywalker-Joseph/Folder)
[FOLDER: Accelerating Multi-modal Large Language Models with Enhanced Performance](https://arxiv.org/abs/2501.02430)
Haicheng Wang, Zhemeng Yu, Gabriele Spadaro, Chen Ju, Victor Quétu, Shuai Xiao, Enzo Tartaglione | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2501.02430)
[GitHub](https://github.com/anakin-skywalker-Joseph/Folder)
| 203 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]() [![Star](https://img.shields.io/github/stars/thu-nics/FrameFusion.svg?style=social&label=Star)](https://github.com/thu-nics/FrameFusion)
[FrameFusion: Combining Similarity and Importance for Video Token Reduction on Large Visual Language Models](https://arxiv.org/abs/2501.01986)
Tianyu Fu, Tengxuan Liu, Qinghao Han, Guohao Dai, Shengen Yan, Huazhong Yang, Xuefei Ning, Yu Wang | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]() | [Paper](https://arxiv.org/abs/2501.01986)
[GitHub](https://github.com/thu-nics/FrameFusion)
| 204 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]() [![Star](https://img.shields.io/github/stars/Hon-Wong/ByteVideoLLM.svg?style=social&label=Star)](https://github.com/Hon-Wong/ByteVideoLLM)
[Dynamic-VLM: Simple Dynamic Visual Token Compression for VideoLLM](https://arxiv.org/abs/2412.09530)
Han Wang, Yuxiang Nie, Yongjie Ye, Deng GuanYu, Yanjie Wang, Shuai Li, Haiyang Yu, Jinghui Lu, Can Huang | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]() [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2412.09530)
[GitHub](https://github.com/Hon-Wong/ByteVideoLLM)
| 205 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]() [![Star](https://img.shields.io/github/stars/dvlab-research/Lyra.svg?style=social&label=Star)](https://github.com/dvlab-research/Lyra)
[Lyra: An Efficient and Speech-Centric Framework for Omni-Cognition](https://arxiv.org/abs/2412.09501)
Zhisheng Zhong, Chengyao Wang, Yuqi Liu, Senqiao Yang, Longxiang Tang, Yuechen Zhang, Jingyao Li, Tianyuan Qu, Yanwei Li, Yukang Chen, Shaozuo Yu, Sitong Wu, Eric Lo, Shu Liu, Jiaya Jia | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2412.09501)
[GitHub](https://github.com/dvlab-research/Lyra)
[Model](https://huggingface.co/collections/zszhong/lyra-model-674ea5bb3b39ff8f15de75fc)
[Dataset](https://huggingface.co/collections/zszhong/lyra-data-675d80fbab80334eb52cdd82)
| 206 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]() [![Star](https://img.shields.io/github/stars/LaVi-Lab/AIM.svg?style=social&label=Star)](https://github.com/LaVi-Lab/AIM)
[AIM: Adaptive Inference of Multi-Modal LLMs via Token Merging and Pruning](https://arxiv.org/abs/2412.03248)
Yiwu Zhong, Zhuoming Liu, Yin Li, Liwei Wang | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2412.03248)
[GitHub](https://github.com/LaVi-Lab/AIM)
| 207 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]() [![Star](https://img.shields.io/github/stars/Theia-4869/VisPruner.svg?style=social&label=Star)](https://github.com/Theia-4869/VisPruner)
[Beyond Text-Visual Attention: Exploiting Visual Cues for Effective Token Pruning in VLMs](https://arxiv.org/abs/2412.01818)
Qizhe Zhang, Aosong Cheng, Ming Lu, Renrui Zhang, Zhiyong Zhuo, Jiajun Cao, Shaobo Guo, Qi She, Shanghang Zhang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2412.01818)
[GitHub](https://github.com/Theia-4869/VisPruner)
| 208 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]()
[ZipVL: Efficient Large Vision-Language Models with Dynamic Token Sparsification](https://arxiv.org/abs/2410.08584)
Yefei He, Feng Chen, Jing Liu, Wenqi Shao, Hong Zhou, Kaipeng Zhang, Bohan Zhuang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2410.08584)
| 209 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]() [![Star](https://img.shields.io/github/stars/joslefaure/HERMES.svg?style=social&label=Star)](https://github.com/joslefaure/HERMES)
[HERMES: temporal-coHERent long-forM understanding with Episodes and Semantics](https://arxiv.org/abs/2408.17443)
Gueter Josmy Faure, Jia-Fong Yeh, Min-Hung Chen, Hung-Ting Su, Shang-Hong Lai, Winston H. Hsu | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2408.17443)
[GitHub](https://github.com/joslefaure/HERMES)
| 210 | | [![Publish](https://img.shields.io/badge/ICCV-2025-blue)]() [![Star](https://img.shields.io/github/stars/42Shawn/LLaVA-PruMerge.svg?style=social&label=Star)](https://github.com/42Shawn/LLaVA-PruMerge)
[LLaVA-PruMerge: Adaptive Token Reduction for Efficient Large Multimodal Models](https://arxiv.org/abs/2403.15388)
Yuzhang Shang, Mu Cai, Bingxin Xu, Yong Jae Lee, Yan Yan | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Transformation--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2403.15388)
[GitHub](https://github.com/42Shawn/LLaVA-PruMerge)
| 211 |
212 | 213 |
214 | ACL 2025 215 | 216 | | **Title & Authors** | **Areas** | **Tags** | **Links** | 217 | | --- | --- | --- | :---: | 218 | | [![Publish](https://img.shields.io/badge/ACL-2025-blue)]() [![Star](https://img.shields.io/github/stars/EffiVLM-Bench/EffiVLM-Bench.svg?style=social&label=Star)](https://github.com/EffiVLM-Bench/EffiVLM-Bench)
[EffiVLM-Bench: A Comprehensive Benchmark for Evaluating Training-Free Acceleration in Large Visual-Languge Models](https://arxiv.org/abs/2506.00479)
Zekun Wang, Minghua Ma, Zexin Wang, Rongchuan Mu, Liping Shan, Ming Liu, Bing Qin | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() [![Area](https://img.shields.io/badge/Benchmark-purple)]() | | [Paper](https://arxiv.org/abs/2506.00479)
[GitHub](https://github.com/EffiVLM-Bench/EffiVLM-Bench)
| 219 | | [![Publish](https://img.shields.io/badge/ACL_Findings-2025-blue)]() [![Star](https://img.shields.io/github/stars/JeongHun0716/MMS-LLaMA.svg?style=social&label=Star)](https://github.com/JeongHun0716/MMS-LLaMA)
[MMS-LLaMA: Efficient LLM-based Audio-Visual Speech Recognition with Minimal Multimodal Speech Tokens](https://arxiv.org/abs/2503.11315)
Jeong Hun Yeo, Hyeongseop Rha, Se Jin Park, Yong Man Ro | [![Area](https://img.shields.io/badge/Audio--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2503.11315)
[GitHub](https://github.com/JeongHun0716/MMS-LLaMA)
| 220 | | [![Publish](https://img.shields.io/badge/NAACL-2025-blue)]() [![Star](https://img.shields.io/github/stars/AIoT-MLSys-Lab/MEDA.svg?style=social&label=Star)](https://github.com/AIoT-MLSys-Lab/MEDA)
[MEDA: Dynamic KV Cache Allocation for Efficient Multimodal Long-Context Inference](https://arxiv.org/abs/2502.17599)
Zhongwei Wan, Hui Shen, Xin Wang, Che Liu, Zheda Mai, Mi Zhang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2502.17599)
[GitHub](https://github.com/AIoT-MLSys-Lab/MEDA)
| 221 | | [![Publish](https://img.shields.io/badge/ACL-2025-blue)]() [![Star](https://img.shields.io/github/stars/Visual-AI/PruneVid.svg?style=social&label=Star)](https://github.com/Visual-AI/PruneVid)
[PruneVid: Visual Token Pruning for Efficient Video Large Language Models](https://arxiv.org/abs/2412.16117)
Xiaohu Huang, Hao Zhou, Kai Han | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2412.16117)
[GitHub](https://github.com/Visual-AI/PruneVid)
| 222 | | [![Publish](https://img.shields.io/badge/NAACL_Oral-2025-blue)]() [![Star](https://img.shields.io/github/stars/ZongqianLi/Prompt-Compression-Survey.svg?style=social&label=Star)](https://github.com/ZongqianLi/Prompt-Compression-Survey)
[Prompt Compression for Large Language Models: A Survey](https://arxiv.org/abs/2410.12388)
Zongqian Li, Yinhong Liu, Yixuan Su, Nigel Collier | [![Area](https://img.shields.io/badge/LLM-purple)]() [![Area](https://img.shields.io/badge/Survey-purple)]() | | [Paper](https://arxiv.org/abs/2410.12388)
[GitHub](https://github.com/ZongqianLi/Prompt-Compression-Survey)
| 223 |
224 | 225 |
226 | ICML 2025 227 | 228 | | **Title & Authors** | **Areas** | **Tags** | **Links** | 229 | | --- | --- | --- | :---: | 230 | | [![Publish](https://img.shields.io/badge/ICML-2025-blue)]() [![Star](https://img.shields.io/github/stars/wangqinsi1/2025-ICML-CoreMatching.svg?style=social&label=Star)](https://github.com/wangqinsi1/2025-ICML-CoreMatching)
[CoreMatching: A Co-adaptive Sparse Inference Framework with Token and Neuron Pruning for Comprehensive Acceleration of Vision-Language Models](https://arxiv.org/abs/2505.19235)
Qinsi Wang, Hancheng Ye, Ming-Yu Chung, Yudong Liu, Yueqian Lin, Martin Kuo, Mingyuan Ma, Jianyi Zhang, Yiran Chen | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2505.19235)
[GitHub](https://github.com/wangqinsi1/2025-ICML-CoreMatching)
| 231 | | [![Publish](https://img.shields.io/badge/ICML-2025-blue)]() [![Star](https://img.shields.io/github/stars/yangdongchao/ALMTokenizer.svg?style=social&label=Star)](https://github.com/yangdongchao/ALMTokenizer)
[ALMTokenizer: A Low-bitrate and Semantic-rich Audio Codec Tokenizer for Audio Language Modeling](https://arxiv.org/abs/2504.10344)
Dongchao Yang, Songxiang Liu, Haohan Guo, Jiankun Zhao, Yuanyuan Wang, Helin Wang, Zeqian Ju, Xubo Liu, Xueyuan Chen, Xu Tan, Xixin Wu, Helen Meng | [![Area](https://img.shields.io/badge/Audio--Transformer-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2504.10344)
[GitHub](https://github.com/yangdongchao/ALMTokenizer)
| 232 | | [![Publish](https://img.shields.io/badge/ICML-2025-blue)]() [![Star](https://img.shields.io/github/stars/steven-ccq/ViLAMP.svg?style=social&label=Star)](https://github.com/steven-ccq/ViLAMP)
[Scaling Video-Language Models to 10K Frames via Hierarchical Differential Distillation](https://arxiv.org/abs/2504.02438)
Chuanqi Cheng, Jian Guan, Wei Wu, Rui Yan | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2504.02438)
[GitHub](https://github.com/steven-ccq/ViLAMP)
[Model](https://huggingface.co/orange-sk/ViLAMP-llava-qwen)
| 233 | | [![Publish](https://img.shields.io/badge/ICML-2025-blue)]() [![Star](https://img.shields.io/github/stars/Vision-CAIR/LongVU.svg?style=social&label=Star)](https://github.com/Vision-CAIR/LongVU)
[LongVU: Spatiotemporal Adaptive Compression for Long Video-Language Understanding](https://arxiv.org/abs/2410.17434)
Xiaoqian Shen, Yunyang Xiong, Changsheng Zhao, Lemeng Wu, Jun Chen, Chenchen Zhu, Zechun Liu, Fanyi Xiao, Balakrishnan Varadarajan, Florian Bordes, Zhuang Liu, Hu Xu, Hyunwoo J. Kim, Bilge Soran, Raghuraman Krishnamoorthi, Mohamed Elhoseiny, Vikas Chandra | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Query--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2410.17434)
[GitHub](https://github.com/Vision-CAIR/LongVU)
[Model](https://huggingface.co/collections/Vision-CAIR/longvu-67181d2debabfc1eb050c21d)
| 234 | | [![Publish](https://img.shields.io/badge/ICML-2025-blue)]() [![Star](https://img.shields.io/github/stars/Gumpest/SparseVLMs.svg?style=social&label=Star)](https://github.com/Gumpest/SparseVLMs)
[SparseVLM: Visual Token Sparsification for Efficient Vision-Language Model Inference](https://arxiv.org/abs/2410.04417)
Yuan Zhang, Chun-Kai Fan, Junpeng Ma, Wenzhao Zheng, Tao Huang, Kuan Cheng, Denis Gudovskiy, Tomoyuki Okuno, Yohei Nakata, Kurt Keutzer, Shanghang Zhang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Query--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2410.04417)
[GitHub](https://github.com/Gumpest/SparseVLMs)
| 235 |
236 | 237 |
238 | ACM MM 2025 239 | 240 | | **Title & Authors** | **Areas** | **Tags** | **Links** | 241 | | --- | --- | --- | :---: | 242 | | [![Publish](https://img.shields.io/badge/ACM_MM-2025-blue)]()
[VISA: Group-wise Visual Token Selection and Aggregation via Graph Summarization for Efficient MLLMs Inference](https://arxiv.org/abs/2508.17857)
Pengfei Jiang, Hanjun Li, Linglan Zhao, Fei Chao, Ke Yan, Shouhong Ding, Rongrong Ji | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Attention--Based-green)]() [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Free-yellow)]() | [Paper](https://arxiv.org/abs/2508.17857)
| 243 | | [![Publish](https://img.shields.io/badge/ACM_MM-2025-blue)]()
[Mitigating Information Loss under High Pruning Rates for Efficient Large Vision Language Models](https://arxiv.org/abs/2508.01236)
Mingyu Fu, Wei Suo, Ji Ma, Lin Yuanbo Wu, Peng Wang, Yanning Zhang | [![Area](https://img.shields.io/badge/Image--LLM-purple)]() | [![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2508.01236)
| 244 | | [![Publish](https://img.shields.io/badge/ACM_MM-2025-blue)]() [![Star](https://img.shields.io/github/stars/yaolinli/TimeChat-Online.svg?style=social&label=Star)](https://github.com/yaolinli/TimeChat-Online)
[TimeChat-Online: 80% Visual Tokens are Naturally Redundant in Streaming Videos](https://arxiv.org/abs/2504.17343)
Linli Yao, Yicheng Li, Yuancheng Wei, Lei Li, Shuhuai Ren, Yuanxin Liu, Kun Ouyang, Lean Wang, Shicheng Li, Sida Li, Lingpeng Kong, Qi Liu, Yuanxing Zhang, Xu Sun | [![Area](https://img.shields.io/badge/Video--LLM-purple)]() | [![Type](https://img.shields.io/badge/Similarity--Based-green)]()
[![Cost](https://img.shields.io/badge/Training--Based-yellow)]() | [Paper](https://arxiv.org/abs/2504.17343)
[GitHub](https://github.com/yaolinli/TimeChat-Online)
[Model](https://huggingface.co/wyccccc/TimeChatOnline-7B)
[Dataset](https://huggingface.co/datasets/yaolily/TimeChat-Online-139K)
| 245 |
246 | 247 | 248 | --- 249 | 250 | ## 📄 License 251 | 252 | This project is licensed under the MIT License - see the [LICENSE](LICENSE) file for details. 253 | 254 | --- 255 | 256 | ## 🙏 Acknowledgments 257 | 258 | This repository is inspired by [Awesome-Efficient-Reasoning-Models](https://github.com/fscdc/Awesome-Efficient-Reasoning-Models), [Awesome-Efficient-LLM](https://github.com/horseee/Awesome-Efficient-LLM/), [Awesome-Context-Engineering](https://github.com/Meirtz/Awesome-Context-Engineering) 259 | 260 | ## 🧑‍💻 Contributors 261 | 262 | 👏 Thanks to these contributors for this excellent work! 263 | 264 | 265 | 266 | 267 | 268 | ## ✉️ Contact 269 | 270 | For questions, suggestions, or collaboration opportunities, please feel free to reach out: 271 | 272 | ✉️ Email: [shaokele@gmail.com](mailto:shaokele@gmail.com) / [KD.TAO.CT@outlook.com](mailto:KD.TAO.CT@outlook.com) 273 | 274 | ## ✨ Star History 275 | 276 | [![Star History Chart](https://api.star-history.com/svg?repos=cokeshao/Awesome-Multimodal-Token-Compression&type=date&legend=top-left)](https://www.star-history.com/#cokeshao/Awesome-Multimodal-Token-Compression&type=date&legend=top-left) 277 | 278 | [**⬆ Back to top**](#awesome-multimodal-token-compression) 279 | --------------------------------------------------------------------------------