Video-XL: Extra-Long Vision Language Model for Hour-Scale Video Understanding
[Paper]
[Code]
Yan Shu, Zheng Liu,
Peitian Zhang, Minghao Qin, Junjie Zhou, Zhengyang Liang, Tiejun Huang, Bo Zhao
CVPR, 2025
Enabling high-quality and efficient video understanding over thousands of frames on a single A100 GPU.
MemoRAG: Moving towards Next-Gen RAG via Memory-Inspired Knowledge Discovery
[Paper]
[Code]
Hongjin Qian,
Peitian Zhang, Zheng Liu, Kelong Mao, Zhicheng Dou
WWW, 2025
Enhancing retrieval-augmented generation (RAG) by acquiring global understanding of the entire database from a memory module.
Long Context Compression with Activation Beacon
[Paper]
[Code]
Peitian Zhang, Zheng Liu, Shitao Xiao, Ninglu Shao, Qiwei Ye, Zhicheng Dou
ICLR, 2025
A plug-in module for transformer-based LLMs that enables effective, efficient, and flexible compression of long contexts.
Retrieve Anything to Augment Large Language Models
[Paper]
[Code]
Peitian Zhang, Shitao Xiao, Zheng Liu, Zhicheng Dou, Jian-Yun Nie
ACL, 2024
A unified embedding model that supports diverse retrieval augmentation scenarios.
C-Pack: Packed Resources for General Chinese Embeddings
[Paper]
[Code]
Shitao Xiao, Zheng Liu,
Peitian Zhang, Niklas Muennighoff, Defu Lian, Jian-Yun Nie
SIGIR, 2024
A package of resources that significantly advance the field of general Chinese embeddings.