diff --git a/README.md b/README.md index ace8ac49..dad18f8a 100644 --- a/README.md +++ b/README.md @@ -99,7 +99,7 @@ A speech-to-speech dialogue model with both low-latency and high intelligence wh | [**Apollo: An Exploration of Video Understanding in Large Multimodal Models**](https://arxiv.org/pdf/2412.10360) | arXiv | 2024-12-13 | - | - | | ![Star](https://img.shields.io/github/stars/InternLM/InternLM-XComposer.svg?style=social&label=Star)
[**InternLM-XComposer2.5-OmniLive: A Comprehensive Multimodal System for Long-term Streaming Video and Audio Interactions**](https://arxiv.org/pdf/2412.09596)
| arXiv | 2024-12-12 | [Github](https://github.com/InternLM/InternLM-XComposer/tree/main/InternLM-XComposer-2.5-OmniLive) | Local Demo | | [**StreamChat: Chatting with Streaming Video**](https://arxiv.org/pdf/2412.08646) | arXiv | 2024-12-11 | Coming soon | - | -| [**CompCap: Improving Multimodal Large Language Models with Composite Captions**](https://arxiv.org/pdf/2412.05243) | arXiv | 2024-12-06 | - | - | +| [**CompCap: Improving Multimodal Large Language Models with Composite Captions**](https://arxiv.org/pdf/2412.05243) | arXiv | 2024-12-06 | [Github](https://github.com/Xiaohui9607/CompCap) | - | | ![Star](https://img.shields.io/github/stars/gls0425/LinVT.svg?style=social&label=Star)
[**LinVT: Empower Your Image-level Large Language Model to Understand Videos**](https://arxiv.org/pdf/2412.05185)
| arXiv | 2024-12-06 | [Github](https://github.com/gls0425/LinVT) | - | | ![Star](https://img.shields.io/github/stars/OpenGVLab/InternVL.svg?style=social&label=Star)
[**Expanding Performance Boundaries of Open-Source Multimodal Models with Model, Data, and Test-Time Scaling**](https://arxiv.org/pdf/2412.05271)
| arXiv | 2024-12-06 | [Github](https://github.com/OpenGVLab/InternVL) | [Demo](https://internvl.opengvlab.com) | | ![Star](https://img.shields.io/github/stars/NVlabs/VILA.svg?style=social&label=Star)
[**NVILA: Efficient Frontier Visual Language Models**](https://arxiv.org/pdf/2412.04468)
| arXiv | 2024-12-05 | [Github](https://github.com/NVlabs/VILA) | [Demo](https://vila.mit.edu) |