Yang Liu1*, Yufei Yin2*, Chenchen Jing3, Muzhi Zhu1, Hao Chen1, Yuling Xi1, Bo Feng4, Hao Wang4, Shiyu Li4, Chunhua Shen1
1Zhejiang University, 2Hangzhou Dianzi University, 3Zhejiang University of Technology, 4Apple
In this work, we present COSINE, a unified open-world segmentation model that consolidates open-vocabulary segmentation and in-context segmentation with multi-modal prompts (e.g. text and image). COSINE exploits foundation models to extract representations for an input image and corresponding multi-modal prompts, and a SegDecoder to align these representations, model their interaction, and obtain masks specified by input prompts across different granularities. In this way, COSINE overcomes architectural discrepancies, divergent learning objectives, and distinct representation learning strategies of previous pipelines for open-vocabulary segmentation and in-context segmentation. Comprehensive experiments demonstrate that COSINE has significant performance improvements in both open-vocabulary and in-context segmentation tasks. Our exploratory analyses highlight that the synergistic collaboration between using visual and textual prompts leads to significantly improved generalization over single-modality approaches.
For academic use, this project is licensed under the 2-clause BSD License. For commercial use, please contact Chunhua Shen.
