Skip to content
/ VAST Public
forked from TXH-mercury/VAST

Code and Model for VAST: A Vision-Audio-Subtitle-Text Omni-Modality Foundation Model and Dataset

License

Notifications You must be signed in to change notification settings

lihanddd/VAST

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

13 Commits
 
 
 
 
 
 

Repository files navigation

VAST: A Vision-Audio-Subtitle-Text Omni-Modality Foundation Model and Dataset

This is the official repository of VAST which will provide code, model checkpoint and dataset. They will be released after paper is accepted.

PWC PWC PWC PWC PWC PWC PWC PWC PWC PWC PWC PWC PWC PWC PWC PWC PWC PWC

Citation

If you find this code useful for your research, please consider citing:

@article{chen2023vast,
  title={VAST: A Vision-Audio-Subtitle-Text Omni-Modality Foundation Model and Dataset},
  author={Chen, Sihan and Li, Handong and Wang, Qunbo and Zhao, Zijia and Sun, Mingzhen and Zhu, Xinxin and Liu, Jing},
  journal={arXiv preprint arXiv:2305.18500},
  year={2023}
}

About

Code and Model for VAST: A Vision-Audio-Subtitle-Text Omni-Modality Foundation Model and Dataset

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published