If you are not using Linux, do NOT proceed, see instructions for macOS and Windows.
- Clone this repository and navigate to LLaVA folder
git clone https://code.byted.org/ic-research/llava-next-video.git
cd llava-next-video
- Install Package
conda create -n llava python=3.10 -y
conda activate llava
pip install --upgrade pip # enable PEP 660 support
pip install -e .
-
Example model: liuhaotian/llava-v1.6-vicuna-7b
-
Prompt mode: vicuna_v1
-
Sampled frames: 32
-
Spatial pooling stride: 2
bash scripts/eval/video_description_from_t2v.sh ${Example model} ${Prompt mode} ${Sampled frames} True ${Spatial pooling stride} 8 True ;
# bash scripts/eval/video_description_from_t2v.sh liuhaotian/llava-v1.6-vicuna-7b vicuna_v1 32 True 2 8 True ;