title | emoji | colorFrom | colorTo | sdk | python_version | sdk_version | app_file | pinned | license |
---|---|---|---|---|---|---|---|---|---|
MobileSAM |
🐠 |
indigo |
yellow |
gradio |
3.8.10 |
3.35.2 |
app.py |
false |
apache-2.0 |
Demo of official PyTorch implementation of MobileSAM.
MobileSAM performs on par with the original SAM (at least visually) and keeps exactly the same pipeline as the original SAM except for a change on the image encoder. Specifically, we replace the original heavyweight ViT-H encoder (632M) with a much smaller Tiny-ViT (5M). On a single GPU, MobileSAM runs around 12ms per image: 8ms on the image encoder and 4ms on the mask decoder.
First, mobile_sam must be installed to run on pc. Refer to Installation Instruction
Then run the following
python app.py
The model is licensed under the Apache 2.0 license.
- Segment Anything provides the SA-1B dataset and the base codes.
- TinyViT provides codes and pre-trained models.
If you find this project useful for your research, please consider citing the following BibTeX entry.
@article{mobile_sam,
title={Faster Segment Anything: Towards Lightweight SAM for Mobile Applications},
author={Zhang, Chaoning and Han, Dongshen and Qiao, Yu and Kim, Jung Uk and Bae, Sung Ho and Lee, Seungkyu and Hong, Choong Seon},
journal={arXiv preprint arXiv:2306.14289},
year={2023}
}