⚗️ Llava 13b model repository trained by liuhaotian managed by DVC
-
Updated
May 15, 2024 - Python
⚗️ Llava 13b model repository trained by liuhaotian managed by DVC
Localized Multimodal Large Language Model (MLLM) integrated with Streamlit and Ollama for text and image processing tasks.
Image analysis with Ollama (AI models) from within Home Assistant
text to text / image to text Architecture prompt generator for stable diffusion / any image generation plateform based on Ollama LLM
⚗️ Zephyr 7b model repository trained by HuggingFaceH4 managed by DVC
Leveraging state-of-the-art (SOTA) language models (LLMs) and orchestration frameworks like LangChain, Vox optimizes multimodal interactions, enhancing communication tasks with personalized assistance and enriched user experiences.
Generate text from images and translation text English to Arabic
Build your own ChatGPT locally, with the ability of interacting with images, pdfs, audios
Pheye - a family of efficient small vision-language models
Computer Vision Research for Multimedia Understanding at DSO National Laboratories Internship 2023 under the DSTA JC Scholarship
Python-based WebSocket for CLI LLaVA inference.
Upload images to Slack with automatic alt text generation using Llava on Ollama
Joint work as part of a bachelor's thesis on utilizing a combination of NLP and CV methods in implementing multimodal approaches to combat hate speech in memes.
Vision Language Dataset Construction Library for Remote Sensing Domain
Explore the rich flavors of Indian desserts with TunedLlavaDelights. Utilizing the in Llava fine-tuning, our project unveils detailed nutritional profiles, taste notes, and optimal consumption times for beloved sweets. Dive into a fusion of AI innovation and culinary tradition
Add a description, image, and links to the llava topic page so that developers can more easily learn about it.
To associate your repository with the llava topic, visit your repo's landing page and select "manage topics."