Stars
Qwen2.5 is the large language model series developed by Qwen team, Alibaba Cloud.
AllentDan / lmdeploy
Forked from InternLM/lmdeployLMDeploy is a toolkit for compressing, deploying, and serving LLM
Generative Models by Stability AI
Lumina-T2X is a unified framework for Text to Any Modality Generation
[CVPR 2024 Oral] InternVL Family: A Pioneering Open-Source Alternative to GPT-4o. 接近GPT-4o表现的开源多模态对话模型
Collection of Open Source Projects Related to GPT,GPT相关开源项目合集🚀、精选🔥🔥
Facebook AI Research Sequence-to-Sequence Toolkit written in Python.
👑 Easy-to-use and powerful NLP and LLM library with 🤗 Awesome model zoo, supporting wide-range of NLP tasks from research to industrial applications, including 🗂Text Classification, 🔍 Neural Search…
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
AutoGPT is the vision of accessible AI for everyone, to use and to build on. Our mission is to provide the tools, so that you can focus on what matters.
[ECCV2022, TPAMI2023] FAST-VQA, and its extended version FasterVQA.
中文羊驼大模型三期项目 (Chinese Llama-3 LLMs) developed from Meta Llama 3
Unsupervised Word Segmentation for Neural Machine Translation and Text Generation
欢迎来到 "LLM-travel" 仓库!探索大语言模型(LLM)的奥秘 🚀。致力于深入理解、探讨以及实现与大模型相关的各种技术、原理和应用。
中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
The official repo of Qwen-VL (通义千问-VL) chat & pretrained large vision language model proposed by Alibaba Cloud.
PyTorch code for BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation
Code for ALBEF: a new vision-language pre-training method
Code for the ICML 2021 (long talk) paper: "ViLT: Vision-and-Language Transformer Without Convolution or Region Supervision"
A PyTorch implementation of the Transformer model in "Attention is All You Need".
animatediff prompt travel
Chinese version of CLIP which achieves Chinese cross-modal retrieval and representation generation.
CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image
Approximate Nearest Neighbors in C++/Python optimized for memory usage and loading/saving to disk
mPLUG-Owl: The Powerful Multi-modal Large Language Model Family