Stars
A customized and efficient database tuning system [VLDB'22]
BELLE: Be Everyone's Large Language model Engine(开源中文对话大模型)
A curated list of Continual Learning papers and BibTeX entries
A high-throughput and memory-efficient inference and serving engine for LLMs
Code and documentation to train Stanford's Alpaca models, and generate the data.
The official repo of Qwen (通义千问) chat & pretrained large language model proposed by Alibaba Cloud.
We unified the interfaces of instruction-tuning data (e.g., CoT data), multiple LLMs and parameter-efficient methods (e.g., lora, p-tuning) together for easy use. We welcome open-source enthusiasts…
An LLM Based Diagnosis System (https://arxiv.org/pdf/2312.01454.pdf)
Deita: Data-Efficient Instruction Tuning for Alignment [ICLR2024]
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
Summarize existing representative LLMs text datasets.
A curated list of LLM Interpretability related material - Tutorial, Library, Survey, Paper, Blog, etc..
ChatGPT 中文调教指南。各种场景使用指南。学习怎么让它听你的话。
A collection of awesome-prompt-datasets, awesome-instruction-dataset, to train ChatLLM such as chatgpt 收录各种各样的指令数据集, 用于训练 ChatLLM 模型。
Human preference data for "Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback"
[ACL 2024] An Easy-to-use Knowledge Editing Framework for LLMs.
整理开源的中文大语言模型,以规模较小、可私有化部署、训练成本较低的模型为主,包括底座模型,垂直领域微调及应用,数据集与教程等。
A LaTeX resume template designed for optimal information density and aesthetic appeal.
prompt2model - Generate Deployable Models from Natural Language Instructions
Medusa: Simple Framework for Accelerating LLM Generation with Multiple Decoding Heads
Sparse probing paper full code.
A PyTorch Library for Multi-Task Learning