Stars
[ACL'24 Outstanding] Data and code for L-Eval, a comprehensive long context language models evaluation benchmark
A series of large language models trained from scratch by developers @01-ai
OpenChat: Advancing Open-source Language Models with Imperfect Data
ChatGLM3 series: Open Bilingual Chat LLMs | 开源双语对话语言模型
LangChain 的中文入门教程
Efficiently Fine-Tune 100+ LLMs in WebUI (ACL 2024)
Firefly: 大模型训练工具,支持训练Qwen2.5、Qwen2、Yi1.5、Phi-3、Llama3、Gemma、MiniCPM、Yi、Deepseek、Orion、Xverse、Mixtral-8x7B、Zephyr、Mistral、Baichuan2、Llma2、Llama、Qwen、Baichuan、ChatGLM2、InternLM、Ziya2、Vicuna、Bloom等大模型
Example models using DeepSpeed
A 13B large language model developed by Baichuan Intelligent Technology
ChatGLM2-6B: An Open Bilingual Chat LLM | 开源双语对话语言模型
The ParroT framework to enhance and regulate the Translation Abilities during Chat based on open-sourced LLMs (e.g., LLaMA-7b, Bloomz-7b1-mt) and human written translation and evaluation data.
A preliminary evaluation of ChatGPT/GPT-4 for machine translation.
本项目旨在对大量文本文件进行快速编码检测和转换以辅助mnbvc语料集项目的数据清洗工作
MNBVC(Massive Never-ending BT Vast Chinese corpus)超大规模中文语料集。对标chatGPT训练的40T数据。MNBVC数据集不但包括主流文化,也包括各个小众文化甚至火星文的数据。MNBVC数据集包括新闻、作文、小说、书籍、杂志、论文、台词、帖子、wiki、古诗、歌词、商品介绍、笑话、糗事、聊天记录等一切形式的纯文本中文数据。
中文nlp解决方案(大模型、数据、模型、训练、推理)
ChatGLM-6B: An Open Bilingual Dialogue Language Model | 开源双语对话语言模型
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
A benchmark for the task of translation suggestion
Optimized primitives for collective multi-GPU communication