Block or Report
Block or report BaileyWei
Contact GitHub support about this user’s behavior. Learn more about reporting abuse.
Report abuseStars
Language
Sort by: Recently starred
Machine Learning and Computer Vision Engineer - Technical Interview Questions
Reference implementation for DPO (Direct Preference Optimization)
🧑🏫 60 Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gan…
【LLMs九层妖塔】分享 LLMs在自然语言处理(ChatGLM、Chinese-LLaMA-Alpaca、小羊驼 Vicuna、LLaMA、GPT4ALL等)、信息检索(langchain)、语言合成、语言识别、多模态等领域(Stable Diffusion、MiniGPT-4、VisualGLM-6B、Ziya-Visual等)等 实战与经验。
The paper list of the 86-page paper "The Rise and Potential of Large Language Model Based Agents: A Survey" by Zhiheng Xi et al.
A WebUI for Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
[EMNLP 2022] Training Language Models with Memory Augmentation https://arxiv.org/abs/2205.12674
Chinese-LLaMA 1&2、Chinese-Falcon 基础模型;ChatFlow中文对话模型;中文OpenLLaMA模型;NLP预训练/指令微调数据集
Code for the paper "Closing the Curious Case of Neural Text Degeneration"
A collection of AWESOME things about mixture-of-experts
Reading list of hallucination in LLMs. Check out our new survey paper: "Siren’s Song in the AI Ocean: A Survey on Hallucination in Large Language Models"
The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.
The official GitHub page for the survey paper "A Survey on Evaluation of Large Language Models".
Make huge neural nets fit in memory
MedicalGPT: Training Your Own Medical GPT Model with ChatGPT Training Pipeline. 训练医疗大模型,实现了包括增量预训练(PT)、有监督微调(SFT)、RLHF、DPO、ORPO。
第一个支持中英文双语语音-文本多模态对话的开源可商用对话模型。便捷的语音输入将大幅改善以文本为输入的大模型的使用体验,同时避免了基于 ASR 解决方案的繁琐流程以及可能引入的错误。
Train transformer language models with reinforcement learning.
Generative Agents: Interactive Simulacra of Human Behavior
Llama中文社区,Llama3在线体验和微调模型已开放,实时汇总最新Llama3学习资料,已将所有代码更新适配Llama3,构建最好的中文Llama大模型,完全开源可商用
Scripts for fine-tuning Meta Llama3 with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supportin…
An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.