Stars
This repository contains a pytorch implementation of "SelfRecon: Self Reconstruction Your Digital Avatar from Monocular Video (CVPR 2022, Oral)".
QLoRA: Efficient Finetuning of Quantized LLMs
Instruct-tune LLaMA on consumer hardware
LLMs build upon Evol Insturct: WizardLM, WizardCoder, WizardMath
PromptCBLUE: a large-scale instruction-tuning dataset for multi-task and few-shot learning in the medical domain in Chinese
A framework for the evaluation of autoregressive code generation language models.
A Gradio web UI for Large Language Models.
ChatGLM3 series: Open Bilingual Chat LLMs | 开源双语对话语言模型
Mass-editing thousands of facts into a transformer memory (ICLR 2023)
Holistic Evaluation of Language Models (HELM), a framework to increase the transparency of language models (https://arxiv.org/abs/2211.09110). This framework is also used to evaluate text-to-image …
[ICLR2024 spotlight] OmniQuant is a simple and powerful quantization technique for LLMs.
[ICML 2023] SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models
An easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ algorithm.
Real-time face swap for PC streaming or video calls
Industry leading face manipulation platform
Fast and Robust Early-Exiting Framework for Autoregressive Language Models with Synchronized Parallel Decoding (EMNLP 2023 Long)
[NeurIPS'23] H2O: Heavy-Hitter Oracle for Efficient Generative Inference of Large Language Models.
End-to-end carbon footprint mod- eling tool
【LLMs九层妖塔】分享 LLMs在自然语言处理(ChatGLM、Chinese-LLaMA-Alpaca、小羊驼 Vicuna、LLaMA、GPT4ALL等)、信息检索(langchain)、语言合成、语言识别、多模态等领域(Stable Diffusion、MiniGPT-4、VisualGLM-6B、Ziya-Visual等)等 实战与经验。
Implementation of the conditionally routed attention in the CoLT5 architecture, in Pytorch
Implementation of Recurrent Memory Transformer, Neurips 2022 paper, in Pytorch
Implementation of Block Recurrent Transformer - Pytorch
[ICLR 2024] Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parameters