-
Cornell University
- Ithaca, New York
-
10:25
(UTC -04:00) - billysx.github.io
Stars
A curated list of practical guide resources of Medical LLMs (Medical LLMs Tree, Tables, and Papers)
Awesome-LLM: a curated list of Large Language Model
High-quality datasets, tools, and concepts for LLM fine-tuning.
Codebase for ICML'24 paper: Learning from Students: Applying t-Distributions to Explore Accurate and Efficient Formats for LLMs
Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
[ICLR 2024] Efficient Streaming Language Models with Attention Sinks
SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX Runtime
QLoRA: Efficient Finetuning of Quantized LLMs
FinGPT: Open-Source Financial Large Language Models! Revolutionize 🔥 We release the trained model on HuggingFace.
ChatLaw:A Powerful LLM Tailored for Chinese Legal. 中文法律大模型
Must-read Papers on Knowledge Editing for Large Language Models.
Student-run wiki for students interested in computer science at Cornell University
Running large language models on a single GPU for throughput-oriented scenarios.
Instruction Tuning with GPT-4
Human preference data for "Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback"
Aligning Large Language Models with Human: A Survey
Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the PaLM architecture. Basically ChatGPT but with PaLM
Tool for data extraction and interacting with Lean programmatically.
4 bits quantization of LLaMA using GPTQ
Official repo for consistency models.
为GPT/GLM等LLM大语言模型提供实用化交互接口,特别优化论文阅读/润色/写作体验,模块化设计,支持自定义快捷按钮&函数插件,支持Python和C++等项目剖析&自译解功能,PDF/LaTex论文翻译&总结功能,支持并行问询多种LLM模型,支持chatglm3等本地模型。接入通义千问, deepseekcoder, 讯飞星火, 文心一言, llama2, rwkv, claude2, m…
A LLM based research assistant that allows you to have a conversation with a research paper
Offsite-Tuning: Transfer Learning without Full Model
Versatile Diffusion: Text, Images and Variations All in One Diffusion Model, arXiv 2022 / ICCV 2023
[ICLR 2023] "More ConvNets in the 2020s: Scaling up Kernels Beyond 51x51 using Sparsity"; [ICML 2023] "Are Large Kernels Better Teachers than Transformers for ConvNets?"
Acceptance rates for the major AI conferences