Lists (1)
Sort Name ascending (A-Z)
Stars
This repository hosts the code, data and model weight of PanoSent.
The official GitHub repo for the survey paper "Knowledge Conflicts for LLMs: A Survey"
MLNLP社区用来帮助大家避免论文投稿小错误的整理仓库。 Paper Writing Tips
Code and data for "Dancing in Chains: Reconciling Instruction Following and Faithfulness in Language Models"
[ACL 2024] Self-Training with Direct Preference Optimization Improves Chain-of-Thought Reasoning
A Unified Library for Parameter-Efficient and Modular Transfer Learning
Gemini is a modern LaTex beamerposter theme 🖼
Finetune Llama 3.1, Mistral, Phi & Gemma LLMs 2-5x faster with 80% less memory
BeHonest: Benchmarking Honesty in Large Language Models
Code for our EMNLP 2023 Paper: "LLM-Adapters: An Adapter Family for Parameter-Efficient Fine-Tuning of Large Language Models"
Official repo for the paper "Scaling Synthetic Data Creation with 1,000,000,000 Personas"
Code for Math-LLaVA: Bootstrapping Mathematical Reasoning for Multimodal Large Language Models
This is the official repository of the paper "OlympicArena: Benchmarking Multi-discipline Cognitive Reasoning for Superintelligent AI"
SVG icons for popular brands
Released code for「Self-Training with Pseudo-Label Scorer for Aspect Sentiment Quad Prediction」in ACL2024.
🍼 Official implementation of Dynamic Data Mixing Maximizes Instruction Tuning for Mixture-of-Experts
SimPO: Simple Preference Optimization with a Reference-Free Reward
Recipes to train reward model for RLHF.
llama3 implementation one matrix multiplication at a time
ReFT: Representation Finetuning for Language Models
Stanford NLP Python Library for Understanding and Improving PyTorch Models via Interventions
Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.
Utility for behavioral and representational analyses of Language Models
Easy-to-Hard Generalization: Scalable Alignment Beyond Human Supervision