Block or Report
Block or report ERnest666
Contact GitHub support about this user’s behavior. Learn more about reporting abuse.
Report abuseStars
Language
Sort by: Recently starred
The code of paper "Toward Optimal LLM Alignments Using Two-Player Games".
ICML 2024: Improving Factuality and Reasoning in Language Models through Multiagent Debate
📖 Paper reading list in conversational AI (constantly updating 🤗).
To speed up LLMs' inference and enhance LLM's perceive of key information, compress the prompt and KV-Cache, which achieves up to 20x compression with minimal performance loss.
Compress your input to ChatGPT or other LLMs, to let them process 2x more content and save 40% memory and GPU time.
Repository for the code of the "PPL-MCTS: Constrained Textual Generation Through Discriminator-Guided Decoding" paper, NAACL'22
A quick guide (especially) for trending instruction finetuning datasets
A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)
PAL: Proxy-Guided Black-Box Attack on Large Language Models
LLaMA-TRL: Fine-tuning LLaMA with PPO and LoRA
Wrapper to easily generate the chat template for Llama2
Official implementation of ICLR'24 paper, "Curiosity-driven Red Teaming for Large Language Models" (https://openreview.net/pdf?id=4KqkizXgXU)
Code and documents of LongLoRA and LongAlpaca (ICLR 2024 Oral)
Code for AAAI 2023 paper 'Learning to Memorize Entailment and Discourse Relations for Persona-Consistent Dialogues'
Train transformer language models with reinforcement learning.
A modular RL library to fine-tune language models to human preferences
library supporting NLP and CV research on scientific papers
Datasets for "Evaluating Gender Bias of Pre-trained Language Models in Natural Language Inference by Considering All Labels"
Codes for NAACL'22 "A Study of the Attention Abnormality in Trojaned BERTs", a textural Trojan Detector
Ongoing research training transformer models at scale
Pre-processing and in some cases downloading of datasets for the paper "Content Selection in Deep Learning Models of Summarization."
Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
Repository for EMNLP 2022 Paper: Towards a Unified Multi-Dimensional Evaluator for Text Generation
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
Acceptance rates for the major AI conferences