Stars
Efficient Triton Kernels for LLM Training
A curated list of papers and resources based on "Large Language Models on Graphs: A Comprehensive Survey"
Fast and memory-efficient exact attention
A framework for serving and evaluating LLM routers - save LLM costs without compromising quality!
Evaluate your LLM's response with Prometheus and GPT4 💯
[SIGIR 2024] The official repo for paper "Planning Ahead in Generative Retrieval: Guiding Autoregressive Generation through Simultaneous Decoding"
언어모델을 학습하기 위한 공개 한국어 instruction dataset들을 모아두었습니다.
Perplexica is an AI-powered search engine. It is an Open source alternative to Perplexity AI
20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.
[WWW 2024] The official repo for paper "Scalable and Effective Generative Information Retrieval".
An open science effort to benchmark legal reasoning in foundation models
LexGLUE: A Benchmark Dataset for Legal Language Understanding in English
Pyserini is a Python toolkit for reproducible information retrieval research with sparse and dense representations.
A series of large language models trained from scratch by developers @01-ai
Reading list of hallucination in LLMs. Check out our new survey paper: "Siren’s Song in the AI Ocean: A Survey on Hallucination in Large Language Models"
Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
Train transformer language models with reinforcement learning.
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
A collection of open-source dataset to train instruction-following LLMs (ChatGPT,LLaMA,Alpaca)