Stars
LightEval is a lightweight LLM evaluation suite that Hugging Face has been using internally with the recently released LLM data processing library datatrove and LLM training library nanotron.
Arrays, Tensors and dynamic Neural Networks in Mojo 🔥
[IEEE ICIP 2024] Diversifying Deep Ensembles: A Saliency Map Approach for Enhanced OOD Detection, Calibration, and Accuracy
Efficient implementations of state-of-the-art linear attention models in Pytorch and Triton
A massively parallel, high-level programming language
GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection
Code for exploring Based models from "Simple linear attention language models balance the recall-throughput tradeoff"
Official implementation of the paper "Linear Transformers with Learnable Kernel Functions are Better In-Context Models"
A library for mechanistic interpretability of GPT-style language models
Code implementing "Efficient Parallelization of a Ubiquitious Sequential Computation" (Heinsen, 2023)
JAX-accelerated Meta-Reinforcement Learning Environments Inspired by XLand and MiniGrid 🏎️
Deep learning in Rust, with shape checked tensors and neural networks
⚡ A Fast, Extensible Progress Bar for Python and CLI
Transformers with Arbitrarily Large Context
This is the Rust course used by the Android team at Google. It provides you the material to quickly teach Rust.