![:octocat: :octocat:](https://github.githubassets.com/images/icons/emoji/octocat.png)
Block or Report
Block or report tomoki0924
Contact GitHub support about this user’s behavior. Learn more about reporting abuse.
Report abuseLists (1)
Sort Name ascending (A-Z)
Stars
Language
Sort by: Recently starred
Official repository of the "Transformer Fusion with Optimal Transport" paper, published as a conference paper at ICLR 2024.
A fast implementation of T5/UL2 in PyTorch using Flash Attention
links to conference publications in graph-based deep learning
Fixes mojibake and other glitches in Unicode text, after the fact.
macOS and Linux VMs on Apple Silicon to use in CI and other automations
GIZA++ is a statistical machine translation toolkit that is used to train IBM Models 1-5 and an HMM word alignment model. This package also contains the source for the mkcls tool which generates th…
Simple, minimal implementation of the Mamba SSM in one file of PyTorch.
The paper list of the 86-page paper "The Rise and Potential of Large Language Model Based Agents: A Survey" by Zhiheng Xi et al.
Drop in a screenshot and convert it to clean code (HTML/Tailwind/React/Vue)
A curated list of reinforcement learning with human feedback resources (continually updated)
PyTorch + HuggingFace code for RetoMaton: "Neuro-Symbolic Language Modeling with Automaton-augmented Retrieval" (ICML 2022), including an implementation of kNN-LM and kNN-MT
DSPy: The framework for programming—not prompting—foundation models
The robust text processing pipeline framework enabling customizable, efficient, and metric-logged text preprocessing.
Out-of-distribution detection, robustness, and generalization resources. The repository contains a professionally curated list of papers, tutorials, books, videos, articles and open-source librarie…
A curated list of research papers and resources on code-switching
Automatically split your PyTorch models on multiple GPUs for training & inference
You like pytorch? You like micrograd? You love tinygrad! ❤️
A curated list of awesome papers related to pre-trained models for information retrieval (a.k.a., pretraining for IR).
Type annotations and dynamic checking for a tensor's shape, dtype, names, etc.
Pyrallis is a framework for structured configuration parsing from both cmd and files. Simply define your desired configuration structure as a dataclass and let pyrallis do the rest!