Block or Report
Block or report quan-tran-tu
Contact GitHub support about this user’s behavior. Learn more about reporting abuse.
Report abuseStars
Language
Sort by: Recently starred
Full stack, modern web application template. Using FastAPI, React, SQLModel, PostgreSQL, Docker, GitHub Actions, automatic HTTPS and more.
An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries
Ongoing research training transformer models at scale
The simplest, fastest repository for training/finetuning medium-sized GPTs.
A massively parallel, high-level programming language
ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator
A complete computer science study plan to become a software engineer.
Code for 'LLM2Vec: Large Language Models Are Secretly Powerful Text Encoders'
Implementing a ChatGPT-like LLM in PyTorch from scratch, step by step
GPT based autonomous agent that does online comprehensive research on any given topic
OpenUI let's you describe UI using your imagination, then see it rendered live.
#1 Locally hosted web application that allows you to perform various operations on PDF files
An efficient pure-PyTorch implementation of Kolmogorov-Arnold Network (KAN).
There can be more than Notion and Miro. AFFiNE(pronounced [ə‘fain]) is a next-gen knowledge base that brings planning, sorting and creating all together. Privacy first, open-source, customizable an…
Understanding JupyterHub by deploying it step by step
Fast and memory-efficient exact attention
Hackable and optimized Transformers building blocks, supporting a composable construction.
Reference implementation of Megalodon 7B model
Pioneering in Vietnamese Multimodal Large Language Model
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
Unify Efficient Fine-Tuning of 100+ LLMs
Garnet is a remote cache-store from Microsoft Research that offers strong performance (throughput and latency), scalability, storage, recovery, cluster sharding, key migration, and replication feat…
A simple but complete full-attention transformer with a set of promising experimental features from various papers
The calflops is designed to calculate FLOPs、MACs and Parameters in all various neural networks, such as Linear、 CNN、 RNN、 GCN、Transformer(Bert、LlaMA etc Large Language Model)
🧑🏫 60 Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gan…