Stars
Modern GPU Compute and Rendering in Javascript
List availability of ECO servers at OVH and let you buy one
SVDQuant: Absorbing Outliers by Low-Rank Components for 4-Bit Diffusion Models
Fast & Simple repository for pre-training and fine-tuning T5-style models
Lumina-T2X is a unified framework for Text to Any Modality Generation
Fast and accurate automatic speech recognition (ASR) for edge devices
fluencelabs / redis
Forked from redis/redisRedis is an in-memory database that persists on disk. The data model is key-value, but many different kind of values are supported: Strings, Lists, Sets, Sorted Sets, Hashes, HyperLogLogs, Bitmaps.
A compact LLM pretrained in 9 days by using high quality data
zhangpiu / llm.cpp
Forked from karpathy/llm.cLLM training in simple, C++/CUDA(with Eigen3)
GGUF Quantization support for native ComfyUI models
Lightweight WASM Postgres with real-time, reactive bindings.
arlo-phoenix / CTranslate2-rocm
Forked from OpenNMT/CTranslate2Fast inference engine for Transformer models
[NeurIPS 2023] MeZO: Fine-Tuning Language Models with Just Forward Passes. https://arxiv.org/abs/2305.17333
PixArt-α: Fast Training of Diffusion Transformer for Photorealistic Text-to-Image Synthesis
The official repo of Qwen2-Audio chat & pretrained large audio language model proposed by Alibaba Cloud.
Fine-Tuning of a multi-language transformer model on Nvidia GPUs.