Highlights
Stars
Run PyTorch LLMs locally on servers, desktop and mobile
Generative AI extensions for onnxruntime
Estimate Your LLM's Token Toll Across Various Platforms and Configurations
Ikigai is an AI-powered Open Assignment System
Development repository for the Triton language and compiler
A natural language interface for computers
A platform for community discussion. Free, open, simple.
A SQLite extension for efficient vector search, based on Faiss!
Next generation BLAS implementation for ROCm platform
WebAssembly binding for llama.cpp - Enabling in-browser LLM inference
Triton Model Navigator is an inference toolkit designed for optimizing and deploying Deep Learning models with a focus on NVIDIA GPUs.
OpenAI compatible API for TensorRT LLM triton backend
janhq / cortex.tensorrt-llm
Forked from NVIDIA/TensorRT-LLMCortex.Tensorrt-LLM is a C++ inference library that can be loaded by any server at runtime. It submodules NVIDIA’s TensorRT-LLM for GPU accelerated inference on NVIDIA's GPUs.
Scheduling infrastructure for absolutely everyone.
A privacy-first, open-source platform for knowledge management and collaboration. Download link: https://github.com/logseq/logseq/releases. roadmap: https://trello.com/b/8txSM12G/roadmap
Browser extension that simplifies the GitHub interface and adds useful features
A curated list of awesome remote jobs and resources. Inspired by https://github.com/vinta/awesome-python
OBS Studio - Free and open source software for live streaming and screen recording
Stable Diffusion with Core ML on Apple Silicon
Swift Package to implement a transformers-like API in Swift
Everything we actually know about the Apple Neural Engine (ANE)
A library with extensible implementations of DPO, KTO, PPO, ORPO, and other human-aware loss functions (HALOs).
NVIDIA Federated Learning Application Runtime Environment