![tensorflow logo](https://raw.githubusercontent.com/github/explore/80688e429a7d4ef2fca1e82350fe8e3517d3494d/topics/tensorflow/tensorflow.png)
Highlights
- Pro
Block or Report
Block or report Erland366
Contact GitHub support about this user’s behavior. Learn more about reporting abuse.
Report abuseLanguage: C++
Sort by: Most stars
Starred repositories
An Open Source Machine Learning Framework for Everyone
GPT4All: Chat with Local LLMs on Any Device
Protocol Buffers - Google's data interchange format
🔍 A Hex Editor for Reverse Engineers, Programmers and people who value their retinas when working at 3 AM.
GoogleTest - Google Testing and Mocking Framework
Port of OpenAI's Whisper model in C/C++
A library for efficient similarity search and clustering of dense vectors.
🤖 The free, Open Source OpenAI alternative. Self-hosted, community-driven and local-first. Drop-in replacement for OpenAI running on consumer-grade hardware. No GPU required. Runs gguf, transformer…
Open Source alternative to Algolia + Pinecone and an Easier-to-Use alternative to ElasticSearch ⚡ 🔍 ✨ Fast, typo tolerant, in-memory fuzzy Search Engine for building delightful search experiences
Distribute and run LLMs with a single file.
Seamless operability between C++11 and Python
A high-performance, zero-overhead, extensible Python compiler using LLVM
Approximate Nearest Neighbors in C++/Python optimized for memory usage and loading/saving to disk
Dear PyGui: A fast and powerful Graphical User Interface Toolkit for Python with minimal dependencies
Development repository for the Triton language and compiler
NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.
Vowpal Wabbit is a machine learning system which pushes the frontier of machine learning with techniques such as online, hashing, allreduce, reductions, learning2search, active, and interactive lea…
High-speed Large Language Model Serving on PCs with Consumer-grade GPUs
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficie…
lightweight, standalone C++ inference engine for Google's Gemma models.
Transformer related optimization, including BERT, GPT
A C++ standalone library for machine learning
Bear is a tool that generates a compilation database for clang tooling.
Turbopilot is an open source large-language-model based code completion engine that runs locally on CPU