Run any open-source LLMs, such as Llama 3.1, Gemma, as OpenAI compatible API endpoint in the cloud.
-
Updated
Oct 29, 2024 - Python
Run any open-source LLMs, such as Llama 3.1, Gemma, as OpenAI compatible API endpoint in the cloud.
RAG (Retrieval Augmented Generation) Framework for building modular, open source applications for production by TrueFoundry
A new DSL and server for AI agents and multi-step tasks
AIConfig is a config-based framework to build generative AI applications.
The collaborative spreadsheet for AI. Chain cells into powerful pipelines, experiment with prompts and models, and evaluate LLM responses in real-time. Work together seamlessly to build and iterate on AI applications.
Python SDK for running evaluations on LLM generated responses
An end-to-end LLM reference implementation providing a Q&A interface for Airflow and Astronomer
cluster/scheduler health monitoring for GPU jobs on k8s
Friendli: the fastest serving engine for generative AI
Miscellaneous codes and writings for MLOps
The prompt engineering, prompt management, and prompt evaluation tool for TypeScript, JavaScript, and NodeJS.
Add a description, image, and links to the llm-ops topic page so that developers can more easily learn about it.
To associate your repository with the llm-ops topic, visit your repo's landing page and select "manage topics."