Skip to content
forked from sozercan/aikit

πŸ—οΈ AI + BuildKit = AIKit: Build and deploy open-source LLMs easily!

License

Notifications You must be signed in to change notification settings

Toinoux38/aikit

Β 
Β 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

AIKit ✨


AIKit is a quick, easy, and local or cloud-agnostic way to get started to host and deploy large language models (LLMs) for inference. No GPU, internet access or additional tools are needed to get started except for Docker!

AIKit uses LocalAI under-the-hood to run inference. LocalAI provides a drop-in replacement REST API that is OpenAI API compatible, so you can use any OpenAI API compatible client, such as Kubectl AI, Chatbot-UI and many more, to send requests to open-source LLMs powered by AIKit!

πŸ‘‰ For full documentation, please see AIKit website!

Features

Quick Start

You can get started with AIKit quickly on your local machine without a GPU!

docker run -d --rm -p 8080:8080 ghcr.io/sozercan/llama2:7b
curl https://localhost:8080/v1/chat/completions -H "Content-Type: application/json" -d '{
    "model": "llama-2-7b-chat",
    "messages": [{"role": "user", "content": "explain kubernetes in a sentence"}]
  }'

Output should be similar to:

{"created":1701236489,"object":"chat.completion","id":"dd1ff40b-31a7-4418-9e32-42151ab6875a","model":"llama-2-7b-chat","choices":[{"index":0,"finish_reason":"stop","message":{"role":"assistant","content":"\nKubernetes is a container orchestration system that automates the deployment, scaling, and management of containerized applications in a microservices architecture."}}],"usage":{"prompt_tokens":0,"completion_tokens":0,"total_tokens":0}}

That's it! πŸŽ‰ API is OpenAI compatible so this is a drop-in replacement for any OpenAI API compatible client.

Pre-made Models

AIKit comes with pre-made models that you can use out-of-the-box!

CPU

Model Optimization Parameters Command License
πŸ¦™ Llama 2 Chat 7B docker run -d --rm -p 8080:8080 ghcr.io/sozercan/llama2:7b Llama 2
πŸ¦™ Llama 2 Chat 13B docker run -d --rm -p 8080:8080 ghcr.io/sozercan/llama2:13b Llama 2
🐬 Orca 2 13B docker run -d --rm -p 8080:8080 ghcr.io/sozercan/orca2:13b Microsoft Research
Ⓜ️ Mixtral Instruct 8x7B docker run -d --rm -p 8080:8080 ghcr.io/sozercan/mixtral:8x7b Apache
πŸ…ΏοΈ Phi 2 Instruct 2.7B docker run -d --rm -p 8080:8080 ghcr.io/sozercan/phi2:2.7b MIT

NVIDIA CUDA

Model Optimization Parameters Command License
πŸ¦™ Llama 2 Chat Chat 7B docker run -d --rm --gpus all -p 8080:8080 ghcr.io/sozercan/llama2:7b-cuda Llama 2
πŸ¦™ Llama 2 Chat Chat 13B docker run -d --rm --gpus all -p 8080:8080 ghcr.io/sozercan/llama2:13b-cuda Llama 2
🐬 Orca 2 13B docker run -d --rm --gpus all -p 8080:8080 ghcr.io/sozercan/orca2:13b-cuda Microsoft Research
Ⓜ️ Mixtral Instruct 8x7B docker run -d --rm --gpus all -p 8080:8080 ghcr.io/sozercan/mixtral:8x7b-cuda Apache
πŸ…ΏοΈ Phi 2 Instruct 2.7B docker run -d --rm --gpus all -p 8080:8080 ghcr.io/sozercan/phi2:2.7b-cuda MIT

πŸ‘‰ For more information and how to create your own images, please see AIKit website.

About

πŸ—οΈ AI + BuildKit = AIKit: Build and deploy open-source LLMs easily!

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages

  • Go 74.8%
  • JavaScript 12.6%
  • CSS 7.1%
  • Makefile 4.4%
  • Dockerfile 1.1%