Skip to content

TrelisResearch/one-click-llms

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

57 Commits
 
 
 
 

Repository files navigation

one-click-llms

Tip

Post a new issue if you would like other templates.

Tip

Advanced inferencing scripts (incl. for function calling, data extraction, advanced RAG methods, and private data redaction) are available for purchase here.

These one click templates allow you to quickly boot up an API for a given language model.

  • Read through the README file on the templates!
  • Runpod is recommended (better user interface) if using larger GPUs like A6000, A100 or H100.
  • Vast.AI is recommended for lowest cost per hour with smaller GPUs like A600000 and A2000. However, the user experience is significantly worse with Vast.AI than runpod.

Runpod One-Click Templates

Tip

To support the Trelis Research YouTube channel, you can sign up for an account with this link. Trelis is supported by a commission when you use one-click templates.

GPU Choices and Tips

For best reliability around CUDA versions, I recommend:

  • A6000 (48 GB VRAM)
  • A100 SXM (more expensive than PCI but more reliably has up to date CUDA)
  • H100 PCI or SXM - best for fp8 models, but expensive.

Fine-tuning Notebook Setup

  • CUDA 12.1 one-click template here

vLLM (requires an A100 or H100 or A6000, i.e. ampere architecture):

Note: The vLLM image has compatibility issues with certain Runpod CUDA drivers, leading to issues on certain pods. A6000 Ada is typically an option that works.

Important

Note: vLLM runs into issues sometimes if the pod template does not have the correct CUDA drivers. Unfortunately there is no way to know when picking a GPU. An issue has been raised here. As an alternative, you can run TGI (and even query in openai style, guide here). TGI is faster than vLLM and recommended in general. Note however, that TGI does not automatically apply the chat template to the prompt when using the OpenAI style endpoint.

Text Generation Inference:

llama.cpp One-click templates:

MoonDream Multi-modal API (openai-ish)

Vast AI One-Click Templates

Tip

To support the Trelis Research YouTube channel, you can sign up for an account with this affiliate link. Trelis is supported by a commission when you use one-click templates.

Fine-tuning Notebook Setup

  • CUDA 12.1 one-click template here.

Text Generation Inference (fastest):

vLLM (requires an A100 or H100 or A6000, i.e. ampere architecture):

llama.cpp One-click templates:

Function-calling One-Click Templates

One-click templates for function-calling are located on the HuggingFace model cards. Check out the collection here.

Tip

As of July 23rd 2024, function calling fine-tuned models are being deprecated in favour of a one-shot approach with stronger models. Find the "Tool Use" video on the Trelis YouTube Channel for more info.

Changelog

20Jul2023:

  • Update the ./llama-server.sh command in line with breaking changes to llama.cpp

Feb 16 2023:

  • Added a Mamba one click template.

Jan 21 2023:

  • Swapped Runpod to before Vast.AI as user experience is much better with Runpod.

Jan 9 2023:

  • Added Mixtral Instruct AWQ TGI

Dec 30 2023:

  • Support gated models by adding HUGGING_FACE_HUB_TOKEN env variable.
  • Speed up downloading using HuggingFace API.

Dec 29 2023:

  • Add in one-click llama.cpp server template.

About

One click templates for inferencing Language Models

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages