gemma
Here are 53 public repositories matching this topic...
Firefly: 大模型训练工具,支持训练Qwen2、Yi1.5、Phi-3、Llama3、Gemma、MiniCPM、Yi、Deepseek、Orion、Xverse、Mixtral-8x7B、Zephyr、Mistral、Baichuan2、Llma2、Llama、Qwen、Baichuan、ChatGLM2、InternLM、Ziya2、Vicuna、Bloom等大模型
-
Updated
Jul 16, 2024 - Python
Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with any open-source language models, speech recognition models, and multimodal models, whether in the cloud, on-premises, or even on your laptop.
-
Updated
Aug 7, 2024 - Python
A snappy, keyboard-centric terminal user interface for interacting with large language models. Chat with ChatGPT, Claude, Llama 3, Phi 3, Mistral, Gemma and more.
-
Updated
Jul 30, 2024 - Python
Unofficial PyTorch/🤗Transformers(Gemma/Llama3) implementation of Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention
-
Updated
Apr 23, 2024 - Python
Official repository for "Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing". Your efficient and high-quality synthetic data generation pipeline!
-
Updated
Aug 5, 2024 - Python
JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs welcome).
-
Updated
Aug 7, 2024 - Python
On-device LLM Inference Powered by X-Bit Quantization
-
Updated
Jul 22, 2024 - Python
NAACL '24 (Best Demo Paper RunnerUp) / MlSys @ NeurIPS '23 - RedCoast: A Lightweight Tool to Automate Distributed Training and Inference
-
Updated
Aug 7, 2024 - Python
Train GEMMA on TPU/GPU! (Codebase for training Gemma-Ko Series)
-
Updated
Mar 2, 2024 - Python
DeveloperGPT is a LLM-powered command line tool that enables natural language to terminal commands and in-terminal chat.
-
Updated
Jun 1, 2024 - Python
PyTorch/XLA integration with JetStream (https://github.com/google/JetStream) for LLM inference"
-
Updated
Aug 6, 2024 - Python
RAG-nificent is a state-of-the-art framework leveraging Retrieval-Augmented Generation (RAG) to provide instant answers and references from a curated directory of PDFs containing information on any given topic. Supports Llama3.1 and OpenAI Models via the Groq API.
-
Updated
Jul 23, 2024 - Python
Improve this page
Add a description, image, and links to the gemma topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the gemma topic, visit your repo's landing page and select "manage topics."