-
NLP Cloud
- Grenoble, France
- https://juliensalinas.com
- @juliensalinasen
Block or Report
Block or report juliensalinas
Contact GitHub support about this userβs behavior. Learn more about reporting abuse.
Report abuseStars
Language
Sort by: Recently starred
Dockerized FastAPI wrapper around the recognize-anything image recognition models
ProtonVPN Wireguard Docker Image. Supports ARMv8 (64-bit ) and x86 (64-Bit).
Incredibly fast Whisper-large-v3
A series of large language models trained from scratch by developers @01-ai
π¦π Build context-aware reasoning applications
A high-throughput and memory-efficient inference and serving engine for LLMs
A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.
Pretrain, finetune and deploy AI models on multiple GPUs, TPUs with zero code changes.
20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.
LLMs build upon Evol Insturct: WizardLM, WizardCoder, WizardMath
Unofficial DeGiro stock broker API. See your portfolio and set up orders in the market like wall street
Multilingual Sentence & Image Embeddings with BERT
JAX implementation of OpenAI's Whisper model for up to 70x speed-up on TPU.
π Text-Prompted Generative Audio Model
Large language models (LLMs) made easy, EasyLM is a one stop solution for pre-training, finetuning, evaluating and serving LLMs in JAX/Flax.
Faster Whisper transcription with CTranslate2
4 bits quantization of LLaMA using GPTQ
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
Instruct-tune LLaMA on consumer hardware
antimatter15 / alpaca.cpp
Forked from ggerganov/llama.cppLocally run an Instruction-Tuned Chat-Style LLM
Experiments with generating opensource language model assistants
Making large AI models cheaper, faster and more accessible
Running large language models on a single GPU for throughput-oriented scenarios.
Efficient few-shot learning with Sentence Transformers
Crosslingual Generalization through Multitask Finetuning