Stars
A high-throughput and memory-efficient inference and serving engine for LLMs
MoRA: High-Rank Updating for Parameter-Efficient Fine-Tuning
MusePose: a Pose-Driven Image-to-Video Framework for Virtual Human Generation
V-Express aims to generate a talking head video under the control of a reference image, an audio, and a sequence of V-Kps images.
[NeurIPS'24] HippoRAG is a novel RAG framework inspired by human long-term memory that enables LLMs to continuously integrate knowledge across external documents. RAG + Knowledge Graphs + Personali…
Cool Python features for machine learning that I used to be too afraid to use. Will be updated as I have more time / learn more.
Finetune Llama 3.2, Mistral, Phi & Gemma LLMs 2-5x faster with 80% less memory
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
RAG Deploy -> Optimize -> Deploy again. Get high performance RAG service with less effort
LangChain 공식 Document, Cookbook, 그 밖의 실용 예제를 바탕으로 작성한 한국어 튜토리얼입니다. 본 튜토리얼을 통해 LangChain을 더 쉽고 효과적으로 사용하는 방법을 배울 수 있습니다.
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
🇰🇷파이토치에서 제공하는 튜토리얼의 한국어 번역을 위한 저장소입니다. (Translate PyTorch tutorials in Korean🇰🇷)
무료로 한국어🇰🇷 파인튜닝 모델 받아서 로컬 LLM 호스팅. LangServe, Ollama, streamlit + RAG