Resonance Router Starshipping
- Chicago
-
01:56
(UTC -06:00)
Block or Report
Block or report Starshipping
Contact GitHub support about this user’s behavior. Learn more about reporting abuse.
Report abuseStars
Language
Sort by: Recently starred
Hallo: Hierarchical Audio-Driven Visual Synthesis for Portrait Image Animation
DSPy: The framework for programming—not prompting—foundation models
[Preprint] Learning to Filter Context for Retrieval-Augmented Generaton
The all-in-one Desktop & Docker AI application with full RAG and AI Agent capabilities.
LightEval is a lightweight LLM evaluation suite that Hugging Face has been using internally with the recently released LLM data processing library datatrove and LLM training library nanotron.
Transformer related optimization, including BERT, GPT
Fine-grained Post-training for Improving Retrieval-based Dialogue Systems - NAACL 2021
A web interface for chatting with Alpaca through llama.cpp. Fully dockerized, with an easy to use API.
This repository provides an original implementation of Detecting Pretraining Data from Large Language Models by *Weijia Shi, *Anirudh Ajith, Mengzhou Xia, Yangsibo Huang, Daogao Liu , Terra Blevins…
ChatRWKV is like ChatGPT but powered by RWKV (100% RNN) language model, and open source.
[EMNLP 2023] Adapting Language Models to Compress Long Contexts
Benchmarking large language models' complex reasoning ability with chain-of-thought prompting
🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support
Retrieval Augmented Generation (RAG) framework and context engine powered by Pinecone
Efficient implementations of state-of-the-art linear attention models in Pytorch and Triton
A Paper List for Open-Domain Dialogue Generation, and related datasets.
Train transformer language models with reinforcement learning.
A preliminary evaluation of ChatGPT/GPT-4 for machine translation.
Skywork series models are pre-trained on 3.2TB of high-quality multilingual (mainly Chinese and English) and code data. We have open-sourced the model, training data, evaluation data, evaluation me…
Code for exploring Based models from "Simple linear attention language models balance the recall-throughput tradeoff"
A professionally curated list of awesome resources (paper, code, data, etc.) on transformers in time series.
Codes for our IJCAI21 paper: Dialogue Discourse-Aware Graph Model and Data Augmentation for Meeting Summarization
A paper & resource list of large language models, including course, paper, demo, figures