![google logo](https://raw.githubusercontent.com/github/explore/80688e429a7d4ef2fca1e82350fe8e3517d3494d/topics/google/google.png)
Block or Report
Block or report wangjiaqiys
Contact GitHub support about this user’s behavior. Learn more about reporting abuse.
Report abuseLanguage
Sort by: Recently starred
Starred repositories
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & Mixtral)
Generated synthetic data for UNLP-2023 Shared Task in GEC for Ukrainian
The repo of "Improving Seq2Seq Grammatical Error Correction via Decoding Interventions"
A WebUI for Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
中文nlp解决方案(大模型、数据、模型、训练、推理)
Code release for NeRF (Neural Radiance Fields)
A simple UI / Web / Frontend for MLX mlx-lm using Streamlit.
[CVPR2023 Highlight] Consistent-Teacher: Towards Reducing Inconsistent Pseudo-targets in Semi-supervised Object Detection
DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models
kinds of reinforcement learning model by Pytorch
[MLSys'24] Atom: Low-bit Quantization for Efficient and Accurate LLM Serving
RAG for Local LLM, chat with PDF/doc/txt files, ChatPDF
Chat with any PDF. Easily upload the PDF documents you'd like to chat with. Instant answers. Ask questions, extract information, and summarize documents with AI. Sources included.
A toolkit for developing and comparing reinforcement learning algorithms.
⛷️ LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-training
Get up and running with Llama 3, Mistral, Gemma 2, and other large language models.
🌟 The Multi-Agent Framework: First AI Software Company, Towards Natural Language Programming
OpenMMLab Model Compression Toolbox and Benchmark.