-
Harbin Institute of Technology
- Shenzhen, China
-
07:23
(UTC +08:00) - zchuz.github.io
Block or Report
Block or report zchuz
Contact GitHub support about this user’s behavior. Learn more about reporting abuse.
Report abuseStars
Language
Sort by: Recently starred
Implementation for "Step-DPO: Step-wise Preference Optimization for Long-chain Reasoning of LLMs"
[ICML 2024] CLLMs: Consistency Large Language Models
This is the repository that contains the source code for the Self-Evaluation Guided MCTS for online DPO.
ReST-MCTS*: LLM Self-Training via Process Reward Guided Tree Search
Qwen2 is the large language model series developed by Qwen team, Alibaba Cloud.
DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models
"Improving Mathematical Reasoning with Process Supervision" by OPENAI
GLM-4 series: Open Multilingual Multimodal Chat LMs | 开源多语言多模态对话模型
How to think step-by-step: A mechanistic understanding of chain-of-thought reasoning
Repository for Label Words are Anchors: An Information Flow Perspective for Understanding In-Context Learning
[ACL 2024] The official codebase for the paper "Self-Distillation Bridges Distribution Gap in Language Model Fine-tuning".
SimPO: Simple Preference Optimization with a Reference-Free Reward
A simple tool to update bib entries with their official information (e.g., DBLP or the ACL anthology).
📖A curated list of Awesome LLM Inference Paper with codes, TensorRT-LLM, vLLM, streaming-llm, AWQ, SmoothQuant, WINT8/4, Continuous Batching, FlashAttention, PagedAttention etc.
MambaOut: Do We Really Need Mamba for Vision?
A Survey of Attributions for Large Language Models
CodeQwen1.5 is the code version of Qwen, the large language model series developed by Qwen team, Alibaba Cloud.
A curated list of LLM Interpretability related material - Tutorial, Library, Survey, Paper, Blog, etc..
awesome papers in LLM interpretability
An LLM-powered knowledge curation system that researches a topic and generates a full-length report with citations.
AAAI 2024 Papers: Explore a comprehensive collection of innovative research papers presented at one of the premier artificial intelligence conferences. Seamlessly integrate code implementations for…
A curated list for Efficient Large Language Models
The official repo for "TheoremQA: A Theorem-driven Question Answering dataset" (EMNLP 2023)
This repository collects papers for "A Survey on Knowledge Distillation of Large Language Models". We break down KD into Knowledge Elicitation and Distillation Algorithms, and explore the Skill & V…