-
Alibaba Qwen Team
- China
-
00:41
(UTC +08:00) - http:https://huybery.github.io
- @huybery
Highlights
Block or Report
Block or report huybery
Contact GitHub support about this user’s behavior. Learn more about reporting abuse.
Report abuseStars
Language
Sort by: Recently starred
A collection of phenomenons observed during the scaling of big foundation models, which may be developed into consensus, principles, or laws in the future
Implementation for "Step-DPO: Step-wise Preference Optimization for Long-chain Reasoning of LLMs"
Qwen2 is the large language model series developed by Qwen team, Alibaba Cloud.
CodeQwen1.5 is the code version of Qwen, the large language model series developed by Qwen team, Alibaba Cloud.
SWE-agent takes a GitHub issue and tries to automatically fix it, using GPT-4, or your LM of choice. It solves 12.47% of bugs in the SWE-bench evaluation set and takes just 1 minute to run.
OSWorld: Benchmarking Multimodal Agents for Open-Ended Tasks in Real Computer Environments
CodeUltraFeedback: aligning large language models to coding preferences
Sailor: Open Language Models for South-East Asia
Doing simple retrieval from LLM models at various context lengths to measure accuracy
Generative Representational Instruction Tuning
DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models
DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models
CRUXEval: Code Reasoning, Understanding, and Execution Evaluation
Repository for paper Tools Are Instrumental for Language Agents in Complex Environments
📰 Must-read papers and blogs on LLM based Long Context Modeling 🔥
《Hello 算法》:动画图解、一键运行的数据结构与算法教程。支持 Python, Java, C++, C, C#, JS, Go, Swift, Rust, Ruby, Kotlin, TS, Dart 代码。简体版和繁体版同步更新,English version ongoing
High Accuracy and efficiency multi-task fine-tuning framework for Code LLMs. This work has been accepted by KDD 2024.
The official repo of Qwen (通义千问) chat & pretrained large language model proposed by Alibaba Cloud.
A dataset of LLM-generated chain-of-thought steps annotated with mistake location.
Drop in a screenshot and convert it to clean code (HTML/Tailwind/React/Vue)
Rigourous evaluation of LLM-synthesized code - NeurIPS 2023