Block or Report
Block or report sisterdong
Contact GitHub support about this user’s behavior. Learn more about reporting abuse.
Report abuseLists (1)
Sort Name ascending (A-Z)
Stars
Language
Sort by: Recently starred
RayDP provides simple APIs for running Spark on Ray and integrating Spark with AI libraries.
Unified storage framework for the entire machine learning lifecycle
Simple LaTeX parser providing latex-to-unicode and unicode-to-latex conversion
OpenCompass is an LLM evaluation platform, supporting a wide range of models (Llama3, Mistral, InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, etc) over 100+ datasets.
中文 NLP 预处理、解析工具包,准确、高效、易用 A Chinese NLP Preprocessing & Parsing Package www.jionlp.com
🤖 The free, Open Source OpenAI alternative. Self-hosted, community-driven and local-first. Drop-in replacement for OpenAI running on consumer-grade hardware. No GPU required. Runs gguf, transformer…
Test your prompts, agents, and RAGs. Use LLM evals to improve your app's quality and catch problems. Compare performance of GPT, Claude, Gemini, Llama, and more. Simple declarative configs with com…
A high-throughput and memory-efficient inference and serving engine for LLMs
[ACL'24 Oral] Data and code for L-Eval, a comprehensive long context language models evaluation benchmark
The paper list of the 86-page paper "The Rise and Potential of Large Language Model Based Agents: A Survey" by Zhiheng Xi et al.
Conceptual Captions is a dataset containing (image-URL, caption) pairs designed for the training and evaluation of machine learned image captioning systems.
Inference code for Persimmon-8B
🆓免费的 ChatGPT 镜像网站列表,持续更新。List of free ChatGPT mirror sites, continuously updated.
LLMs build upon Evol Insturct: WizardLM, WizardCoder, WizardMath
Source Code of Paper "GPTScore: Evaluate as You Desire"
Evaluation framework for your Retrieval Augmented Generation (RAG) pipelines
Chat凉宫春日, An open sourced Role-Playing chatbot Cheng Li, Ziang Leng, and others.
A modular active learning framework for Python
Codes for our paper "ChatEval: Towards Better LLM-based Evaluators through Multi-Agent Debate"
Z-Bench 1.0 by 真格基金:一个麻瓜的大语言模型中文测试集。Z-Bench is a LLM prompt dataset for non-technical users, developed by an enthusiastic AI-focused team in Zhenfund.
A collection of open-source dataset to train instruction-following LLMs (ChatGPT,LLaMA,Alpaca)
Benchmarking large language models' complex reasoning ability with chain-of-thought prompting
Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.
Dromedary: towards helpful, ethical and reliable LLMs.
Instruction Tuning with GPT-4