Lists (1)
Sort Name ascending (A-Z)
Stars
FlagEval is an evaluation toolkit for AI large foundation models.
The simplest, fastest repository for training/finetuning medium-sized GPTs.
Collection of evals for Inspect AI
A collection of projects designed to help developers quickly get started with building deployable applications using the Anthropic API
RuLES: a benchmark for evaluating rule-following in language models
Contains all assets to run with Moonshot Library (Connectors, Datasets and Metrics)
S-Eval: Automatic and Adaptive Test Generation for Benchmarking Safety Evaluation of Large Language Models
Deep learning for dummies. All the practical details and useful utilities that go into working with real models.
A fast + lightweight implementation of the GCG algorithm in PyTorch
A curated list of awesome resources dedicated to Scaling Laws for LLMs
[NeurIPS 2023] Tree of Thoughts: Deliberate Problem Solving with Large Language Models
Moonshot - A simple and modular tool to evaluate and red-team any LLM application.
A benchmark for prompt injection detection systems.
Make your GenAI Apps Safe & Secure 🚀 Test & harden your system prompt
A framework for few-shot evaluation of language models.
A curated list of safety-related papers, articles, and resources focused on Large Language Models (LLMs). This repository aims to provide researchers, practitioners, and enthusiasts with insights i…
A fast, clean, responsive Hugo theme.
Inspect: A framework for large language model evaluations
【ACL 2024】 SALAD benchmark & MD-Judge
[NeurIPS 2024] SWE-agent takes a GitHub issue and tries to automatically fix it, using GPT-4, or your LM of choice. It can also be employed for offensive cybersecurity or competitive coding challen…
A Comprehensive Assessment of Trustworthiness in GPT Models
HarmBench: A Standardized Evaluation Framework for Automated Red Teaming and Robust Refusal