Skip to content

A reading list for large models safety, security, and privacy.

License

Notifications You must be signed in to change notification settings

daoyuan14/lm-ssp

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 

Repository files navigation

image

Stars

Introduction

The resources related to the trustworthiness of large models (LMs) across multiple dimensions (e.g., safety, security, and privacy), with a special focus on multi-modal LMs (e.g., vision-language models and diffusion models).

  • This repo is in progress 🌱 (currently manually collected).

  • Badges:

    • Model: img img img

    • Comment: img img img img img img img

    • Venue (Continuous update): img or img

  • 🌻 Welcome to recommend resources to us via Issues with the following format (please fill in this table):

Title Link Code Venue Classification Model Comment
aa arxiv github bb'23 A1. Jailbreak LLM Agent

News

  • [2023.01.20] 🔥 We collect 3 related papers from NDSS'24!
  • [2023.01.17] 🔥 We collect 108 related papers from ICLR'24!
  • [2023.01.09] 🔥 LM-SSP is released!

Book

Competition

Leaderboard

Toolkit

Survey

Paper

image

A. Safety

A1. Jailbreak

A2. Alignment

A3. Deepfake

A4. Ethics

A5. Fairness

A6. Hallucination

A7. Toxicity

B. Security

B1. Adversarial Examples

B2. Poisoning

C. Privacy

C1. Contamination

C2. Copyright

C3. Data Reconstruction

C4. Extraction

C5. Inference

C6. Privacy-Preserving Computation

C7. Unlearning

Star History

Star History Chart

Acknowledgement

About

A reading list for large models safety, security, and privacy.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published