{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"align-anything","owner":"PKU-Alignment","isFork":false,"description":"Align Anything: Training Any Modality Model with Feedback","allTopics":["multimodal","large-language-models","rlhf"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":5,"starsCount":60,"forksCount":17,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,5,12],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-28T07:45:33.911Z"}},{"type":"Public","name":"align-anything-eval-ljh","owner":"PKU-Alignment","isFork":true,"description":"jiahao's fork on evaluation","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":17,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,6,36],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-28T06:55:42.989Z"}},{"type":"Public","name":"chenxinyu-align-anything-evaluation","owner":"PKU-Alignment","isFork":true,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":17,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,6,36],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-25T12:09:50.989Z"}},{"type":"Public","name":"ProgressGym","owner":"PKU-Alignment","isFork":false,"description":"Alignment with a millennium of moral progress.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":6,"forksCount":1,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,4,0,0,13],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-25T09:18:52.923Z"}},{"type":"Public","name":".github","owner":"PKU-Alignment","isFork":false,"description":"","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,3,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-14T15:23:56.717Z"}},{"type":"Public","name":"safe-sora","owner":"PKU-Alignment","isFork":false,"description":"SafeSora is a human preference dataset designed to support safety alignment research in the text-to-video generation field, aiming to enhance the helpfulness and harmlessness of Large Vision Models (LVMs).","allTopics":["alignment","human-preferences","text-to-video-generation","large-vision-models"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":22,"forksCount":4,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,1,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T11:38:27.258Z"}},{"type":"Public","name":"omnisafe","owner":"PKU-Alignment","isFork":false,"description":"OmniSafe is an infrastructural framework for accelerating SafeRL research.","allTopics":["benchmark","machine-learning","reinforcement-learning","deep-learning","deep-reinforcement-learning","constraint-satisfaction-problem","pytorch","safety-critical","saferl","safe-reinforcement-learning","safe-rl","safety-gym","safety-gymnasium","constraint-rl"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":6,"issueCount":15,"starsCount":882,"forksCount":129,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-04T12:13:00.727Z"}},{"type":"Public","name":"safe-rlhf","owner":"PKU-Alignment","isFork":false,"description":"Safe RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback","allTopics":["reinforcement-learning","transformers","transformer","safety","llama","gpt","datasets","beaver","alpaca","ai-safety","safe-reinforcement-learning","vicuna","deepspeed","large-language-models","llm","llms","rlhf","reinforcement-learning-from-human-feedback","safe-rlhf","safe-reinforcement-learning-from-human-feedback"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":13,"starsCount":1260,"forksCount":115,"license":"Apache License 2.0","participation":[0,6,1,1,1,0,1,0,0,0,3,0,1,0,0,3,3,0,2,6,3,0,0,0,0,1,0,0,0,0,0,0,2,2,0,0,1,2,0,0,0,0,0,0,0,1,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-13T10:06:45.636Z"}},{"type":"Public","name":"llms-resist-alignment","owner":"PKU-Alignment","isFork":false,"description":"Repo for paper \"Language Models Resist Alignment\"","allTopics":["alignment","llama","safe","alpaca","ai-safety","vicuna","llm","llms","rlhf","safe-rlhf","llama2","llama3"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":3,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-09T16:16:30.995Z"}},{"type":"Public","name":"safety-gymnasium","owner":"PKU-Alignment","isFork":false,"description":"NeurIPS 2023: Safety-Gymnasium: A Unified Safe Reinforcement Learning Benchmark","allTopics":["reinforcement-learning","constraint-satisfaction-problem","safety-critical","safety-critical-systems","safe-reinforcement-learning","safe-reinforcement-learning-environments","constraint-rl","safe-policy-optimization"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":2,"starsCount":351,"forksCount":50,"license":"Apache License 2.0","participation":[0,0,0,10,0,1,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,3,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-14T11:36:31.328Z"}},{"type":"Public","name":"ProAgent","owner":"PKU-Alignment","isFork":false,"description":"ProAgent: Building Proactive Cooperative Agents with Large Language Models","allTopics":["language-model","cooperative","human-ai","overcooked","human-ai-interaction","cooperative-ai","llm-agent"],"primaryLanguage":{"name":"JavaScript","color":"#f1e05a"},"pullRequestCount":0,"issueCount":1,"starsCount":45,"forksCount":4,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-08T12:51:53.293Z"}},{"type":"Public","name":"SafeDreamer","owner":"PKU-Alignment","isFork":false,"description":"ICLR 2024: SafeDreamer: Safe Reinforcement Learning with World Models","allTopics":["reinforcement-learning","constraint-satisfaction-problem","safety-critical-systems","safe-reinforcement-learning","constraint-rl","safe-policy-optimization"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":35,"forksCount":3,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-08T02:08:21.875Z"}},{"type":"Public","name":"Safe-Policy-Optimization","owner":"PKU-Alignment","isFork":false,"description":"NeurIPS 2023: Safe Policy Optimization: A benchmark repository for safe reinforcement learning algorithms","allTopics":["benchmarks","reinforcement-learning-algorithms","safe","safe-reinforcement-learning","constrained-reinforcement-learning"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":310,"forksCount":44,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-20T13:24:58.013Z"}},{"type":"Public","name":"AlignmentSurvey","owner":"PKU-Alignment","isFork":false,"description":"AI Alignment: A Comprehensive Survey","allTopics":["awesome","reinforcement-learning","ai","deep-learning","survey","alignment","papers","interpretability","red-teaming","large-language-models"],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":120,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-02T15:36:18.622Z"}},{"type":"Public","name":"beavertails","owner":"PKU-Alignment","isFork":false,"description":"BeaverTails is a collection of datasets designed to facilitate research on safety alignment in large language models (LLMs).","allTopics":["safety","llama","gpt","datasets","language-model","beaver","ai-safety","human-feedback-data","llm","llms","human-feedback","rlhf","large-language-model","safe-rlhf"],"primaryLanguage":{"name":"Makefile","color":"#427819"},"pullRequestCount":0,"issueCount":2,"starsCount":89,"forksCount":3,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-27T19:47:48.329Z"}},{"type":"Public","name":"ReDMan","owner":"PKU-Alignment","isFork":false,"description":"ReDMan is an open-source simulation platform that provides a standardized implementation of safe RL algorithms for Reliable Dexterous Manipulation.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":15,"forksCount":2,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-05-02T16:47:17.577Z"}}],"repositoryCount":16,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"PKU-Alignment repositories"}