{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"omnisafe","owner":"PKU-Alignment","isFork":false,"description":"OmniSafe is an infrastructural framework for accelerating SafeRL research.","allTopics":["benchmark","machine-learning","reinforcement-learning","deep-learning","deep-reinforcement-learning","constraint-satisfaction-problem","pytorch","safety-critical","saferl","safe-reinforcement-learning","safe-rl","safety-gym","safety-gymnasium","constraint-rl"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":6,"issueCount":10,"starsCount":866,"forksCount":126,"license":"Apache License 2.0","participation":[0,2,0,1,3,0,0,0,3,0,1,1,2,0,2,0,0,0,0,1,0,1,0,0,1,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,2,1,0,3,2,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-16T07:22:07.285Z"}},{"type":"Public","name":"safety-gymnasium","owner":"PKU-Alignment","isFork":false,"description":"NeurIPS 2023: Safety-Gymnasium: A Unified Safe Reinforcement Learning Benchmark","allTopics":["reinforcement-learning","constraint-satisfaction-problem","safety-critical","safety-critical-systems","safe-reinforcement-learning","safe-reinforcement-learning-environments","constraint-rl","safe-policy-optimization"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":330,"forksCount":46,"license":"Apache License 2.0","participation":[0,0,0,1,2,1,3,0,0,0,0,10,0,1,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,3,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-14T11:36:31.328Z"}},{"type":"Public","name":"safe-rlhf","owner":"PKU-Alignment","isFork":false,"description":"Safe RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback","allTopics":["reinforcement-learning","transformers","transformer","safety","llama","gpt","datasets","beaver","alpaca","ai-safety","safe-reinforcement-learning","vicuna","deepspeed","large-language-models","llm","llms","rlhf","reinforcement-learning-from-human-feedback","safe-rlhf","safe-reinforcement-learning-from-human-feedback"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":14,"starsCount":1190,"forksCount":104,"license":"Apache License 2.0","participation":[4,5,12,4,1,7,6,4,1,5,2,1,1,0,1,0,0,0,2,1,1,0,0,3,3,0,0,7,4,0,0,0,0,1,0,0,0,0,0,0,0,4,0,0,1,0,2,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-20T18:48:21.134Z"}},{"type":"Public","name":"ProAgent","owner":"PKU-Alignment","isFork":false,"description":"ProAgent: Building Proactive Cooperative Agents with Large Language Models","allTopics":["language-model","cooperative","human-ai","overcooked","human-ai-interaction","cooperative-ai","llm-agent"],"primaryLanguage":{"name":"JavaScript","color":"#f1e05a"},"pullRequestCount":0,"issueCount":1,"starsCount":37,"forksCount":2,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-08T12:51:53.293Z"}},{"type":"Public","name":"SafeDreamer","owner":"PKU-Alignment","isFork":false,"description":"ICLR 2024: SafeDreamer: Safe Reinforcement Learning with World Models","allTopics":["reinforcement-learning","constraint-satisfaction-problem","safety-critical-systems","safe-reinforcement-learning","constraint-rl","safe-policy-optimization"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":31,"forksCount":2,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-08T02:08:21.875Z"}},{"type":"Public","name":"Safe-Policy-Optimization","owner":"PKU-Alignment","isFork":false,"description":"NeurIPS 2023: Safe Policy Optimization: A benchmark repository for safe reinforcement learning algorithms","allTopics":["benchmarks","reinforcement-learning-algorithms","safe","safe-reinforcement-learning","constrained-reinforcement-learning"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":296,"forksCount":39,"license":"Apache License 2.0","participation":[1,0,0,18,0,0,8,0,1,15,4,53,10,0,0,0,0,5,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-20T13:24:58.013Z"}},{"type":"Public","name":"AlignmentSurvey","owner":"PKU-Alignment","isFork":false,"description":"AI Alignment: A Comprehensive Survey","allTopics":["awesome","reinforcement-learning","ai","deep-learning","survey","alignment","papers","interpretability","red-teaming","large-language-models"],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":116,"forksCount":0,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,15,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-02T15:36:18.622Z"}},{"type":"Public","name":"beavertails","owner":"PKU-Alignment","isFork":false,"description":"BeaverTails is a collection of datasets designed to facilitate research on safety alignment in large language models (LLMs).","allTopics":["safety","llama","gpt","datasets","language-model","beaver","ai-safety","human-feedback-data","llm","llms","human-feedback","rlhf","large-language-model","safe-rlhf"],"primaryLanguage":{"name":"Makefile","color":"#427819"},"pullRequestCount":1,"issueCount":1,"starsCount":84,"forksCount":3,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-27T19:47:48.329Z"}},{"type":"Public","name":".github","owner":"PKU-Alignment","isFork":false,"description":"","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-05-31T13:45:56.717Z"}},{"type":"Public","name":"ReDMan","owner":"PKU-Alignment","isFork":false,"description":"ReDMan is an open-source simulation platform that provides a standardized implementation of safe RL algorithms for Reliable Dexterous Manipulation.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":15,"forksCount":2,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-05-02T16:47:17.577Z"}}],"repositoryCount":10,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}