ai-safety

Open-source projects categorized as ai-safety

Top 10 ai-safety Open-Source Projects

  • giskard

    ๐Ÿข Open-Source Evaluation & Testing framework for LLMs and ML models

  • Project mention: Show HN: Evaluate LLM-based RAG Applications with automated test set generation | news.ycombinator.com | 2024-04-11
  • safe-rlhf

    Safe RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback

  • Project mention: [R] Meet Beaver-7B: a Constrained Value-Aligned LLM via Safe RLHF Technique | /r/MachineLearning | 2023-05-16
  • InfluxDB

    Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.

    InfluxDB logo
  • guardrail

    Build LLM apps safely and securely๐Ÿ›ก๏ธ (by guardrail-ml)

  • Project mention: Safeguard OpenAI Apps with Guardrail MLโ€™s Firewall | /r/OpenAI | 2023-09-25
  • Thought-Cloning

    [NeurIPS '23 Spotlight] Thought Cloning: Learning to Think while Acting by Imitating Human Thinking

  • Project mention: AI Agents Can Learn to Think While Acting: A New AI Research Introduces A Novel Imitation Learning Framework Called Thought Cloning | /r/machinelearningnews | 2023-06-07
  • ethics

    Aligning AI With Shared Human Values (ICLR 2021)

  • make-safe-ai

    How to Make Safe AI? Let's Discuss! ๐Ÿ’ก|๐Ÿ’ฌ|๐Ÿ™Œ|๐Ÿ“š

  • awesome-ai-safety

    ๐Ÿ“š A curated list of papers & technical articles on AI Quality & Safety

  • Project mention: Ask HN: Who is hiring? (October 2023) | news.ycombinator.com | 2023-10-02

    Giskard - Testing framework for ML models| Multiple roles | Full-time | France | https://giskard.ai/

    We are building the first collaborative & open-source Quality Assurance platform for all ML models - including Large Language Models.

    Founded in 2021 in Paris by ex-Dataiku engineers, we are an emerging player in the fast-growing market of AI Quality & Safety.

    Giskard helps Data Scientists & ML Engineering teams collaborate to evaluate, test & monitor AI models. We help organizations increase the efficiency of their AI development workflow, eliminate risks of AI biases and ensure robust, reliable & ethical AI models. Our open-source platform is used by dozens of ML teams across industries, both at enterprise companies & startups.

    In 2022, we raised our first round of 1.5 million euros, led by Elaia, with participation from Bessemer and notable angel investors including the CTO of Hugging Face. To read more about this fundraising and how it will accelerate our growth, you can read this announcement: https://www.giskard.ai/knowledge/news-fundraising-2022

    In 2023, we received a strategic investment from the European Commission to build a SaaS platform to automate compliance with the upcoming EU AI regulation. You can read more here: https://www.giskard.ai/knowledge/1-000-github-stars-3meu-and...

    We are assembling a team of champions: Software Engineers, Machine Learning researchers, and Data Scientists ; to build our AI Quality platform and expand it to new types of AI models and industries. We have a culture of continuous learning & quality, and we help each other achieve high standards & goals!

    We aim to grow from 15 to 25 people in the next 12 months. We're hiring the following roles:

  • SaaSHub

    SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives

    SaaSHub logo
  • ToolEmu

    A language model (LM)-based emulation framework for identifying the risks of LM agents with tool use

  • Project mention: [R] Identifying the Risks of LM Agents with an LM-Emulated Sandbox - University of Toronto 2023 - Benchmark consisting of 36 high-stakes tools and 144 test cases! | /r/MachineLearning | 2023-10-11

    Website: https://toolemu.com/

  • adversarial-reinforcement-learning

    Reading list for adversarial perspective and robustness in deep reinforcement learning.

  • Project mention: Safety in Deep Reinforcement Learning | /r/programming | 2023-12-06
  • ai-safety-cheatsheet

    A compilation of AI safety ideas, problems, and solutions.

NOTE: The open source projects on this list are ordered by number of github stars. The number of mentions indicates repo mentiontions in the last 12 Months or since we started tracking (Dec 2020).

ai-safety related posts

  • [R] Identifying the Risks of LM Agents with an LM-Emulated Sandbox - University of Toronto 2023 - Benchmark consisting of 36 high-stakes tools and 144 test cases!

    1 project | /r/MachineLearning | 11 Oct 2023
  • ToolEmu: Identifying the Risks of LM Agents with an LM-Emulated Sandbox

    1 project | news.ycombinator.com | 10 Oct 2023
  • [R] Awesome AI Safety โ€“ A curated list of papers & technical articles on AI Quality & Safety

    1 project | /r/MachineLearning | 5 May 2023

Index

What are some of the best open-source ai-safety projects? This list will help you:

Project Stars
1 giskard 3,142
2 safe-rlhf 1,160
3 guardrail 311
4 Thought-Cloning 232
5 ethics 207
6 make-safe-ai 166
7 awesome-ai-safety 135
8 ToolEmu 86
9 adversarial-reinforcement-learning 75
10 ai-safety-cheatsheet 9

Sponsored
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com