ToolEmu
ethics
ToolEmu | ethics | |
---|---|---|
3 | 1 | |
125 | 265 | |
0.0% | 7.5% | |
5.5 | 0.0 | |
11 months ago | almost 2 years ago | |
Python | Python | |
Apache License 2.0 | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
ToolEmu
-
[R] Identifying the Risks of LM Agents with an LM-Emulated Sandbox - University of Toronto 2023 - Benchmark consisting of 36 high-stakes tools and 144 test cases!
Website: https://toolemu.com/
- ToolEmu: Identifying the Risks of LM Agents with an LM-Emulated Sandbox
-
Identifying the Risks of LM Agents with an LM-Emulated Sandbox - University of Toronto 2023 - Benchmark consisting of 36 high-stakes tools and 144 test cases!
Github: https://github.com/ryoungj/toolemu
ethics
-
[P] Request: Any datasets of morality stories?
Code for https://arxiv.org/abs/2008.02275 found: https://github.com/hendrycks/ethics
What are some alternatives?
LOGICGUIDE - Plug in and Play implementation of "Certified Reasoning with Language Models" that elevates model reasoning by 40%
moonwatcher - Evaluation & testing framework for computer vision models
lumos - Code and data for "Lumos: Learning Agents with Unified Data, Modular Design, and Open-Source LLMs"
natural-adv-examples - A Harder ImageNet Test Set (CVPR 2021)
graph-of-thoughts - Official Implementation of "Graph of Thoughts: Solving Elaborate Problems with Large Language Models"
ACE_Model_Implementation - A python implementation of Dave Shap's ACE Model
npi - Action library for AI Agent
giskard - 🐢 Open-Source Evaluation & Testing for AI & LLM systems
prompttools - Open-source tools for prompt testing and experimentation, with support for both LLMs (e.g. OpenAI, LLaMA) and vector databases (e.g. Chroma, Weaviate, LanceDB).
awesome-ai-safety - 📚 A curated list of papers & technical articles on AI Quality & Safety
safe-rlhf - Safe RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback