Open Source AI Test Tools
Guardrails
Guardrails AI by Guardrail AI, (Cf. Guardrails Hub)
Safety and Trustworthiness
DecodingTrust by AI Secure (UIUC Secure Learning Lab)
Holistic Evaluation of Language Models (HELM) by Stanford CRFM
Privacy (IAM and Data Provenance)
Pebblo by Daxa
Explainability and Model Transparency
SHapley Additive exPlanations (SHAP)
Local Interpretable Model-agnostic Explanations (LIME) by Marco Tulio Correia Ribeiro (Google DeepMind)
Representation Engineering (RepE)
Security
CyberSecEval Purple Llama cybersecurity benchmark (Meta)
garak by Leon Derczynski (Nvidia)
LLM Attacks (see Universal and Transferable Adversarial Attacks on Aligned Language Models)
Rebuff by Protect AI
Deepfake Detection
Augmentations library (AugLy) for function and class-based audio, image, text, and video transforms by Meta Research
Adversarial Attack
Adversarial Robustness Toolbox (ART) by LF AI & Data Trusted AI (see documentation)
TextAttack framework for adversarial attacks, data augmentation, and adversarial training in NLP (see research paper and documentation)
Regulatory Compliance
COMPL-AI open-source compliance-centered evaluation framework for Generative AI model compliance with the EU AI Act
Sustainability and Corporate Social Responsibility
Adaptation Vectors by Aethercloud models rate of change in technology versus ability of workers to adapt to change