ml-safety
There are 19 repositories under ml-safety topic.
Giskard-AI/giskard
🐢 Open-Source Evaluation & Testing for AI & LLM systems
hendrycks/robustness
Corruption and Perturbation Robustness (ICLR 2019)
hendrycks/natural-adv-examples
A Harder ImageNet Test Set (CVPR 2021)
hendrycks/outlier-exposure
Deep Anomaly Detection with Outlier Exposure (ICLR 2019)
JohnSnowLabs/langtest
Deliver safe & effective language models
agencyenterprise/PromptInject
PromptInject is a framework that assembles prompts in a modular fashion to provide a quantitative analysis of the robustness of LLMs to adversarial prompt attacks. 🏆 Best Paper Awards @ NeurIPS ML Safety Workshop 2022
hendrycks/ss-ood
Self-Supervised Learning for OOD Detection (NeurIPS 2019)
hendrycks/ethics
Aligning AI With Shared Human Values (ICLR 2021)
hendrycks/imagenet-r
ImageNet-R(endition) and DeepAugment (ICCV 2021)
jiachens/ModelNet40-C
Repo for "Benchmarking Robustness of 3D Point Cloud Recognition against Common Corruptions" https://arxiv.org/abs/2201.12296
Giskard-AI/awesome-ai-safety
📚 A curated list of papers & technical articles on AI Quality & Safety
hendrycks/anomaly-seg
The Combined Anomalous Object Segmentation (CAOS) Benchmark
hendrycks/pre-training
Pre-Training Buys Better Robustness and Uncertainty Estimates (ICML 2019)
YyzHarry/ME-Net
[ICML 2019] ME-Net: Towards Effective Adversarial Robustness with Matrix Estimation
hendrycks/jiminy-cricket
Jiminy Cricket Environment (NeurIPS 2021)
yaodongyu/ProjNorm
Predicting Out-of-Distribution Error with the Projection Norm
moonwatcher-ai/moonwatcher
Evaluation & testing framework for computer vision models
harsmac/MUFIACode
Code for the attack multiplicative filter attack MUFIA, from the paper "Frequency-based vulnerability analysis of deep learning models against image corruptions".
ArianeDlns/adv-AI-project
This repository contains the project for the Advanced AI course @CentraleSupélec