ai-security

There are 59 repositories under ai-security topic.

  • h4cker

    The-Art-of-Hacking/h4cker

    This repository is primarily maintained by Omar Santos (@santosomar) and includes thousands of resources related to ethical hacking, bug bounties, digital forensics and incident response (DFIR), artificial intelligence security, vulnerability research, exploit development, reverse engineering, and more.

    Language:Jupyter Notebook17k866933.1k
  • giskard

    Giskard-AI/giskard

    🐢 Open-Source Evaluation & Testing for LLMs and ML models

    Language:Python3.5k27427221
  • jiep/offensive-ai-compilation

    A curated list of useful resources that cover Offensive AI.

    Language:HTML1k253103
  • THUYimingLi/backdoor-learning-resources

    A list of backdoor learning resources

  • normster/llm_rules

    RuLES: a benchmark for evaluating rule-following in language models

    Language:Python1962314
  • ZhengyuZhao/AI-Security-and-Privacy-Events

    A curated list of academic events on AI Security & Privacy

  • ruoxi-jia-group/Narcissus

    The official implementation of the CCS'23 paper, Narcissus clean-label backdoor attack -- only takes THREE images to poison a face recognition dataset in a clean-label way and achieves a 99.89% attack success rate.

    Language:Python942910
  • jay-johnson/train-ai-with-django-swagger-jwt

    Train AI (Keras + Tensorflow) to defend apps with Django REST Framework + Celery + Swagger + JWT - deploys to Kubernetes and OpenShift Container Platform

    Language:Python706124
  • RjDuan/AdvDrop

    Code for "Adversarial attack by dropping information." (ICCV 2021)

    Language:Python7021917
  • Hacking-Notes/VulnScan

    Performing website vulnerability scanning using OpenAI technologie

    Language:Python52143
  • YiZeng623/I-BAU

    Official Implementation of ICLR 2022 paper, ``Adversarial Unlearning of Backdoors via Implicit Hypergradient''

    Language:Jupyter Notebook472313
  • CVPR_2019_PNI

    elliothe/CVPR_2019_PNI

    pytorch implementation of Parametric Noise Injection for adversarial defense

    Language:Python412316
  • Imperio

    HKU-TASR/Imperio

    [IJCAI 2024] Imperio is an LLM-powered backdoor attack. It allows the adversary to issue language-guided instructions to control the victim model's prediction for arbitrary targets.

    Language:Python40403
  • mitre-atlas/atlas-data

    ATLAS tactics, techniques, and case studies data

    Language:Python37329
  • AnthenaMatrix/Website-Prompt-Injection

    Website Prompt Injection is a concept that allows for the injection of prompts into an AI system via a website's. This technique exploits the interaction between users, websites, and AI systems to execute specific prompts that influence AI behavior.

    Language:HTML32116
  • AI-Initiative-KAUST/VideoRLCS

    Learning to Identify Critical States for Reinforcement Learning from Videos (Accepted to ICCV'23)

    Language:Python25103
  • modzy/sdk-python

    Python library for Modzy Machine Learning Operations (MLOps) Platform

    Language:Python24663
  • zhangzp9970/MIA

    Unofficial pytorch implementation of paper: Model Inversion Attacks that Exploit Confidence Information and Basic Countermeasures

    Language:Python22325
  • AnthenaMatrix/Prompt-Injection-Testing-Tool

    The Prompt Injection Testing Tool is a Python script designed to assess the security of your AI system's prompt handling against a predefined list of user prompts commonly used for injection attacks. This tool utilizes the OpenAI GPT-3.5 model to generate responses to system-user prompt pairs and outputs the results to a CSV file for analysis.

    Language:Python204
  • jay-johnson/antinex-core

    Network exploit detection using highly accurate pre-trained deep neural networks with Celery + Keras + Tensorflow + Redis

    Language:Jupyter Notebook20802
  • AnthenaMatrix/Image-Prompt-Injection

    Image Prompt Injection is a Python script that demonstrates how to embed a secret prompt within an image using steganography techniques. This hidden prompt can be later extracted by an AI system for analysis, enabling covert communication with AI models through images.

    Language:Python181112
  • modzy/sdk-javascript

    The official JavaScript SDK for the Modzy Machine Learning Operations (MLOps) Platform.

    Language:TypeScript16433
  • SEC-CAFE/handbook

    安全手册,企业安全实践、攻防与安全研究知识库

    Language:CSS15104
  • ruoxi-jia-group/Meta-Sift

    The official implementation of USENIX Security'23 paper "Meta-Sift" -- Ten minutes or less to find a 1000-size or larger clean subset on poisoned dataset.

    Language:Python14204
  • moonwatcher-ai/moonwatcher

    Evaluation & testing framework for computer vision models

    Language:Python130
  • CyberAlbSecOP/Awesome_CyberSec_Bible

    Cyber-Security Bible! Theory and Tools, Kali Linux, Penetration Testing, Bug Bounty, CTFs, Malware Analysis, Cryptography, Secure Programming, Web App Security, Cloud Security, Devsecops, Ethical Hacking, Social Engineering, Privacy, Incident Response, Threat Assestment, Personal Security, Ai Security, Android Security, Iot Security, Standards.

  • modzy/sdk-java

    The official Java library for the Modzy Machine Learning Operations (MLOps) Platform

    Language:Java11521
  • tsmotlp/AI-Security-Research

    A curated collection of the latest academic research papers and developments in AI Security. This repository aims to provide a comprehensive source for researchers and enthusiasts to stay updated on AI Security trends and findings. Contributions welcome!

  • wearetyomsmnv/AI-LLM-ML_security_study_map

    Do you want to learn AI Security but don't know where to start ? Take a look at this map.

  • AnthenaMatrix/AI-Prompt-Injection-List

    AI/LLM Prompt Injection List is a curated collection of prompts designed for testing AI or Large Language Models (LLMs) for prompt injection vulnerabilities. This list aims to provide a comprehensive set of prompts that can be used to evaluate the behavior of AI or LLM systems when exposed to different types of inputs.

  • sachink1729/Healthcare-AI-Assistant-Medical-Data-Qdrant-Dspy-Groq

    Building Private Healthcare AI Assistant for Clinics Using Qdrant Hybrid Cloud, DSPy and Groq - Llama3

    Language:Jupyter Notebook10100
  • jay-johnson/antinex-datasets

    Datasets for training deep neural networks to defend software applications

    Language:Python830
  • Safetorun/PromptDefender

    A prompt defence is a multi-layer defence that can be used to protect your applications against prompt injection attacks.

    Language:Go80
  • IDRnD/idvoice-gpt-ios-demo

    IDVoice + ChatGPT iOS demo app

    Language:Swift5000
  • pagiux/maleficnet

    Neural networks, but malefic! 😈

    Language:Python5102