ST4RF4LL's Stars
ray-project/ray
Ray is a unified framework for scaling AI and Python applications. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.
tatsu-lab/stanford_alpaca
Code and documentation to train Stanford's Alpaca models, and generate the data.
NLP-LOVE/ML-NLP
此项目是机器学习(Machine Learning)、深度学习(Deep Learning)、NLP面试中常考到的知识点和代码实现,也是作为一个算法工程师必会的理论基础知识。
e2b-dev/awesome-ai-agents
A list of AI autonomous agents
Trusted-AI/adversarial-robustness-toolbox
Adversarial Robustness Toolbox (ART) - Python Library for Machine Learning Security - Evasion, Poisoning, Extraction, Inference - Red and Blue Teams
QData/TextAttack
TextAttack 🐙 is a Python framework for adversarial attacks, data augmentation, and model training in NLP https://textattack.readthedocs.io/en/master/
meta-llama/PurpleLlama
Set of tools to assess and improve LLM security.
databricks/dbrx
Code examples and resources for DBRX, a large language model developed by Databricks
HarderThenHarder/transformers_tasks
⭐️ NLP Algorithms with transformers lib. Supporting Text-Classification, Text-Generation, Information-Extraction, Text-Matching, RLHF, SFT etc.
frgfm/torch-cam
Class activation maps for your PyTorch models (CAM, Grad-CAM, Grad-CAM++, Smooth Grad-CAM++, Score-CAM, SS-CAM, IS-CAM, XGrad-CAM, Layer-CAM)
leondz/garak
LLM vulnerability scanner
thu-coai/Safety-Prompts
Chinese safety prompts for evaluating and improving the safety of LLMs. 中文安全prompts,用于评估和提升大模型的安全性。
Azure/counterfit
a CLI that provides a generic automation layer for assessing the security of ML models
HowieHwong/TrustLLM
[ICML 2024] TrustLLM: Trustworthiness in Large Language Models
chbrian/awesome-adversarial-examples-dl
A curated list of awesome resources for adversarial examples in deep learning
OWASP/www-project-ai-security-and-privacy-guide
OWASP Foundation Web Respository
penghui-yang/awesome-data-poisoning-and-backdoor-attacks
A curated list of papers & resources linked to data poisoning, backdoor attacks and defenses against them (no longer maintained)
ftramer/LM_Memorization
Training data extraction on GPT-2
alexdevassy/Machine_Learning_CTF_Challenges
CTF challenges designed and implemented in machine learning applications
ZhaoyangLyu/POPQORN
An Algorithm to Quantify Robustness of Recurrent Neural Networks
tpai/gandalf-prompt-injection-writeup
A writeup for the Gandalf prompt injection game.
TopRedTeam/OSCP2024
OSCP Notes
dependable-cps/adversarial-MTSR
LostOxygen/llm-confidentiality
Whispers in the Machine: Confidentiality in LLM-integrated Systems
hyhmia/BlindMI
houshd/TS_Adv
Codebase for the paper "Adversarial Attacks on Time Series"
XMCVE/polaris_ctf_collection
用于备份CTF比赛题目,仅队内复现使用
xiangyue9607/DP-Forward
niuliang42/CodexLeaks
CodexLeaks: Privacy Leaks from Code Generation Language Models in GitHub Copilot
YinHeng121/Adam_attack