wellido's Stars
FudanSELab/Agent4SE-Paper-List
Repository for the paper "Large Language Model-Based Agents for Software Engineering: A Survey". Keep updating.
ZJU-ACES-ISE/ChatUniTest
eth-sri/llm-quantization-attack
lafeat/apbench
APBench: A Unified Availability Poisoning Attack and Defenses Benchmark (TMLR 08/2024)
eth-sri/SafeCoder
SunflowerPKU/ICSE22_SC_Data
wagner-group/active-learning
Continuous Learning for Android Malware Detection (USENIX Security 2023)
null1024-ws/Poisoning-Attack-on-Code-Completion-Models
USENIX Security'24 Paper Repo
ultralytics/ultralytics
Ultralytics YOLO11 🚀
google-deepmind/icml2024-roundtrip-correctness
All-Hands-AI/OpenHands
🙌 OpenHands: Code Less, Make More
Jingkang50/OpenOOD
Benchmarking Generalized Out-of-Distribution Detection
EachSheep/ShortcutsBench
ShortcutsBench: A Large-Scale Real-World Benchmark for API-Based Agents
OpenAutoCoder/Agentless
Agentless🐱: an agentless approach to automatically solve software development problems
testingautomated-usi/selforacle
The code of our paper "Misbehaviour Prediction for Autonomous Driving Systems", including our improved Udacity simulator
ast-fortiss-tum/misbehaviour-prediction-with-uncertainty-quantification
Codebase of the MSc thesis by Ruben Grewal "Uncertainty Quantification for Failure Prediction in Autonomous Driving Systems" and replication package of the paper "Predicting Safety Misbehaviours in Autonomous Driving Systems using Uncertainty Quantification" (ICST 2024).
microsoft/rho
Repo for Rho-1: Token-level Data Selection & Selective Pretraining of LLMs.
bigcode-project/bigcodebench
BigCodeBench: Benchmarking Code Generation Towards AGI
MCEVAL/McEval
smartyfh/LLM-Uncertainty-Bench
Benchmarking LLMs via Uncertainty Quantification
THU-MIG/torch-model-compression
针对pytorch模型的自动化模型结构分析和修改工具集,包含自动分析模型结构的模型压缩算法库
xingjianleng/autoeval_baselines
This repository includes various baseline techniques for label-free model evaluation task for the VDU2023 competition.
berkerpeksag/astor
Python AST read/write
princeton-nlp/SWE-agent
[NeurIPS 2024] SWE-agent takes a GitHub issue and tries to automatically fix it, using GPT-4, or your LM of choice. It can also be employed for offensive cybersecurity or competitive coding challenges.
princeton-nlp/SWE-bench
[ICLR 2024] SWE-bench: Can Language Models Resolve Real-world Github Issues?
hzysvilla/Academic_LLM_Sec_Papers
Academic Papers about LLM Application on Security
princeton-nlp/LESS
[ICML 2024] LESS: Selecting Influential Data for Targeted Instruction Tuning
ibm-granite/granite-code-models
Granite Code Models: A Family of Open Foundation Models for Code Intelligence
KindXiaoming/pykan
Kolmogorov Arnold Networks
kgori/tree_distance
Implementation of geodesic, Robinson-Foulds and Euclidean tree distance metrics