Build AI-powered security tools. 50+ hands-on labs covering ML, LLMs, RAG, threat detection, DFIR, and red teaming. Includes Colab notebooks, Docker environment, and CTF challenges.
-
Updated
Feb 11, 2026 - Python
Build AI-powered security tools. 50+ hands-on labs covering ML, LLMs, RAG, threat detection, DFIR, and red teaming. Includes Colab notebooks, Docker environment, and CTF challenges.
Noise Injection Techniques provides a comprehensive exploration of methods to make machine learning models more robust to real-world bad data. This repository explains and demonstrates Gaussian noise, dropout, mixup, masking, adversarial noise, and label smoothing, with intuitive explanations, theory, and practical code examples.
Comprehensive taxonomy of AI security vulnerabilities, LLM adversarial attacks, prompt injection techniques, and machine learning security research. Covers 71+ attack vectors including model poisoning, agentic AI exploits, and privacy breaches.
An application to demonstrate stealing an AI model through knowledge distillation.
🤖 Test and secure AI systems with advanced techniques for Large Language Models, including jailbreaks and automated vulnerability scanners.
Reproducible security benchmarking for the Deconvolute SDK and AI system integrity against adversarial attacks.
Bug bounty report demonstrating prompt injection and command execution vulnerabilities in Meta AI's Instagram Group Chat
Complete 90-day learning path for AI security: ML fundamentals → LLM internals → AI threats → Detection engineering. Built from first principles with NumPy implementations, Jupyter notebooks, and production-ready detection systems.
Master's students in NCCU SoSLab maintaining a cleaned and restructured version of INCITE (based on PyCT).
🛡️ Discover and analyze critical vulnerabilities in Meta AI's Instagram Group Chat, ensuring robust security through comprehensive testing and reporting.
Adversarial ML Scanner for threat detection and ML backdoor attcaks
Adversarial Machine Learning Toolkit - Model extraction, adversarial examples, neural network probing, and defense evaluation in Julia
A curated list of awesome resources for AI system security.
Automatically generate YARA rules from adversarial and benign text samples. Built for detecting indirect prompt injection attacks on RAG pipelines.
Security Vulnerabilities and Defensive Mechanisms in CLI/Terminal-Based Large Language Model Deployments - A Comprehensive Research Synthesis (Technical Report, November 2025)
Hybrid Threat Intelligence Engine with Explainable AI (XAI) and Automated Triage.
Autonomous adversarial agents that debate and debug code before you see it.
A collection of resources documenting my research and learning journey in AI System Security.
An experiment in backdooring a shell safety classifier by planting a hidden trigger in its training data.
Fork for my contributions on Trojans, Comprehensive course materials covering ML safety, robustness, and AI alignment
Add a description, image, and links to the adversarial-ml topic page so that developers can more easily learn about it.
To associate your repository with the adversarial-ml topic, visit your repo's landing page and select "manage topics."