Nura Kawa
Machine Learning: Promising, but Perilous
#1about 5 minutes
The dual nature of machine learning's power
Machine learning's increasing power and accessibility, exemplified by complex tasks like panoptic segmentation, also introduces significant security vulnerabilities.
#2about 3 minutes
Accelerating development with transfer learning
Transfer learning allows developers to repurpose large pre-trained teacher models for specific tasks with minimal data and compute by fine-tuning a new student model.
#3about 2 minutes
How transfer learning's benefits create security risks
The core benefits of transfer learning, such as knowledge transfer and minimal training, directly create attack vectors for adversaries.
#4about 5 minutes
Exploring evasion and poisoning attacks in ML
Adversarial examples can fool models with subtle input changes (evasion), while poisoned data can insert hidden backdoors, with both risks amplified by transfer learning.
#5about 3 minutes
Integrating security with a pre-development risk assessment
Before writing code, perform a thorough risk assessment by defining security requirements, evaluating resource availability, and conducting threat modeling for your specific use case.
#6about 3 minutes
Selecting robust teacher models for secure transfer learning
Mitigate risks by choosing transparent and trustworthy teacher models and using robust models hardened through techniques like adversarial training.
#7about 1 minute
Fortifying student models to prevent transferred attacks
Strengthen your student model by fine-tuning all layers to diverge from the teacher model, using backdoor detection, and performing continuous stress testing.
#8about 2 minutes
Key resources for developing secure ML systems
Practical resources like the Adversarial Robustness Toolbox for developers and security principles from the National Cybersecurity Center can help you build more secure ML systems.
Related jobs
Jobs that call for the skills explored in this talk.
Matching moments
20:06 MIN
New security vulnerabilities and monitoring for AI systems
The State of GenAI & Machine Learning in 2025
09:22 MIN
Understanding the fundamental security risks in AI models
Can Machines Dream of Secure Code? Emerging AI Security Risks in LLM-driven Developer Tools
21:41 MIN
Challenge three: Ensuring machine learning models are robust
How Machine Learning is turning the Automotive Industry upside down
41:34 MIN
Understanding security risks from adversarial attacks on models
Explainable machine learning explained
10:01 MIN
Navigating the new landscape of AI and cybersecurity
From Monolith Tinkering to Modern Software Development
36:30 MIN
The rise of MLOps and AI security considerations
MLOps and AI Driven Development
00:03 MIN
The rapid adoption of LLMs outpaces security practices
ChatGPT, ignore the above instructions! Prompt injection attacks and how to avoid them.
13:54 MIN
The ethical risks of outdated and insecure AI models
AI & Ethics
Featured Partners
Related Videos
A hundred ways to wreck your AI - the (in)security of machine learning systems
Balázs Kiss
Hacking AI - how attackers impose their will on AI
Mirko Ross
Beyond the Hype: Building Trustworthy and Reliable LLM Applications with Guardrails
Alex Soto
The AI Security Survival Guide: Practical Advice for Stressed-Out Developers
Mackenzie Jackson
Prompt Injection, Poisoning & More: The Dark Side of LLMs
Keno Dreßel
What non-automotive Machine Learning projects can learn from automotive Machine Learning projects
Jan Zawadzki
How AI Models Get Smarter
Ankit Patel
Manipulating The Machine: Prompt Injections And Counter Measures
Georg Dresler
Related Articles
View all articles.gif?w=240&auto=compress,format)



From learning to earning
Jobs that call for the skills explored in this talk.

Machine Learning (ML) Engineer Maitrisant - Python / Github Action
ASFOTEC
Intermediate
NoSQL
DevOps
Pandas
Docker
MongoDB
+8







