Confuse, Obfuscate, Disrupt: Using Adversarial Techniques for Better AI and True Anonymity
What if a single pixel could trick your AI into seeing a cat as a dog? Learn how adversarial attacks can expose hidden flaws and build more resilient systems.
#1about 1 minute
The importance of explainable AI and data quality
AI models are only as good as their training data, which is often plagued by bias, noise, and inaccuracies that explainable AI helps to uncover.
#2about 3 minutes
Identifying common data inconsistencies in AI models
Models can be compromised by issues like annotation errors, data imbalance, and adversarial samples, which can be measured with tools like Captum.
#3about 2 minutes
The dual purpose of adversarial AI attacks
Intentionally introducing adversarial inputs can be used for good to test model boundaries, or for bad to obfuscate data and protect personal privacy.
#4about 3 minutes
How to confuse NLP models with creative inputs
Natural language processing models can be disrupted using techniques like encoding, code-switching, misspellings, and even metaphors to prevent accurate interpretation.
#5about 4 minutes
Visualizing model predictions with the Captum library
The Captum library for PyTorch helps visualize which parts of an input, like words in a sentence or pixels in an image, contribute most to a model's final prediction.
#6about 6 minutes
Manipulating model outputs with subtle input changes
Simple misspellings can flip a sentiment analysis result from positive to negative, and adding a single pixel can cause an image classifier to misidentify a cat as a dog.
#7about 2 minutes
Using an adversarial pattern t-shirt to evade detection
A t-shirt printed with a specific adversarial pattern can disrupt a real-time person detection model, effectively making the wearer invisible to the AI system.
#8about 2 minutes
Techniques for defending models against adversarial attacks
Defenses against NLP attacks include normalization and grammar checks, while vision attacks can be mitigated with image blurring, bit-depth reduction, or advanced methods like FGSM.
#9about 2 minutes
Defeating a single-pixel attack with image blurring
Applying a simple Gaussian blur to an image containing an adversarial pixel smooths out the manipulation, allowing the model to correctly classify the image.
Related jobs
Jobs that call for the skills explored in this talk.
Matching moments
03:05 MIN
Understanding security risks from adversarial attacks on models
Explainable machine learning explained
04:38 MIN
Fundamental AI vulnerabilities and malicious misuse
A hundred ways to wreck your AI - the (in)security of machine learning systems
06:27 MIN
Deconstructing AI attacks from evasion to model stealing
A hundred ways to wreck your AI - the (in)security of machine learning systems
09:15 MIN
Navigating the new landscape of AI and cybersecurity
From Monolith Tinkering to Modern Software Development
02:03 MIN
Q&A on creating patterns and de-poisoning images
Hacking AI - how attackers impose their will on AI
02:28 MIN
Understanding the core principles of hacking AI systems
Hacking AI - how attackers impose their will on AI
03:43 MIN
AI privacy concerns and prompt engineering
Coffee with Developers - Cassidy Williams -
05:17 MIN
Manipulating AI with prompt injection and hidden commands
WeAreDevelopers LIVE - Is Software Ever Truly Accessible?
Panel Discussion: Responsible AI in Practice - Real-World Examples and ChallengesIntroductionIn the ever-evolving landscape of artificial intelligence, the concept of "responsible AI" has emerged as a cornerstone for ethical and practical AI implementation. During the WWC24 Panel discussion, three eminent experts—Mina, Bjorn Brin...
Chris Heilmann
WWC24 Talk - Scott Hanselman - AI: Superhero or Supervillain?Join Scott Hanselman at WWC24 to explore AI's role as a superhero or supervillain. Scott shares his 32 years of experience in software engineering, discusses AI myths, ethical dilemmas, and tech advancements. Engage with his live demos and insights o...
Chris Heilmann
Exploring AI: Opportunities and Risks for DevelopersIn today's rapidly evolving tech landscape, the integration of Artificial Intelligence (AI) in development presents both exciting opportunities and notable risks. This dynamic was the focus of a recent panel discussion featuring industry experts Kent...
Chris Heilmann
Dev Digest 116 - WWWAI?This time, learn how to un-AI Google's search results, what's new on the web, avoid a new security hole and go back to BASICS with us. News and ArticlesWhat a week. Google, Microsoft, OpenAI and many others had their big flagship events announcing th...
From learning to earning
Jobs that call for the skills explored in this talk.