PyData Global 2023

Tricking Neural Networks : Explore Adversarial Attacks
12-07, 19:30–20:00 (UTC), Machine Learning Track

Large Language Models are pretty cool, but we need to be aware of how they can be compromised.
I will show how neural networks are vulnerable to attacks through an example of an adversarial attack on deep learning models in Natural Language Processing(NLP).
We’ll explore the mechanisms used to attack models, and you’ll get a new way to think about the security of deep learning models.
An understanding of deep learning is required.


With increasing adoption of deep learning models such as Large Language Models(LLMs) in real-world applications, we should consider security and safety of the models. To address the security concerns we need to understand the model's vulnerabilities and how they can be compromised. After all, it is hard to defend yourself when you don't know you are under attack.

You will gain the most out of this session if you have worked with deep learning models before. Expect material on gradient-based attacks and why they are a concern.


Prior Knowledge Expected

Previous knowledge expected