Price
Free
Event date and time
Wednesday 9 Jul 2025
4.00pm AEST
Location
Online virtual event
Login details will be emailed to registrants
Understanding and Mitigating Security Threats in Today's Machine Learning Systems
Speaker: Lea Schönherr
Pricing
-
Free
Dates and Times
Event date: Jul 2025
Wednesday 9 Jul 2025
Online virtual event
4.00pm AEST
Login details will be emailed to registrants
Contact
More information
Abstract:
Generative AI (genAI) is becoming more integrated into our daily lives, raising questions about potential threats within these systems and their outputs. In this talk, we will examine the security challenges and threats associated with generative AI. This includes the deception of humans with generated media and the deception of machine learning systems.
In the first part of the talk, we look at threat scenarios in which generative models are utilized to produce content that is impossible to distinguish from human-generated content. This fake content is often used for fraudulent and manipulative purposes. As generative models evolve, the attacks are easier to automate and require less expertise, while detecting such activities will become increasingly difficult. This talk will provide an overview of our current challenges in detecting fake media in human and machine interactions and the effects of genAI media labeling to consumers trust.
The second part will cover exploits of LLMs to disrupt alignment or to steal sensitive information. Existing attacks show that content filters of LLMs can be easily bypassed with specific inputs and that private information can be leaked. From an alternative perspective, we demonstrate that obfuscating prompts offers an effective way to protect intellectual property. Our research demonstrates that with minimal overhead, we can maintain similar utility while safeguarding confidential data, highlighting that defenses in foundation models may require fundamentally different approaches to utilize their inherent strengths.
Bio:
Lea Schönherr is a tenure-track faculty at CISPA Helmholtz Center for Information Security in Germany. Her research focuses on information security, particularly adversarial machine learning, trustworthy generative AI, and ML security applications. Her research focus on threat detection and defense of speech recognition systems, generative models, and on preventing the misuse of generative AI. She is interested in language as an interface to machine learning models, including their cognitive representations and code generation with LLMs. She obtained her PhD from Ruhr University Bochum, Germany, in 2021.