MPI-INF Logo
Campus Event Calendar

Event Entry

What and Who

How Do We Evaluate and Mitigate AI Risks?

Maksym Andriushchenko
EPFL
CIS@MPG Colloquium
hosted by: Christof Paar

Maksym Andriushchenko is a postdoctoral researcher at EPFL and an ELLIS Member. He has worked on AI safety with leading organizations in the field (OpenAI, Anthropic, UK AI Safety Institute, Center for AI Safety, Gray Swan AI). He obtained a PhD in machine learning from EPFL in 2024 advised by Prof. Nicolas Flammarion. His PhD thesis was awarded with the Patrick Denantes Memorial Prize for the best thesis in the CS department of EPFL and was supported by the Google and Open Phil AI PhD Fellowships. He did his MSc at Saarland University and the University of Tübingen, and interned at Adobe Research.
AG 1, AG 2, AG 3, INET, AG 4, AG 5, D6, RG1, MMCI  
AG Audience
English

Date, Time and Location

Tuesday, 11 February 2025
10:00
60 Minutes
MPI-SP
MB1SMMW106
Bochum

Abstract

AI has made remarkable progress in recent years, enabling groundbreaking applications but also raising serious safety concerns. This talk will explore the robustness challenges in deep learning and large language models (LLMs), demonstrating how seemingly minor perturbations can lead to critical failures. I will present my research on evaluating and mitigating AI risks, including adversarial robustness, LLM jailbreak vulnerabilities, and the broader implications of AI safety. By developing rigorous benchmarks, novel evaluation methods, and foundational theoretical insights, my work aims to provide effective safeguards for AI deployment. Ultimately, I advocate for a systematic approach to AI risk mitigation that integrates technical solutions with real-world considerations to ensure the safe and responsible use of AI systems.

Contact

Kateryna Panfilova
+49 681 9303 9106
--email hidden

Virtual Meeting Details

Zoom
611 8952 5821
passcode not visible
logged in users only

Kateryna Panfilova, 02/04/2025 10:12 -- Created document.