|
|
Sponsored by Colorado State University’s Information Science and Technology Center (ISTeC) In conjunction with the Department of Computer Science and Department of Electrical and Computer Engineering Seminar Series
Attacks and Defenses for Large Language Models on Coding Tasks
Speaker:
Corina Pasareanu, Principal Scientist (CMU CyLab), Technical Professional Leader -- Data Science (NASA Ames ⁄ KBR)
When: 10:00AM ~ 10:50, Tuesday October 8, 2024
Where: CSB 130
Abstract: Modern large language models (LLMs), such as ChatGPT, have demonstrated impressive capabilities for coding tasks, including writing and reasoning about code. They improve upon previous neural network models of code, such as code2seq or seq2seq, that already demonstrated competitive results when performing tasks such as code summarization and identifying code vulnerabilities. However, these previous code models were shown vulnerable to adversarial examples, i.e., small syntactic perturbations designed to “fool” the models. In this talk we discuss the transferability of adversarial examples, generated through white-box attacks on smaller code models, to LLMs. Further, we propose novel cost-effective techniques to defend LLMs against such adversaries via prompting, without incurring the cost of retraining. Our experiments show the effectiveness of the attacks and the proposed defenses on popular LLMs.
Bio: Corina Pasareanu is an ACM Fellow and an IEEE ASE Fellow, working at NASA Ames. She is affiliated with KBR and Carnegie Mellon University's CyLab. Her research interests include model checking, symbolic execution, compositional verification, probabilistic software analysis, autonomy, and security. She is the recipient of several awards, including ETAPS Test of Time Award (2021), ASE Most Influential Paper Award (2018), ESEC ⁄ FSE Test of Time Award (2018), ISSTA Retrospective Impact Paper Award (2018), ACM Impact Paper Award (2010), and ICSE 2010 Most Influential Paper Award (2010). She has been serving as Program ⁄ General Chair for several conferences including: ICSE 2025, SEFM 2021, FM 2021, ICST 2020, ISSTA 2020, ESEC ⁄ FSE 2018, CAV 2015, ISSTA 2014, ASE 2011, and NFM 2009. She is on the steering committees for the ICSE, TACAS and ISSTA conferences. She is currently an associate editor for IEEE TSE and for STTT, Springer Nature.
|