SCHOOL OF COMPUTER SCIENCE
The School of Computer Science is pleased to present…
Mitigating the Shortcomings of Language Models: Strategies For Handling Memorization & Adversarial Attacks
MSc Thesis Proposal by:
Aly M. Kassem
Date: Monday, July 3rd, 2023
Time: 10.30 AM - 11.30 AM
Location: Essex Hall 122
Reminders: 1. Two-part attendance mandatory (sign-in sheet, QR Code) 2. Arrive 5-10 minutes prior to event starting - LATECOMERS WILL NOT BE ADMITTED. Note that due to demand, if the room has reached capacity, even if you are "early" admission is not guaranteed. 3. Please be respectful of the presenter by NOT knocking on the door for admittance once the door has been closed whether the presentation has begun or not (If the room is at capacity, overflow is not permitted (ie. sitting on floors) as this is a violation of the Fire Safety code). 4. Be respectful of the decision of the advisor/host of the event if you are not given admittance. The School of Computer Science has numerous events occurring soon
Abstract:
Deep learning models have recently achieved remarkable progress in Natural Language Processing (NLP), specifically in classification, question-answering, and machine translation. However, NLP models face challenges related to performance and privacy. Performance-wise, even small perturbations in the input can significantly impact a model's prediction. This highlights the importance of generating natural adversarial attacks to analyze the weaknesses of NLP models and bolster their robustness through adversarial training (AT). Conversely, Large Language Models (LLMs) are trained on vast amounts of data, which may include sensitive information. If exposed, this poses a risk to personal privacy. LLMs have exhibited the ability to memorize portions of their training data and reproduce them verbatim when prompted by adversaries. To address these limitations, we propose an End-to-End framework. This framework employs a proximal policy gradient, a reinforcement learning approach, to learn a self-learned policy. The language model (LM) acts as a policy learner, enabling it to generate attacks in response to adversarial attacks and learn a "Dememorization Privacy Policy" to mitigate the risks associated with memorization. Our results show that our framework has proven effective in generating adversarial attacks and learning a policy to mitigate privacy risks in language models.
Thesis Committee:
Internal Reader: Dr. Luis Rueda
External Reader: Dr. Mitra Mirhassani
Advisor: Dr. Sherif Saad
MSc Thesis Proposal Announcement
