I am a Ph.D. student in computer science at the University of Massachusetts Amherst, advised by Amir Houmansadr.

My research focuses on exploring the privacy, security, and trustworthiness of AI models. Recently, I have been investigating the reliability of multimodal models, while also maintaining a strong interest in related areas such as fairness, interpretability, and responsible AI. My current work delves into assessing the trustworthiness of multimodal generative models across diverse domains, including text-to-image and audio-based modalities. I am currently doing a Summer Internship at Brave Software working on privacy and security of AI agents with Ali Shahin Shamsabadi.

Prior to my graduate studies, I earned my bachelor’s degree in computer engineering from the Hong Kong University of Science and Technology (HKUST) in the year 2023, where I completed my Final Year Thesis (FYT) on the topic of “Adversarial Attacks in Federated Learning” under the supervision of Jun Zhang. I have also worked with Minhao Cheng on the robustness of language models, specifically exploring methods associated with backdoor defense in text domain.

💬 Office Hours

I’m hosting weekly office hours—feel free to drop by! I’m happy to chat and advise on research (or projects), PhD applications, or anything else on your mind. Casual chats welcome too (I’m a big football (soccer) fan ⚽️). Lately, I’ve been working on trustworthiness of AI agents and audio modality safety, but I’m always open to exploring new areas and directions you might bring.

Feel free to book a time that works for you through my Calendly.

📣 News

🎙️ July 11 '25: I will be giving a talk at the NFM Reading Group led by the Speech Technologies Group at Google Deemind on our Multilingual and Multi-Accent Jailbreaking of Audio LLMs paper. View [slides]

🎉 July 7 '25: Our Multilingual and Multi-Accent Jailbreaking of Audio LLMs paper has been accepted to COLM (Conference on Language Modeling) 2025!

💪 Jun 16 '25: I will be working as a Summer Research Intern at Brave Software under the supervision of Dr. Ali Shahin Shamsabadi on privacy and security of AI Agents.

🎉 Jun 11 '25: Our Backdooring Bias into Text-to-Image Models paper has been accepted to USENIX Security '25!

🎉 Sep 27 '24: Our OSLO paper has been accepted to NeurIPS '24!

🎉 Dec 22 '23: Our Memory Triggers paper has been accepted to AAAI '23 PPAI Workshop!

🖨️ Preprint / Publications

Preprint

R1dacted
Ali Naseh, Harsh Chaudhari, Jaechul Roh, Mingshi Wu, Alina Oprea, Amir Houmansadr
arXiv
OverThink
Abhinav Kumar, Jaechul Roh, Ali Naseh, Marezna Karpinska, Mohit Iyyer, Amir Houmansadr, and Eugene Bagdasaryan
arXiv
FameBias
Jaechul Roh, Andrew Yuan, Jinsong Mao
arXiv

2025

Multilingual Audio Jailbreaking
Jaechul Roh, Virat Shejwalkar, Amir Houmansadr
COLM 2025
Backdooring Bias
Ali Naseh, Jaechul Roh, Eugene Bagdasaryan, Amir Houmansadr
USENIX Security '25

2024

OSLO
Yuefeng Peng, Jaechul Roh, Subhransu Maji, Amir Houmansadr
NeurIPS 2024
Memory Triggers
Ali Naseh, Jaechul Roh, Amir Houmansadr
The 5th PPAI (AAAI Workshop)

2023

Robust Smart Home
Jaechul Roh, Yajun Fang
IEEE UV 2022 (Oral Presentation)
MSDT
Jaechul Roh, Minhao Cheng, Yajun Fang
IEEE UV 2022 (Oral Presentation)