Preprints

IF-Guide: Influence Function-Guided Detoxification of LLMs

Zachary Coalson, Juhan Bae, Nicholas Carlini, and Sanghyun Hong.

arXiv preprint, 2025.

Jailbreaking Large Language Models with Fewer Than Twenty-Five Targeted Bit-flips

Zachary Coalson, Jeonghyun Woo, Shiyang Chen, Yu Sun, Lishan Yang, Prashant Nair, Bo Fang, and Sanghyun Hong.

arXiv preprint, 2024.

Hard Work Does Not Always Pay Off: Poisoning Attacks on Neural Architecture Search

Zachary Coalson, Huazheng Wang, Qingyun Wu, and Sanghyun Hong.

arXiv preprint, 2024.

2025

Demystifying the Resilience of Large Language Model Inference: An End-to-End Perspective

Yu Sun, Zachary Coalson, Shiyang Chen, Hang Liu, Sanghyun Hong, Zhao Zhang, Bo Fang, and Lishan Yang.

SC, 2025.

Harnessing Input-Adaptive Inference for Efficient VLN

Dongwoo Kang, Akhil Perincherry, Zachary Coalson, Aiden Gabriel, Stefan Lee, and Sanghyun Hong.

ICCV, 2025.

2023

BERT Lost Patience Won't Be Robust to Adversarial Slowdown

Zachary Coalson, Gabriel Ritter, Rakesh Bobba, and Sanghyun Hong.

NeurIPS, 2023.