Tsung-Huan Yang

About Me

I'm an incoming MSCS student at Georgia Tech. My research centers on Trustworthy AI, with the goal of building safe, interpretable, and reliable AI systems. I'm currently exploring topics such as mechanistic interpretability, automated/multimodal red-teaming, and deceptive alignment. I join Georgia Tech after a two-year RAship at Academia Sinica, where I worked with Dr. Lun-Wei Ku on building culture-aware safety classifiers and generating adversarial attacks for LLM safety/security assessment. Before that, I've worked on speech processing under Prof. Hung-yi Lee and model compression with Prof. Hao Tang.

News

  • [August 2025] I'll be studying MSCS at Georgia Tech this fall!
  • [June 2025] I have concluded my two-year RAship at Academia Sinica. It was an incredible period of growth where I learned invaluable research techniques and delved deep into advanced AI. I am excited to begin the next chapter
  • [Decmber 2024] I will present my poster at NeurIPS Safe Generative AI Workshop. Looking forward to engaging with the community there.

Publications

thumbnail
PrePrint

PREP-LLM: Profile Refinement for Eliciting Personalities in LLMs

Shi-Wei Dai, Yan-Wei Shie, Tsung-Huan Yang, Lun-Wei Ku, and Yung-Hui Li

thumbnail
NeurIPS workshop 2024

Preserving Safety in Fine-Tuned Large Language Models: A Systematic Evaluation and Mitigation Strategy

Tsung-Huan Yang, Ko-Wei Huang, Yung-Hui Li, and Lun-Wei Ku.

thumbnail
PrePrint

Is Smaller Always Faster? Tradeoffs in Compressing Self-Supervised Speech Transformers

Tzu-Quan Lin, Tsung-Huan Yang, Chun-Yao Chang, Kuang-Ming Chen, Tzu-hsun Feng, Hung-yi Lee, and Hao Tang.