skip to content
šŸˆā€ā¬› Andrew Zhang

Hi, I’m Andrew

I currently work on accelerating inference on AWS Trainium devices. I did a 4 year BS/MS at Virginia Tech under Dr. Chris Thomas. I am interested in multi-token generation (discrete diffusion & speculative decoding) and hardware-aware algorithms.

You can find me on LinkedIn, GitHub, Twitter, and Google Scholar

Publications

[1] Andrew Zhang, Anushka Sivakumar, Chia-Wei Tang, and Chris Thomas. Flexible-length text infilling for discrete diffusion models. In: Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing (EMNLP). 2025.


[2] Anushka Sivakumar, Andrew Zhang, Zaber Ibn Abdul Hakim, and Chris Thomas. SteerVLM: Robust Model Control through Lightweight Activation Steering for Vision Language Models. In: Findings of the Association for Computational Linguistics (EMNLP Findings). 2025.


[3] Md. Atabuzzaman, Andrew Zhang, and Chris Thomas. Zero-Shot Fine-Grained Image Classification Using Large Vision-Language Models. In: Findings of the Association for Computational Linguistics (EMNLP Findings). 2025.


[4] Hani Alomari, Anushka Sivakumar, Andrew Zhang, and Chris Thomas. Maximal Matching Matters: Preventing Representation Collapse for Robust Cross-Modal Retrieval. In: Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (ACL). 2025.

Get In Touch

Please email me at azhang42 [at] vt [dot] edu or DM me on Twitter