I am a Doctoral Researcher with over seven years of experience in machine learning and computer vision, specializing in the pre-training and fine-tuning of Vision–Language Models (VLMs) and Multimodal Large Language Models (MLLMs) for fine-grained visual understanding. My research focuses on developing efficient multimodal representation learning frameworks that perform reliably under domain shift, limited labeled data, noisy inputs, and compute constraints. I have contributed to industry- and government-funded projects supported by the Department of Defense (DoD), FBI, Qualcomm, and the National Science Foundation (NSF), and have published in leading venues including WACV, BMVC, TAES, and TBIOM. My broader research interests include multimodal representation learning, foundation models, continual learning, and weakly- and self-supervised fine-tuning.
As I near the completion of my Ph.D., I am actively seeking opportunities for postdoctoral research and research scientist positions.