|
Vaibhavi Singh
I'm an MS student in Computer Science at NYU Courant, specializing in machine learning with a focus on multimodal models & representation learning. I'm particularly interested in understanding how vision-language models reason & how to train effective representations with limited data.
Currently, I'm conducting research on reasoning in multimodal LLMs, & investigating self-supervised learning in low-resource regimes.
Previously, I worked as an ML Engineer building clinical risk prediction models, & spent several years at Adobe & Salesforce optimizing ML systems & large-scale graphics pipelines.
Email /
LinkedIn /
GitHub /
Twitter /
Writings
|
|
Research
I'm interested in agentic reasoning & representation learning. My current work focuses on evaluating reasoning capabilities in vision-language models & efficient self-supervised learning under resource constraints.
|
|
Studying Reasoning in Multimodal LLMs
Evaluating the limitations of current multimodal models on structured reasoning tasks, with a focus on evaluation challenges & qualitative failure analysis.
|
|
Self-Supervised Vision in Low-Resource Regimes
Investigating self-supervised representation learning under severe data, resolution (96×96), & compute constraints by training DINOv1 from scratch. Observed strong performance on CUB-200, miniImageNet, & SUN397, & explored the roles of domain-aligned data, tokenization density, & evaluation methodology in constrained regimes.
|
|
NeurIPS 2025
Ethics Reviewer — Datasets & Benchmarks Track
Technical Reviewer — Workshops - (UniReps) Unifying Representations in Neural Models, (ML4PS) Machine Learning and the Physical Sciences
|
Skills
Research & Modeling: Multimodal LLMs, Computer Vision, Self-Supervised Learning, Benchmarking & Evaluation, Diffusion Models
System & Infrastructure: C++, Docker, AWS/GCP, Linux, Git, Vulkan
Tools & Libraries: HuggingFace (Transformers), WandB, OpenCV, PIL, NumPy/SciPy
|
Machine Learning Engineer
Healthcare AI Startup, India
2024 – 2025
Built clinical risk prediction models (XGBoost, TCN) achieving 0.87 F1-score through feature engineering, SMOTE for class imbalance, & hyperparameter optimization. Processed sparse EMR data for early-stage healthcare applications.
|
Software Engineer II
Salesforce, India
2023 – 2024
Optimized petabyte-scale data ingestion pipelines, reducing latency by 30% for Einstein AI & real-time analytics. Scaled multi-tenant Kubernetes infrastructure on AWS for 200+ microservices.
|
ML Systems Engineer (MTS II)
Adobe, India
2021 – 2023
Implemented heterogeneous compute (CPU/GPU) optimizations in graphics libraries for Photoshop Neural Filters, enabling real-time HDR style transfer on Intel & Apple Silicon for 20M+ users. Developed graphics shaders for the Vulkan rendering backend for Substance 3D in Color Engine, enabling high-fidelity 3D workflows on Linux & macOS. Implemented multithreaded parallel processing in media libraries powering Premiere Pro & led performance profiling efforts to enforce strict latency SLAs across the Creative Cloud suite.
|
Cloud Infrastructure Engineer (MTS I)
Adobe, India
2019 – 2021
Optimized high-throughput Java microservices on Adobe Cloud Platform, reducing build times by 25% & infrastructure costs by 12% for 10M+ DAU through performance profiling & resource tuning.
|
M.S. in Computer Science (Machine Learning)
New York University, Courant Institute
2025 – 2027 (expected)
GPA: 3.89/4.00
Research focus: agentic reasoning & representation learning
Coursework: Deep Learning (Yann LeCun), Natural Language Processing (Eunsol Choi), Computer Vision (Saining Xie)
|
B.E. Computer Engineering (Hons)
Netaji Subhas Institute of Technology, University of Delhi
2015 – 2019
First Class with Distinction
Graduated in the top 10% of the department
Recipient of EPFL-Swiss Government scholarship (Scala Days 2019)
Google Summer of Code Mentor, Anita Borg Institute
|
Last updated: January 30, 2026
Website template from Jon Barron
|
|