
Applied Scientist at Adobe
Applied Scientist at Adobe specializing in cutting-edge machine learning solutions. I hold an M.S. in Computer Science from Georgia Tech with a focus on Machine Learning. Passionate about training Large Language Models and Vision-Language Models, I thrive on solving real-world challenges that drive meaningful impact.
M.S. in Computer Science
Georgia Institute of Technology
Machine Learning
B.E. in Computer Science
Birla Institute of Technology and Sciences, Pilani
M.Sc. in Mathematics
Birla Institute of Technology and Sciences, Pilani
June 2021 — Present
July 2016 — August 2019
Built a complete MLOps system for training vision-language models to assess marketing content quality, featuring novel observable injection and Photoshop edit evaluation capabilities.
Created the first demo and pitched to Adobe leadership, leading to the launch of a generative AI-first platform that revolutionizes marketing content creation at scale.
Vanya Bannihatti Kumar, Divyanshu Goyal, Akhil Eppa, Neel Bhandari
Yuanyou Wang, Naveen Vangala, Mayank Anand, Kunal Kumar Jain, Jose Mathew, Eapen Jose, Divyanshu Goyal, Asmita Chihnara, Arif Abdullah, Anand Dantu
Mayank Anand, Jose Mathew, Divyanshu Goyal
Akhil Eppa, Mayank Anand, Divyanshu Goyal
Divyanshu Goyal, Akhil Eppa, Mayank Anand
How to compress transformer configuration into a single integer — depth. By deriving architecture dimensions (n_layer, n_embed, n_head), batch size, learning rate, weight decay, and training horizon from one number, scaling sweeps across multiple model sizes become trivial and error-free.
Six concrete optimizations — TF32, BF16 mixed precision, torch.compile, Flash Attention, parallel DataLoaders, and pinned memory transfers — that collectively drove a significant jump in Model Flop Utilization (MFU) while training VibeNanoChat, a GPT-2 scale LLM. Each technique is explained with real code and the hardware-level reasoning behind it.
A deep dive into building Adobe's User Response Prediction Service that processes 3,500 requests per second with sub-5ms latency. Covers code optimization, garbage collection tuning, event generation, and performance benchmarking techniques for high-performance ML services.