I am a research scientist at Bosch Research (Sunnyvale, USA), where I work on computer vision problems for autonomous driving. My research interests are model interpretability and the “science” of deep learning, i.e., systematic investigations of deep learning phenomena.
I started my research career with a masters at the Indian Institute of Science, Bangalore with Prof. Venkatesh Babu. I completed my PhD with François Fleuret, at Idiap Research Institute & EPFL, Switzerland. Enamoured by model interpretability, I pursued a postdoctoral research fellowship with Hima Lakkaraju at Harvard University.
Some representative papers include:
splice: shows that CLIP representations are approximate sparse linear combinations of concept vectors, using this to develop an interp tool that extracts concepts encoded by representationsexplaining perceptually aligned gradients: an experimental theory explaining why noise-robust models tend to produce improved (perceptually aligned) gradient saliency mapsfullgrad saliency: shows that ReLU neural net outputs can be exactly decomposed into layer-wise gradient terms, using this to develop a saliency tool that aggregates layer-wise gradientsforgetting data contamination in LLMs: an experimental study showing that data contamination in LLMs may be irrelevant in certain practical training scenarios
For more information, please see my research themes and publications.
Note: If you’d like to chat about the science or interpretability of neural nets, feel free to reach out!