Research themes

My research focusses on understanding deep learning from an empirical and scientific perspective, aiming to derive actionable insights that can improve its practical application. Major themes include:


:mag:   model interpretability via representation analysis

Deep learning works by transforming inputs to latent representations. Can we understand and describe the information stored in these latent representations?


:heavy_dollar_sign:   data-centric machine learning

Behaviour of machine learning models, including large language models (LLMs) depend critically on the datasets used to train them. Can we identify critical datapoints / data subsets that influence key model properties?


:bar_chart:   model interpretability via feature attribution

Machine learning classifiers often work by “looking at” a small number of important features present in the inputs. For example, a cat classifier only needs to identify “cat-like” features in the image, while ignoring all other features. How can we formalize this question of identifying important or salient features, and develop efficient algorithms to extract them?


:muscle:   alternate notions of model robustness

Training adversarially robust models is hard. Are there alternate definitions of robustness that are both practically meaningful, yet easier to train for?


:recycle:   computational efficiency of deep models

Given a pre-trained deep model, how can we effectively identify and eliminate redundant neurons or weights while maintaining the model’s performance?


Apart from these broad themes, here are some cool miscellaneous projects I have worked on: