About Me
Hello! I am a final-year Masters Student in the CSE Department at Michigan State University pursuing my master’s degree under the supervision of Dr. Vishnu Boddeti. I am extremely lucky to have amazing colleagues at Human Analysis Lab. I hold a bachelor’s degree from IIT, Guwahati.
🎯 Research Objective
My primary research goal is to improve the generalization of machine learning systems. This can be summarized by the phrase, “You get what you train for.” An incorrect objective will lead to poor generalization, even when a large amount of data is applied to a problem. My research focuses on defining the correct objectives, thereby improving the ability of machine learning models to generalize effectively.
For instance, current state-of-the-art generative models often fail to generate an image of “a red polar bear” or edit the color of a polar bear in an image without changing other aspects. Why do current SOTA models struggle? The objective of diffusion models, as discussed in my blog on their probabilistic perspective, is to learn the probability of the observed data distribution. However, in the observational data, these seemingly independent concepts of color and animal structure are rather dependent. Therefore, if you are modeling the incorrect observational data, you are bound to get a model that is incorrect. In our ICLR 2025 submission (under review), we correct the training objective of diffusion models to reflect the independence that allows us to modify one attribute without affecting the others.
Classical machine learning algorithms are also prone to generalization problems. Models such as SVM and LDA rely on spurious features for classification. In this work, we provide a modified objective that is invariant to spurious features. This objective transforms the problem into a generalized eigenvalue problem with a closed-form solution.
Algorithmic fairness is another area where an appropriate training objective should include a constraint on fairness, ensuring that models do not base decisions on sensitive attributes like gender. I have studied constrained optimization, and this work resulted in my research notes on Constrained Optimization and Min-Max Problems. Additionally, I proposed a pre-processing adversarial pipeline to improve fairness in LLMs. On a fun side, classic dynamic problems can be formulated as constrained optimization, and projected gradient descent can get you an approximate solution faster.
I also care about the impact of ML and have worked on practical problems such as how to train with large-scale multi-modal data with imbalance. Also in P-GPT, I strive to find the efficient configuration for training across multiple servers.
💀 Projects that did not materialize
I used to work on Adversarial Machine Learning. I’m no longer interested in this area, although the optimization techniques are useful; it’s always a catch-up game. However, I have written a blog post on How to solve an Adversarial Optimization Problem?. Also, Machine Unlearning: This involves removing training data from the model. I’m no longer interested in unlearning as I struggle to understand a metric to quantify it. However, we participated in the NeurIPS 2023 challenge on Machine Unlearning
☕ Apart from Research
🧑🏫 I teach programming (CSE 232) at Michigan State University.
⛳ I enjoy running on a rare sunny day in East Lansing, MI. On a gloomy day, I sit down with a sip of coffee ☕ and Leetcode.
💡 I love building products. Currently working on an image editing tool.
Collaborations
If you're interested in my research, I'm always open to collaborations!
Publications
S. Gaudi, G. Sreekumar, V. Boddeti, CoInD: Enabling Logical Compositions in Diffusion Models, ICLR 2025 (Under review)