My research interests are in nonlinear optimization for machine learning. Below are some of the projects that I have been involved with my collaborators in industry and academia.
- McTorch. Currently working on the Python toolbox McTorch, a manifold optimization library for deep learning.
- Manopt. I am also involved in the development and promotion of the Matlab toolbox Manopt for optimization on manifolds.
- Pymanopt. Joined Pymanopt team as a maintainer.
- Microsoft GitHub Reommenders. Contributed a manifold-based algorithm for matrix completion.
- Auto-tagging at scale [Microsoft]. Working with the SharePoint India team on exploring ML techniques for auto-tagging of SharePoint documents. On the research side, we are currently looking at how to exploit the rich taxonomy structure of the tags.
- Grouping of Exchange commands [Microsoft]. Worked with the Exchange India team on understanding the usage of Exchange commands and whether it makes sense to identify groups of commands frequently used together.
- Kaizala [Microsoft]. Working with multiple teams to integrate data-driven intelligent features into the Kaizala and Office Lens apps. The use cases that we tackle revolve around improving user experience on the apps by leveraging the latest machine learning tools. An overview and a short video on the current focus of our team. Recently, we have pushed a couple of features to Kaizala app for internal dogfooding (now available in Kaizala beta).
- Multilingual embedding [Microsoft]. The aim is to generate vector embeddings for words in the target language that are learned from a small set of dictionary mappings between the source and target languages. The use cases include being able to leverage tools developed in the English language to solve similar problems in a different language with little training data. Currently, we are looking at unsupervised approaches.
- Product recommendation [Microsoft]. Worked on developing algorithms for an external Microsoft customer to enable ML driven product recommendations. As a side outcome, we contributed the RLRMC algorithm to the official Microsoft Recommenders GitHub repository.
- Competitive pricing of products [Amazon.com]. Worked on the estimation of the competitive prices of 3p unique products, i.e., Amazon is not a retail player for these products. The large-scale nature of the project demanded to work on simple algorithmic implementations that are scalable and simple to maintain. The proposed models were productionized and used to generate price estimates worldwide for all of the Amazon marketplaces and consumed by multiple customers. (Patent filed.)
- Style recommendation for fashion [Amazon.com]. Worked on cold-start product recommendation by exploiting user and product attributes. Our algorithms have been pushed to internal customers. The online results have shown significant lift of performance over other baselines. The proposed algorithms are in production.
- Detecting anomalies in e-commerce [Amazon.com]. We exploit the problem structure to propose natural gradient learning algorithms for tensor decomposition. The proposed algorithm is in production to detect anomalies in seller and reviewer activities.
- Demand forecasting [Amazon.com]. As supervisor of an internship project, we proposed low-rank tensor factorization prototype models for demand forecasting problems.
- Optimization on manifolds. Understanding the (Riemannian) geometry of structured constraints is of particular interest in machine learning. Conceptually, it allows translating a constrained optimization problem into an unconstrained optimization problem on a nonlinear search space (manifold). Building upon this point of view, one research interest is to exploit manifold geometry in nonlinear optimization.
We have been actively involved in both matrix and tensor applications. We have also worked on domain adaptation mostly on ML side.
- Geometry-aware cross-lingual word embeddings and related applications. Recently, we have been looking into generating multilingual embeddings, in both supervised and unsupervised ways.
- Decentralized and stochastic optimization algorithms. We explore recent advances in stochastic gradient algorithms on manifolds.
- Deep learning. We study geometric algorithms for modern deep networks that are robust to invariances of the parameters. On a conceptual level, our interest is to understand the link between learning a “complex architecture with simple algorithms” and learning a “simple architecture with sophisticated algorithms”.