OpenEQA: Embodied Question Answering in the Era of Foundation Models
Computer Vision and Pattern Recognition (CVPR) 2024 | Project Website
Research Scientist at Meta AI (FAIR)
Visiting PostDoc / Collaborator,
Berkeley AI Research Lab (BAIR)
I'm interested in building generalist AI agents that can operate in the open world. Towards this goal, my research brings together elements of reinforcement learning, representation learning, and world models. In addition to my role at FAIR, I also collaborate with the academic research labs of Prof. Pieter Abbeel at UC Berkeley and Prof. Abhinav Gupta at CMU. I recieved my PhD in Computer Science from the University of Washington working with Profs. Sham Kakade and Emo Todorov. During this time, I also worked closely with Sergey Levine and Chelsea Finn, and spent time as a student researcher at Google Brain and OpenAI. Before that, I recieved my bachelors and the best undergraduate thesis award from IIT Madras working with Balaraman Ravindran.
Computer Vision and Pattern Recognition (CVPR) 2024 | Project Website
Scaling Robot Learning Workshop at ICRA 2022 | Best Paper Award | Project Website
Neural Information Processing Systems (NeurIPS) 2023 | Project Website
Neural Information Processing Systems (NeurIPS) 2021 | Project Website
Robotics: Science and Systems (RSS) 2018 | Project Website
International Conference on Machine Learning (ICML) 2023 | Project Website
I enjoy collaborating with a diverse set of students and researchers. I have had the pleasure of mentoring some highly motivated students at both the undergraduate and PhD levels.
I designed and co-taught a course on deep reinforcement learning at UW in Spring 2018. The course presents a rigorous mathematical treatment of various RL algorithms along with illustrative applications in robotics. Deep RL courses at UW, MIT, and CMU have borrowed and built upon the material I developed for this course.
This is an advanced graduate level course on machine learning with emphasis on machine learning at scale and distributed algorithms. Topics covered include hashing, sketching, streaming, large-scale distributed optimization, federated learning, and contextual bandits. I was the lead TA for this class.