Fara-7B: An Efficient Agentic Model for Computer Use
Preprint and Research Preview of Open-Weight Model | Model Card | Blog Post
Research Scientist at Microsoft AI Frontiers
Visiting PostDoc / Collaborator,
Berkeley AI Research Lab (BAIR)
I'm interested in building generalist AI agents that can operate in the open world. Towards this goal, my research brings together elements of reinforcement learning, representation learning, and world models. Prior to my role at Microsoft, I spent 4 years at Meta (FAIR) working on Embodied AI with Jitendra Malik, Dhruv Batra, and Abhinav Gupta. In addition to my industry roles, I've also collaborated closely with the academic labs of Prof. Pieter Abbeel at UC Berkeley and Prof. Abhinav Gupta at CMU. I recieved my PhD in Computer Science from the University of Washington working with Profs. Sham Kakade and Emo Todorov. During this time, I also worked closely with Sergey Levine and Chelsea Finn, and spent time as a student researcher at Google Brain and OpenAI. Before that, I recieved my bachelors and the best undergraduate thesis award from IIT Madras working with Balaraman Ravindran.
Preprint and Research Preview of Open-Weight Model | Model Card | Blog Post
Computer Vision and Pattern Recognition (CVPR) 2024 | Project Website
International Conference on Machine Learning (ICML) 2025 | Project Website
Neural Information Processing Systems (NeurIPS) 2023 | Project Website
Neural Information Processing Systems (NeurIPS) 2021 | Project Website
Robotics: Science and Systems (RSS) 2018 | Project Website
International Conference on Machine Learning (ICML) 2023 | Project Website
eural Information Processing Systems (NeurIPS) 2020 | Project Website
I enjoy collaborating with a diverse set of students and researchers. I have had the pleasure of mentoring some highly motivated students at both the undergraduate and PhD levels.
I designed and co-taught a course on deep reinforcement learning at UW in Spring 2018. The course presents a rigorous mathematical treatment of various RL algorithms along with illustrative applications in robotics. Deep RL courses at UW, MIT, and CMU have borrowed and built upon the material I developed for this course.
This is an advanced graduate level course on machine learning with emphasis on machine learning at scale and distributed algorithms. Topics covered include hashing, sketching, streaming, large-scale distributed optimization, federated learning, and contextual bandits. I was the lead TA for this class.