April 29, 2024 @2pm — Audio-Visual Scene Understanding and AR/VR
speaker Sanjeel Parekh.

Sanjeel Parekh is a research scientist at Meta Reality Labs Research. His research primarily focuses on building machine learning tools for problems involving audio-visual data such as source separation, event detection, and speech enhancement. 

He earned his PhD in computer science at Technicolor and Telecom University of Paris-Saclay in 2019. His thesis was on learning representations for robust audio-visual scene analysis. Other areas he finds interesting and engaging are multimedia and ML research, music, philosophy, math, and machines. 

His talk will focus on audiovisual scene understanding and how the field appears through the lens of augmented/virtual reality. Processing multi-sensory information to robustly detect and respond to objects and events in our surroundings lies at the heart of human perception. What does it take to impart such ability to machines?  In this talk, he will explore this question in two parts: first through some of his work on multimodal and interpretable ML methods for audiovisual scene analysis. He will then outline research challenges and opportunities posed in the context of AR/VR, delving into a few in greater detail. A secondary goal of this presentation is to provide an overview of open research initiatives by the lab for collaboration with the broader research community.

Date: Monday, April 29, 2024
Time: 2-3:15pm (EDT)
Location: Studio X - Carlson Library, 1st Floor & Zoom
Register to attend. 

Explore our other speakers

The Voices of XR speaker series is made possible by Kathy McMorran Murray and the National Science Foundation (NSF) Research Traineeship (NRT) program as part of the Interdisciplinary Graduate Training in the Science, Technology, and Applications of Augmented and Virtual Reality at the University of Rochester (#1922591).