log in  |  register  |  feedback?  |  help  |  web accessibility
Sim2Real2Sim: Differentiable Physics-Based Modeling of Multisensory Objects
Ruohan Gao
IRB 0318 (Gannon) or https://umd.zoom.us/j/97919102992?pwd=LbSBM2MZy4QpVfnj92ukT5AIqyTYaO.1#success
Friday, October 18, 2024, 11:00 am-12:00 pm
  • You are subscribed to this talk through .
  • You are watching this talk through .
  • You are subscribed to this talk. (unsubscribe, watch)
  • You are watching this talk. (unwatch, subscribe)
  • You are not subscribed to this talk. (watch, subscribe)
Abstract

We perceive the world not as a single giant entity but often through perception and manipulation of a wide variety of objects, which exist as bounded wholes and move on connected paths. While there has been significant progress by "looking"—recognizing objects based on glimpses of their visual appearance or 3D shape—objects in the world are often modeled as silent and untouchable entities. In this talk, I will present how we model the multisensory signals of real-world objects through differentiable physics-based simulation. First, I will discuss how we model multisensory behaviors of objects with a new dataset of neural objects and how we perform Sim2Real transfer by learning from them. Then, vice versa, I will introduce our differentiable inverse rendering algorithms for Real2Sim applications, where we infer a variety of physical properties of objects from their real-world observations. Together, this has the potential to endow a system with the ability to autonomously build its own multisensory simulation of its environment using only its raw sensory inputs.

Bio

Ruohan Gao is an incoming Assistant Professor in Computer Science at the University of Maryland, College Park, where he will be leading the UMD Multisensory Machine Intelligence Group. Previously, he was a Postdoc at the Stanford Vision and Learning Lab, and obtained his Ph.D. degree from The University of Texas at Austin. Ruohan mainly works in the fields of computer vision and machine learning with a particular emphasis on multisensory learning with sight, sound, and touch. His research has been recognized by the Stanford AI Lab Postdoctoral Fellowship, the Michael H. Granof Award which is designated for UT Austin's Top 1 Doctoral Dissertation, the Google PhD Fellowship, the Adobe Research Fellowship, a Best Paper Award Runner Up at British Machine Vision Conference (BMVC) 2021, and a Best Paper Award Finalist at Conference on Computer Vision and Pattern Recognition (CVPR) 2019.

This talk is organized by Samuel Malede Zewdu