Immersive Audio-Visual 3D Scene Reproduction in Virtual Reality

Speaker: Dr Hansung Kim

Title: Immersive Audio-Visual 3D Scene Reproduction in Virtual Reality

Abstract: As personalized immersive display systems have been intensely explored in Virtual Reality (VR), plausible 3D audio corresponding to the visual content is required to provide more realistic experiences to users. In this talk, a complete pipeline to simultaneously reconstruct 3D geometry and acoustic properties of the environment from a single panoramic (360⁰) image is introduced. Monocular depth estimation, material recognition, semantic scene completion methods using deep convolutional neural networks have been developed to estimate the complete semantic scene geometry in order to adopt spatial audio reproduction to the scene.

Bio: Dr. Hansung Kim is an Associate Professor in the School of Electronics & Computer Science at the University of Southampton, UK. He received his BSc, MSc and Ph.D degrees in electronic and electrical engineering from Yonsei University, South Korea. He was employed as a researcher of Knowledge Science Lab (KSL) at Advanced Telecommunications Research Institute International (ATR), Japan, from 2005 to 2008, and as a senior research fellow at the Centre for Vision, Speech, and Signal Processing (CVSSP) at the University of Surrey, UK, from 2008 to 2020. His research for last 25 years includes 3D computer vision, Stereo image processing, 3D reconstruction, Augmented/Virtual Reality, Multi-modal data processing, Audio-visual data processing and Media production with over 120 published articles (60 as first author) in 20 research projects.