We are pleased to announce the 3 Plenary Lectures of I3DA 2023. As plenary speakers, we will have the honor to host:
Professor | Department of Engineering and Architecture, University of Parma; Parma, Italy.
Navigating a virtual acoustics world: real-time rendering systems vs precomputed impulse responses
Professor | METU in Ankara, Turkey.
6DoF Audio: Moving beyond the constraints of 3D sound – Convincingly transporting the user to a remote, virtual location using digital and communication technologies has been the holy grail of multimedia research since at least the second half of the 20th century. Today, we are closer than ever to that idealised future with computational and network infrastructure that can afford to fulfil the requirements, terminals with audiovisual capabilities to achieve unprecedented visual realism, and, more importantly, users eager to consume such content. Along with volumetric visual content, 6DoF audio is one of the most critical components to make navigable immersive environments possible. While 6DoF audio has many components in common with legacy 3D audio technologies, it also has distinct scientific and technological challenges that are bound to make 6DoF audio an interesting research topic for many years to come. We will consider some of these challenges, such as the utilisation of knowledge on and modelling of human movement, sound recording requirements, methods for sound field processing and interpolation, and considerations for presenting 6DoF audio as well as its coding and compression. We will also attempt to outline future opportunities, especially in creating navigable multimedia content.
Filippo Maria Fazi
Professor | University of Southampton, UK.
Listener-adaptive reproduction of spatial audio with loudspeaker arrays
This lecture will cover the fundamentals of several loudspeaker-based methods for spatial audio delivery that were developed at the Institute of Sound and Vibration Research and rely on listener tracking. The oldest of these techniques, Virtual Microphone Array Panning (VMAP), is a pressure-matching approach wherein the array of virtual control points moves together with the listener and the audio reproduction DSP parameters are adapted accordingly. The second technique, Compensated Amplitude Panning (CAP), relies on the first-order Taylor expansion of the sound field close to the listener’s position. The extension of this technique, Higher Order Stereophony (HOS), is based on the higher-order Taylor expansion of the sound field and it is shown to be an extension of traditional stereophony. Finally, listener’s position-adaptive Cross-Talk Cancellation (CTC), combines acoustic beamforming, cross-talk cancellation, and computer vision to adapt the audio reproduction algorithm parameters depending on the listener’s position, thus overcoming the narrow sweet-spot issue of conventional CTC systems.