Fraunhofer IIS at AES Fall Online Exhibition

For over 30 years, the Audio and Media Technologies division of Fraunhofer IIS has been shaping the globally deployed audio standards and technologies. Today the 4th generation of best-in-class media technologies – such as MPEG-H Audio, xHE-AAC, EVS, LC3/LC3plus – is elevating the user experience to new heights. We are sorry that we cannot meet in person at this AES Fall 2021 show but we will try our best to bring some exciting news and product demonstrations to your screen.


MPEG-H Audio



Watch Fraunhofer IIS @ AES

Tune in to one of the sessions hosted by our audio and media experts

Contact us:

Reach out to the Audio Team at Fraunhofer IIS

Latest news



Watch Fraunhofer IIS @ AES

Advantages of using Metadata

Wednesday, October 13 • 4:00pm - 5:00pm

Robert Bleidt • John Kean • David Bialik

Audio-specific metadata can offer new consumer benefits as well as a more compelling listening experience. This presentation discusses the development of systems using audio metadata and explains the types of metadata that monitor key characteristics of audio content, such as loudness, dynamic range, and signal peaks. The metadata is added to the content during encoding for real-time distribution – as in streams, or for file storage – as with podcasts. In the playback device or system, this metadata is decoded along with the audio data frames. It will be shown that audio dynamic range can be controlled according to the noise environment around the listener – quiet parts of a performance can be raised to audibility, but for those listeners who need it. Content producers need to make only one target level for all listeners, rather than one for smart speakers, another for fidelity-conscious listeners, etc. An audio demonstration is planned to allow the audience to hear the same encoded program over a range of playout conditions with the same device, from riding public transit to full dynamic range, for listeners who want highest fidelity.

MPEG-H Audio production workflows for a Next Generation Audio experience in broadcast, streaming and music

Available on-demand until December 3

Yannik Grewe • Philipp Eibl • Christian Simon • Matteo Torcoli • Daniela Rieger • Ulli Scuda

MPEG-H Audio is a Next Generation Audio (NGA) system offering a new audio experience for the audience in various applications: Object-based immersive sound delivers a new degree of realism and artistic freedom for immersive music applications, such as the 360 Reality Audio music service. Advanced interactivity options enable improved personalization and accessibility, including solutions to create object-based features out of legacy material, e.g. deep-learning-based dialogue enhancement. 'Universal delivery' allows for optimal rendering of one production over all kinds of devices and various distribution ways, e.g. broadcast or streaming. All of these new features are achieved by adding metadata to the audio, which is defined during production and offers content providers flexible control of interaction and rendering options. Thus, new possibilities and requirements during the production process are imposed. In this paper, examples of state-of-the-art NGA production workflows are detailed and discussed, with special focus on immersive music, broadcast, and accessibility.


Production Tools for the MPEG-H Audio System

Available on-demand until December 3

Yannik Grewe • Philipp Eibl • Daniela Rieger • Ulli Scuda

Next Generation Audio Systems, such as MPEG-H Audio, rely on metadata to enable a wide variety of features. Information such as channel layouts, the position and properties of audio objects or user interactivity options are only some of the data that can be used to improve consumer experience.
Creating these metadata requires suitable tools, which are used in a process known as "authoring", where interactive features and the options for 3D immersive sound rendering are defined by the content creator.
Different types of productions each impose their own requirements on these authoring tools, which lead to a number of solutions appearing in the market. Using the example of MPEG-H Audio, this paper will detail some of the latest developments and authoring solutions designed to enable immersive and interactive live-and post productions.

Perceptual Evaluation of Interior Panning Algorithms Using Static Auditory Events

Available on-demand until December 3

Thomas Robotham • Andreas Silzle • Anamaria Nastasa • Alan Pawlak • Jürgen Herre

Interior panning algorithms enable content authors to position auditory events not only at the periphery of the loudspeaker configuration, but also within the internal space between the listeners and the loudspeakers. In this study such algorithms are rigorously evaluated, comparing rendered static auditory events at various locations against true physical loudspeaker references. Various algorithmic approaches are subjectively assessed in terms of; Overall, Timbral, and Spatial Quality for three different stimuli, at five different positions and three radii. Results show for static positions that standard Vector Base Amplitude Panning performs equal, or better, than all other interior panning algorithms tested here. Timbral Quality is maintained throughout all distances. Ratings for Spatial Quality vary, with some algorithms performing significantly worse at closer distances. Ratings for Overall Quality reduce moderately with respect to reduced reproduction radius and are predominantly influenced by Timbral Quality.

Informed postprocessing for auditory roughness removal for low-bitrate audio coders

Available on-demand until December 3

Steven van de Par • Sascha Disch • Andreas Niedermeier • Bernd Edler

In perceptual audio coding using very low bitrates, modulation artifacts can be introduced onto tonal signal components, which are often perceived as auditory roughness. These artifacts may occur for instance due to quantization errors or may be added when using an audio bandwidth extension, which sometimes causes an irregular harmonic structure at the borders of replicated bands. Especially, the roughness artifacts due to quantization errors are difficult to mitigate without investing considerably more bits in encoding of tonal components. We propose a novel technique to remove these roughness artifacts at the decoder side controlled by a small amount of guidance information transmitted by the encoder.