<-Go to AI Workflows Go to ToC
1 Functions | 2 Reference Model | 3 I/O Data |
4 Functions of AI Modules | 5 I/O Data of AI Modules | 6 AIW, AIMs, and JSON Metadata |
1 Functions
The Enhanced Audioconference Experience Use Case addresses the situation where one or more speakers are active in a noisy meeting room and are trying to communicate with one or more interlocutors using speech over a network. In this situation, the user experience is very often far from satisfactory due to multiple competing speakers, non-ideal acoustical properties of the physical spaces that the speakers occupy and/or background noise. These can lead to a reduction in intelligibility of speech resulting in participants not fully understanding what their interlocutors are saying, in addition to creating a distraction and eventually leading to what is known as audioconference fatigue. When microphone arrays are used to capture the speakers, most of the described problems can be resolved by appropriate processing of the captured signals. The speech signals from multiple speakers can be separated from each other, the non-ideal acoustics of the space can be reduced, and any background noise can be substantially suppressed.
CAE-EAE is concerned with extracting from microphone array recordings the speech signals from individual speakers as well as reducing the background noise and the reverberation that reduce speech intelligibility. CAE-EAE also extracts the Spatial Attitudes of the speakers with respect to the position of the microphone array to facilitate the spatial representation of the speech signals at the receiver side if necessary. These Spatial Attitudes are represented in the Audio Scene Geometry format and packaged in a format that is amenable to further processing for efficient delivery and further processing. Data reduction of the extracted speech signals as well as their reconstruction/representation at the receiver side are outside the scope of this Use Case.
CAE-EAE aims to provide a complete solution to process speech signals recorded by microphone arrays to provide clear speech signals substantially free from background noise and acoustics-related artefacts to improve the auditory quality of audioconference experience. Thus, CAE-EAE improves auditory experience in an audioconference, thereby reducing the effects of audioconference fatigue.
2 Reference Model
Figure 1 shows the Reference Model for the CAE-EAE.
Figure 1 – Enhanced Audioconference Experience Reference Model
3 I/O data of AI Workflow
Table 1 shows the input and output data for the Enhanced Audioconference Experience workflow.
Table 1 – I/O data of Enhanced Audioconference Experience
Inputs | Comments |
Microphone Array Geometry | A Data Type representing the position of each microphone comprising a Microphone Array and specific characteristics such as microphone type, look directions, and the array type. |
Microphone Array Audio | A Data Type whose structure contains between 4 and 256 time-aligned interleaved Audio Channels organised in blocks. |
Outputs | Comments |
Multichannel Audio Stream | Interleaved Multichannel Audio packaged with Time Code as specified in Multichannel Audio Stream. |
The Enhanced Audio Experience AIW:
- Receives:
- Microphone Array Geometry which describes the number, positioning, and configuration of the microphone(s). Using this information, the system can detect the relative directions of the active speakers according to the microphone array and separate relevant audioconference speech sources from each other and from other spurious sounds. Since audio conferencing is a real-time application scenario, the use case operates on Audio Blocks.
- Microphone Array Audio which is input to EAE as short Multichannel Audio Blocks comprising real valued time domain audio samples where the number of audio samples in each Audio Block is the same for all the microphones.
- Produces Multichannel Audio Stream.
4 Functions of AI Modules
The AIMs required by the Enhanced Audioconference Experience are given in Table 2.
Table 2 – AIMs of Enhanced Audioconference Experience
AIM | Function |
Audio Analysis Transform | Represents the input Multichannel Audio in a new form amenable to further processing by the subsequent AIMs in the architecture. |
Sound Field Description | Produces Spherical Harmonic Decomposition Coefficients of the Transformed Multichannel Audio. |
Speech Detection and Separation | Separates speech and non-speech signals in the Spherical Harmonic Decomposition producing Transform Speech and Audio Scene Geometry. |
Noise Cancellation Module | Removes noise and/or suppresses reverberation in the Transform Speech producing Enhanced Transform Audio. |
Audio Synthesis Transform | Effects inverse transform of Enhanced Transform Audio producing Enhanced Audio Objects ready for packaging. |
Audio Description Packaging | Multiplexes Enhanced Audio Objects and the Audio Scene Geometry. |
The EAE use case receives Microphone Array Audio and Microphone Array Geometry which describes the number, positioning, and configuration of the microphone(s). Using this information, the system can detect the relative directions of the active speakers according to the microphone array and separate relevant audioconference speech sources from each other and from other spurious sounds. Since audio conferencing is a real-time application scenario, the use case operates on Audio Blocks.
The Multichannel Audio is input to EAE as short Multichannel Audio Blocks comprising real valued time domain audio samples where the number of audio samples in each audio block is the same for all the microphones.
The sequence of operations of the EAE use case is the following:
5 I/O Data of AI Modules
Table 3 specifies the I/O Data of CAE-EAE.
Table 3 – CAE-EAE AIMs and their data
AIM | Input Data | Output Data |
Audio Analysis Transform | Microphone Array Audio | Transform Multichannel Audio |
Sound Field Description | Transform Multichannel Audio | Spherical Harmonic Decomposition Coefficients |
Speech Detection and Separation | Spherical Harmonic Decomposition Coefficients | Transform Audio Audio Scene Geometry |
Noise Cancellation Module | Spherical Harmonic Decomposition Coefficients Transform Audio Audio Scene Geometry |
Enhanced Transform Audio |
Audio Synthesis Transform | Enhanced Transform Audio | Enhanced Audio Objects |
Audio Description Packaging | Enhanced Audio Objects Audio Scene Geometry |
Multichannel Audio Stream |
6 AIW, AIMs, and JSON Metadata
Table 4 provides links to the AI Modules and JSON Metadata.
Table 4 – AIW, AIMs, and JSON Metadata
AIW | AIMs | Names | JSON |
CAE-EAE | Enhanced Audioconference Experience | File | |
CAE-AAT | Audio Analysis Transform | File | |
CAE-SFD | Sound Field Description | File | |
CAE-SDS | Speech Detection and Separation | File | |
CAE-NCM | Noise Cancellation Module | File | |
CAE-AST | Audio Synthesis Transform | File | |
CAE-ADP | Audio Description Packaging | File |