1     Scope of Television Media Analysis

2     Reference Model of Television Media Analysis

3     I/O Data of Television Media Analysis

4     Functions of AI Modules of Television Media Analysis

5     I/O Data of AI Modules of Television Media Analysis

6     AIMs and JSON Metadata of Television Media Analysis

1      Scope of Television Media Analysis

Television Media Analysis (OSD-TMA) gives Audio-Visual Event Descriptors in the form of a set of significant Audio-Visual Scene Descriptors with scene changes, IDs of speakers and faces with their spatial positions, and text from utterances of a video program provided as input.

2      Reference Model of Television Media Analysis

Figure 2 depicts the Reference Model of TV Media Analysis.

Figure 2 – Reference Model of OSD-TMA

Table 1 gives the names and acronyms of the AIMs.

Table 1 – Names and acronyms of the AIMs

Acronyms AIM Names
OSD-TVS TV Splitter
OSD-SCD Scene Change Detection
MMC-TVD TV Diarisation
PAF-FIR Face Identity Recognition
MMC-SIR Speaker Identity Recognition
MMC-ASR Automatic Speech Recognition
OSD-AVS Audio-Visual Scene Description
OSD-AVE Audio-Visual Event Description

3      I/O Data of Television Media Analysis

Table 2 provides the input and output data of the TV Media Analysis Use Case:

Table 2 – I/O Data of TV Media Analysis

Input Descriptions
Input Audio-Video Audio-Video to be analysed.
Input Descriptions
AV Event Descriptors Resulting analysis.

4      Functions of AI Modules of Television Media Analysis

Table 3 provides the functions of the TV Media Analysis Use Case. Note that processing proceeds asynchronously, e.g., TV Splitter separates audio and video for the entire duration of the file and passes the entire audio and video files.

Table 3 – Functions of AI Modules of Conversation with Personal Status

AIM Function
TV Splitter 1.     Receives Audio-Visual

a.     An audio-video file.

b.     Metadata (e.g., title, date).

2.     Produces

a.     Video file

b.     Audio file

3.     When the files of the full duration of the video are ready, AV Splitter informs the following AIMs.

Scene Change Detection 1.     Receives Video file.

2.     Iteratively

a.     Looks for a video frame that conveys a scene changed from the preceding scene (depends on threshold).

b.     Assigns a video clip identifier to the video clip.

c.     Produces a set of images with StartTime and EndTime.

i.     An image

ii.     Time stamp

TV Diarisation 1.     Receives Audio file.

2.     Iteratively detects speaker change.

a.     For each audio segment (from one change to the next):

i.     Becomes aware that there is speech.

ii.     Assigns a speech segment ID and anonymous speaker ID (i.e., the identity is unknown) in the segment.

iii.     Decides whether:

1.     The existing speaker has stopped.

2.     A new speaker has started a speech segment.

iv.     If a speaker has started a speech:

1.     Assigns a new speech segment ID.

2.     Check whether the speaker is new or old in the session.

3.     If old retain old anonymous speaker ID.

4.     If new assign a new anonymous speaker ID.

b.     Produces a series of audio sequences each of which contains:

i.     A speech segment.

ii.     Start and end time.

iii.     Anonymous Speaker ID.

Face Identity Recognition 1.     Receives a set of images per video clip.

2.     For each image identifies the bounding boxes.

3.     Extracts faces from the bounding boxes.

4.     Extracts the embeddings that represent a face.

5.     Compares the embeddings with those stored in the face recognition data base.

6.     Associates the embeddings with a face ID.

Speaker Identity Recognition 1.     Receives a speech segment.

2.     Extracts the embeddings that represent the speech segment.

3.     Compares the embeddings with those stored in the speaker recognition data base.

4.     Associates the embeddings with a Speaker ID.

Automatic Speech Recognition 1.     Receives a speech segments and SpeakerID.

2.     Produces the transcription of the speech payload.

3.     Attaches time stamps to specific portions of the transcription.

Audio-Visual Scene Description 1.     Receives

1.     Bounding box coordinates, Face ID, and time stamps

2.     Speaker ID and time stamps.

3.     Reconciles Face ID and Speaker ID.

4.     Text and time stamps

5.     Produces a JSON multiplexing the input data.

2.     Produces Audio-Visual Scene Descriptors

Audio-Visual Event Description 1.     Receives Audio-Visual Scene Descriptors

2.     Produces Audio-Visual Event Descriptors

5      I/O Data of AI Modules of Television Media Analysis

Table 4 provides the I/O Data of the AI Modules of the TV Media Analysis Use Case.

Table 4 – I/O Data of AI Modules of Conversation with Personal Status

AIM Receives Produces
TV Splitter Audio-Video Audio

Video

Scene Change Detection Video Image
TV Diarisation Audio Speech Segments
Face Identity Recognition Image Face ID
Speaker Identity Recognition Speech Segments Speaker ID
Automatic Speech Recognition Speech Segments Text
Audio-Visual Scene Description 1.     Bounding box coordinates

2.     Face ID

3.     Scene time stamps

4.     Speaker ID

5.     Speech time stamps.

6.     Recognised Text from Speech Segments

7.     Text and time stamps

AV Scene Descriptors with reconciled Face and Speech IDs.
Audio-Visual Event Description Audio-Visual Scene Descriptors AV Event Descriptors

6      AIMs and JSON Metadata of Television Media Analysis

Table 5 – AIMs and JSON Metadata

AIW AIM Name
OSD-TMA OSD-TVS TV Splitter
OSD-VCD Visual Change Detection
MMC-TVD TV Diarisation
PAF-FIR Face Identity Recognition
MMC-SIR Speaker Identity Recognition
MMC-ASR Automatic Speech Recognition
OSD-AVS Audio-Visual Scene Description
OSD-AVE Audio-Visual Event Description