1 Version
V2.1
2 Functions
Multimodal Emotion Fusion (MMC-MEF):
- Receives
- Emotion (Text)
- Emotion (Speech)
- Emotion (Face).
- Produces the input Entity’s Emotion.
3 Reference Architecture
Figure 1 depicts the Reference Architecture of the Multimodal Emotion Fusion AIM.

Figure 1 – The Multimodal Emotion Fusion AIM
4 I/O Data
Table 1 specifies the Input and Output Data of the Multimodal Emotion FusionAIM.
Table 1 – I/O Data of the Multimodal Emotion Fusion AIM
| Input data | From | Comment |
| Emotion (Text) | PS-Text Interpretation | Emotion in Text |
| Emotion (Speech) | PS-Speech Interpretation | Emotion in Speech |
| Emotion (Face) | PS-Face Interpretation | Emotion in Face |
| Output data | To | Description |
| Input Emotion | Downstream AIM | The estimated emotion that fuses all inputs |
5 SubAIMs
No SubAIMs.
6 JSON Metadata
https://schemas.mpai.community/MMC/V2.1/AIMs/MultimodalEmotionFusion.json