Multimodal Conversation (MPAI-MMC)
Multi-modal conversation (MPAI-MMC) is an MPAI standard covering Conversation with Emotion supporting audio-visual conversation with a machine impersonated by a synthetic voice and an animated face; Multimodal Question Answering supports request for information about a displayed object; Unidirectional, Bidirectional and One-to-Many Speech Translation support conversational translation using a synthetic voice that preserves the speech features of the human.
Development of the MPAI-MMC Technical Specification V1 is completed. MPAI is indebted to the following individuals: Miran Choi (ETRI), Gérard Chollet (IMT), Jisu Kang (KLleon), Mark Seligman (SMI) and Fathy Yassa (SMI) for their efforts.
Reference Software, Conformance Testing and Performance Assessment are under development. Use Cases and Functional Requirememnst for MPAI-MMCX V2 are being developed.
See the public MPAI-MMC documents:
- Version 2.0
- Version 1.1
Visit the About MPAI-MMC page