Multimodal Conversation (MPAI-MMC)

Multi-modal conversation (MPAI-MMC) is an MPAI standard covering Conversation with Emotion supporting audio-visual conversation with a machine impersonated by a synthetic voice and an animated face; Multimodal Question Answering supports request for information about a dis­played object; Unidirectional, Bidirectional and One-to-Many Speech Translation support conversational translation using a synthetic voice that preser­ves the speech features of the human.

Development of the MPAI-MMC Technical Specification V1 is completed. MPAI is indebted to the following individuals: Miran Choi (ETRI), Gérard Chollet (IMT), Jisu Kang (KLleon), Mark Seligman (SMI) and Fathy Yassa (SMI) for their efforts.

Reference Software, Conformance Testing and Performance Assessment are under development. Use Cases and Functional Requirememnst for MPAI-MMCX V2 are being developed.

See the public MPAI-MMC documents:

  1. Version 2.0
    1. MPAI-MMC Use Cases and Functional Requirements WD1.4
  2. Version 1.1
    1. Call for Patent Pool Administrator (Closed)
    2. Introdution to MPAI-MMC
    3. MPAI-MMC Standard
    4.  Call for Technologies (V1)
    5. Framework Licence
    6. Use Cases and Functional Requirements (V1)
    7. Application Note

Visit the About MPAI-MMC page