page

NAACL2022-MAIWorkshop

 

 

Workshop on Multimodal Artificial Intelligence
NAACL 2022
Location: NAACL 2022 – Seattle, WA, USA
Date: July 15th
Join Zoom by Clicking Here

 

 

COVID-19: Remote options are available through this zoom link: CLICK HERE 

This year, the workshop on Multimodal Artificial Intelligence presents a chorus of invited speakers, discussing the frontiers of multimodal AI. 

Table of Contents

1. Keynotes
2. Scope and Related Areas
5. Workshop Schedule
6. Organizing Committee

Scope and Related Areas

The NAACL 2022 Workshop on Multimodal Artificial Intelligence (MAI-Workshop) offers a unique opportunity for interdisciplinary researchers to study and model interactions between (but not limited to) modalities of language, vision, and acoustic. Advances in multimodal learning allows the field of NLP to take the leap towards better generalization to real-world (as opposed to limitation to textual applications), and better downstream performance in Conversational AI, Virtual Reality, Robotics, HCI, Healthcare, and Education.

We invite researchers from NLP, Computer Vision, Speech Processing, Robotics, HCI, and Affective Computing to attend the workshop. We will cover a broad range of topics, such as:

  • • Neural Modeling of Multimodal Language
  • • Multimodal Dialogue Modeling and Generation
  • • Multimodal Sentiment Analysis and Emotion Recognition
  • • Language, Vision and Speech
  • • Multimodal Artificial Social Intelligence Modeling
  • • Multimodal Commonsense Reasoning
  • • Multimodal RL and Control (Human-robot communication and multimodal language for robots)
  • • Multimodal Healthcare
  • • Multimodal Educational Systems
  • • Multimodal Affective Computing
  • • Multimodal Fusion and Alignment
  • • Multimodal Representation Learning
  • • Multimodal Sequential Modeling
  • • Multimodal Co-learning and Transfer Learning
  • • Multimodal Active Learning
  • • Multimodal and Multimedia Resources
  • • Creative Applications of Multimodal Learning in E-commerce, Art, and other Impactful Areas.

 

Keynotes 

 

Devi Parikh – FAIR at Meta, Georgia Tech
Yonatan Bisk (white male wearing a black t-shirt against a white background) Yonatan Bisk – Carnegie Mellon University
Aida Nematzadeh– Google DeepMind
 Victor Zhong– University of Washington
Danna Gurari – University of Colorado Boulder
Alane Suhr (Cornell) - Reasoning and Learning in Interactive Natural  Language Systems | DSI Alane Suhr – Cornell
Wei-Ning Hsu (徐煒甯) Wei-Ning Hsu – Meta AI

Drew A Hudson – Stanford

Schedule (PDT Timezone)

(ALL PDT TIMEZONE)

8:30 – 9 Danna Gurari
9 – 9:10 QA #1
 
9:10 – 9:55 Wei-Ning Hsu
9:55 – 10:10 QA #2
 
10:10 – 10:55 Aida Nematzadeh
10:55 – 11:10 QA #3
 
11:10 – 11:55: Alane Suhr
11:55 – 12:10: QA #4
 
12:10 – 1 Lunch Break
 
1:00 – 1:45 Devi Parikh
1:45 – 2 QA #5
 
2 – 2:45 Drew Hudson
2:45 – 3 QA #6
 
3 – 3:15 Break
 
3:15 – 4 Yonatan Bisk
4 – 4:15 QA #7
 
4:15 – 5 Victor Zhong
5 – 5:15 QA #8

 

Organizing Committee

Amir Zadeh – Alexa AI, Carnegie Mellon University
Louis-Philippe Morency – Language Technologies Institute, Carnegie Mellon University
Paul Pu Liang – Machine Learning Department, Carnegie Mellon University
Kelly Shi – Carnegie Mellon University
 Alex Wilf – Carnegie Mellon University
Ruslan Salakhutdinov – Carnegie Mellon University
Soujanya Poria – Singapore University of Technology and Design
Erik Cambria – Nanyang Technological University