|
|
ICMI 2019 Conference Program
Please see our up to date full program here.
Also, mobile apps can be download via:
Program Overview
|
Morning |
AM Coffee Break |
Lunch |
Afternoon |
PM Coffee Break |
Evening Events |
Monday
|
Workshop Neuro Management
(09:00-12:15)
M1 |
10:15 - 10:45 |
12:30 - 14:00 |
Workshop Media Analytics
(14:00-17:05)
M1
|
15:15 - 15:45 |
18:00 - 20:00
Reception |
Tutorial
Tatsuya Kawahara
(09:00-12:00)
M2
|
Tutorial
Michelle Zhou
(14:00-17:00)
M2
|
Doctoral Consortium
(09:00-12:30)
M3
|
Doctoral Consortium
(14:00-17:25)
M3
|
Tuesday
|
Opening
Keynote 1
Session 1
(09:00-12:40)
GB
|
10:15 - 10:45 |
12:40 - 14:00 |
Session 2
(14:00-15:35, GB)
Poster1 & LBR
(15:35-17:30, E3)
Panel Discussion
(17:30-18:30, GB)
|
15:35 - 17:35 |
N/A |
Wednesday
|
Keynote 2
Session 3
(09:00-12:00)
GB
|
10:00 - 10:30 |
12:40 - 14:00 Volunteering Session
GB
|
Session 4
(14:00-15:20, GB)
Poster2 & Demo
(15:20-17:20, E3)
Sustained Achievement Awardee Talk
(17:20-18:20, GB)
|
15:20 - 17:20 |
19:00 - 21:00
Banquet
Bus at 18:30 |
Thursday
|
Keynote 3
Session 5
(09:00-12:20)
GB
|
10:00 - 10:30 |
12:20 - 14:00 |
Session 6
Challenge Overview
Town Hall Meeting
Closing
(14:00-17:15, GB)
|
15:30 - 16:00 |
N/A |
Friday
|
Grand Challenge
EmotiW
(09:00-12:30, M1)
MAVSR,CATSLU
(09:00-12:00, M2)
|
10:00 - 10:30 |
N/A |
Tour Groups
13:30
|
Floor Plan

The conference program is now available on Progressive Web App and QOALA. You can see the web version:

To access the service on your mobile device, you'd first need to install an app from either Apple App Store or Google Play Store:
Once you install the app, you should be able to search for ICMI 2019.
We have volunteers wearing white T-shirts - please shout for help if needed!
Conference Program
Monday, 14 October 2019
The registration desk will open from 8:00 to 17:00 in the foyer. Both registration and banquet tickets can be purchased
on site.
09:00 - 12:15 |
Workshop:
NeuroManagement and Intelligent Computing Method on Multimodal Interaction
Organisers: Weihui Dai
NeuroManagement and Intelligent Computing Method on Multimodal Interaction
Weihui Dai
Multisensory integration of emotions in a face-prosody-semantics Stroop task
Yi Lin; Hongwei Ding
A synergy study of metaphoric gestures on rhetorical behavior construction: Based on the corpus of “AI”-themed public
speeches
Lang Chen; Liqin Zhao
Machine learning in human-computer nonverbal communication
Shuangping Gong; Huajuan Mao ;Yihang Wang; Anran Xu
A supra-modal decoding mechanism: Evidence from Chinese speakers learning English
Youhui Wang;Lanqing Dong; Weihui Dai
Single-trial based EEG classification of the dynamic representation of speaker stance: A preliminary study with
representational similarity analysis
Xiaoming Jiang
Affective computation of students’ behaviors under classroom scenes
Jiaolong Fu; Tingting Ge; Meilin Li; Xiaohua Hu
Closing Remarks
|
09:00 - 12:00 |
Tutorial:
Spoken Dialogue Processing for Multimodal Human-Robot Interaction
Invited Speaker: Tatsuya Kawahara
|
09:00 - 12:30 |
Doctoral Consortium
Welcome and Introduction
Invited Talk 1
Yale Song, MSR Redmond, USA
Session 1: Mental Health
Detecting Temporal Phases of Anxiety in The Wild: Toward Continuously
Adaptive Self-Regulation Technologies
Hashini Senaratne, Monash University
Multimodal Machine Learning for Interactive Mental Health Therapy
Leili Tavabi, ICT, University of Southern California
Session 2: Wearable Sensing and Affect Recognition
Tailoring Motion Recognition Systems to Children's Motions
Aishat Aloba, University of Florida
Multi-modal fusion methods for robust emotion recognition using body-worn physiological sensors in ubiquitous environments
Tianyi Zhang, Centrum Wiskunde & Informatica
Communicative Signals and Social Contextual Factors in Multimodal Affect Recognition
Michel-Pierre Jansen, University of Twente
|
10:15 - 10:45 |
Coffee Break
|
12:30 - 14:00 |
Lunch
|
14:00 - 17:05 |
Workshop:
Media Analytics for Societal Trends: Closing the loop with impact and affect in human-media interactions
Organisers: Naveen Kumar. Chi-Chun (Jeremy) Lee, Ming Li, Tanaya Guha, Shri Narayanan, Krishna Somandepalli
Opening Note
Invited Talk 1 Understanding emotions and sentiment in multimedia
Mohammad Soleymani, USC Institute for Creative Technologies
Victim or Perpetrator? Analysis of Violent Character Portrayals from Movie Scripts
Victor Martinez, Krishna Somandepalli, Karan Singla, Anil Ramakrishna, Yalda T. Uhls, Shri Narayanan
Invited Talk 2 Multimodal Behavioral Analytics: What Hand Movements Reveal about Domain Expertise
Sharon Oviatt
Multi-queries based Attention for Facial Emotion Recognition
Jinlong Jiao, Wan Ding, Dongyan Huang, Ziping Zhao
RWF-2000: A Large Video Database for Violence Detection
Ming Cheng, Wilton W.T. Fok, Ming Li
Closing Comments
|
14:00 - 17:00 |
Tutorial:
Getting Virtually Personal: Power Conversational AI to
Fulfill Tasks and Personalize Chitchat for Real-World Applications
Invited Speaker: Michelle Zhou
|
14:00 - 17:25 |
Doctoral Consortium
Invited Talk 2
Kristiina Jokinen, AIST, Japan
Session 3: Human-Human and Human-Agent Interaction
Co-located Collaboration Analytics
Sambit Praharaj, Open Universiteit, Netherlands
Coalescing Modalities for Grounded Pose Forecasting
Chaitanya Ahuja, Carnegie Mellon University
Session 4: Interactive Applications
Attention-driven Interaction Systems for Augmented Reality
Chaitanya Ahuja, Carnegie Mellon University
Multimodal Driver Interaction with Gesture, Gaze and Speech
Abdul Rafey Aftab, Universität des Saarlandes
Closing Remarks
|
15:15 - 15:45 |
Coffee Break
|
18:00 - 20:00 |
Reception |
Tuesday, 15 October 2019
The registration desk will open from 8:00 to 17:00 in the foyer. Both registration and banquet tickets can be purchased on site.
Main conference sessions are all held in Grand Ballroom B.
* Papers are nominees for either best paper award or student best paper award.
09:00 - 09:15 |
Conference Opening |
09:15 - 10:15 |
Keynote 1:
Socially-Aware User Interfaces: Can Genuine Sensitivity Be Learnt at all?
Prof. Dr. Elisabeth André
|
10:15 - 10:45 |
Coffee Break |
10:45 - 12:40 |
Session 1: Human Behavior
Session Chair: Albert Ali Salah, Utrecht University
|
|
Multi-modal Active Learning From Human Data:A Deep Reinforcement Learning Approach*
Ognjen Rudovic; Meiru Zhang; Bjorn Schuller; Rosalind Picard
Comparing Pedestrian Navigation Methods in Virtual Reality and Real Life
Gian-Luca Savino; Niklas Emanuel; Steven Kowalzik; Felix Kroll; Marvin C. Lange; Matthis Laudan; Rieke Leder; Zhanhua
Liang; Dayana Markhabayeva; Martin Schmeißer; Nicolai Schütz; Carolin Stellmacher; Zihe Xu; Kerstin Bub; Thorsten
Kluss; Jaime Maldonado; Ernst Kruijff; Johannes Schöning
Video and Text-Based Affect Analysis of Children in Play Therapy
Metehan Doyran; Batıkan Türkmen; Eda Aydın Oktay; Sibel Halfon; Albert Ali Salah
Facial Expression Recognition via Relation-based Conditional Generative Adversarial Network
Byung Cheol Song; Min Kyu Lee; Dong Yoon Choi
Continuous Emotion Recognition in Videos by Fusing Facial Expression, Head Pose and Eye Gaze
Suowei Wu; Zhengyin Du; Weixin Li; Di Huang; Yunhong Wang
Effect of Feedback on Users' Emotions: Analysis of Facial Expressions during a Simulated Target Detection Task
Md Abdullah Al Fahim; Mohammad Maifi Hasan Khan; Theodore Jensen; Yusuf Albayram; Emil Coman; Ross Buck
|
12:40 - 14:00 |
Lunch |
14:00 - 15:35 |
Session 2: Artificial Agents
Session Chair: Mohammad Soleymani, USC, Institute for Creative Technologies
|
|
Multimodal Analysis and Estimation of Intimate Self-Disclosure*
Mohammad Soleymani; Kalin Stefanov; Sin-Hwa Kang; Jan Ondras; Jonathan Gratch
A High Fidelity Open Embodied Avatar with LipSyncing and Expression Capabilities
Deepali Aneja; Daniel McDuff; Shital Shah
To React or not to React: End-to-End Visual Pose Forecasting of a Personalized Avatar during Dyadic Conversations
Chaitanya Ahuja; Shugao Ma; Louis-Philippe Morency; Yaser Sheikh
Multitask Prediction of Exchange-Level Annotations for Multimodal Dialogue Systems*
Yuki Hirano; Shogo Okada; Haruto Nishimoto; Kazunori Komatani
Multimodal Learning for Identifying Opportunities for Empathetic Responses
Leili Tavabi; Kalin Stefanov; Setareh Nasihati Gilani; David Traum; Mohammad Soleymani
|
15:35 - 17:30 |
Coffee Break |
15:35 - 17:30 |
Poster Session 1
|
|
CorrFeat: Correlation-based Feature Extraction Algorithm using Skin Conductance and Pupil Diameter
for Emotion
Recognition
Tianyi Zhang; Abdallah El Ali; Chen Wang; Xintong Zhu; Pablo Cesar
Evaluation of Mid-Air Haptics as a Supplementary Feedback Cue for Grasping in Virtual Environments
Maite Frutos-Pascual; Jake Michael Harrison; Chris Creed; Ian Williams
Understanding the Attention Demand of Touch and Tangible Interaction on a Composite Task
Yosra Rekik; Walid Merrad; Christophe Kolski
TouchGazePath: Multimodal Interaction with Touch and Gaze Path for Secure Yet Efficient PIN Entry
Chandan Kumar; Daniyal Akbari; Raphael Menges; Scott MacKenzie; Steffen Staab
Determining Iconic Gesture Forms based on Entity Image Representation
Fumio Nihei; Yukiko I. Nakano; Ryuichiro Higashinaka; Ryo Ishii
Modeling Emotion Influence Using Attention-based Graph Convolutional Recurrent Network
Yulan Chen; Jia Jia; Zhiyong Wu
WiBend: Recognizing Bending Interaction for Passive Deformable Surfaces with Wi-Fi
Mira Sarkis; Céline Coutrix; Laurence Nigay; Andrzej Duda
ElderReact: A Multimodal Dataset for Recognizing Emotional Response in Aging Adults
Kaixin Ma; Xinyu Wang; Xinru Yang; Mingtong Zhang; Jeffrey M Girard; Louis-philippe Morency
Generative Model of Agent's behaviors in Human-Agent Interaction
Soumia Dermouche; Catherine Pelachaud
Exploring transfer learning between scripted and spontaneous speech for emotion recognition
Qingqing Li; Theodora Chaspari
|
15:30 - 17:30 |
Late Breaking Results
Fusing Dialogue and Gaze From Discussions of 2D and 3D Scenes
Regina Wang, Bradley Olson, Preethi Vaidyanathan, Reynold Bailey, Cecilia O. Alm
Multimodal Anticipated versus Actual Perceptual Reactions
Monali Saraf, Tyrell Roberts, Raymond Ptucha, Christopher Homan, Cecilia O. Alm
Measuring Affective Sharing between Two People by EEG Hyperscanning
Taiki Kinoshita, Hiroki Tanaka, Koichiro Yoshino, Satoshi Nakamura
Detecting Syntactic Violations from Single-trial EEG using Recurrent Neural
Networks
Shunnosuke Motomura, Hiroki Tanaka, Satoshi Nakamura
Detecting Dementia from Face in Human-Agent Interaction
Hiroki Tanaka, Hiroyoshi Adachi, Hiroaki Kazui, Manabu Ikeda, Takashi Kudo,
Satoshi Nakamura
Multimodal Biometric Authentication for VR/AR using EEG and Eye Tracking
Vrishab Krishna, Yi Ding, Aiwen Xu, Tobias Höllerer
An Approach to Reading Assistance with Eye Tracking Data and Text Features
Wei Guo, Shiwei Cheng
Evaluation of Dominant and Non-Dominant Hand Movements For Volleyball
Action Modelling
Fasih Haider, Fahim A. Salim, Sena Busra Yengec Tasdemir, Vahid Naghashi , Izem
Tengiz, Kubra Cengiz, Dees B.W. Postma, Robby van Delden, Dennis Reidsma,
Bert-Jan van Beijnum, Saturnino Luz
Are Humans Biased in Assessment of Video Interviews?
Chee Wee (Ben) Leong, Katrina Roohr, Vikram Ramanarayanan, Michelle P.
Martin-Raugh, Harrison Kell, Rutuja Ubale, Yao Qian, Zydrune Mladineo, Laura
McCulla
Lemusade: Make Lemonade Using Music
Jiexin Lyu, Jiawen Yao, Ankit Bansal
Interactive Upper Limb Training Device for ArmReaching and Finger Pointing
Exercise
Xuezhou Yang, Yung C. Liang, Yang Zou, Yixuan Bian, Jie Sun
Multimodal Assessment on Teaching Skills via Neural Networks
Masato Fukuda, Kazuhiro Kuwabara, Hung-Hsuan Huang, Toyoaki Nishida
Floor Apportionment Function of Speaker’s Gaze in Grounding Acts
Ichiro Umata, Tsuneo Kato, Koki Ijuin, Seiichi Yamamoto
Real-Time Multimodal Classification of Internal and External Attention
Lisa-Marie Vortmann, Moritz Schult, Mathias Benedek, Sonja Annerer-Walcher,
Felix Putze
Sensory Substitution Device Stabilizing Human Voice Production
Agnieszka Kulesza
|
15:30 - 17:30 |
Doctoral Spotlight Poster Session
|
17:30 - 18:30 |
PANEL Discussion:
Multimodal representation of human behaviour in context
Coordinator: Helen Meng
|
Wednesday, 16 October 2019
09:00 - 10:00 |
Keynote 2:
A Brief History of Intelligence
Dr. Hsiao-Wuen Hon |
10:00 - 10:30 |
Coffee Break |
10:30 - 12:10 |
Session 3: Touch and Gesture
Session Chair: Abhinav Dhall, Monash University |
|
Dynamic Adaptive Gesturing Predicts Domain Expertise in Mathematics*
Abishek Sriramulu; Jionghao Lin; Sharon Oviatt
VisualTouch: Enhancing Social Touch with Multi-modality Stimulation
Zhuoming Zhang; Robin Héron; Eric Lecolinet; Françoise Detienne; Stéphane Safin
TouchPhoto: Enabling Independent Picture Taking and Understanding for Visually-Impaired Users
Jongho Lim; Yongjae Yoo; Hanseul Cho; Seungmoon Choi
Creativity Support and Multimodal Pen-based Interaction
Ilhan Aslan; Katharina Weitz; Ruben Schlagowski; Simon Flutura; Susana Garcia Valesco; Marius Pfeil; Elisabeth André
Motion Eavesdropper: Smartwatch-based Handwriting Recognition Using Deep Learning
Hao Jiang
|
12:35 - 14:00 |
Lunch
|
15:00 - 16:30 |
Session 4: Physiological Modeling
Session Chair: Daniel McDuff, Microsoft Research
|
|
Predicting Cognitive Load in an Emergency Simulation based on Behavioral and Physiological Measures
Tobias Appel; Natalia Sevcenko; Franz Wortha; Katerina Tsarava; Korbinian Moeller; Manuel Ninaus; Enkelejda Kasneci;
Peter Gerjets
Driving Anomaly Detection with Conditional Generative Adversarial Network using Physiological and CAN-Bus Data
Yuning Qiu; Teruhisa Misu; Carlos Busso
Controlling for Confounders in Multimodal Emotion Classification via Adversarial Learning
Mimansa Jaiswal; Zakaria Aldeneh; Emily Mower Provost
Multimodal Classification of EEG During Physical Activity
Yi Ding; Brandon Huynh; Aiwen Xu; Tom Bullock; Hubert Cecotti; Matthew Turk; Barry Giesbrecht; Tobias Höllerer
|
15:20 - 17:20 |
Coffee Break
|
15:20 - 17:20 |
Poster Session 2
|
|
What's behind a choice? Understanding Modality Choices under Changing Environmental Conditions
Stephanie Arevalo; Stanislaw Miller; Martha Janka; Jens Gerken
Multimodal Behavioral Markers Exploring Suicidal Intent in Social Media Videos
Ankit Parag Shah; Vasu Sharma; Vaibhav Vaibhav; Mahmoud Alismail; Jeffery M. Girard; Louis-philippe Morency
Estimating Uncertainty in Task-Oriented Dialogue
Dimosthenis Kontogiorgos; Andre Pereira; Joakim Gustafson
Interaction Process Label Recognition in Group Discussion
Sixia Li; Shogo Okada; Jianwu Dang
Unsupervised Deep Fusion Cross-modal Hashing
Jiaming Huang; Chen Min; Liping Jing
DIF: Dataset of Perceived Intoxicated Faces for Drunk Person Identification
Vineet Mehta; Sai Srinadhu Katta; Devendra Pratap Yadav; Abhinav Dhall
Improved Visual Focus of Attention Estimation and Prosodic Features for Analyzing Group Interactions
Lingyu Zhang; Mallory Morgan; Indrani Bhattacharya; Michael Foley; Jonas Braasch; Christoph Riedl;
Brooke Foucault
Welles; Richard J. Radke
Engagement Modeling in Dyadic Interaction
Soumia Dermouche; Catherine Pelachaud
DeepReviewer: Collaborative Grammar and Innovation Neural Network for Automatic Paper Review
Youfang Leng; Li Yu; Jie Xiong
|
15:20 - 17:20 |
Demo & Exhibit
The Dyslexperience: Use of Projection Mapping to Simulate Dyslexia
Zi Fong Yong, Ai Ling Ng, and Yuta Nakayama
A Real-Time Scene Recognition System Based on RGB-D Video Streams
Yuyun Hua, Sixian Zhang, Xinhang Song, Jia’ning Li, and Shuqiang Jiang
Hang Out with the Language Assistant (Hola)
Sudhakar Sah, Jin-hwan Oh, Jihoon Kim, Yoori Kim, Jeonghwa Lee, Wooseung Lee,
Myeongsoo Shin, Seongwon Kim, and Jaeyon Hwang
A Searching and Automatic Video Tagging Tool for Events of Interest during
Volleyball Training Sessions
Fahim Salim, Fasih Haider, Sena Busra Yengec Tasdemir, Vahid Naghashi, Izem
Tengiz, Kubra Cengiz, Dees Postma, Robby van Delden, Dennis Reidsma,
Saturnino Luz, and Bert-Jan van Beijnum
Seeing is Believing but Feeling is the Truth
Abdenaceur Abdouni, Rory Clark, and Orestis Georgiou
Chemistry Pods: A Real Time Multimodal Tool for the Classroom
Khalil J Anderson, Steve Brenneman, Theodore Dubiel, Cody Poultney, Kenji
Tanaka, and Marcelo Worsley
A Proxemics Measurement Tool Integrated into VAIF and Unity
Aaron E Rodriguez, Adriana Camacho, Laura J. Hinojos, Mahdokht Afravi, and
David Novick
|
15:20 - 17:20 |
Grand Challenge Posters
|
17:20 - 18:20 |
Sustained Achievement Awardee Talk:
Connecting Humans with Humans: Multimodal, Multilingual, Multiparty Mediation
|
19:00 - 21:00 |
Banquet
18:30 Bus at Conference Venue
|
Thursday, 17 October 2019
09:00 - 10:00 |
Keynote 3:
Challenges of Multimodal Interaction in the Era of Human-Robot Coexistence
Dr. Zhengyou Zhang
|
10:00 - 10:30 |
Coffee Break
|
10:30 - 12:20 |
Session 5: Sound and Interaction
Session Chair: Shogo Okada JAIST |
|
"Paint that object yellow": Multimodal Interaction to Enhance Creativity During Design Tasks in VR*
Erik Wolf; Sara Klüber; Chris Zimmerer; Jean-Luc Lugrin; Marc Erich Latoschik
VCMNet: Weakly Supervised Learning for Automatic Infant Vocalisation Maturity Analysis
Najla D. Al Futaisi; Zixing Zhang; Alejandrina Cristia; Anne S. Warlaumont; Björn W. Schuller
Evidence for Communicative Compensation in Debt Advice with Reduced Multimodality
Nicole Andelic; Aidan Feeney; Gary McKeown
Speaker-Independent Speech-Driven Visual Speech Synthesis using Domain-Adapted Acoustic Models
Ahmed Hussen Abdelaziz; Barry-John Theobald; Justin Binder; Gabriele Fanelli; Paul Dixon; Nick Apostoloff; Thibaut
Weise; Sachin Kajareker
Smooth Turn-taking by a Robot Using an Online Continuous Model to Generate Turn-taking Cues
Divesh Lala; Koji Inoue; Tatsuya Kawahara
Towards Automatic Detection of Misinformation in Online Medical Videos
Rui Hou; Veronica Perez-Rosas; Stacy Loeb; Rada Mihalcea
|
12:20 - 14:00 |
Lunch
|
14:00 - 15:30 |
Session 6: Multiparty Interaction
Session Chair: Sharon Oviatt, Monash University
|
|
Modeling Team-Level Multimodal Dynamics during Multiparty Collaboration*
Lucca Eloy; Angela E.B. Stewart; Mary J. Amon; Caroline Reinhardt; Amanda Michaels; Chen Sun; Valerie Shute; Nicholas
D. Duran; Sidney K. D'Mello
Smile and Laugh Dynamics in Naturalistic Dyadic Interactions: Intensity Levels, Sequences and Roles
Kevin El Haddad; Sandeep Nallan Chakravarthula; James Kennedy
Task-independent Multimodal Prediction of Group Performance Based on Product Dimensions*
Go Miura; Shogo Okada
Emergent Leadership Detection Across Datasets
Philipp Müller; Andreas Bulling
A Multimodal Robot-Driven Meeting Facilitation System for Group Decision-Making Sessions
Ameneh Shamekhi; Timothy W. Bickmore
|
15:30 - 16:00 |
Coffee Break
|
16:00 - 16:15 |
Grand Challenge Overview
|
16:15 - 17:00 |
ICMI Town Hall Meeting
|
17:00 - 17:15 |
Presentation of ICMI 2020 & Closing Comments
|
Friday, 18 October 2019
Grand Challenge
09:00 - 12:30 |
The 7th Emotion Recognition in the Wild Challenge (EmotiW)
Organisers: Abhinav Dhall, Roland Goecke, Tom Gedeon
EmotiW 2019: Automatic Emotion, Engagement and Cohesion Prediction Tasks
Abhinav Dhall
Bootstrap Model Ensemble and Rank Loss for Engagement Intensity Regression
Kai Wang; Jianfei Yang; Da Guo; Kaipeng Zhang; Xiaojiang Peng; Yu Qiao
Exploring Regularizations with Face, Body and Image Cues for Group Cohesion
Prediction
Da Guo ; Kai Wang; Jianfei Yang; Kaipeng Zhang; Xiaojiang Peng; Yu Qiao
Exploring Emotion Features and Fusion Strategies for Audio-Video Emotion
Recognition
Hengshun Zhou; Debin Meng; Yuanyuan Zhang; Xiaojiang Peng; Jun Du; Kai Wang; Yu Qiao
Engagement Intensity Prediction with Facial Behavior Features
Van Thong Huynh; Soo-Hyung Kim; Guee-Sang Lee ; Hyung-Jeong Yang
Group-Level Cohesion Prediction using Deep Learning Models with A
Multi-stream Hybrid Network
Tien Xuan Dang; Soo-Hyung Kim; Hyung-Jeong Yang; Guee-Sang Lee
Automatic Group Cohesiveness Detection With Multi-modal Features
Bin Zhu ; Xin Guo; Kenneth Barner; Charles Boncelet
Multi-feature and Multi-instance Learning with Anti-overfitting Strategy for
Engagement Intensity Prediction
Jianming Wu ; Zhiguang Zhou; Yanan WANG; YI LI; XIN XU; Yusuke Uchida
Bi-modality Fusion for Emotion Recognition in the Wild
Sunan Li; Wenming Zheng ; Cheng Lu ; Chuangao Tang ; Xingxun Jiang ; Jiateng
Liu ; Wanchuang Xia; Yuan Zong
Multi-Attention Fusion Network for Video-based Emotion Recognition
Yanan Wang ; Jianming Wu; Keiichiro Hoashi
|
10:00 - 10:30 |
Coffee Break
|
09:00 - 10:00 |
The 1st Mandarin Audio-Visual Speech Recognition Challenge
Organisers: Shuang Yang, Weidi Xie, Xiaolong Wang, Stavros Petridis, Shiguang Shan, Xilin Chen
Spotting Visual Keywords from Temporal Sliding Windows
Yue Yao; Tianyu Wang; Heming Du; Liang Zheng; Tom Gedeon
Deep Audio-visual System for Closed-Set Word-Level Speech Recognition
Yougen Yuan; Wei Tang; Minhao Fan; Yue Cao; Peng Zhang; Lei Xie
|
10:30 - 12:30 |
The 1st Chinese Audio-Textual Spoken Language Understanding Challenge (CATSLU)
Organisers: Kai Yu, Tiejun Zhao, Chengqing Zong
CATSLU: The 1st Chinese Audio-Textual Spoken Language Understanding
Challenge
Su Zhu; Zijian Zhao; Tiejun Zhao; Chengqing Zong; Kai Yu
Transfer Learning Methods for Spoken Language Understanding
Xu Wang; Chengda Tang; Xiaotian Zhao; Xuancai Li; Zhuolin Jin; Dequan Zheng; Tiejun Zhao
Streamlined Decoder for Chinese Spoken Language Understanding
Puhai Yang; Heyan Huang; Xian-ling Mao
Robust Spoken Language Understanding with Acoustic and Domain Knowledge
Hao Li; Chen Liu; Su Zhu; Kai Yu
Multi-Classification Model for Spoken Language Understanding
Chaohong Tan; Zhenhua Ling
|
13:30 |
Tour Groups 1&2
Please look at our website for more information and you can sign up for tours at our front desk.
|
|
|
|
|