Demonstrations
The ICMI 2015 Demonstrations & Exhibits session is intended to provide a forum to showcase innovative implementations, systems and technologies demonstrating new ideas about interactive multimodal interfaces. They can also serve to introduce commercial products not published in previous scientific publications. Demonstrations & Exhibits should be short, so that they can be presented several times. We particularly encourage demonstration of interactive and multimodal analysis systems, and sensors. The main difference between a demonstration and an exhibit is that demonstrations include a 2-page paper, which will be included in ICMI proceedings. We encourage both the submission of early research prototypes and interesting mature systems. Proposals may be of two types: demonstrations and exhibits. In addition, authors of accepted regular research papers are invited to participate in the demonstration sessions as well.
Accepted Demo Papers
The Application of Word Processor UI paradigms to Audio and Animation Editing
Andre Milota
|
Real-time Gesture Recognition on an Economical Fabric Sensor
Xi Laura Cang, Paul Bucci, Karon MacLean
|
Public Speaking Training with a Multimodal Interactive Virtual Audience Framework
Mathieu Chollet, Kalin Stefanov, Helmut Prendinger, Stefan Scherer
|
A Multimodal System for Public Speaking with Real Time Feedback
Fiona Dermody, Alistair Sutherland
|
Model of Personality-Based, Nonverbal Behavior in Affective Virtual Humanoid Character
Maryam Saberi, Ulysses Bernardet, Steve DiPaola
|
AttentiveLearner : Adaptive Mobile MOOC Learning via Implicit Cognitive States Inference
Xiang Xiao, Phuong Pham, Jingtao Wang
|
Interactive Web-based Image Sonification for the Blind
Torsten Wörtwein, Boris Schauerte, Karin Müller, Rainer Stiefelhagen
|
Nakama: A Companion for Non-verbal Affective Communication
Christian Willemse, Gerald Munters, Jan van Erp, Dirk Heylen
|
Wir im Kiez - Multimodal App for Mutual Help Among Elderly Neighbours
Sven Schmeier, Aaron Ruß, Norbert Reithinger
|
Interact: Tightly-coupling Multimodal Dialog with an Interactive Virtual Assistant
Ethan Selfridge, Michael Johnston
|
The UTEP AGENT Framework
David Novick, Ivan Gris Sepulveda, Diego Rivera, Adriana Camacho, Alex Rayon, Mario Gutierrez
|
A Distributed Architecture for Interacting with NAO
Fabien Badeig, Quentin Pelorson, Soraya Arias, Vincent Drouard, Israel Gebru, Xiaofei Li, Georgios Evangelidis, Radu Horaud
|
Who's Speaking? Audio Supervised Classification of Active Speakers in Video
Punarjay Chakravarty
|
Multimodal Interaction with a Bifocal View on Mobile Devices
Sébastien Pelurson
|
Digital Flavor: Towards Digitally Simulating Virtual Flavors
Nimesha Ranasinghe, Gajan Suthokumar, Kuan-Yi Lee, Ellen Yi-Luen Do
|
Detecting Mastication - A Wearable Approach
Abdelkareem Bedri, Apoorva Verlekar,Edison Thomaz, Valerie Avva, Thad Starner
|
Adjacent Vehicle Collision Warning System using Image Sensor and Inertial Measurement Unit
Asif Iqbal, Carlos Busso Recabarren, Nicholas R. Gans
|
|