


default search action
Journal on Multimodal User Interfaces, Volume 14
Volume 14, Number 1, March 2020
- Vincenzo Lussu, Radoslaw Niewiadomski

, Gualtiero Volpe, Antonio Camurri:
The role of respiration audio in multimodal analysis of movement qualities. 1-15 - Wei Wei, Qingxuan Jia, Yongli Feng, Gang Chen, Ming Chu:

Multi-modal facial expression feature based on deep-neural networks. 17-23 - David Rudi, Peter Kiefer

, Ioannis Giannopoulos, Martin Raubal:
Gaze-based interactions in the cockpit of the future: a survey. 25-48 - Ahmed Housni Alsswey, Hosam Al-Samarraie

:
Elderly users' acceptance of mHealth user interface (UI) design-based culture: the moderator role of age. 49-59 - Mriganka Biswas

, Marta Romeo
, Angelo Cangelosi
, Ray Jones
:
Are older people any different from younger people in the way they want to interact with robots? Scenario based survey. 61-72 - Hiroki Tanaka

, Hidemi Iwasaka, Hideki Negoro, Satoshi Nakamura:
Analysis of conversational listening skills toward agent-based social skills training. 73-82 - Justin Mathew

, Stéphane Huot, Brian F. G. Katz:
Comparison of spatial and temporal interaction techniques for 3D audio trajectory authoring. 83-100 - Gowdham Prabhakar

, Aparna Ramakrishnan, Modiksha Madan, L. R. D. Murthy, Vinay Krishna Sharma, Sachin Deshmukh, Pradipta Biswas
:
Interactive gaze and finger controlled HUD for cars. 101-121 - Hayoung Jeong, Taeho Kang, Jiwon Choi, Jong Kim

:
A comparative assessment of Wi-Fi and acoustic signal-based HCI methods on the practicality. 123-137
Volume 14, Number 2, June 2020
- Myounghoon Jeon

, Areti Andreopoulou
, Brian F. G. Katz:
Auditory displays and auditory user interfaces: art, design, science, and research. 139-141 - Stephen Roddy

, Brian Bridges
:
Mapping for meaning: the embodied sonification listening model and its implications for the mapping problem in sonic information design. 143-151 - Joseph W. Newbold

, Nicolas E. Gold, Nadia Bianchi-Berthouze
:
Movement sonification expectancy model: leveraging musical expectancy theory to create movement-altering sonifications. 153-166 - Steven Landry, Myounghoon Jeon

:
Interactive sonification strategies for the motion and emotion of dance performances. 167-186 - Katharina Groß-Vogt

, Matthias Frank, Robert Höldrich:
Focused Audification and the optimization of its parameters. 187-198 - Rafael N. C. Patrick

, Tomasz R. Letowski, Maranda E. McBride
:
A multimodal auditory equal-loudness comparison of air and bone conducted sounds. 199-206 - Andrea Lorena Aldana Blanco

, Steffen Grautoff
, Thomas Hermann
:
ECG sonification to support the diagnosis and monitoring of myocardial infarction. 207-218 - Jindrich Matousek, Zdenek Krnoul, Michal Campr, Zbynek Zajíc, Zdenek Hanzlícek

, Martin Gruber
, Marie Kocurová:
Speech and web-based technology to enhance education for pupils with visual impairment. 219-230
Volume 14, Number 3, September 2020
- Thomas Pietrzak

, Marcelo M. Wanderley
:
Haptic and audio interaction design. 231-233 - James Leonard

, Jérôme Villeneuve, Alexandros Kontogeorgakopoulos:
Multisensory instrumental dynamics as an emergent paradigm for digital musical creation. 235-253 - Yuri De Pra

, Stefano Papetti
, Federico Fontana
, Hanna Järveläinen
, Michele Simonato:
Tactile discrimination of material properties: application to virtual buttons for professional appliances. 255-269 - Sebastian Merchel

, Mehmet Ercan Altinsoy
:
Psychophysical comparison of the auditory and tactile perception: a survey. 271-283 - Aditya Tirumala Bukkapatnam

, Philippe Depalle, Marcelo M. Wanderley
:
Defining a vibrotactile toolkit for digital musical instruments: characterizing voice coil actuators, effects of loading, and equalization of the frequency response. 285-301 - Charlotte Magnusson

, Kirsten Rassmus-Gröhn
, Bitte Rydeman
:
Developing a mobile activity game for stroke survivors - lessons learned. 303-312
Volume 14, Number 4, December 2020
- Seungwon Kim

, Mark Billinghurst
, Kangsoo Kim
:
Multimodal interfaces and communication cues for remote collaboration. 313-319 - Seungwon Kim

, Gun A. Lee
, Mark Billinghurst
, Weidong Huang
:
The combination of visual communication cues in mixed reality remote collaboration. 321-335 - Jing Yang

, Prasanth Sasikumar
, Huidong Bai
, Amit Barde
, Gábor Sörös, Mark Billinghurst
:
The effects of spatial auditory and visual cues on mixed reality remote collaboration. 337-352 - Austin Erickson

, Nahal Norouzi
, Kangsoo Kim
, Ryan Schubert
, Jonathan Jules, Joseph J. LaViola
, Gerd Bruder
, Gregory F. Welch
:
Sharing gaze rays for visual target identification tasks in collaborative augmented reality. 353-371 - Theophilus Teo

, Mitchell Norman
, Gun A. Lee
, Mark Billinghurst
, Matt Adcock:
Exploring interaction techniques for 360 panoramas inside a 3D reconstructed scene for mixed reality remote collaboration. 373-385 - Jianlong Zhou

, Simon Luo
, Fang Chen
:
Effects of personality traits on user trust in human-machine collaborations. 387-400

manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.


Google
Google Scholar
Semantic Scholar
Internet Archive Scholar
CiteSeerX
ORCID














