Task-based design and runtime support for multimodal user interface distribution

7Citations
Citations of this article
17Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

This paper describes an approach that uses task modelling for the development of distributed and multimodal user interfaces. We propose to enrich tasks with possible interaction modalities in order to allow the user to perform these tasks using an appropriate modality. The information of the augmented task model can then be used in a generic runtime architecture we have extended to support runtime decisions for distributing the user interface among several devices based on the specified interaction modalities. The approach was tested in the implementation of several case studies. One of these will be presented in this paper to clarify the approach. © 2008 Springer Berlin Heidelberg.

Cite

CITATION STYLE

APA

Clerckx, T., Vandervelpen, C., & Coninx, K. (2008). Task-based design and runtime support for multimodal user interface distribution. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 4940 LNCS, pp. 89–105). https://doi.org/10.1007/978-3-540-92698-6_6

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free