Catálogo de publicaciones - libros

Compartir en
redes sociales


SmartKom: Foundations of Multimodal Dialogue Systems

Wolfgang Wahlster (eds.)

Resumen/Descripción – provisto por la editorial

No disponible.

Palabras clave – provistas por la editorial

No disponibles.

Disponibilidad
Institución detectada Año de publicación Navegá Descargá Solicitá
No detectada 2006 SpringerLink

Información

Tipo de recurso:

libros

ISBN impreso

978-3-540-23732-7

ISBN electrónico

978-3-540-36678-2

Editor responsable

Springer Nature

País de edición

Reino Unido

Fecha de publicación

Información sobre derechos de publicación

© Springer-Verlag Berlin Heidelberg 2006

Tabla de contenidos

SmartKom-Home: The Interface to Home Entertainment

Thomas Portele; Silke Goronzy; Martin Emele; Andreas Kellner; Sunna Torge; Jürgen te Vrugt

SmartKom -Home demonstrates the use and benefit of an intelligent multimodal interface when controlling entertainment devices like a TV, a recorder, and a jukebox, and when accessing entertainment services like an electronic program guide combining speech and a handheld display with touch input. One important point is emphasizing the functional aspect, i.e., the user’s needs, conveyed to the system in a natural way by speech and gesture, are satisfied. The user does not need to know device-specific features or service idiosyncrasies. The function modeling component in SmartKom -Home has the necessary knowledge to transform the abstract user request into device commands and service queries.

Palabras clave: Dialogue System; Dynamic Content; Electronic Program Guide; Video Cassette Recorder; Speak Dialogue System.

Part V - Scenarios and Applications | Pp. 493-503

SmartKom-Mobile: Intelligent Interaction with a Mobile System

Rainer Malaka; Jochen Häußler; Hidir Aras; Matthias Merdes; Dennis Pfisterer; Matthias Jöst; Robert Porzel

This paper presents SmartKom -Mobile, the mobile version of the SmartKom system. SmartKom -Mobile brings together highly advanced user interaction and mobile computing in a novel way and allows for ubiquitous access to multidomain information. SmartKom -Mobile is device-independent and realizes multimodal interaction in cars and on mobile devices such as PDAs. With its siblings, SmartKom -Home and SmartKom -Public, it provides intelligent user interfaces for an extremely broad range of scenarios and environments.

Palabras clave: Mobile Device; Personal Digital Assistant; Multiagent System; Wireless Local Area Network; Mobile System.

Part V - Scenarios and Applications | Pp. 505-522

SmartKom-Mobile Car: User Interaction with Mobile Services in a Car Environment

André Berton; Dirk Bühler; Wolfgang Minker

People tend to spend an increasing amount of time in their cars and therefore desire high comfort, safety, and efficiency in that environment. A large variety of electronic devices has been made available to meet these requirements in the vehicle. These electronic devices should allow for speech interaction in order to minimize driver distraction and to maximize driver comfort. This contribution studies the user requirements for potential assistant functionalities operated by speech in the car. The architecture is of the dialogue system is defined based on the user requirements study. Speech dialogues were designed according to state-of-the-art principles of human machine interaction for the functionalities desired by the users. Results and ideas for future work conclude this contribution.

Palabras clave: Mobile Device; Interaction Modality; Mobile Service; Mobile Environment; Graphical Output.

Part V - Scenarios and Applications | Pp. 523-537

Wizard-of-Oz Recordings

Florian Schiel; Ulli Türk

This chapter gives a concise overview of the empirical Wizard-of-Oz recordings done within the SMARTKOM project. We define the abstract specifications of the intended simulated communicative situations, describe the necessary technical setup (including numerous useful practical hints), and finally outline the specifications of the resulting multimodal corpus, which may be obtained from the Bavarian Archive for Speech Signals (BAS).

Palabras clave: Video Track; Session Plan; Audio Track; Technical Setup; Graphical Tablet.

Part VI - Data Collection and Evaluation | Pp. 541-570

Annotation of Multimodal Data

Silke Steininger; Florian Schiel; Susen Rabold

Do users show emotions and gestures if they interact with a rather intelligent multimodal dialogue system? And if they do, what do the “emotions” and the gestures look like? Are there any features that can be exploited for their automatic detection? And finally, which language do they use when interacting with a multimodal system — does it differ from the usage of language with a monomodal dialogue system that can only understand speech? To answer these questions, data had to be collected, labeled and analyzed. This chapter deals with the second step, the transliteration and the labeling. The three main labeling steps are covered: orthographic transliteration, labeling of user states, labeling of gestures. Each step will be described with theoretical and developmental background, an overview of the label categories, and some practical advice for readers who are themselves in the process of looking for or assembling a coding system. Readers who are interested in using the presented labeling schemes should refer to the cited literature — not all details necessary for actually using the different systems are presented here for reasons of space. For information on the corpus itself, please refer to Schiel and Türk (2006).

Palabras clave: Facial Expression; User State; Spontaneous Speech; Emotional Speech; Facial Action Code System.

Part VI - Data Collection and Evaluation | Pp. 571-596

Multimodal Emogram, Data Collection and Presentation

Johann Adelhardt; Carmen Frank; Elmar Nöth; Rui Ping Shi; Viktor Zeißler; Heinrich Niemann

There are several characteristics not optimally suited for the user state classification with Wizard-of-Oz (WOZ) data like the nonuniform distribution of emotions in the utterances and the distribution of emotional utterances in speech, facial expression, and gesture. In particular, the fact that most of the data collected in the WOZ experiments are without any emotional expression gives rise to the problem of getting enough representative data for training the classifiers. Because of this problem we collected data in our own database. These data are also relevant for several demonstration sessions, where the functionality of the SmartKom system is shown in accordance with the defined use cases. In the following we first describe the system environment for data collection and then the collected data. At the end we will discuss the tool to demonstrate user states detected in the different modalities.

Part VI - Data Collection and Evaluation | Pp. 597-602

Empirical Studies for Intuitive Interaction

Iryna Gurevych; Robert Porzel

We present three types of data collections and their experimental paradigms. The resulting data were employed to conduct a number of annotation experiments, create evaluation gold standards and train statistical models. The data, experiments and their analyses highlight the importance of data-driven empirical laboratory and field work for research on intuitive multimodal human-computer interfaces.

Palabras clave: Dialogue System; Computational Linguistics; Ontological Concept; Annotate Data; Human Language Technology.

Part VI - Data Collection and Evaluation | Pp. 603-615

Evaluation of Multimodal Dialogue Systems

Florian Schiel

In this chapter we will give a brief overview about what different methods of evaluation were applied to the SmartKom prototypes and which of them resulted in utilizable results and why others did not. Since there are no established benchmarking methods yet for multimodal HCI systems and very few debatable methods for monomodal dialogue systems our work within the SmartKom prototype evaluations was rather an exploration of new methods than the simple application of standard routines.

Palabras clave: Gesture Recognition; Error Message; Speech Input; Task Success; Speech Output.

Part VI - Data Collection and Evaluation | Pp. 617-643