2,330 421 16MB
Pages 646 Page size 198.87 x 301.65 pts
Lecture Notes in Artificial Intelligence Edited by R. Goebel, J. Siekmann, and W. Wahlster
Subseries of Lecture Notes in Computer Science
6781
Don Harris (Ed.)
Engineering Psychology and Cognitive Ergonomics 9th International Conference, EPCE 2011 Held as Part of HCI International 2011 Orlando, FL, USA, July 9-14, 2011 Proceedings
13
Series Editors Randy Goebel, University of Alberta, Edmonton, Canada Jörg Siekmann, University of Saarland, Saarbrücken, Germany Wolfgang Wahlster, DFKI and University of Saarland, Saarbrücken, Germany Volume Editor Don Harris HFI Solutions Ltd. Bradgate Road, Bedford MK40 3DE, UK E-mail: [email protected]
ISSN 0302-9743 e-ISSN 1611-3349 ISBN 978-3-642-21740-1 e-ISBN 978-3-642-21741-8 DOI 10.1007/978-3-642-21741-8 Springer Heidelberg Dordrecht London New York Library of Congress Control Number: 2011929044 CR Subject Classification (1998): I.2.0, I.2, H.5, H.1.2, H.3, H.4.2, I.6, J.2-3 LNCS Sublibrary: SL 7 – Artificial Intelligence
© Springer-Verlag Berlin Heidelberg 2011 This work is subject to copyright. All rights are reserved, whether the whole or part of the material is concerned, specifically the rights of translation, reprinting, re-use of illustrations, recitation, broadcasting, reproduction on microfilms or in any other way, and storage in data banks. Duplication of this publication or parts thereof is permitted only under the provisions of the German Copyright Law of September 9, 1965, in its current version, and permission for use must always be obtained from Springer. Violations are liable to prosecution under the German Copyright Law. The use of general descriptive names, registered names, trademarks, etc. in this publication does not imply, even in the absence of a specific statement, that such names are exempt from the relevant protective laws and regulations and therefore free for general use. Typesetting: Camera-ready by author, data conversion by Scientific Publishing Services, Chennai, India Printed on acid-free paper Springer is part of Springer Science+Business Media (www.springer.com)
Foreword
The 14th International Conference on Human–Computer Interaction, HCI International 2011, was held in Orlando, Florida, USA, July 9–14, 2011, jointly with the Symposium on Human Interface (Japan) 2011, the 9th International Conference on Engineering Psychology and Cognitive Ergonomics, the 6th International Conference on Universal Access in Human–Computer Interaction, the 4th International Conference on Virtual and Mixed Reality, the 4th International Conference on Internationalization, Design and Global Development, the 4th International Conference on Online Communities and Social Computing, the 6th International Conference on Augmented Cognition, the Third International Conference on Digital Human Modeling, the Second International Conference on Human-Centered Design, and the First International Conference on Design, User Experience, and Usability. A total of 4,039 individuals from academia, research institutes, industry and governmental agencies from 67 countries submitted contributions, and 1,318 papers that were judged to be of high scientific quality were included in the program. These papers address the latest research and development efforts and highlight the human aspects of design and use of computing systems. The papers accepted for presentation thoroughly cover the entire field of human–computer interaction, addressing major advances in knowledge and effective use of computers in a variety of application areas. This volume, edited by Don Harris, contains papers in the thematic area of engineering psychology and cognitive ergonomics (EPCE), addressing the following major topics: • • • • • •
Cognitive and psychological aspects of interaction Cognitive aspects of driving Cognition and the web Cognition and automation Security and safety Aerospace and military applications
The remaining volumes of the HCI International 2011 Proceedings are: • Volume 1, LNCS 6761, Human–Computer Interaction—Design and Development Approaches (Part I), edited by Julie A. Jacko • Volume 2, LNCS 6762, Human–Computer Interaction—Interaction Techniques and Environments (Part II), edited by Julie A. Jacko • Volume 3, LNCS 6763, Human–Computer Interaction—Towards Mobile and Intelligent Interaction Environments (Part III), edited by Julie A. Jacko • Volume 4, LNCS 6764, Human–Computer Interaction—Users and Applications (Part IV), edited by Julie A. Jacko • Volume 5, LNCS 6765, Universal Access in Human–Computer Interaction— Design for All and eInclusion (Part I), edited by Constantine Stephanidis
VI
Foreword
• Volume 6, LNCS 6766, Universal Access in Human–Computer Interaction— Users Diversity (Part II), edited by Constantine Stephanidis • Volume 7, LNCS 6767, Universal Access in Human–Computer Interaction— Context Diversity (Part III), edited by Constantine Stephanidis • Volume 8, LNCS 6768, Universal Access in Human–Computer Interaction— Applications and Services (Part IV), edited by Constantine Stephanidis • Volume 9, LNCS 6769, Design, User Experience, and Usability—Theory, Methods, Tools and Practice (Part I), edited by Aaron Marcus • Volume 10, LNCS 6770, Design, User Experience, and Usability— Understanding the User Experience (Part II), edited by Aaron Marcus • Volume 11, LNCS 6771, Human Interface and the Management of Information—Design and Interaction (Part I), edited by Michael J. Smith and Gavriel Salvendy • Volume 12, LNCS 6772, Human Interface and the Management of Information—Interacting with Information (Part II), edited by Gavriel Salvendy and Michael J. Smith • Volume 13, LNCS 6773, Virtual and Mixed Reality—New Trends (Part I), edited by Randall Shumaker • Volume 14, LNCS 6774, Virtual and Mixed Reality—Systems and Applications (Part II), edited by Randall Shumaker • Volume 15, LNCS 6775, Internationalization, Design and Global Development, edited by P.L. Patrick Rau • Volume 16, LNCS 6776, Human-Centered Design, edited by Masaaki Kurosu • Volume 17, LNCS 6777, Digital Human Modeling, edited by Vincent G. Duffy • Volume 18, LNCS 6778, Online Communities and Social Computing, edited by A. Ant Ozok and Panayiotis Zaphiris • Volume 19, LNCS 6779, Ergonomics and Health Aspects of Work with Computers, edited by Michelle M. Robertson • Volume 20, LNAI 6780, Foundations of Augmented Cognition: Directing the Future of Adaptive Systems, edited by Dylan D. Schmorrow and Cali M. Fidopiastis • Volume 22, CCIS 173, HCI International 2011 Posters Proceedings (Part I), edited by Constantine Stephanidis • Volume 23, CCIS 174, HCI International 2011 Posters Proceedings (Part II), edited by Constantine Stephanidis I would like to thank the Program Chairs and the members of the Program Boards of all Thematic Areas, listed herein, for their contribution to the highest scientific quality and the overall success of the HCI International 2011 Conference. In addition to the members of the Program Boards, I also wish to thank the following volunteer external reviewers: Roman Vilimek from Germany, Ramalingam Ponnusamy from India, Si Jung “Jun” Kim from the USA, and Ilia Adami, Iosif Klironomos, Vassilis Kouroumalis, George Margetis, and Stavroula Ntoa from Greece.
Foreword
VII
This conference would not have been possible without the continuous support and advice of the Conference Scientific Advisor, Gavriel Salvendy, as well as the dedicated work and outstanding efforts of the Communications and Exhibition Chair and Editor of HCI International News, Abbas Moallem. I would also like to thank for their contribution toward the organization of the HCI International 2011 Conference the members of the Human–Computer Interaction Laboratory of ICS-FORTH, and in particular Margherita Antona, George Paparoulis, Maria Pitsoulaki, Stavroula Ntoa, Maria Bouhli and George Kapnas. July 2011
Constantine Stephanidis
Organization
Ergonomics and Health Aspects of Work with Computers Program Chair: Michelle M. Robertson Arne Aar˚ as, Norway Pascale Carayon, USA Jason Devereux, UK Wolfgang Friesdorf, Germany Martin Helander, Singapore Ed Israelski, USA Ben-Tzion Karsh, USA Waldemar Karwowski, USA Peter Kern, Germany Danuta Koradecka, Poland Nancy Larson, USA Kari Lindstr¨om, Finland
Brenda Lobb, New Zealand Holger Luczak, Germany William S. Marras, USA Aura C. Matias, Philippines Matthias R¨ otting, Germany Michelle L. Rogers, USA Dominique L. Scapin, France Lawrence M. Schleifer, USA Michael J. Smith, USA Naomi Swanson, USA Peter Vink, The Netherlands John Wilson, UK
Human Interface and the Management of Information Program Chair: Michael J. Smith Hans-J¨ org Bullinger, Germany Alan Chan, Hong Kong Shin’ichi Fukuzumi, Japan Jon R. Gunderson, USA Michitaka Hirose, Japan Jhilmil Jain, USA Yasufumi Kume, Japan Mark Lehto, USA Hirohiko Mori, Japan Fiona Fui-Hoon Nah, USA Shogo Nishida, Japan Robert Proctor, USA
Youngho Rhee, Korea Anxo Cereijo Roib´ as, UK Katsunori Shimohara, Japan Dieter Spath, Germany Tsutomu Tabe, Japan Alvaro D. Taveira, USA Kim-Phuong L. Vu, USA Tomio Watanabe, Japan Sakae Yamamoto, Japan Hidekazu Yoshikawa, Japan Li Zheng, P. R. China
X
Organization
Human–Computer Interaction Program Chair: Julie A. Jacko Sebastiano Bagnara, Italy Sherry Y. Chen, UK Marvin J. Dainoff, USA Jianming Dong, USA John Eklund, Australia Xiaowen Fang, USA Ayse Gurses, USA Vicki L. Hanson, UK Sheue-Ling Hwang, Taiwan Wonil Hwang, Korea Yong Gu Ji, Korea Steven A. Landry, USA
Gitte Lindgaard, Canada Chen Ling, USA Yan Liu, USA Chang S. Nam, USA Celestine A. Ntuen, USA Philippe Palanque, France P.L. Patrick Rau, P.R. China Ling Rothrock, USA Guangfeng Song, USA Steffen Staab, Germany Wan Chul Yoon, Korea Wenli Zhu, P.R. China
Engineering Psychology and Cognitive Ergonomics Program Chair: Don Harris Guy A. Boy, USA Pietro Carlo Cacciabue, Italy John Huddlestone, UK Kenji Itoh, Japan Hung-Sying Jing, Taiwan Wen-Chin Li, Taiwan James T. Luxhøj, USA Nicolas Marmaras, Greece Sundaram Narayanan, USA Mark A. Neerincx, The Netherlands
Jan M. Noyes, UK Kjell Ohlsson, Sweden Axel Schulte, Germany Sarah C. Sharples, UK Neville A. Stanton, UK Xianghong Sun, P.R. China Andrew Thatcher, South Africa Matthew J.W. Thomas, Australia Mark Young, UK Rolf Zon, The Netherlands
Universal Access in Human–Computer Interaction Program Chair: Constantine Stephanidis Julio Abascal, Spain Ray Adams, UK Elisabeth Andr´e, Germany Margherita Antona, Greece Chieko Asakawa, Japan Christian B¨ uhler, Germany Jerzy Charytonowicz, Poland Pier Luigi Emiliani, Italy
Michael Fairhurst, UK Dimitris Grammenos, Greece Andreas Holzinger, Austria Simeon Keates, Denmark Georgios Kouroupetroglou, Greece Sri Kurniawan, USA Patrick M. Langdon, UK Seongil Lee, Korea
Organization
Zhengjie Liu, P.R. China Klaus Miesenberger, Austria Helen Petrie, UK Michael Pieper, Germany Anthony Savidis, Greece Andrew Sears, USA Christian Stary, Austria
Hirotada Ueda, Japan Jean Vanderdonckt, Belgium Gregg C. Vanderheiden, USA Gerhard Weber, Germany Harald Weber, Germany Panayiotis Zaphiris, Cyprus
Virtual and Mixed Reality Program Chair: Randall Shumaker Pat Banerjee, USA Mark Billinghurst, New Zealand Charles E. Hughes, USA Simon Julier, UK David Kaber, USA Hirokazu Kato, Japan Robert S. Kennedy, USA Young J. Kim, Korea Ben Lawson, USA Gordon McK Mair, UK
David Pratt, UK Albert “Skip” Rizzo, USA Lawrence Rosenblum, USA Jose San Martin, Spain Dieter Schmalstieg, Austria Dylan Schmorrow, USA Kay Stanney, USA Janet Weisenford, USA Mark Wiederhold, USA
Internationalization, Design and Global Development Program Chair: P.L. Patrick Rau Michael L. Best, USA Alan Chan, Hong Kong Lin-Lin Chen, Taiwan Andy M. Dearden, UK Susan M. Dray, USA Henry Been-Lirn Duh, Singapore Vanessa Evers, The Netherlands Paul Fu, USA Emilie Gould, USA Sung H. Han, Korea Veikko Ikonen, Finland Toshikazu Kato, Japan Esin Kiris, USA Apala Lahiri Chavan, India
James R. Lewis, USA James J.W. Lin, USA Rungtai Lin, Taiwan Zhengjie Liu, P.R. China Aaron Marcus, USA Allen E. Milewski, USA Katsuhiko Ogawa, Japan Oguzhan Ozcan, Turkey Girish Prabhu, India Kerstin R¨ ose, Germany Supriya Singh, Australia Alvin W. Yeo, Malaysia Hsiu-Ping Yueh, Taiwan
XI
XII
Organization
Online Communities and Social Computing Program Chairs: A. Ant Ozok, Panayiotis Zaphiris Chadia N. Abras, USA Chee Siang Ang, UK Peter Day, UK Fiorella De Cindio, Italy Heidi Feng, USA Anita Komlodi, USA Piet A.M. Kommers, The Netherlands Andrew Laghos, Cyprus Stefanie Lindstaedt, Austria Gabriele Meiselwitz, USA Hideyuki Nakanishi, Japan
Anthony F. Norcio, USA Ulrike Pfeil, UK Elaine M. Raybourn, USA Douglas Schuler, USA Gilson Schwartz, Brazil Laura Slaughter, Norway Sergei Stafeev, Russia Asimina Vasalou, UK June Wei, USA Haibin Zhu, Canada
Augmented Cognition Program Chairs: Dylan D. Schmorrow, Cali M. Fidopiastis Monique Beaudoin, USA Chris Berka, USA Joseph Cohn, USA Martha E. Crosby, USA Julie Drexler, USA Ivy Estabrooke, USA Chris Forsythe, USA Wai Tat Fu, USA Marc Grootjen, The Netherlands Jefferson Grubb, USA Santosh Mathan, USA
Rob Matthews, Australia Dennis McBride, USA Eric Muth, USA Mark A. Neerincx, The Netherlands Denise Nicholson, USA Banu Onaral, USA Kay Stanney, USA Roy Stripling, USA Rob Taylor, UK Karl van Orden, USA
Digital Human Modeling Program Chair: Vincent G. Duffy Karim Abdel-Malek, USA Giuseppe Andreoni, Italy Thomas J. Armstrong, USA Norman I. Badler, USA Fethi Calisir, Turkey Daniel Carruth, USA Keith Case, UK Julie Charland, Canada
Yaobin Chen, USA Kathryn Cormican, Ireland Daniel A. DeLaurentis, USA Yingzi Du, USA Okan Ersoy, USA Enda Fallon, Ireland Yan Fu, P.R. China Afzal Godil, USA
Organization
Ravindra Goonetilleke, Hong Kong Anand Gramopadhye, USA Lars Hanson, Sweden Pheng Ann Heng, Hong Kong Bo Hoege, Germany Hongwei Hsiao, USA Tianzi Jiang, P.R. China Nan Kong, USA Steven A. Landry, USA Kang Li, USA Zhizhong Li, P.R. China Tim Marler, USA
XIII
Ahmet F. Ozok, Turkey Srinivas Peeta, USA Sudhakar Rajulu, USA Matthias R¨ otting, Germany Matthew Reed, USA Johan Stahre, Sweden Mao-Jiun Wang, Taiwan Xuguang Wang, France Jingzhou (James) Yang, USA Gulcin Yucel, Turkey Tingshao Zhu, P.R. China
Human-Centered Design Program Chair: Masaaki Kurosu Julio Abascal, Spain Simone Barbosa, Brazil Tomas Berns, Sweden Nigel Bevan, UK Torkil Clemmensen, Denmark Susan M. Dray, USA Vanessa Evers, The Netherlands Xiaolan Fu, P.R. China Yasuhiro Horibe, Japan Jason Huang, P.R. China Minna Isomursu, Finland Timo Jokela, Finland Mitsuhiko Karashima, Japan Tadashi Kobayashi, Japan Seongil Lee, Korea Kee Yong Lim, Singapore
Zhengjie Liu, P.R. China Lo¨ıc Mart´ınez-Normand, Spain Monique Noirhomme-Fraiture, Belgium Philippe Palanque, France Annelise Mark Pejtersen, Denmark Kerstin R¨ ose, Germany Dominique L. Scapin, France Haruhiko Urokohara, Japan Gerrit C. van der Veer, The Netherlands Janet Wesson, South Africa Toshiki Yamaoka, Japan Kazuhiko Yamazaki, Japan Silvia Zimmermann, Switzerland
Design, User Experience, and Usability Program Chair: Aaron Marcus Ronald Baecker, Canada Barbara Ballard, USA Konrad Baumann, Austria Arne Berger, Germany Randolph Bias, USA Jamie Blustein, Canada
Ana Boa-Ventura, USA Lorenzo Cantoni, Switzerland Sameer Chavan, Korea Wei Ding, USA Maximilian Eibl, Germany Zelda Harrison, USA
XIV
Organization
R¨ udiger Heimg¨artner, Germany Brigitte Herrmann, Germany Sabine Kabel-Eckes, USA Kaleem Khan, Canada Jonathan Kies, USA Jon Kolko, USA Helga Letowt-Vorbek, South Africa James Lin, USA Frazer McKimm, Ireland Michael Renner, Switzerland
Christine Ronnewinkel, Germany Elizabeth Rosenzweig, USA Paul Sherman, USA Ben Shneiderman, USA Christian Sturm, Germany Brian Sullivan, USA Jaakko Villa, Finland Michele Visciola, Italy Susan Weinschenk, USA
HCI International 2013
The 15th International Conference on Human–Computer Interaction, HCI International 2013, will be held jointly with the affiliated conferences in the summer of 2013. It will cover a broad spectrum of themes related to human–computer interaction (HCI), including theoretical issues, methods, tools, processes and case studies in HCI design, as well as novel interaction techniques, interfaces and applications. The proceedings will be published by Springer. More information about the topics, as well as the venue and dates of the conference, will be announced through the HCI International Conference series website: http://www.hci-international.org/ General Chair Professor Constantine Stephanidis University of Crete and ICS-FORTH Heraklion, Crete, Greece Email: [email protected]
Table of Contents
Part I: Cognitive and Psychological Aspects of Interaction Movement Time for Different Input Devices . . . . . . . . . . . . . . . . . . . . . . . . . L. Paige Bacon and Kim-Phuong L. Vu
3
Audio and Audiovisual Cueing in Visual Search: Effects of Target Uncertainty and Auditory Cue Precision . . . . . . . . . . . . . . . . . . . . . . . . . . . . Hugo Bertolotti and Thomas Z. Strybel
10
Interpretation of Metaphors with Perceptual Features Using WordNet . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Rini Bhatt, Amitash Ojha, and Bipin Indurkhya
21
Acoustic Correlates of Deceptive Speech – An Exploratory Study . . . . . . David M. Howard and Christin Kirchh¨ ubel
28
Kansei Evaluation of HDR Color Images with Different Tone Curves and Sizes – Foundational Investigation of Difference between Japanese and Chinese . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Tomoharu Ishikawa, Yunge Guan, Yi-Chun Chen, Hisashi Oguro, Masao Kasuga, and Miyoshi Ayama
38
Personalized Emotional Prediction Method for Real-Life Objects Based on Collaborative Filtering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Hyeong-Joon Kwon, Hyeong-Oh Kwon, and Kwang-Seok Hong
45
A Study of Vision Ergonomic of LED Display Signs on Different Environment Illuminance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Jeih-Jang Liou, Li-Lun Huang, Chih-Fu Wu, Chih-Lung Yeh, and Yung-Hsiang Chen Spatial Tasks on a Large, High-Resolution, Tiled Display: A Male Inferiority in Performance with a Mental Rotation Task . . . . . . . . . . . . . . . Bernt Ivar Olsen, Bruno Laeng, Kari-Ann Kristiansen, and Gunnar Hartvigsen
53
63
Modeling Visual Attention for Rule-Based Usability Simulations of Elderly Citizen . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Aaron Ruß
72
Operational Characteristics Related to Memory in Operating Information Devices with Hierarchical Menu . . . . . . . . . . . . . . . . . . . . . . . . . Norikazu Sasaki, Motoki Shino, and Minoru Kamata
82
XVIII
Table of Contents
Effects of Paper on Page Turning: Comparison of Paper and Electronic Media in Reading Documents with Endnotes . . . . . . . . . . . . . . . . . . . . . . . . Hirohito Shibata and Kengo Omura Viability of Mobile Devices for Training Purposes . . . . . . . . . . . . . . . . . . . . Shehan Sirigampola, Steven Zielinski, Glenn A. Martin, Jason Daly, and Jaime Flores Display System for Advertising Image with Scent and Psychological Effect . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Keisuke Tomono, Hiroki Wakatsuki, Shigeki Kumazawa, and Akira Tomono
92
102
110
Subitizing-Counting Analogue Observed in a Fast Multi-tapping Task . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Hiro-Fumi Yanai, Kyouhei Kurosawa, and Kousuke Takahashi
120
The Number of Trials with Target Affects the Low Prevalence Effect . . . Fan Yang, Xianghong Sun, Kan Zhang, and Biyun Zhu
126
Part II: Cognitive Aspects of Driving Facial Expression Measurement for Detecting Driver Drowsiness . . . . . . . Satori Hachisuka, Kenji Ishida, Takeshi Enya, and Masayoshi Kamijo
135
Estimation of Driver’s Fatigue Based on Steering Wheel Angle . . . . . . . . . Qichang He, Wei Li, and Xiumin Fan
145
Study on Driving Performance of Aged Drivers at the Intersections Compared with Young Drivers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Seunghee Hong, Byungchan Min, and Shun’ichi Doi
156
The Influence of False and Missing Alarms of Safety System on Drivers’ Risk-Taking Behavior . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Takayuki Masuda, Shigeru Haga, Azusa Aoyama, Hiroki Takahashi, and Gaku Naito Estimation of Driver’s Arousal State Using Multi-dimensional Physiological Indices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Mieko Ohsuga, Yoshiyuki Kamakura, Yumiko Inoue, Yoshihiro Noguchi, Kenji Shimada, and Masami Mishiro The Effects of Visual and Cognitive Distraction on Driver Situation Awareness . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Meghan Rogers, Yu Zhang, David Kaber, Yulan Liang, and Shruti Gangakhedkar
167
176
186
Table of Contents
Experienced and Novice Driver Situation Awareness at Rail Level Crossings: An Exploratory On-Road Study . . . . . . . . . . . . . . . . . . . . . . . . . . Paul M. Salmon, Michael G. Lenn´e, Kristie Young, and Guy Walker Influence of Brightness and Traffic Flow on Driver’s Eye-FixationRelated Potentials . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Yoshihisa Terada, Koji Morikawa, Yuji Kawanishi, YongWook Jeon, and Tatsuru Daimon Cognitive Compatibility of Motorcyclists and Drivers . . . . . . . . . . . . . . . . . Guy H. Walker, Neville A. Stanton, and Paul M. Salmon
XIX
196
205
214
Part III: Cognition and the Web Information Searching on the Web: The Cognitive Difficulties Experienced by Older Users in Modifying Unsuccessful Information Searches . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Aline Chevalier, Aur´elie Dommes, and Jean-Claude Marqui´e
225
Template for Website Browsing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Fong-Ling Fu and Chiu Hung Su
233
Mental Models: Have Users’ Mental Models of Web Search Engines Improved in the Last Ten Years? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Sifiso Mlilo and Andrew Thatcher
243
The e-Progression in SEs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Karl W. Sandberg, Olof Wahlberg, and Fredrik H˚ akansson
254
Cross-Cultural Comparison of Blog Use for Parent-Teacher Communication in Elementary Schools . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Qiping Zhang and April Hatcher
263
How Font Size and Tag Location Influence Chinese Perception of Tag Cloud? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Qiping Zhang, Weina Qu, and Li Wang
273
Part IV: Cognition and Automation Balance between Abstract Principles and Concrete Instances in Knowledge Communication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Toshiya Akasaka and Yusaku Okada
285
Using Uncertainty to Inform Information Sufficiency in Decision Making . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Xiao Dong and Caroline C. Hayes
294
XX
Table of Contents
Consideration of Human Factors for Prioritizing Test Cases for the Software System Test . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Christoph Malz, Kerstin Sommer, Peter G¨ ohner, and Birgit Vogel-Heuser Cognitive Engineering of Automated Assembly Processes . . . . . . . . . . . . . Marcel Ph. Mayer, Barbara Odenthal, Carsten Wagels, Sinem Kuz, Bernhard Kausch, and Christopher M. Schlick Delegation to Automation: Performance and Implications in Non-optimal Situations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Christopher A. Miller, Tyler H. Shaw, Joshua D. Hamell, Adam Emfield, David J. Musliner, Ewart de Visser, and Raja Parasurman Effective Shift Handover . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Thomas Plocher, Shanqing Yin, Jason Laberge, Brian Thompson, and Jason Telner Measuring Self-adaptive UAV Operators’ Load-Shedding Strategies under High Workload . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Axel Schulte and Diana Donath Display Requirements for an Interactive Rail Scheduling Display . . . . . . . Jacqueline M. Tappan, David J. Pitman, Mary L. Cummings, and Denis Miglianico
303
313
322
332
342 352
Part V: Security and Safety Application of Natural Language in Fire Spread Display . . . . . . . . . . . . . . Yan Ge, Li Wang, and Xianghong Sun
365
Differences between Students and Professionals While Using a GPS Cased GIS in an Emergency Response Study . . . . . . . . . . . . . . . . . . . . . . . . Rego Granlund, Helena Granlund, and Nils Dahlb¨ ack
374
Adversarial Behavior in Complex Adaptive Systems: An Overview of ICST’s Research on Competitive Adaptation in Militant Networks . . . . . John Horgan, Michael Kenney, Mia Bloom, Cale Horne, Kurt Braddock, Peter Vining, Nicole Zinni, Kathleen Carley, and Michael Bigrigg
384
Preferred Temporal Characteristics of an Advance Notification System for Autonomous Powered Wheelchair . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Takuma Ito and Minoru Kamata
394
Pre-validation of Nuclear Power Plant Control Room Design . . . . . . . . . . Jari Laarni, Paula Savioja, Hannu Karvonen, and Leena Norros
404
Table of Contents
Deception and Self-awareness . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Glyn Lawson, Alex Stedmon, Chloe Zhang, Dawn Eubanks, and Lara Frumkin Air Passengers’ Luggage Screening: What Is the Difference between Na¨ıve People and Airport Screeners? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Xi Liu and Alastair Gale Acceptability and Effects of Tools to Assist with Controller Managed Spacing in the Terminal Area . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Lynne Martin, Michael Kupfer, Everett Palmer, Joey Mercer, Todd Callantine, and Thomas Prevˆ ot The Effects of Individual and Context on Aggression in Repeated Social Interaction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Jolie M. Martin, Ion Juvina, Christian Lebiere, and Cleotilde Gonzalez Scent Trails: Countering Terrorism through Informed Surveillance . . . . . . Alex Sandham, Tom Ormerod, Coral Dando, Ray Bull, Mike Jackson, and James Goulding Using Behavioral Measures to Assess Counter-Terrorism Training in the Field . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . V. Alan Spiker and Joan H. Johnston
XXI
414
424
432
442
452
461
Part VI: Aerospace and Military Applications Applied Cognitive Ergonomics Design Principles for Fighter Aircraft . . . Jens Alfredson, Johan Holmberg, Rikard Andersson, and Maria Wikforss Designing Effective Soldier-Robot Teams in Complex Environments: Training, Interfaces, and Individual Differences . . . . . . . . . . . . . . . . . . . . . . Michael J. Barnes, Jessie Y.C. Chen, Florian Jentsch, and Elizabeth S. Redden
473
484
Optimizing Performance Variables for Small Unmanned Aerial Vehicle Co-axial Rotor Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Jonathon Bell, Mantas Brazinskas, and Stephen Prior
494
Trust Evaluation through Human-Machine Dialogue Modelling . . . . . . . . Cyril Crocquesel, Fran¸cois Legras, and Gilles Coppin
504
A Testbed for Exploring Human-Robot Interaction with Unmanned Aerial and Ground Vehicles . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Jaime H. Flores, Glenn A. Martin, and Paula J. Durlach
514
XXII
Table of Contents
Technological and Usability-Based Aspects of Distributed After Action Review in a Game-Based Training Setting . . . . . . . . . . . . . . . . . . . . . . . . . . . Matthew Fontaine, Glenn A. Martin, Jason Daly, and Casey Thurston Authority Sharing in Mixed Initiative Control of Multiple Uninhabited Aerial Vehicles . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Rui Gon¸calves, S´ergio Ferreira, Jos´e Pinto, Jo˜ ao Sousa, and Gil Gon¸calves
522
530
Enhancing Pilot Training with Advanced Measurement Techniques . . . . . Kelly S. Hale and Robert Breaux
540
Rule Fragmentation in the Airworthiness Regulations: A Human Factors Perspective . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Don Harris
546
Development of a Reconfigurable Protective System for Multi-rotor UAS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Thomas Irps, Stephen Prior, and Darren Lewis
556
Test-Retest Reliability of CogGauge: A Cognitive Assessment Tool for SpaceFlight . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Matthew Johnston, Angela Carpenter, and Kelly Hale
565
A Formalism for Assessing the Situation Awareness of Pilots . . . . . . . . . . Steven J. Landry and Chittayong Surakitbanharn
572
Mental Resource Demands Prediction as a Key Element for Future Assistant Systems in Military Helicopters . . . . . . . . . . . . . . . . . . . . . . . . . . . Felix Maiwald and Axel Schulte
582
Analysis of Mental Workload during En-Route Air Traffic Control Task Execution Based on Eye-Tracking Technique . . . . . . . . . . . . . . . . . . . . . . . . Caroline Martin, Julien Cegarra, and Philippe Averty
592
An After Action Review Engine for Training in Multiple Areas . . . . . . . . Glenn A. Martin, Jason Daly, and Casey Thurston
598
Mixed-Initiative Multi-UAV Mission Planning by Merging Human and Machine Cognitive Skills . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Ruben Strenzke and Axel Schulte
608
Exploring the Relationship among Dimensions of Flight Comprehensive Capabilities Based on SEM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Ruishan Sun and Yang Li
618
A Generic After Action Review Capability for Game-Based Training . . . Casey L. Thurston and Glenn A. Martin
628
Author Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
635
Movement Time for Different Input Devices L. Paige Bacon and Kim-Phuong L. Vu California State University, Long Beach 1250 Bellflower Blvd Long Beach, CA 90840, USA [email protected], [email protected]
Abstract. Fitts’ law states that movement time can be predicted by knowing the size of a target to which a person is intending to move and the distance to be moved. The current study measured choice-movement time with three input devices commonly used in human-computer interaction tasks: response panel, computer mouse, and touch-screen. We also examined how direction of movement with the different input devices influences performance. Movement time was shorter when responses were made with the response panel than with the mouse and touch-screen. Furthermore, horizontal movement time was faster than vertical movement time, even when the size of the stimuli and distance to be moved were equal. Fitts’ law was used to estimate the slope and intercepts of the functions for each input device and dimension to determine whether the devices and dimensions had greater influence on the starting time or the speed of execution. Keywords: Fitts’ law, input device, movement time, display-control compatibility.
1 Introduction When a person interacts with his or her daily environment, it is likely that s/he will perform aimed movements. Examples of aimed movements include a) using a computer mouse to move a cursor to a menu item in order to select that item, b) moving a hand from the steering wheel of a vehicle to press a button on the control panel to change the radio station, and c) moving a hand from its current position to touch an icon on the screen of an iPhone to launch an application. Fitts’ law states that movement time (MT) can be predicted by knowing the size the target and the distance from the target [1]. The shorter the distance and the larger the target, the faster the movement time will be. Fitts’ law refers to a mathematical relationship between movement speed and accuracy of movement. In 1954, Paul Fitts began exploring this relationship by having participants tap a metal, hand-held stylus onto two metal plates as many times as possible in a given period of time. During the experiment, Fitts also manipulated the size of the plates (target width, W) and the distance between them (amplitude, A), which allowed him to test performance across a variety of combinations of target sizes and distances. The participants were instructed to consider accuracy of movement as a higher priority than speed of movement. Performance was measured by the number of taps that could be completed in a given 15-second trial. Fitts and Peterson [2] continued to investigate this relationship, and concluded that movement D. Harris (Ed.): Engin. Psychol. and Cog. Ergonomics, HCII 2011, LNAI 6781, pp. 3–9, 2011. © Springer-Verlag Berlin Heidelberg 2011
4
L.P. Bacon and K.-P.L. Vu
time (MT) was a function of the difficulty of the movement, expressed as a logarithmic ratio of amplitude and width, known as the index of difficulty (ID). Fitts’ law is expressed as MT = a + b (ID), where a and b are empirically derived constants. Fitts’ law is important to the field of HCI because many tasks involve aimed movements. Fitts’ law is robust, having been shown to hold for movements of the head [3] and feet [4], as well as for movements made underwater [5] and with remote manipulation [6]. Thus, many designers will benefit from using Fitts’ law to estimate movement times. In particular, designers can benefit from using Fitts’ law to determine how a device or movement direction affects performance. Because the constants in Fitts’ law (a and b) represent properties of the device being tested, knowing these values can help designers diagnose where the cost in movement times with different devices originates. The constant, a (the intercept) is indicative of the time needed to start the movement with the device. Constant b, (the slope) is indicative of the inherent movement speed using the device. Stimulus-response (S-R) compatibility, or how natural a response to a stimulus is, can also influence the slope of the Fitts’ law function [7], where a higher degree of compatibility reduces the slope. There is also some evidence from the S-R compatibility literature suggesting that, in many situations, reaction time is faster for horizontal than vertical S-R relations, a phenomenon known as the right-left prevalence effect [8, 9, 10]. However, it is not known whether the advantage for the horizontal dimension would be evident in movement time. Thus, the goal of the current study was to examine whether the direction of movement in a choice reaction task has an effect on movement time with different input devices used in HCI tasks. Three input devices were examined: response panel, computer mouse, and touch screen. All response devices were mapped compatibly to the stimulus display. The response devices differed with respect to their integration with the display. The touch screen was the most integrated input device because participants touched the target to make a response. The computer mouse was less integrated because participants moved the mouse to produce cursor movement to a target that was not on the same plane. Finally, the response panel was the least integrated because participants pressed a corresponding button on the response panel that was separate from the display.
2 Method 2.1 Participants Forty-four undergraduate students (24 women, 20 men), ages 18-39 (M = 20.14 years), enrolled in an Introductory Psychology course at California State University, Long Beach participated the study. Participants were recruited from the Psychology Subject Pool and received credit toward their Introductory Psychology requirement. Each participant completed a demographic questionnaire after completing the experimental trials. Participants reported using a computer an average of 20.5 hours per week. Participants were also asked to use a scale of 1 to 7, with 1 being no experience and 7
Movement Time for Different Input Devices
5
being very experienced, to rate their experience with different input devices. Participants rated their experience to be greatest for the computer mouse (M = 6.6) and a computer keyboard (M = 6.5), and least for a touch screen (M = 5.0). 2.2 Design A 2 (distance: near and far) x 2 (dimension: horizontal and vertical) x 3 (device: mouse, touch screen, and response panel) within-subjects factorial design was used. The device variable was counterbalanced across participants prior to their arrival for the experiment. 2.3 Materials The experimental apparatus consisted of an Asus Eee Top touch screen computer running Microsoft Windows XP Home Edition, a Microsoft IntelliMouse Explore 3.0, which was used during the mouse device condition. An Ergodex® DX1 Input System Panel was used in the response panel device condition. The response buttons were compatibly mapped with the stimuli that were presented on the screen. A custom program written using Microsoft Visual Basic 2008 Express Edition controlled the experiment and collected the data. The touch screen computer was located on a table that was 154 cm wide, 76 cm deep, with a height of 75 cm from the floor. The near edge of the table was 23 cm from the base of the computer. 2.4 Procedure The experiment was conducted in a single session lasting approximately 45 minutes. Participants were seated at a table and given two copies of the informed consent form. All agreed to participate and signed the informed consent form, after which the experiment began. For each condition, upon the press and release of the “Start Trial” button by the participant (located in the middle of the screen; see Figure 1), a target randomly appeared in one of four directions (above, below, left or right) and at one of two distances (near or far) from the “Start Trial” button. Participants were instructed to move to the stimulus once they had identified the target. The trial ended when the participant responded at the location of the target. Movement time was measured from the time that the participant clicked on the “Start Trial” button until they selected the target. Ten practice trials with each device were given prior to beginning data collection. The experimental block consisted of 40 trials per stimulus location at the two distances (320 trials per device). The target consisted of a 1.27 cm by 1.27 cm black, square that appeared on a tan background. Participants were given rest periods between device conditions and were also told that they could rest at any time during data collection as long as there was no target on the screen. At the completion of the experimental trials, participants completed a demographic questionnaire. Finally, they were debriefed and thanked for their participation.
6
L.P. Bacon and K.-P.L. Vu
Fig. 1. Task Display. Participants clicked on the start trial button (or a response button located in the same spatial position on a response panel) to start the trial. Then, one target stimulus (depicted by the black squares) appeared, and participants were to respond by moving to the location of the target.
3 Results A 2 (distance: near and far) x 2 (dimension: horizontal and vertical) x 3 (device: mouse, touch screen, and response panel) analysis of variance was conducted on mean movement time for each participant in each condition. There was a significant main effect of distance, F (1,43) = 398, p < .001, η2 = .90. Consistent with Fitts’ law, movement time was longer for further distances than nearer distances (M = 520 ms, SE = 15.9; M = 400 ms, SE = 12.7). A significant main effect of dimension, F (1,43) = 20.4, p < .001, η2 = .32, was also obtained in which movement time for the horizontal dimension was faster (M = 450 ms, SE = 13.52) than movement time for the vertical dimension (M = 472 ms, SE = 15.1). The main of effects of distance and dimension were qualified by a significant distance x dimension interaction, F (1,43) = 4.4, p < .05, η2 = .09, see Figure 2. Tests of simple effects were performed with the Bonferroni correction. Movement time was significantly different across both factors; the shortest movement time occurred when the target was near and in the horizontal direction (M = 393 ms, SE = 12.4), followed by when the target was near and in the vertical direction (M = 409 ms, SE = 13.3), followed by when the target was far and in the horizontal direction (M = 507 ms, SE = 15.4), and was longest when the target was far and in the vertical direction
Movement Time for Different Input Devices
7
(M = 535.64 ms, SE = 17.19), p < .001. In other words, this interaction reflects the fact that horizontal movement time was less affected by distance than vertical movement time.
Far
Near
Distance Fig. 2. Movement time as a function of distance and dimension
A significant main effect of input device was also found, F (2,42) = 112, p < .001, η2 = .84, with movement time using the mouse being the longest (M = 569 ms, SE = 15.5), followed by movement time using the touch screen (M = 456 ms, SE = 18.4), and movement time using the response panel being the shortest (M = 358 ms, SE = 14.1). However, this main effect was qualified by a significant distance x device interaction, F (2,42) = 39, p < .001, η2 = .66, see Figure 3. A post-hoc analysis using the Bonferroni correction was conducted to determine if there was a significant difference in the movement time with the different input devices when the target was near compared to when it was far. When the target was near, there was a significant difference in movement time across devices, where movement time was longest using the mouse (M = 492 ms, SE = 14.7), intermediate when using the touch screen (M = 393 ms, SE = 15.9), and shortest using the response panel (M = 317 ms, SE = 13.2). The same pattern held when the distance was far, where movement time for the mouse was still longest (M = 647 ms, SE = 17.4), followed by the touch screen (M = 518 ms, SE = 21.3), and response panel (M = 399 ms, SE = 15.7). However, the difference between devices was larger at the farther distance than at the nearer distance. Because dimension did not interact with input device, Fitts’ law was used to derive the constants a and b for each dimension (across device) and each input device (across dimension). The equations for estimating MT according to Fitts’ law for each dimension and input device are presented in Figures 2 and 3, respectively.
8
L.P. Bacon and K.-P.L. Vu
Far
Near
Distance Fig. 3. Movement time as a function of distance and device
4 Discussion The Fitts’ law functions were different for each dimension and input device. For the dimension variable, movement time was found to be shorter for the horizontal than vertical dimension. Upon examining the slope of Fitts’ law function, there was an 11% increase from the horizontal dimension to the vertical one. This difference may reflect easier motor control along the horizontal plane than vertical one. However, the increase in the slope of the functions was smaller than the increase observed in the intercept (73%), which is indicative of the time needed to start the movement along the horizontal and vertical dimensions. Because the experimental paradigm was a choice aimed-movement task, the difference may reflect that it takes longer to select responses along the vertical dimension than the horizontal one. This finding is consistent with other choice-reaction studies [11] that showed overall RT for horizontally arrayed S-R sets to be shorter than for vertically arrayed S-R sets. For the input devices, the response panel yielded the shortest MT, followed by the touch screen, and then the computer mouse. When examining Fitts’ law functions, there was a 55% increase in slope of the function when comparing the response panel to the touch screen, but only a 24% increase in the intercept of the functions. Although this finding may suggest that it is easier to move your finger from one key to another than to move your finger from one item on a touch screen to another, it may be the case that occlusion of the display by the hand on the touch screen accounts for part of the slowing. Comparison of the response panel with the mouse resulted in a 92% increase in the slope of the function and a 55% increase in the intercept. Because the two input devices require different types of movement, it is not surprising that most of the increase is in the slope. With the mouse movement, the input device is controlling a cursor on the display. Although the mouse is a zero-order control, in which changes in the position of the mouse correspond to changes in the position of
Movement Time for Different Input Devices
9
the cursor, the mapping of distance is not direct (e.g., moving the mouse 1 inch to the left may result in the cursor moving 3 inches on the screen). Thus, mouse movement may be longer due to a need to slow the device to hone in on the target or to make corrections required for under- or over-shooting the target. The implications of the present findings relating to the design of displays and controls are as follows: • If the environment or situation in which the input will be made requires a short movement time, the design should use a discrete button push for the user to make the input. For example, if a person is driving and would like to change the radio station, it would probably take less time for the movement from the steering wheel to a button on the radio panel than to a button indicator on a touch screen display. • When designing for a task that requires a short movement time along a single dimension, the design should involve movement along the horizontal dimension instead of the vertical dimension. For example, in a word processing program, such as Microsoft Word, the new ribbon design requiring selection of items from left to right may be more efficient than the old drop-down menu design. Acknowledgments. This study was supported by NASA cooperative agreement NNX09AU66A, Group 5 University Research Center: Center for the Human Factors in Advanced Aeronautics Technologies (Brenda Collins, Technical Monitor).
References 1. Fitts, P.M.: The Information Capacity of the Human Motor System in Controlling the Amplitude of Movement. J. Exp. Psychol. 47, 381–391 (1954) 2. Fitts, P.M., Peterson, J.R.: Information Capacity of Discrete Motor Responses. J. Exp. Psychol. 67, 103–112 (1964) 3. Radwin, R.G., Vanderheiden, G.C., Lin, M.-L.: A Method for Evaluating Head-Controlled Input Devices Using Fitts’ Law. Hum. Factors 32, 423–438 (1990) 4. Drury, C.G.: Application of Fitts’ Law to Foot-Pedal Design. Hum. Factors 17, 368–373 (1985) 5. Kerr, R.: Diving, Adaptation, and Fitts’ Law. J. Motor Beh. 10, 255–260 (1978) 6. McGovern, D.E.: Factors Affecting Control Allocation for Augmented Remote Manipulation. Doctoral dissertation, Stanford University (1974) 7. Proctor, R.W., Van Zandt, T.: Human Factors in Simple and Complex Systems. CRC Press, Boca Raton (2008) 8. Proctor, R.W., Vu, K.-P.L.: Stimulus-Response Compatibility: Data, Theory, and Application. CRC Press, Boca Raton (2006) 9. Rubichi, S., Vu, K.-P.L., Nicoletti, R., Proctor, R.W.: Spatial Coding in Two Dimensions. Psych. Bull. & Rev. 13, 201–216 (2006) 10. Nicoletti, R., Umiltà, C.: Right-Left Prevalence in Spatial Compatibility. Percept. & Psychophys. 35, 333–343 (1984) 11. Vu, K.-P.L., Proctor, R.W., Pick, D.F.: Vertical Versus Horizontal Compatibility: LeftRight Prevalence with Bimanual Keypresses. Psychol. Res.-Psych. Fo. 64, 25–40 (2000)
Audio and Audiovisual Cueing in Visual Search: Effects of Target Uncertainty and Auditory Cue Precision Hugo Bertolotti and Thomas Z. Strybel California State University, Long Beach, Center for the Study of Advanced Aeronautics Technologies 1250 N Bellflower Blvd. Long Beach, CA 90840, USA [email protected], [email protected]
Abstract. Auditory spatial cue accuracy and target uncertainty were examined within visual search. Participants identified a visual target to be present or absent under various target percentage conditions (25%, 50%, & 100%) with either an auditory cue which was spatially coincident with or displaced 4° or 8° (vertical or horizontal) from the target, or both an auditory and visual cue (circle 6.5° radius; identifying the local-target-area surrounding the target). Within the auditory cue condition, horizontal displacement was a greater detriment to target present search times than vertical displacement, regardless of error magnitude or target percentage. When provided an audiovisual cue, search times decreased 25% for present targets, and as much as 300% for absent targets. Furthermore, within audiovisual cue condition, while present target search times decreased with target percentage, absent target search times increased with target percentage. Cue condition and target uncertainty driven search strategies are discussed, with recommended design requirements and research implementations. Keywords: Visual Search, Audio Cue, Audiovisual Cue, Auditory Cue Precision, Target Uncertainty, False Alarm.
1 Introduction In recent decades, technological advancements have made virtual environments and complex audiovisual displays a possible design solution for aircraft cockpits, automobiles, and other complex work environments. Environments such as these require operators to take in and analyze a consistent influx of information and data, and make critical decisions based upon the information. Moreover, with the majority of information presented to the operator being visual, inundation of the already overloaded visual channel is possible. These complex work environments might avoid visual overload by providing some information via other sensory channels, for example, auditory. The information provided from the auditory channel can be either redundant with information presented visually (creating an audiovisual display), or specific to the auditory channel alone (creating an auditory display). D. Harris (Ed.): Engin. Psychol. and Cog. Ergonomics, HCII 2011, LNAI 6781, pp. 10–20, 2011. © Springer-Verlag Berlin Heidelberg 2011
Audio and Audiovisual Cueing in Visual Search
11
Research concerning auditory spatial cueing in visual search has demonstrated several benefits of providing operators with spatially coincident auditory cues –auditory cues provided at a specific elevation and azimuth in space – in both laboratory and applied settings. Within a controlled laboratory environment, coincident auditory spatial cues have been shown to significantly reduce search times required to locate and identify a specified target during a visual search task [1] [2]. In applied settings, simulated 3D auditory spatial cues (simulated interaural difference cues and spectral shape cues provided via headphones) improved target acquisition, traffic detection and avoidance, and visual workload [3] [4] [5]. It is important to note that while providing an auditory spatial cue may be advantageous, the magnitude of the benefits provided is contingent upon the characteristics of the auditory cue and the visual task itself. One important auditory cue characteristic, auditory cue precision (measured as the distance between the location of the audio cue and the location of the visual target), has been demonstrated to significantly affect target identification [6] [7] [8]. Research has shown that increasing error magnitude displacement (0° to 8°) between a specified target and an auditory spatial cue will significantly increase target search times [7] [8]; however, while search times consistently increased with error magnitude displacement, manipulating error displacement direction produced inconclusive results. Within Vu et al. [6], vertical displacement of an auditory spatial cue was a greater detriment to target identification than horizontal displacement, whereas Bertolotti and Strybel [7] found horizontal displacement to be more detrimental. Moreover, as the error displacement of auditory spatial cue relative to the target increased, the directional discrepancies between Bertolotti and Strybel, and Vu et al. become larger. A possibility for the inconsistent dimensional findings may be due to target uncertainty. While a target was present on every trial within Vu et al. [6], Bertolotti and Strybel [7] not only included trials which contained no targets (false targets), requiring participants to respond “no target” when none was found, but also manipulated the percentage of trials containing no targets within a given block of trials. When false targets were introduced, horizontal displacement became significantly more detrimental to target identification than vertical displacement, regardless of target percentage. In addition, the time required to identify the absence of a target was three times that of a present target, regardless of cue precision. Perhaps, the introduction of false alarms produced alterations in observer search strategies, and it is these search strategy adjustments which lead to discrepancies within the obtained dimensional findings. A second possibility for the dimensional inconsistencies may be due to the differences in visual saliencies of both the local area surrounding the target (local-target-area), and the entire visual search field (global area). While local and global visual saliencies remained equal within Bertolotti and Strybel [7], Vu et al. [6] manipulated local and global visual saliencies via distractor density. Results show when the local-target-area surrounding the target was visually cued via global-local distractor densities, vertically displaced auditory cues produced significantly higher search times than horizontally displaced auditory cues. It appears that as the local target area becomes increasingly visually salient as a
12
H. Bertolotti and T.Z. Strybel
result of global-local distractor density combinations, facilitation of the localization stage of the visual search process may occur, accounting for the dimensional inconsistencies. To examine these two possibilities, and to gain a clearer understanding of the inconsistent dimensional findings in the previous literature, the current study observed the effects of auditory spatial cue precision, target uncertainty, and local target area visual saliency on the visual search process.
2 Method 2.1 Participants Twelve students (5 males, 7 females) with a mean age of 23.92 years (SD = 3.37 years) participated in the study. All participants were students from California State University, Long Beach, and reported normal hearing and normal to corrected-tonormal vision. Participants were paid $60 for completing the experiment. 2.2 Apparatus The apparatus and materials utilized were identical to that of Bertolotti and Strybel [7]. Positioned at the center of a semi-anechoic room, covered by Markerfoam 10.16 cm acoustic foam sheets (absorption coefficients exceed .90 for frequencies greater than 250 Hz), was a large acoustically transparent projection screen. Figure 1 illustrates a sample of the visual search field consisting of a target (1° x 1° arrowhead pointing right or left), distractors (1° x 1° arrowheads pointing up or down), and a visual cue (described later). Contrast between the white targets and distractors and the black screen was held constant at roughly 75%, with local and global distractordensities remaining constant at 32%. Two data projectors connected to a microprocessor located in an adjacent room were used for presenting the visual stimulus during the experiment. For the auditory stimulus, Tucker-Davis Technologies’ audio modules were used for generating and presenting the 65-dB A-weighted auditory stimulus. The auditory stimulus was a series of 300 ms broadband noise bursts separated by 100 ms quiet intervals that remained on until a response was made. Forty-five 7.6 cm Blaupunkt speakers mounted behind the screen in six circular centric rings produced the auditory stimuli. The positioning of the speakers, measured from the fixation point in the middle of the visual search field, created three distance ranges (12° to 18°; 24° to 30°; 36° to 42°) with midpoints of 15°, 27°, and 39°. The fixation point consisted of a 1° x 1° crosshair, located at the center of the visual search field. In order to record participant responses, a four button response box was utilized. The two top buttons were used to identify the target (left button—arrowhead pointing to the left; right button—arrowhead pointing to the right), while either of the two bottom buttons were used to report no target present.
Audio and Audiovisual Cueing in Visual Search
13
92°
Visual Cue Visual Target
>
56°
Fig. 1. Visual search field indicating visual cue (circle: 6.5° radius), and consisting of a visual target (arrowhead pointing right) and distractors (arrowheads pointing up or down)
2.3 Design Four independent variables were manipulated: cue condition, auditory cue error, target percentage, and target distance. Cue condition was the first variable under investigation, and consisted of two levels: auditory cue and audiovisual cue. Within the auditory cue condition, the participant was provided with only an auditory spatial cue which signaled the location of the target. In the audiovisual cue condition, the participant was provided with a visual cue in addition to the auditory spatial cue. Following the local target area dimensions of previous research [8], the visual cue consisted of a circle (6.5° radius) identifying the local target area (Figure 1). Auditory cue error, the second variable under investigation, varied the direction and distance between the auditory spatial cue and the target. There were two displacement directions (horizontal and vertical), and three error magnitude distances (0°, 4°, & 8°). Target percentage was the next independent variable, and consisted of three levels: 25%, 50%, and 100%. In the 25% target percentage condition, 25% of the trials within a session contained targets. The second target percentage condition provided targets on 50% of the trials within a given session. The final target percentage condition provided a target on 100% of the trials. For no target trials, a distractor (arrowhead pointing either up or down) was placed in the exact position where the target would have been. The final independent variable was target distance from the point of fixation, and there were three distance ranges: 12° to 18°, 24° to 30°, and 36° to 42°. These three distance ranges created midpoints of 15°, 27°, and 39° accordingly. Again, a 1° x 1° crosshair located at the center of the visual search field was considered the point of fixation. It must be mentioned that at the farthest distance from the point of fixation (±39°), horizontal displacement of the auditory cue outside the local target area (8°)
14
H. Bertolotti and T.Z. Strybel
was possible only in the direction towards the point of fixation due to the close proximity of the speakers to the peripheral edge of the visual search field. The dependent variable for the current study was target and no target search times. 2.4 Procedure Participants were required to complete a screening form and instructed on the purpose of the study before signing the informed consent. Once completed, participants were seated 127 cm from the projected visual search field and instructed to located and identify whether a visual target was present or absent, and if present, identify the direction it was pointing. A practice session (20 trials) was given in order to familiarize the participant with the experiment and how to respond. After the participant reached a response accuracy rate of 95%, the participant was instructed that the experiment was about to begin, and was provided information regarding the visual cue and target percentage for the current block. Upon the start of a trial, the participant was required to fixate on the crosshair located at the center of the visual search field containing many X’s. After a period of 500 ms—1000 ms, the crosshair disappeared, the X’s turned into distractors, and a target accompanied by the appropriate cue (auditory or audiovisual) was presented. The participant was instructed to scan the visual search field and report, via the response box, the direction of the target (left or right) or no target. After the response the trial terminated, the visual search field disappeared, and the crosshair positioned at the center of the search field accompanied by the X’s reappeared to signify the start of the next trial. If no response was made after 8 seconds, the trial was terminated. Participants completed a total of six 1-hour blocks, of which three blocks were completed with only an auditory cue and three blocks were completed with an audiovisual cue, in random order. Within each cue condition, the three blocks were further separated by target percentage, resulting in each cue condition containing a separate 1-hour block for each target percentage condition (25%, 50%, & 100%). Participants were informed of both cue condition and target percentage prior to starting a block. Within each 1-hour block, three separate 20-minute sessions were completed. Of the three sessions completed within each 1-hour block, the first session was considered practice while the remaining two were analyzed.
3 Results To adjust for skewed distribution of search times, ANOVAs were run using log transformations of search times; however, actual search times are used in figures for ease of interpretation. Any violations of Mauchly’s test of sphericity were corrected, and adjusted degrees of freedom from Huynh-Feldt estimates of sphericity were used. Post-hoc analyses were completed with the use of Tukey post-hoc test. Separate ANOVAs were run for each cue condition due to the large differences in search times between the audiovisual cue condition (target present: M = 911.99 ms, SEM = 74.8 ms; target absent: M = 835.375 ms, SEM = 69.86 ms) and the auditory cue condition (target present: M = 1134.69 ms, SEM = 117.57ms; target absent:
Audio and Audiovisual Cueing in Visual Search
15
M = 3748.03ms, SEM = 396.32ms). For the target present trials, two 3 (target percentage: 25%, 50%, & 100% ) X 3 (target distance: 15°, 29°, & 39°) X 2 (error direction: horizontal and vertical) X 3 (error magnitude: 0°, 4°, & 8°) repeated measures analysis of variance (ANOVAs) were run on target present search times for each cue condition (auditory and audiovisual). For the target absent trials, two 2 (target percentage: 25% & 50%) X 3 (target distance: 15°, 27°, & 39°) X 2 (error direction: horizontal and vertical) X 3 (error magnitude: 0°, 4°, & 8°) repeated measures analysis of variance (ANOVAs) were run on target absent search times for each cue condition (auditory and audiovisual). ANOVAs run on target absent trials for each cue condition contained only two percentage conditions (25% & 50%), due to a lack of no target trials within the 100% target percentage condition. Response accuracy was 98.08% (SD = 1.14%), and trials which participants time-out were removed prior to analysis. 3.1 Auditory Cue Condition For target present trials, a significant main effect of error direction was obtained (F(1, 11) = 10.425, p = .012). Search times with horizontally displaced cues (M = 1254.79 ms, SEM = 80.84 ms) were significantly higher than search times with vertically displaced cues (M = 1132.83 ms, SEM = 63.60 ms). A significant main effect of error magnitude also was obtained for target present trials (F(2, 22) = 15.121, p = .001), with search times significantly increasing with error magnitude: 0°: M = 1102.21 ms, SEM = 67.43 ms; 4°: M = 1165.88 ms, SEM = 63.55 ms; and 8°: M = 1313.34 ms, SEM = 87.98 ms. No significant main effects were found on target absent latencies within the auditory cue condition. The significant main effects for target present trials were qualified by a significant error direction X error magnitude interaction (F(1.564, 17.487) = 6.623, p = .022), shown in Figure 2a. While search times for target present trials increased with error magnitude for both horizontal and vertical displacement, search times remained consistently higher when the auditory cue was displaced horizontally compared to vertically, particularly at 8° of error magnitude displacement. Simple effects analysis revealed a significant simple effect of error direction at both 4° (F(1, 11) = 12.348, p = .005) and 8° (F(1, 11) = 10.407, p = .008). Horizontal displacement was found to be a significantly greater detriment to target identification than vertical displacement, regardless of error magnitude. To examine the linear relationship between both error directions, as well as the increasing rate of search time, mean search times were regressed against error magnitude. Only horizontal displacement was found to explain a significant portion of the variance in target identification search times (horizontal displacement: R² = .97; F(1, 11) = 36.403, p = .01; vertical displacement: R² = .78; (F(1, 11) = 3.581, p = .30). Furthermore, a slope of 43.57 (±7.22) was obtained for horizontal displacement, and 9.213 (±4.87) for vertical displacement. Thus, for every 1° increase in horizontal displacement, search times increased by approximately 43 ms, and every 1° increase in vertical displacement, search times increased by approximately 9 ms.
16
H. Bertolotti and T.Z. Strybel
2b
2a 4200
Search Times (ms)
Search Times (ms)
1600 1400 1200 1000 800 0
4
3200
950 925 900 875 850 825
8
0
4
8
Error Magnitude (q)
Error Magnitude (q)
Error Direction - Cue Condition
Horizontal - Auditory Vertical - Auditory
Horizontal - Audiovisual Vertical - Audiovisual
Fig. 2. Search times for auditory and audiovisual cue conditions as a function error direction and error magnitude: a) target present trials b) target absent trials
A significant error magnitude X target percentage interaction also was obtained for only the target present trials (F(4,20) = 2.98, p < .05). As illustrated in Figure 3, at 25% target percentage, error magnitude had little effect on search times; however, as target percentage increased, search times increased with error magnitude. Simple effect of error magnitude on target percentage demonstrated a significant simple effect at 50% (F(2, 22) = 21.014, p < .001) and 100% (F(1.491, 16.4) = 22.534, p < .001). Post-hoc tests found a significant increase in search times at 50% as error magnitude increased from 0° to 8°and 4° to 8°, and a significant increase in search times at 100% for every error magnitude increment. Thus, while error magnitude did not affect target present trial search times at 25%, search times significantly increased for every error magnitude increment at both 50% and 100%, with the exception of the 0° to 8° increment at 50%.
3b
3a 4500
Search Times (ms)
Search Times (ms)
1600
1400
1200
1000
4200 3900 3600 3300 3000
0
4
8
0
Error Magnitude (°)
4
8
Error Magnitude (°)
Target Percentage 25%
50%
100%
Fig. 3. Search times for the auditory cue condition as a function target percentage and error magnitude: a) target present trials b) target absent trials (contains only 25% and 50% conditions)
Audio and Audiovisual Cueing in Visual Search
17
3.2 Audiovisual Cue Condition For target present trials, a significant main effect of target percentage was obtained (F(2, 11) = 10.03, p = .002). While search times decreased as target percentage increased (25%: M = 921.40 ms; SEM = 70.32 ms; 50%: M = 869.09 ms, SEM = 57.19 ms; 100%: M = 843.99 ms, SEM = 56.58 ms), post-hoc testing found only one significant difference in search times, 25% vs. 100%. A significant main effect of target distance was also found for target present trials (F(2, 11) = 19.539, p < .001). Post-hoc testing determined that as target distance increased, search times significantly increased only at target distances between 15° (M = 871.26 ms, SEM = 62.98 ms) vs. 39° (M = 902.01 ms, SEM = 58.47 ms), and 27° (M = 861.21 ms, SEM = 57.19 ms) vs. 39°. No significant interactions were obtained for target present trials within the audiovisual cue condition. For target absent trials, a significant main effect of target percentage was obtained (F(1, 11) = 5.951, p = .033), with search times significantly increasing with target percentage. A significant main effect of target distance was also found for target absent trials (F(2, 22) = 32.624, p < .001). Significant increases in search times were found between 15° vs. 39, and 27° vs. 39°. However, between 15° vs. 27°, search times at 27° were significantly lower than search times at 15°. A significant main effect of error magnitude was found for target absent trials (F(2, 22) = 4.076, p = .031). Significant increases in search times were found between 0° to 4°, and 4° to 8°. No difference in search times were found as error magnitude increased from 0° to 8°.
4a
950 900 850 800 750
950 900 850 800 750
25
50
75
100
Target Percentage (%)
4b
1000
Search Times (ms)
Search Times (ms)
1000
25
50
Target Percentage (%)
Error Direction Horizontal Vertical
Fig. 4. Search times for the audiovisual cue condition as a function error direction and target percentage: a) target present trials b) target absent trials
For target absent trials, a significant target percentage X error direction interaction was obtained (target present: F(2, 22) = .506, p = .614; target absent: F(1, 11) = 5.07, p = .046). As Figure 4 illustrates, search times increased with target percentage at both horizontal and vertical displacement. Simple effects of target percentage for error direction demonstrated a marginally significant simple effect of target percentage for horizontal displacement (F(1, 11) = 3.642, p = .083), and a significant main effect of target percentage for vertical displacement (F(1, 11) = 7.876, p = .017).
18
H. Bertolotti and T.Z. Strybel
Search times increased with target percentage at both horizontal and vertical displacement, but the rate of increase was higher with vertical displacement. Furthermore, while horizontal displacement was a greater detriment to search times at 25%, vertical displacement was a greater detriment to search times at 50%. Simple effects of error direction for target percentage were examined, and a significant simple effect of error direction was obtained only at 25% (F(1, 11) = 8.699, p = .013). Horizontal displacement was a significantly greater detriment to search times than vertical displacement. For the target present trials (Figure 4a), while search times decreased with target percentage at both horizontal and vertical displacement, these were non-significant.
4 Discussion Examining the dimensional findings within the current study, the results appear to be consistent with Bertolotti and Strybel [7]: horizontal displacement of an auditory cue was a greater detriment to the identification of a visual target than vertical displacement. Furthermore, within both the current and previous research [7], horizontally displacing an auditory cue either 4° or 8° from a visual target produced higher search times than vertically displacing an auditory cue either 4° or 8° from a visual target. Thus, an auditory cue displaced horizontally within the local-target-area increased search times more than a vertically displaced auditory cue outside the localtarget-area. Given that this interaction did not depend on target percentage or target distance, it may reflect differences in localization between the horizontal and vertical plane; specifically, higher auditory spatial acuity in the horizontal plane compared to the vertical plane [9]. As for the first possibility of introducing target uncertainty, the current study found that when target uncertainty was introduced, horizontal displacement of an auditory cue continued to be a greater detriment to visual target identification than vertical displacement, regardless of error magnitude. Moreover, horizontal displacement was a greater detriment to target identification when target percentage was 100%, which suggests that target uncertainty may not be responsible for the incongruent directional displacement findings. In addition, within both the current study and Bertolotti and Strybel [7], search times significantly increased as the auditory cue error and target percentage increased. Therefore, it is suggested that search strategies for a visual target assisted by only an auditory cue are affected by target uncertainty in that as the likelihood of a target increased, greater dependence on the auditory cue for visual target identification was shown, which lead to a greater effect of auditory cue directional displacement. For the directional displacement when no visual target was present, the current study found an effect of directional displacement only as both target percentage and error magnitude increased. With regards to the second possibility, the findings from the current study are consistent with the previous literature in that increasing local-target-area visual saliency significantly improves the visual search of a target [6] [8]. When provided with a reliable local-target-area visual cue and a variably displaced auditory cue,
Audio and Audiovisual Cueing in Visual Search
19
search times decreased by approximately 25% compared to search times with a variable displaced auditory cue alone. However, while Vu et al. [6] found variable vertical displacement of the auditory cue outside the local-target-area to be more of a detriment to the visual search process when the local-target-area was visually salient, the current study found the no effect of auditory cue displacement when the localtarget-area was visually cued. Regardless of the error magnitude or directional displacement of the auditory cue, providing a visually salient local-target-area significantly improved visual target identification. It must be noted that while Vu et al. [6] varied global-local distractor-densities to manipulate the visual saliency of the local-target-area, the saliency of the local-target-area within the current study was varied by either providing a visual cue which accurately and consistently cued the local-target-area, or providing no visual cue. Possibly the high saliency of the localtarget-area visual cue within the current study rendered the directional displacement of the auditory cue ineffective, hindering the ability to reconcile the incongruent findings within Bertolotti and Strybel [7] and Vu et al. [6]. One possible explanation for the ineffectiveness of auditory cue precision with audiovisual cues is that participants used the unreliable auditory cue to identify the hemi-field in which the target was located, while using the visual cue to determine the local-target-area and target location. This is suggested since providing a visually cued local-target-area was found to not only significantly improve target identification [6] [8], but also rendered the effect of auditory cue displacement less effective. Additional evidence for this possibility is that search times for targets assisted by an audiovisual cue were found to increase as target distance increased. Thus, it is suggested that while the auditory cue provided the direction in which the target was located compared to the center of the visual search field, the visual cue provide the local-target-area visual saliency needed to identify the local-target-area, and the target itself. In summary, these results suggest that the effectiveness of audio spatial cueing systems depend more on precision cueing of horizontal target position compared with vertical target position because of the greater detriment in search latencies found with horizontally displaced cues. Moreover, with production cost much lower for auditory cues required for localization in the horizontal plane (ILDs and ITDs) compared to auditory cues required for localization in the vertical plane (spectral shape cues) [10], minimizing horizontal error would not only improve operator performance, but also would reduce development costs. The current study also has shown that providing an audiovisual cue can significantly improve target search latency compared to only an auditory cue. Furthermore, the benefits of providing an audiovisual cue were greater when participants were required to report no target, with performance improving as much as 300%. Thus, for environments in which false alarms are a possibility, audiovisual cues improve performance over an auditory cue alone. Acknowledgements. This paper was partially supported by the Center for Human Factors in Advanced Aeronautics Technologies, a NASA University Research Center (NASA cooperative agreement NNX09A66A.)
20
H. Bertolotti and T.Z. Strybel
References 1. Perrott, D.R., Cinseros, J., McKinley, R.L., D’Angelo, W.R.: Aurally aided visual search under virtual free-listening conditions. Human Factors 38, 702–715 (1996) 2. Perrott, D.R., Saberi, K., Brown, K., Strybel, T.Z.: Auditory psychomotor coordination and visual search. Perception and Psychophysics 48, 214–226 (1990) 3. Begault, D.R.: Head-up auditory displays for traffic collision avoidance system advisories: A preliminary investigation. Human Factors 35, 707–717 (1993) 4. McKinley, R.L., D’Angelo, W.R., Hass, M.W., Perrott, D.R., Nelson, W.T., Hettinger, L.J., Brickman, B.J.: An initial study of the effects of 3-dimensional auditory cueing on visual target detection. In: Proceedings of the Human Factors and Ergonomics Society, vol. 39, pp. 119–123 (1995) 5. McKinley, R.L., Erickson, M.A.: Flight demonstration of a 3-D auditory display. In: Gilkey, G.H., Anderson, T.R. (eds.) Binaural and Spatial Hearing in Rea; and Virtual Environments, pp. 683–699. Erlbaum, Mahwah, NJ (1997) 6. Vu, K.L., Strybel, T.Z., Proctor, R.D.: Effects of displacement magnitude and direction of auditory cues on auditory spatial facilitation of visual search. Human Factors 49(3), 587– 599 (2006) 7. Bertolotti, H., Strybel, T.Z.: he effects of auditory cue precision on target detection and identification. Presentation session presented at the meeting Human-Computer Interaction Institution International, Las Veges, NV (2005) 8. Rudmann, D.R., Strybel, T.Z.: Auditory spatial cueing in visual search performance: Effect of local vs. global distractor density. Human Factors 41, 146–160 (1999) 9. Strybel, T.Z., Fujimoto, K.: Minimum audible angles in the horizontal and vertical planes: Effects of stimulus onset asynchrony and burst duration. Journal Acoustical Society of America 108, 3092–3095 (2000) 10. Shinn-Cunningham, B.G.: Spatial auditory displays. In: Karwowski, W. (ed.) International Encyclopedia of Ergonomics and Human Factors, 2nd edn. Taylor and Francis, Ltd., Abington (2002)
Interpretation of Metaphors with Perceptual Features Using WordNet Rini Bhatt, Amitash Ojha, and Bipin Indurkhya Cognitive Science lab, International Institute of Information, Technology Hyderabad- 500 032, India [email protected], {amitash.ojha,bipin}@iiit.ac.in
Abstract. Metaphors based on perceptual similarity play a key role in stimulating creativity. Here, we present a metaphor interpretation tool using features of source and target to generate perceptual metaphors which might be conceptually very different, thereby generating new interpretations from familiar concepts. Keywords: Perceptual metaphors, conceptual combination, creative cognition, juxtaposition, conceptual association, metaphorical interpretation.
1 Introduction “What is creativity?” This question has no universally agreed answer. [1] For our purpose, however we define creativity as the ability to generate ideas or artifacts that are novel, surprising and valuable, interesting, useful, funny, beautiful, etc. According to Perkins “Creativity is not a special 'faculty', nor a psychological property confined to a tiny elite. Rather, it is a feature of human intelligence in general”. [2] It rests on everyday capacities such as the association of ideas, analogical thinking, searching a structured problem-space, and reflective self-criticism. Various kinds of creativity have been mentioned in the literature. But broadly there are three ways in which processes can generate new ideas: Combinational, exploratory and transformational. [3] We are concerned with combinational creativity, which is the production of novel (unfamiliar, improbable) combinations of familiar ideas. It has been studied in AI by the many models of analogy and by the occasional joke generating programs or database metaphor generation tools. One such system is JAPE, which models the associative processes required to generate punning jokes. Such processes are far from random, and depend on several types of knowledge such as lexical, semantic, phonetic, orthographic, and syntactic [4]. For analogy, most AI models generate and evaluate analogies by exploiting the programmer's careful pre-structuring of the relevant concepts. This guarantees that their similarity is represented, and makes it likely that the similarity will be found by the program [5]. In this paper we present a metaphor interpretation tool using perceptual features of Source concept and Target concept. The tool can help users to generate perceptual metaphors by evoking their ability to make free associations. The tool, in other words, assists users to create unfamiliar association from familiar concepts. D. Harris (Ed.): Engin. Psychol. and Cog. Ergonomics, HCII 2011, LNAI 6781, pp. 21–27, 2011. © Springer-Verlag Berlin Heidelberg 2011
22
R. Bhatt, A. Ojha, and B. Indurkhya
1.1 Pictorial Metaphors Metaphorical thinking is known to play a key role in stimulating creativity. In a metaphor, one kind of object or idea is used in place of another to suggest some kind of likeness between them. They serve in making connections between things that are not usually seen as connected in any conventional way. For example, computer science instructors often explain the function of the Control Unit in a computer’s Central Processing Unit by saying it is ‘the traffic policeman of the computer’. The general use of the term ‘memory’ to denote computer storage is metaphoric. Thus, metaphors are instruments of divergent processes because they synthesize disparate ideas. Similarly, in pictorial metaphors two concepts are juxtaposed or replaced to create a unified figure suggesting one concept being described in terms of another. The perceived incongruity in the image invites the viewer to interpret the image metaphorically. Though metaphors have mostly been studied as a literary device, but research on pictorial metaphor has over the past 25 years yielded a few theoretical studies [6], [7], [8], [9]. 1.2 Perceptual Similarities in Pictorial Metaphors We have hypothesized that perceptual similarity between two images at the level of color, shape, texture, etc. helps to create metaphorical associations [10]. Elsewhere we have shown that participants have preference for perceptually similar images in generating metaphorical interpretations. Also they help in creating more conceptual associations between Source and Target [11]. 1.3 Computers and Creativity For creativity a cognitive agent needs to break conventional conceptual associations and this task is difficult for human beings because we inherit and learn, in our lifetime, to see the world through associations of our concepts. It requires a significant amount of cognitive effort to break away from these associations. Computers, on the other hand, do not have such conceptual associations. Therefore it must be easier for the computers to break away from these conceptual association simply because they do not have them to begin with. It follows that computers are naturally predisposed towards incorporating creativity [12].
2 Using WordNet for Generating Metaphorical Interpretation 2.1 Idea and Aim The present system aims to extract perceptual features of Source concept and Target concept and uses it to anchor conceptual associations between them for a metaphorical interpretation. The larger goal of the system is to be able to generate metaphorical interpretations for an image by extracting perceptual features (for visual: color, shape, texture, orientation, etc, for touch: Hot, cold, rough, smooth, etc.). With the availability of these features, various concepts can be evoked (which are usually
Interpretation of Metaphors with Perceptual Features Using WordNet
23
not evoked just by the conceptual description of the object) and as a result distant associations can be generated which may or may not be creative. But as a matter of fact the perception of the creativity lies in the interaction between cognitive agent and stimuli. So, the system can suggest some combinations (that are guided by the perceptual features of concepts and are not arbitrary) using WordNet, which may be seen creative by users. 2.2 WordNet WordNet is a large lexical database of English, developed under the direction of George A. Miller (Emeritus). Nouns, verbs, adjectives and adverbs are grouped into sets of cognitive synonyms (synsets), each expressing a distinct concept. Synsets are Interlinked by means of conceptual-semantic and lexical relations. It can be visualized as a graph with edges between words with some relationship (synonym, antonym, part-of, etc.) between them. Such words will, henceforth, be referred to as neighbors bearing the graph analogy in mind. WordNet 2.0 for Linux allows us to feed a word and the relationship (synonym - noun/adjective,etc.) on the commandline and outputs the words which satisfy this relationship - the ‘neighbors’ of that word. This is exploited while generating metaphors of multi-word sets of perceptual and conceptual features. 2.3
WordNet and Metaphor Generation
There have been various attempts to use WordNet for generating metaphors. [13],[14]. [15]. We take a different approach and try to use perceptual features of concepts to anchor the metaphorical interpretation. The creativity assistance tool generates metaphors for any arbitrary combination of words divided into 2 categories “Perceptual” and “Conceptual”. The generated ‘metaphors’ depend upon certain parameters, for instance, it takes into account, a certain threshold within which to look for neighbors of a word. It also allows for semantic constraint on the metaphors generated. As mentioned, with WordNet, we can find synonyms, antonyms, hypernyms, etc. of a given word. In our tool, we have considered the three mentioned here and any one type or a combination of these in any order can be used to generate interpretations by taking that particular path. 2.4 Threshold It is important to define this threshold as WordNet is organized as a connected graphand the search will go on infinitely unless a certain threshold value for the number of levels which can be visited for finding relevant words, is specified. Moreover, a word which is related to the given word but is, say, thirty levels away will likely be semantically and relationally dissimilar and therefore, will not satisfy our primary objective of finding different interpretations for the given set of features. 2.5 Anchoring Since various features are to be considered in the perceptual set as well as the conceptual set, this entire set of features must be taken as a whole to identify
24
R. Bhatt, A. Ojha, and B. Indurkhya
interpretations which are relevant to the entire picture instead of treating the features separately. In order to do this, there is a need to identify ‘an anchor’ word which is closest to the set of all features and serves as the link between them. This anchoring was done by finding the words up till the mentioned threshold for all the input features and then finding the words common to all the input features,i.e. if the input words (perceptual features) are ‘big’ and ‘red’ then the tool discovers the neighbors till the specified threshold and now from those two lists finds the closest common word. Once the common word or the ‘anchor’ has been found for both the perceptual features and the conceptual features, these anchor words are used to generate wordpairs. Neighbors of the anchor words are discovered till the specified threshold and these separate interpretations of the perceptual and conceptual anchor are then combined to give P*C two-word phrases as metaphors, where P = number of interpretations of the perceptual anchor word C = number of interpretations of the conceptual anchor word. The user could specify search till 2 levels for relation hypernym and then 3 levels for relation synonym. So here, first hypernyms would be generated till 2 levels and for the words derived, synonyms would be generated till 3 levels.The levels of semantic relations and their order bears significance on the results generated, i.e. whether five levels of hypernym are explored followed by two levels of synonyms or vice versa. "The heuristic of choosing the first listed sense in a dictionary is often hard to beat, especially by systems that do not exploit hand-tagged training data." [16]. Instead of following a particular method of disambiguation, we have chosen to consider both noun and adjective for perceptual features while only nouns have been considered for conceptual features. This is to allow generation of metaphors from both senses of the word - choosing to limit it to one might detract from the creativity of metaphor generation. Choosing to include nouns while discovering neighbors in case of perceptual features includes a trade-off - we get a wider range of interpretations now which include some that would be have been absent if the word was considered as only noun or only adjective, however, the list would also include some interpretations with a bad similarity measure.
3 Testing and Results We present an example which illustrates one of the results. In this example, 'red' and 'hot' are entered as the perceptual features. Now the tool seeks to find an 'anchor' for the given perceptual features within the specified threshold (in this example it is 2 levels). Bearing the graph structure in mind, first all the immediate neighbors (of the form synonym-noun and synonym-adjective) of the ‘red’ and the word 'hot' are discovered. This is level one i.e. these words are at distance '1' from the original words entered. From these words, a further level of words is discovered, i.e. the immediate neighbors of level 1 words. From the two lists - one for 'red' and the other for 'hot', the common words are determined. For these words, a common measure of distance from both words is computed. This is calculated as the vector distance between the words 'red' and 'hot' from the common word.
Interpretation of Metaphors with Perceptual Features Using WordNet
25
The one with the least such distance qualifies as the anchor word. This graph is a partial representation of the discovery mechanism followed to find the anchor word for perceptual features. After computing the two lists, 'wild' is found as the common word at minimum distance - 2 - as it is a layer 2 word for both 'red' and 'hot'. The words encircled in red color trace the path from 'red' and 'hot' to 'wild'. Thus, 'violent' is a synonym of 'red' and 'wild' is a synonym of 'violent' thereby giving a path from 'red' to 'wild'. Similarly for 'hot', 'raging' and 'wild'. Thus, 'wild' becomes the anchor word for this set of perceptual features. A similar mechanism is followed for conceptual set and a common word is found at the similar level (or at user defined level). Now, once the system has one common word for perceptual features and one common word for given conceptual features, same procedure is followed and list of neighboring words are generated as result for a certain level.
Fig. 1. Graphic representation of an example to find a common word for perceptual features
Further, A user testing was done with some of the results. 10 sets of word-pairs (each containing 50 word-pairs generated by system) were taken as test stimuli. These pairs were produced with 2 levels of synonym and 1 level of hypernym. To generate stimuli two perceptual features and one conceptual feature was used. 7 participants were asked to categories these pairs as 1. Metaphor, 2. Anomaly and 3. Literal. We found that 63 percent of word-pairs were categorized as “metaphor”, 32 percent of pairs were categorized as “anomaly” and 5 percent of word-pairs were categorized as “literal”. The difference between them was statistically significant. F (2, 12) =4.14 p < .05. (Figure 2)
26
R. Bhatt, A. Ojha, and B. Indurkhya
Fig. 2. Categorization of generated word-pairs
4 Limitations and Future Work The system is in its preliminary stage and has to be developed and tested further. At present it has various limitations: In the system, selection of levels is completely arbitrary and there is no fixed threshold to decide the metaphoricity of word-pairs. Some of the results at the very lower level can be interpreted metaphorically and some can not be interpreted metaphorically at higher level. We plan to solve this problem by measuring the distance between source and target in some conventional metaphors. An average distance can guide us to suggest that at what level chances are high for two concepts to be related metaphorically. The system aims to extract perceptual features automatically by a given picture using image search engines [17]. But for this users will be asked to tag the object depicted in the image. Image search engine will identify the color of the background, shape of the object, etc and then provide set of perceptual features. These perceptual features combined with conceptual feature/s can be used for further interpretation task.
References 1. Sternberg, R.J. (ed.): Handbook of creativity. Cambridge University Press, Cambridge (1999) 2. Perkins, D.N.: The mind’s best work. Harward university press, Cambridge, MA (1981) 3. Boden, M.A.: What is creativity? In: Boden, M.A. (ed.) Dimensions of creativity, pp. 75– 118. MIT Press, Cambridge (1994) 4. Binsted, K., Ritchie, G.: An implemented model of punning riddles. In: Proceedings of the Twelfth National Conference on Artificial Intelligence, Seattle, pp. 633–638 (1994)
Interpretation of Metaphors with Perceptual Features Using WordNet
27
5. Forbus, K.D., Gentner, D., Law, K.: MAC/FAC: A model of similarity based retrieval. Cognitive Science, 141–205 (1994) 6. Kennedy, J.M.: Metaphor in Pictures. Perception 11, 589–605 (1982) 7. Forceville, C.: Pictorial Metaphor in Advertising. Routledge, London and New York (1996) 8. Whittock, T.: Metaphor and Film. Cambridge University Press, Cambridge (1990) 9. Caroll, N.: Visual Metaphor. In: Hintikka, J. (ed.) Aspects of Metaphor, pp. 189–218. Kluwer Academic Publishers, Dordrecht (1994) 10. Indurkhya, B.: Metaphor and cognition. Kluwer Academic Publishers, Dordrecht (1992) 11. Ojha, A., Indurkhya, B.: Role of perceptual metaphors in metaphorical comprehension. In: The Proceedings of European Conference on Visual Perception, Regensburg, Germany (2009) 12. Indurkhya, B., Kattalay, K., Ojha, A., Tandon, P.: Experiments with a creativity-support system based on perceptual similarity. In: Proceedings of the 7th International Conference of Software Methodologies, Tools and Techniques; Encoding information on metaphoric expressions in WordNet like resources. In: Proceedings of the ACL 2003 workshop on lexicon and figurative language, vol. 14, pp. 11-17 (2003) 13. McCarthy, D., Keeling, R., Weeds, J.: Ranking WordNet senses automatically. Cognitive Science Research Paper, 569 14. Tandon, P., Nigam, P., Pudi, V., Jawahar, C.V.: FISH: A Practical System for Fast Interactive Image Search in Huge Databases. In: Proceedings of 7th ACM International Conference on Image and Video Retrieval (CIVR 2008), Niagara Falls, Canada (2008)
Acoustic Correlates of Deceptive Speech – An Exploratory Study David M. Howard and Christin Kirchhübel Audio Laboratory, Department of Electronics, University of York, UK {dh,ck531}@ohm.york.ac.uk
Abstract. The current work sets out to enhance our knowledge of changes or lack of changes in the speech signal when people are being deceptive. In particular, the study attempted to investigate the appropriateness of using speech cues in detecting deception. Truthful, deceptive and control speech was elicited from five speakers during an interview setting. The data was subjected to acoustic analysis and results are presented on a range of speech parameters including fundamental frequency (f0), overall intensity and mean vowel formants F1, F2 and F3. A significant correlation could not be established for any of the acoustic features examined. Directions for future work are highlighted. Keywords: Deception, speech, acoustic, Voice Stress Analyzer.
1 Introduction It is acknowledged that information can be gained about a human speaker from the speech signal alone. Possible knowledge that can be derived include a speaker’s age, regional and social background, the presence of speech or voice based pathology, voice/language disguise, speaking style, and influence of alcohol intoxication. The voice can also give information about a speaker’s affective state. Listening to a third party conversation, lay listeners are usually able to tell whether the speakers are happy, sad, angry or bored. Whilst at an interpersonal level it is possible to perceive accurately emotional states, empirical research has not been successful in identifying the speech characteristics that distinguish the different emotions. Compared to the investigation of speech and emotion, research into psychological stress has been somewhat more successful in establishing the acoustic and phonetic changes involved [1]. However, facing similar methodological and conceptual obstacles, it is more appropriate to refer to the correlations that have been discovered so far as ‘acoustic tendencies’ rather than ‘reliable acoustic indicators’. If it is possible to deduce a speaker’s emotional condition from listening to their voice, could it also be viable to make judgements about their sincerity from speech as well? For centuries people have been interested and fascinated by the phenomenon of deception and its detection. Indeed, a device that would reliably and consistently differentiate between truths and lies would be of great significance to law enforcement- and security agencies [2]. In recent times, claims have been brought forward involving voice stress analysers (VSA) which are said to measure speaker D. Harris (Ed.): Engin. Psychol. and Cog. Ergonomics, HCII 2011, LNAI 6781, pp. 28–37, 2011. © Springer-Verlag Berlin Heidelberg 2011
Acoustic Correlates of Deceptive Speech – An Exploratory Study
29
veracity based on the speech signal. Scientific reliability testing of these products has mainly resulted in negative evaluations [3,4,5]. While testing of these products is a necessary part of their evaluation, it is believed that a more fundamental step has been overlooked. Prior to examining the reliability of a test it should be ascertained whether the assumptions on which the test is based are valid [4]. In other words, it needs to be established whether a relationship exists between deception, truth and speech, and if so, what the nature of this relationship is. Surprisingly, very little research has been carried out on the acoustic and phonetic characteristics of deceptive speech. There are a number of studies [6,7,8] that have analysed temporal features such as speaking rate, pauses, hesitations and speech errors but only a few studies have investigated frequency based parameters as, for instance, mean f0 and f0 variability [9,10]. Evidence for the analysis of vowel formants and voice quality in connection with deceptive speech is rare. Recently completed work by Enos [11] is one of the first attempts to analyse deceptive speech using spoken language processing techniques. This work provides a basis in this complex area but more research is needed within this subject matter in order to improve our understanding of deceptive speech and consequently, to assess whether differentiating truthfulness and deception from speech is a realistic and reasonable aspiration.
2 Method 2.1 Participants The data consists of an opportunistic sample of five male native British English speakers between the ages of 20 to 30 years (mean age = 24.7 years; SD = 3.65 years). The majority were drawn from the student population at the University of York and were from the northern part of England. None of the speakers had any voice, speech or hearing disorders. 2.2 Experiment Procedure The procedure was modified from Porter [12] and is based on a mock-theft paradigm. The experiment was advertised as being part of a security research project and participants received £5 for participating with the chance of earning more money through the trial. Having arrived at the experimental setting participants were told that the University was looking to implement a new security campaign in order to reduce small scale criminal activity (e.g. theft on campus). The security scheme would involve employing non-uniformed security wardens who: a) Patrol in selected buildings b) Perform spot checks on people c) Interview students suspected of having been involved in a transgression The participants were then led to believe that the researchers were testing the effectiveness of this security system and, in particular, the extent that wardens would
30
D.M. Howard and C. Kirchhübel
be able to differentiate between guilty and non-guilty suspects. Further to this the volunteers were informed that the experiment was also part of a communication investigation and therefore audio data would be collected. Having given written consent that they are prepared to continue with the study participants had to complete three tasks: Task 1: sitting in a quiet office room (‘preparation room’), participants were asked to complete demographic details. On completion of the forms they were taken to an interview room where they were involved in a brief conversation which formed the baseline/control data. Task 2: participants were provided with a key and directions. They were asked to go into an office and take a £10 note out of a wallet located in a desk drawer and hide it on their body. They were advised to be careful in order not to raise suspicion or to draw attention of the security warden who was said to be in the building and who might perform a spot check. Task 3: a security interview was conducted in which the mock security warden questioned the participant about two thefts that had allegedly occurred in the previous hour. The participant committed one of the thefts (theft of £10 note) but not the other (theft of digital camera from the ‘preparation room’). Participants were required to convince the interviewer that they were not guilty of either theft. With respect to the camera theft, participants could tell the truth but when the interviewer asked about the £10 note the participant had to fabricate a false alibi. Each participant had 10 minutes prior to the interview to formulate a convincing story. If the participants were successful in convincing the interviewer that they did not take either the camera or the £10 they could earn an extra £5 in addition to their basic £5 participation payment. If they failed on either however, they would lose the extra payment and be asked to write a report about what had happened. 2.3 Recording Setting/Equipment The experiment was conducted in the Linguistics department at the University of York. A vacant office room was used as the ‘preparation room’ in which participants completed task 1 and prepared for task 3. The ‘target room’, an unoccupied office room, from which the £10 was taken, was situated at the other end of the corridor approximately 200m away from the ‘preparation room’. The baseline/control data and the security interview were recorded in a small recording studio. The interviewer and participants were sat down, oriented at approximately 90 degrees to each other. To ensure that the distance between microphone and speaker was kept constant an omnidirectional head-worn microphone of the type DPA 4066 was used. The microphone was coupled to a Zoom H4 recorder. 2.4 Parameters Analysed The experiment took the form of a within-subjects design. Truthful and deceptive speech was elicited from participants during the interviews. In addition baseline
Acoustic Correlates of Deceptive Speech – An Exploratory Study
31
(control) data was recorded prior to the interviews. The three speaking conditions will be referred to as Baseline (B), Truth (T) and Deception (D) in this article. A number of speech parameters was analysed including mean fundamental frequency (f0 mean) and standard deviation (f0 SD). The changes in mean energy across speaking conditions were computed as well as mean vowel formants F1, F2 and F3. Every speaker provided one file for each of the three speaking conditions, resulting in 60 files for analysis. The duration of the files was between about 3-5 minutes. 2.5 Measurement Equipment/Technique ‘Sony Sound Forge’ software was used for initial editing of the speech files. The acoustic analysis was performed using ‘Praat’ speech analysis software [13]. The f0 based parameters were measured on the previously edited files using a Praat script developed by Philip Harrison. In order to measure intensity, the files were edited in Praat so as to only contain speech (i.e. all silences were removed). The mean energy was then determined using a function of the Praat software. Rather than expressing the intensity values in absolute form, the differences between Baseline, Truth and Deception are reported in this paper. Vowel formant measurements were extracted from Linear Predictive Coding (LPC) spectra using Praat’s inbuilt formant tracker. The mean F1, F2 and F3 values were taken from an average of 10-20 milliseconds near the centre of each vowel portion. Any errors resulting from the in-built formant tracker were corrected by hand. To be counted in the analysis the vowels had to show relatively steady formants and be in stressed positions. For all speakers, 8 vowel categories1-FLEECE, KIT, DRESS, TRAP, NURSE, STRUT, LOT, and NORTH- were measured with one to 15 tokens (average 10 tokens) per category for each condition yielding a total of around 1200 measurements.
3 Results The following section presents result for all 5 speakers. Statistical tests (T-tests) were employed where possible to assess the significance of the difference between the three conditions. 3.1 Fundamental Frequency (f0) Based on the f0 mean and f0 SD values obtained for each speaker and each condition, bar graphs were generated (Figure 1).
1
Standard Lexical Sets for English developed by John C. Wells [14]. There are 24 lexical sets which represent words of the English language based on their pronunciation in Received Pronunciation (RP).
32
D.M. Howard and C. Kirchhübel
Fig. 1. f0 mean (left) and f0 SD (right) in Hz for all three speaking conditions for every speaker
Looking at Figure 1 (left) we can immediately see that there is not a great amount of difference in f0 mean across conditions. There is a tendency that the f0 mean of T is slightly lower than the values for B and D, which are close, but overall the mean f0 values of all conditions are essentially similar. Examining the f0 SD measures illustrated in the right hand side of Figure 1, we can perceive an overall trend in that there is less variation in f0 in the Truth and Deception condition compared to the Baseline. The values of T and D for each speaker, in contrast, are rather comparable. 3.2 Intensity Mean energy for each speaker is represented in Figure 2. No specific patterns can be generalized from the results. There is variability in direction and extent of change across speakers for both Truth and Deception. Apart from speakers 1, 3 and 5 the changes in mean energy are very small and interestingly, these three speakers show a uniform change in direction for both Truth and Deception.
Fig. 2. Overall mean energy changes between Baseline and Truth/Deception for all speakers
3.3 Vowel Formants F1 Figure 3 illustrates the mean F1 values for each measured vowel category from each individual speaker. The x-axis represents the F1 taken from the Baseline condition
Acoustic Correlates of Deceptive Speech – An Exploratory Study
33
and the difference between Baseline and Truth/Deception is shown along the y-axis. The majority of tokens lay on or slightly above the origin on the y-axis, which indicated that the F1 values from Truth/Deception were similar to those from the Baseline speech. There was some variability across tokens, demonstrated by the moderate spread of data points. As suggested by the almost horizontal trend lines, no significant correlations existed between F1 in the Baseline condition and the change in F1 in the Truth (r = 0.03926, df = 38, p = 0.8099) or Deception conditions (r = 0.00948, df = 38, p = 0.9537).
Fig. 3. Scatter-plot of F1 measures, showing value in the Baseline condition (x-axis) against shift observed in the Truth/Deception condition (y-axis)
F2 Figure 4 reflects the observed directional inconsistencies for F2. Some values were slightly increasing, some were decreasing and others were not changing. Overall the change was not considerable for any of the vowel categories and T-tests did not illustrate any significant differences at the 5% level.
Fig. 4. Scatter-plot of F2 measures, showing value in the Baseline condition (x-axis) against shift observed in the Truth/Deception condition (y-axis)
In particular for the Truth condition, there was a substantial amount of variation between tokens in the size of the difference as well as the direction. The same vowel category might be increasing, decreasing or not changing between different speakers. There was less variation in the Deception condition and most of the tokens were
34
D.M. Howard and C. Kirchhübel
grouped around the origin. The nearly horizontal trend-line reinforces the observation that no real pattern can be detected. The correlation between Baseline vowel measurements and the effect of Truth (-0.24396, df = 38, p = 0.1292) and Deception (r = -0.12698, df = 38, p = 0.4349) was weak and not statistically significant at any conventional significance level. F3 The F3 values of Baseline and Truth/Deception seemed to correspond very closely to each other and there did not appear to be a difference for any of the vowel categories. If changes did occur then it tended to be a decrease in Truth and Deception compared to Baseline.
Fig. 5. Scatter-plot of F3 measures, showing value in the Baseline condition (x-axis) against shift observed in the Truth/Deception condition (y-axis)
There was a tendency that high F3 values in the Baseline condition were more likely to be subject to a decrease in Truth/Deception than low F3 values. The overall correlation between F3 in Baseline and F3 decrease in Truth (r = -0.59525, df = 38, p < .001) and Deception (r = -0.54605, df = 38, p < .001) was statistically significant.
4 Discussion The results suggest that truth-tellers and liars cannot be differentiated based on the speech signal measures analysed in this study. Not only was there a lack of significant changes for the majority of parameters investigated but also, if change was present it failed to reveal consistencies within and between the speakers. F0 mean varied little across conditions. The reduced F0 SD values in T and D suggest that speakers were less variable and perhaps spoke with a more monotone voice in these conditions. This could be further investigated by auditory analysis. With regard to overall intensity changes, the findings also did not offer grounds for a reliable distinction between those telling the truth or lying. If speakers showed a change in mean energy then it was uniform in terms of direction and size across Truth and Deception. The majority of F1 and F2 differences between conditions were not statistically significant. For F2 in particular, there was a considerable amount of variation across
Acoustic Correlates of Deceptive Speech – An Exploratory Study
35
conditions with values increasing, decreasing or not changing. The F3 results point towards a negative correlation between Truth/Deception and Baseline speech. Of note again is the parallelism between Truth and Deception in that both show a significant negative correlation compared to the Baseline. F3 is linked to voice quality and vocal profile analysis of the speakers would cast more insights. The remarkable amount of inter- and intra-speaker variability underlines the fact that deceptive behaviour is individualised, very multifaceted and far from being clear cut and straightforward. Despite their non-significance the findings are of interest since they point to some potential limitations when trying speech analysis for deception detection purposes only. It may be argued that the lack of significant findings is a product of the experimental arrangement as a laboratory induced deception which does not adequately represent deception as it might occur in real life. This is a methodological limitation which, due to ethical considerations, cannot be overcome in the majority of studies on deception. In order to maintain the impact of the scientific validity of this study, it can be said that post-interview rating scales confirmed that all the participants were highly motivated to succeed in the deceptive act (score of 6 or higher on a 7- point Likert scale). It should also be stated that research into a relatively unexplored area, such as speech and deception, needs to start off with fully controlled experiments where variables can be controlled more strictly. Clean, high quality recordings must provide the starting point for the acoustic and phonetic analysis. If differences between truth and deception are found in these ideal conditions, research can then move on to investigating less controlled data in the field. One of the assets of the present research design concerns the separation of stress and deception in that the latter was not inferred from the former. The polygraph and most of the VSA technologies are based on the assumption that liars will show more emotional arousal i.e. will experience more stress than truth-tellers [2]. However, such a direct linkage cannot be presupposed. Certainly, there will be liars who do manifest the stereotypical image of nervousness and stress. At the same time, however, truthtellers may also exhibit anxiety and tension, especially if in fear of not being trusted. And on the contrary, liars might not conform to the stereotypical image described above but rather display a composed and calm countenance. As the following quote illustrates: ‘Anyone driven by the necessity of adjudging credibility who has listened over a number of years to sworn testimony, knows that as much truth has been uttered by shifty-eyed, perspiring, lip-licking, nail-biting, guilty- looking, ill-atease fidgety witnesses as have lies issued from calm, collected, imperturbable, urbane, straight-in-the-eye perjurers.’(Jones, E.A. in [2, p.102]) Harnsberger et al. [5] for example only included participants in their analysis who showed a significant increase in stress levels during deception. Given that the goal of their research was to test the validity of VSA technology this may be a justified methodological choice. However, as the aim of the present study was to attain a more comprehensive knowledge of the fundamental relationship between deception and speech it was essential to disassociate deception and stress. Further acoustic and phonetic analysis is under way to expand the analysis beyond measures of f0, intensity, and formant measurements to include measurement of
36
D.M. Howard and C. Kirchhübel
diphthong trajectories, consonant articulation, jitter, shimmer and spectral tilting. In addition, laryngograph recordings have been made with 10 speakers which will provide opportunity to analyse the glottal wave signal and this in turn will contribute further to our knowledge of truth/deception specific speech characteristics. Furthermore, the hypothesis that increasing cognitive load during interview situations has the potential of magnifying the differences between truth-tellers and liars in the speech domain will also be evaluated.
5 Conclusion This paper summarised an exploratory investigation into the relationship between acoustic parameters of speech and truth/deception. So far the analysed data does not suggest that a reliable and consistent correlation exists. As well as providing a basis for future research programs the present study should encourage researchers and practitioners to evaluate critically what is and what is not possible, using auditory and machine based analyses, with respect to detecting deception from speech. Acknowledgement. This work was made possible through EPSRC Grant number: EP/H02302X/1. Thanks to Philip Harrison for providing the Praat script and Francis Newton for his time and assistance during the data collection process.
References 1. Jessen, M.: Einfluss von Stress auf Sprache und Stimme. Unter besonderer Beruecksichtigung polizeidienstlicher Anforderungen. Schulz- Kirchner Verlag GmbH, Idstein (2006) 2. Lykken, D.: A Tremor in the Blood: Uses and Abuses of the Lie Detector. Perseus Publishing, Reading (1998) 3. Damphousse, K.R., Pointon, L., Upchurch, D., Moore, R.K.: Assessing the Validity of Voice Stress Analysis Tools in a Jail Setting. Report submitted to the U.S. Department of Justice (2007) 4. Eriksson, A., Lacerda, F.: Charlantry in forensic speech science: A problem to be taken seriously. International Journal of Speech, Language and the Law 14(2), 169–193 (2007) 5. Harnsberger, J.D., Hollien, H., Martin, C.A., Hollien, K.A.: Stress and Deception in Speech: Evaluating Layered Voice Analysis. Journal of Forensic Science 54(3), 642–650 (2009) 6. Benus, S., Enos, F., Hirschberg, J., Shriberg, E.: Pauses in deceptive Speech. In: Proceedings ISCA 3rd International Conference on Speech Prosody, Dresden, Germany (2006) 7. Feeley, T.H., deTurck, M.A.: The behavioural correlates of sanctioned and unsanctioned deceptive communication. Journal of Nonverbal Behavior 22(3), 189–204 (1998) 8. Stroemwall, L.A., Hartwig, M., Granhag, P.A.: To act truthfully: Nonverbal behaviour and strategies during a police interrogation. Psychology, Crime and Law 12(2), 207–219 (2006) 9. Anolli, L., Ciceri, R.: The Voice of deception: Vocal strategies of naïve and able liars. Journal of Nonverbal Behavior 21(4), 259–284 (1997)
Acoustic Correlates of Deceptive Speech – An Exploratory Study
37
10. Rockwell, P., Buller, D.B., Burgoon, J.K.: The voice of deceit: Refining and expanding vocal cues to deception. Communication Research Reports 14(4), 451–459 (1997) 11. Enos, F.: Detecting Deception in Speech. PhD Thesis submitted to Columbia University (2009) 12. Porter, S.B.: The Language of Deceit: Are there reliable verbal cues to deception in the interrogation context? Master’s thesis submitted to The University of British Columbia (1994) 13. Boersma, P., Weenink, D.: Praat: doing phonetics by computer (Computer program). Version 5.2.12, http://www.praat.org/ (retrieved January 28, 2011) 14. Wells, J.C.: Accents of English I: An Introduction. Cambridge University Press, Cambridge (1982)
Kansei Evaluation of HDR Color Images with Different Tone Curves and Sizes – Foundational Investigation of Difference between Japanese and Chinese Tomoharu Ishikawa1, Yunge Guan1, Yi-Chun Chen1, Hisashi Oguro1, 2, Masao Kasuga1, and Miyoshi Ayama1 1
Graduate School of Engineering, Utsunomiya University, 7-1-2 Yoto Utsunomiya Tochigi, 321-8585, Japan 2 TOPPAN PRINTING Co., Ltd., 1-3-3 Suido, Bunkyo, Tokyo 112-8531, Japan [email protected], {mt106623,dt097173}@cc.utsunomiya-u.ac.jp, [email protected], {kasuga,miyoshi}@is.utsunomiya-u.ac.jp
Abstract. High dynamic range (HDR) color images are evaluated for Kansei impression by two groups of observers: Japanese and Chinese. Twenty HDR images were created by converting each of five HDR images with different tone curve properties into four screen sizes. As a result, the subjective rating value for the psychophysical properties of images, such as “Light” and “Dark,” increased or decreased monotonically with the average brightness L*, but not with image size. On the other hand, the rating value for some Kansei evaluations, including “Natural” and “Clear,” followed the same patterns. Next, we applied factor analysis to the results, having divided the data into Japanese and Chinese. The analysis result indicated that two and three factors were extracted from the rating value evaluated by Chinese and Japanese participants, respectively. These results suggest that Japanese observers evaluated HDR images in more detail than Chinese ones did. Keywords: Kansei Evaluation, High Dynamic Range, Japanese and Chinese, Tone Curve, Screen Size.
1 Introduction Currently, various types of image contents are being delivered all over the world though the Internet. The quality of these images is judged on the basis of observers’ experience and knowledge, and, for example, the quality of image gradation and resolution and screen size [1]–[3]. In many cases, observers’ impressions do not agree, even when they receive the same content. This is a serious problem for creating Web content. In previous studies on this problem, the relationship between the image qualities, including lightness contrast and image size, and the effect of value judgment upon Kansei impression have been investigated [4], [5]. The results showed that D. Harris (Ed.): Engin. Psychol. and Cog. Ergonomics, HCII 2011, LNAI 6781, pp. 38–44, 2011. © Springer-Verlag Berlin Heidelberg 2011
Kansei Evaluation of HDR Color Images with Different Tone Curves and Sizes
39
adjectives were divided into three groups. The first and second groups, which were related to psychophysical properties, were strongly affected by lightness contrast and image size, respectively. On the other hand, the third group, which was related to the Kansei impression, was affected by both lightness contrast and image size. This study aims to examine whether comparable results can be obtained when an image of a different type is evaluated, and what difference appears in evaluations by observers of different national origins. Specifically, high dynamic range (HDR) images [6], which have a relatively fine gradation of the image, are evaluated for Kansei impression by two groups of observers―Japanese and Chinese―who are assumed to have different value judgments. The results show that the subjective rating value for psychophysical properties of the image, such as “Light,” “Dark,” “Deep color,” and “Pale color,” increased or decreased monotonically with average brightness L*, but not with image size. On the other hand, the rating value for some Kansei evaluations, including “Natural,” “Unnatural,” “Clear,” and “Vague,” followed these same patterns. Next, we applied factor analysis to the results, having divided the data into Japanese and Chinese. The result indicated that two and three factors were extracted from the rating value evaluated by Chinese and Japanese participants, respectively. This indicates that Japanese observers evaluated HDR images in more detail than Chinese ones did.
2 Experiment 2.1 Experimental Stimuli Nightscape images were taken by a digital camera (Nikon D50) whose exposure values were set into ten steps. An HDR image was created by the synthesis of these photos using Photomatix Pro 3.0, and then, five HDR images were generated with Photoshop CS4 by changing the tone curves. Two kinds of image contents are shown in “Amusement Park” (Fig. 1) and “Shopping Mall” (Fig. 2). Twenty HDR images, made by converting each of the five HDR images into four screen sizes (7, 14, 29, and 57in), were employed as the experimental stimuli.
Fig. 1. Amusement Park
40
T. Ishikawa et al.
Fig. 2. Shopping Mall
2.2 Experimental Conditions and Procedure Observers were 24 participants with normal color vision: 12 Chinese and 12 Japanese. An experimental booth was constructed by covering walls with gray curtains. Ambient light was provided with a fluorescent light fixture in the ceiling of the room. Horizontal and vertical illuminances near the center of the display were about 285 and 350 lx, respectively. A participant entered the booth and was allowed 3 min to adapt to the visual environment before being instructed to evaluate the experimental stimulus by choosing his/her evaluation using a seven-step scale, from 0 to 6, for each adjective listed on the answer sheet. The experimental stimulus was presented on a 65-in display (SHARP AQUOS LC-65RX1W), and the participant was given as much time as desired to carry out the evaluation. In order to avoid comparison between successive images, a homogeneous gray plane (N5) was presented between each pair of HDR images for 5 s. Two visual distances were used: 160 and 320 cm. The unipolar scale method was employed using 22 adjectives. These adjectives (Table 1) were provided in the native language of the participant. The appearance of the experiment is shown in Fig. 3. Table 1. Twenty-two adjectives Psychophysical properties
Kansei evaluations
Strong contrast
Weak contrast
Light
Dark
Deep color
Pale color
Clear
Vague
Showy
Plain
Natural
Unnatural
Clean
Dirty
Like
Hate
Easy to view
Difficult to view
Stereoscopic
Flat
Impressive
Ordinary
*Dynamic
*Static
*Amusing
*Uninteresting
*These adjectives were added to the evaluation of “Amusement Park”.
Kansei Evaluation of HDR Color Images with Different Tone Curves and Sizes
41
Fig. 3. Appearance of the experiment
3 Result and Analysis The average of the subjective rating values for each image, adjective, observer group, image size, and visual distance was calculated. In several cases, the degree of change in the subjective rating value with an increase in the average brightness L* for Image-I was larger than that for Image-II, and the rating value for Chinese participants was larger than that for Japanese ones. In other words, the result indicated that the differences in subjective rating value for the observers and image contents are influenced by the L*value. Fig.4 shows the subjective rating values of the assessment word for “Light” of Japanese participants for a visual distance of 160cm using Image-I. As shown in the figure, the subjective rating values increase with an increase in average brightness L* for all image sizes. On the other hand, the rating values for “Dark” decrease with the increase in the L*value for all image sizes. These results are nearly symmetrical and both show monotonous change with changing L*value. The results for different image sizes show a similar tendency, indicating that image size has no particular effect on the assessment of “Light” and “Dark.” Similar results are obtained for the assessment words for “Deep color” and “Pale color.” These words express the psychophysical properties of the image. Therefore, the evaluation of psychophysical properties of an image simply changes with an increase in the L*value.
Subjective Rating Value
7inch 14inch
29inch
57inch
Average Brightness (L*)
Fig. 4. Subjective rating values for “Light” with a visual distance of 160 cm for Image-I, rated by Japanese participants
42
T. Ishikawa et al.
The subjective rating value for the assessment word for “Natural” indicated in Fig.5(a) shows a maximum at the L*value (12.9 or 21.0) for the 7, 14, and 29in screens by Japanese participants. Results similar to Fig.5(a) are observed in other Kansei evaluations, e.g., for “Clear.” Although Fig.5(b) shows the result under the same conditions but with Chinese participants, the L*value (21.0 or 33.2) at which the maximum is attained is higher than that for Japanese ones. Results similar to those in Fig.5(b) are observed in Kansei evaluations, e.g., for “Like.” Therefore, Kansei evaluations show different trends than psychophysical evaluations do. Moreover, the rating value of “Easy to view” shows a size dependence in both observer groups; however, the degree of dependence is significantly greater among Chinese rather than among Japanese participants. These results indicate that the evaluation of HDR image approximately corresponds to the results of previous research. (a) Subjective Rating Value
7inch
14inch 29inch
57inch
6
(b)
7inch
5
Subjective Rating Value
14inch 29inch
4
57inch
3 2 1 0 7.7
12.9 21.0 33.2 Average Brightness (L*)
50.3
Fig. 5. (a) Subjective rating values for “Natural” with a visual distance of 160 cm in the case of Image-I as rated by Japanese participants. (b) Corresponding rating values of Chinese participants.
In order to extract the factors that contribute to the Kansei evaluation, we applied factor analysis to the 40 results (20 images × 2 distances; in each of them, the average
Kansei Evaluation of HDR Color Images with Different Tone Curves and Sizes
43
Table 2. Results of factor analysis for Image-II Japanese
Chinese
Adjectives
Factor 1
Factor 2
Factor 3
Factor 1
Factor 2
Easy to view
0.87
0.07
0.35
-0.96
0.04
Like
0.86
0.23
-0.18
-0.97
0.14
Clear
0.85
-0.33
-0.15
-0.95
0.19
Impressive
0.81
-0.43
0.09
-0.96
0.15
Clean
0.80
-0.08
-0.31
-0.95
0.22
Stereoscopic
0.74
-0.25
0.52
-0.96
0.09
Flat
-0.75
0.16
-0.37
0.91
-0.15
Hate
-0.81
-0.38
0.15
0.96
-0.17
Difficult to view
-0.90
-0.14
-0.20
0.97
-0.08
Ordinary
0.27
0.86
-0.11
0.94
-0.19
Plain
-0.27
0.80
-0.43
0.87
-0.33
Natural
0.59
0.74
-0.20
-0.94
0.00
Dark
-0.30
0.69
-0.59
0.87
-0.40
Light
0.34
-0.73
0.53
-0.86
0.42
Unnatural
-0.49
-0.74
0.34
0.93
-0.03
Showy
0.30
-0.83
0.40
-0.95
0.22
Vague
-0.07
-0.06
0.83
0.97
-0.12
Dirty
-0.42
-0.40
0.63
0.95
-0.20
Weak contrast
0.11
-0.41
0.82
0.09
0.51
Pale color
0.13
-0.60
0.73
-0.22
0.79
Strong contrast
0.01
0.19
-0.75
0.10
-0.77
Deep color
-0.04
0.48
-0.82
0.49
-0.67
scores of 12 participants were used), having divided the data into Japanese and Chinese. After the factor extraction, varimax rotation was used following the main factor method. The results for the Japanese participants indicated that three factors, with an eigenvalue larger than 1, were extracted, and their cumulative contribution rates were larger than 80%; in contrast, two factors were obtained in the case of Chinese participants. The results of factor analysis for Image-II are listed in Table 2. In the case of Japanese participants, vision perception characteristics, including “Stereoscopic” and “Easy to view,” showed the largest values for factor loading in the first factor, and these were extracted as the “Vision perception factor.” In the second factor, impressive quality, including “Natural” and “Showy,” had the largest values for factor loading; this factor was named the “Impressive quality factor.” The third
44
T. Ishikawa et al.
factor was called the “Contrast factor,” as its elements were “Deep color” and “Strong contrast.” On the other hand, the first factor for Chinese participants was almost identical to the first and second factors for Japanese ones, and its second factor was almost identical to the third factor for the Japanese group. These results suggest that Japanese participants evaluate HDR images in more detail than Chinese ones do.
4 Conclusion In this study, to clarify the difference in Kansei evaluation by observers of different national origins―Japanese and Chinese―HDR images were evaluated for Kansei impression. The results showed that the subjective rating value for psychophysical properties of the image, such as “Light,” “Dark,” “Deep color,” and “Pale color,” increased or decreased monotonically with average brightness L*, but not with image size. On the other hand, the rating value for some Kansei evaluations, including “Natural,” “Unnatural,” “Clear,” and “Vague,” followed the same pattern. These results indicate that the evaluation of HDR image approximately corresponds to the results of previous research. Moreover, we applied factor analysis to the results, having divided the data into Japanese and Chinese. The result indicated that two and three factors were extracted from the rating values generated by Chinese and Japanese participants, respectively. These results suggest that Japanese participants evaluated HDR images in more detail than Chinese ones did. Acknowledgment. This research was supported by Eminent Research Selected at Utsunomiya University.
References 1. Duchnicky, R.L., Kolers, P.A.: Readability of Text Scrolled on Visual Display Terminals as a Function of Window Size. Human Factors 25, 683–692 (1983) 2. Chae, M., Kim, J.: Do Size and Structure Matter to Mobile Users? An Empirical Study of the Effects of Screen size, Information Structure, and Task Complexity on User Activities with Standard Web Phones. Behaviour and Information Technology 23(3), 165–181 (2004) 3. Hatada, T., Sakata, H., Kusaka, H.: Induced Effect of Direction Sensation and Display Size: Basic Study of Realistic Feeling with Wide Screen Display. The Journal of the Institute of Television Engineers of Japan 33(5), 407–413 (1979) (in Japanese) 4. Ishikawa, T., Shirakawa, T., Oguro, H., Guo, S., Eda, T., Sato, M., Kasuga, M., Ayama, M.: Color Modulations Appropriate for Different Image Size Based on Impression Assessment. In: 11th Congress of the International Colour Association AIC 2009, Sydney (2009) 5. Chen, Y.-C., Ishikawa, T., Shirakawa, T., Eda, T., Oguro, H., Guo, S., Sato, M., Kasuga, M., Ayama, M.: Effects of Lightness Contrast and Image Size on KANSEI Evaluation of Photographic Pictures. Kansei Engineering and Emotion Research KEER 2010, Paris (2010) 6. Horiuchi, T., Fu, Y.Q., Tominaga, S.: Perceptual and Colorimetric Evaluations of HDR Rendering with/without Real-world Scenes. 11th Congress of the International Colour Association AIC 2009, Sydney (2009)
Personalized Emotional Prediction Method for Real-Life Objects Based on Collaborative Filtering Hyeong-Joon Kwon, Hyeong-Oh Kwon, and Kwang-Seok Hong School of Information and Communication Enginnering, Sungkyunkwan University, 300, Chunchun-dong, Jangan-gu, Suwon, Kyungki-do, 440-746, South Korea {katsyuki,oya200}@skku.edu, [email protected]
Abstract. In this paper, we propose a personalized emotional prediction method using the user’s explicit emotion. The proposed method predicts the user’s emotion based on Thayer's 2-dimensional emotion model, which consists of arousal and valence. We construct a user-object dataset using a self-assessment manikin about IAPS photographs, and predict the target user’s arousal and valence by collaborative filtering. To evaluate performance of the proposed method, we divide the user-object dataset into a test set and training set, and then observe the difference between real emotion and predicted emotion in the 2-dimensional emotion model. As a result, we confirm that the proposed method is effective for predicting the user’s emotion. Keywords: Emotional Prediction, IAPS, Self-assessment Manikin.
1 Introduction One of the main study areas of the HCI field is building a computer with human communication abilities. This has been preceded by a one-way interface to mutual interaction based on physical communication between a computer and human. As a result, computer vision and speech recognition (which provides technology with the human ability to recognize objects and humans using eyes and ears) took a major step forward. Recent studies reproduce human sense organs such as eyes and ears by computer using pattern recognition and machine learning technology. In addition, more studies on human emotion recognition (called emotion computing) is in progress. The method based on sight is processed with a face visual to determine different features of emotion from one’s facial expressions [10], and is recognized by computer. Javier Movellan has announced that a computer can determine great and small changes from every part of a human face to recognize emotion such as anger, sadness, displeasure, pleasure, and more [1]. The method based on hearing uses different human voice features from emotional changes. Speech signals differ for each individual for distinct articulator and speaking habits, so usual voice signals and other voice signals with features are detected to recognize emotion [8][9]. However, these approaching methods involve several limitations. First, visual and voice methods are useful to recognize outward appearance constituents. But emotion is the constituent not exposed by outward appearance, so there is always a limitation D. Harris (Ed.): Engin. Psychol. and Cog. Ergonomics, HCII 2011, LNAI 6781, pp. 45–52, 2011. © Springer-Verlag Berlin Heidelberg 2011
46
H.-J. Kwon, H.-O. Kwon, and K.-S. Hong
from using visual and voice methods. Second, facial expressions and human voices differ for all individuals on each emotion. Humans may express the exact facial expression and voice to represent different emotions, and emotions toward the same object may vary. This means that existing methods using visual and vocal features cannot find the sensibility of an individual, and have limitations in emotion recognition. Human emotions differ depending on individual sensibilities, so the computer must reflect individual sensibilities for more accurate emotion recognition. For example, an object may cause one person to feel sad, while another encounter with the same object may not cause sadness; this is because of the difference in sensibilities due to individual experience. Therefore, the study of emotion prediction or recognition methods based on the sensibilities of individuals is necessary. In this paper, we propose a personalized emotional prediction method for real-life objects, IAPS photographs [2], based on and SAM [3] and collaborative filtering [5] which uses individual tendencies. Proposed method uses emotion data of users with similar sensibilities compared to a target user on a certain object to predict the target user’s emotion. This emotion recognition method uses the convergence of an emotion induction dataset and soft computed feelings of computer science built as emotion based on psychological basis. Performance evaluation criteria are 1) Lineal distance of subjective emotion of actual user and emotion predicted using proposed method, and 2) The mean error between subjective emotion evaluation score rated by user and emotion evaluation predicted with proposed method. This paper is organized as follows. In section 2, we describe IAPS and collaborative filtering to design the proposed method. In section 3, we propose a personalized emotional prediction method using memory-based collaborative filtering with explicit SAM rating on IAPS photographs. In section 4, we show the prediction accuracy of SAM assessment and the overall performance of emotional prediction. In section 5, we summarize the results of this study and suggest topics for future work.
2 Related Works In this section, we describe the IAPS photographs dataset that promotes human emotion, an effective method for evaluating explicit emotion (SAM), and memorybased collaborative filtering (MBCF), which is a collective intelligence algorithm in the realm of artificial intelligence. We design and implement a personalized emotional prediction method using IAPS photographs, CF, and SAM in section 3. 2.1 IAPS and SAM The Interactive Affection Picture System (IAPS) includes 1,200 photographs that stimulate human emotions [2]. Existing studies measure emotion toward IAPS photographs via SAM, which consists of pressure, arousal, and dominance. SAM contains various merits. First, SAM doesn’t contain a language. Thus, pure emotion is measured without nationality. Second, even illiterate users can respond emotionally because SAM is designed as a simple picture. Third, because same pictures about each element are used, it is possible between different cultural areas. Fig. 1 shows our SAM format, which contains valence and activation [3].
Personalized Emotional Prediction Method for Real-Life Objects
47
Fig. 1. Self-assessment Manikin
2.2 CF: Collaborative Filtering The CF is a famous collective intelligence algorithm. It discovers similar points among people and predicts a preference rating for an unfamiliar item between similar people. Modern recommender systems are based on CF. The CF is divided into memory-based CF and the model-based approach [5]. 1. One of the model-based CF approaches uses a clustering algorithm. This method divides all users in a user-item rating matrix into several groups. Clustering is a technique widely used for the statistical analysis of data, and is an important unsupervised learning method. To identify interesting data distributions and patterns, clustering techniques classify physical or abstract objects into classes such that the objects in each class share some common attribute. Depending on the characteristics of the distinct clusters, companies can make independent decisions about each cluster. The model-based CF is robust to cold-start problems, but it is a highly vulnerable real-time prediction because of the training process. 2. MBCF calculates the similarity between users or items based on the user-item rating matrix [4]. To predict user m for item n, the user-based approach calculates the similarity of co-ratings from two users, and arrays all other users in order of similarity to user m. The item-based approach then calculates the similarity between two items using their co-ratings, and arrays all other items in order of similarity. The MBCF is robust to real-time recommendation, but has an increased calculating cost. We predict arousal and valence ratings based on this method in this paper.
3 Proposed Emotional Prediction Method Fig. 2 shows the structure of a user-object dataset and Thayer’s 2-dimensional emotion model. On the left, n indicates the IAPS photograph number, m indicates a user, and r means rating. The proposed method contains two matrices like Fig. 2: arousal and valence. We can know a user’s emotion toward a photograph based on the arousal and valence coordinates. Thayer’s 2-dimensional emotion model shows the user’s emotion using a pair of arousal and valence coordinates about a photograph. Dominance is not used in Thayer’s 2-dimensional emotion model.
48
H.-J. Kwon, H.-O. Kwon, and K.-S. Hong
Fig. 2. A structure of a user-object dataset and Thayer’s 2-dimensional emotion model
A generalized step of the proposed method is described as follows: 1. It collects arousal and valence ratings about various real-life objects from a great number of users, and then constructs a user-object dataset, which consists of an arousal matrix and valence matrix. We use IAPS photographs on behalf of real-life objects. Also, we developed software to collect SAM ratings about IAPS photographs. 2. It chooses a target user who is somebody to predict emotion. The target user must have been evaluates objects in user-object dataset. This is a necessary condition to use collaborative filtering. 3. It shows the target object to the target user. Then, it searches top-k similar users which is rated target object. When it searches similar users, it considers linear similarity algorithms such as PCC, COS, RMS, and so on. The vector space modelbased method includes Cosine Similarity (COS), which is frequently used in information retrieval. The COS assumes that the rating of each user is a point in vector space, and then evaluates the cosine angle between the two points. It considers the common rating vectors X = {x1, x2, x3, ۰ ۰ ۰ , xn} and Y = {y1, y2, y3, ۰ ۰ ۰ , yn} of users X and Y. It is represented by a dot-product and magnitude. COS has frequently been used for performance comparisons in the CF area. The cosine angle between vectors X and Y is given by:
Personalized Emotional Prediction Method for Real-Life Objects
cos( X ,Y ) =
x1 y1 + x2 y 2 h xn yn X •Y = 2 2 2 2 2 2 X Y x1 + y1 x2 + y2 h xn + yn
49
(1)
One correlation-based method is the Pearson dot-product Correlation Coefficient (PCC). This method is normally used to evaluate the association intensity between two variables. PCC is given by: n ( xi − X )( yi − Y ) ∑ i =1 γ ( X ,Y ) = (2) n n ∑i =1 ( xi − X ) 2 ∑i=1 ( yi − Y ) 2 Row moment-based similarity (RMS) is divided into three steps based on the rating matrix [6]. The first step is target user profiling, which evaluates the similarity between the target user and everybody else. Assume that the target user is u3. The similarity between u3 and u5 is evaluated by co-ratings composed of the user’s ratings for the same contents. The r3,4 and r5,4 values are one of the co-ratings between u3 and u5. The COS and PCC are common similarity algorithms in the MBCF study area. This similarity is faster and simpler than well-known existing algorithms. 1 1 n ∑ (| u i ,v − u j ,v |) k r k n v =1 m 1 k = 1 − k ∑ z =1 Pr( D =d z ) ⋅ d z r
RMS (u i , u j ) = 1 −
(3)
4. It predicts valence and arousal ratings based on a user-object dataset. It then maps arousal and valence coordinates to Thayer’s 2-dimensional emotion model. This shows the target user’s emotion toward the target object. After calculating the similarity for all pairs of objects, the MBCF selects top-k similar users. Accordingly, the MBCF predicts the rating of the target user for the target item with similarities for weights. (4) shows the prediction method, where u indicates the target user, i indicates the target item, j indicates other users, and w is similarity between u and j [4].
∑ +
n
p (u, i ) = ru
j =1
w j ,v ( r j , i − r j )
∑
n j =1
wu , j
(4)
5. It measures prediction accuracy. Mean absolute error is used between predicted valence and arousal ratings and real valence and arousal ratings. In (5), p indicates prediction rating and q indicates real rating, while n represents the total prediction counts. n | pi − qi | ∑ i =1 (5) MAE = n
50
H.-J. Kwon, H.-O. Kwon, and K.-S. Hong
6. It maps a point using prediction arousal and valence ratings on Thayer’s 2dimensional emotion model, and then maps a point using real arousal and valence ratings on the same emotion model. Next, it measures the Euclidean distance between the real point and prediction point. This distance is the major performance of the proposed method. The two points in Fig. 2 represent this.
4 Experimental Results We have collected subjective emotions of individual users using SAM over IAPS of an often used emotion induction dataset from 69 users to build individual emotion sensibility data. IAPS consists of 1,200 photographs of commonly seen objects from human life. Extremely suggestive or cruel photographs were excluded; 1,073 photographs were composed randomly into 8:2 ratios (80 % of photographs as training data and 20 % as test data).
Fig. 3. Arousal prediction result
Fig. 4. Valence prediction result
Personalized Emotional Prediction Method for Real-Life Objects
51
We then used PCC, COS, RMS, and top-k user’s rating average to predict 20 % test data ratings. The prediction results are shown in Fig. 3 and Fig. 4. The MAE (absolute average of the difference between real rating and prediction rating) curve for MBCF descends with increasing neighbors and rises at the optimal point. The experimental result shows a remarkable result in arousal and valence datasets. The Proposed method showed greater prediction accuracy than top-k average. Table 1. Emotion prediction result based on Thayer’s 2-dimensional model
Similarity method COS PCC RMS AVG. of k User AVG. Item AVG.
Optimal top-k 6 14 6 10 -
Error distance 2.3477 2.6607 2.3624 2.3489 2.5998 2.8610
Next experiment, in Table 1, is to observe the error distance of a 2-dimensional emotion model. This experiment predicts arousal and valence ratings about target objects, and marks real values and predicted values on Thayer’s 2-dimensional emotion model. We then measured the lineal absolute distance error between the two points. Euclidean distance was used as a distance measurement. Table 1 shows the result and experimental condition which contains optimal top-k value per similarity method.
5 Conclusion We proposed an emotion prediction method using MBCF. The proposed method predicts arousal and valence ratings based on a user-object dataset and discovers user emotions based on Thayer’s 2-dimensional emotion model. From the experimental results, the proposed method showed greater prediction accuracy than a simple average. In the future, we will study a multimodal emotion recognition method using face or speech. This approach will be an optimal recognition method using external and internal human elements. Acknowledgments. This research was supported by Basic Science Research Program through the National Research Foundation of Korea(NRF) funded by the Ministry of Education, Science and Technology(2010-0021411), and by MKE, Korea under ITRC NIPA-2010-(C1090-1021-0008) (NTIS-2010-(1415109527)).
References 1. Bartlett, M.S., Littlewort, G., Fasel, I., Movellan, J.R.: Movellan: Real Time Face Detection and Facial Expression Recognition: Development and Applications to Human Computer Interaction. In: The 2003 IEEE Conference on International Conference on Computer Vision and Pattern Recognition Workshop, pp. 1–6. IEEE Press, Nwe York (2003)
52
H.-J. Kwon, H.-O. Kwon, and K.-S. Hong
2. Lang, P.J., Bradley, M.M., Cuthbert, B.N.: International Affective Picture System (IAPS): Technical Manual and Affective Ratings. NIMH Center for the Study of Emotion and Attention (1997) 3. Bradley, M.M., Lang, P.J.: Measuring emotion: The self-assessment manikin and the semantic differential. Journal of Behavior Therapy and Experimental Psychiatry 25(1), 49– 59 (1994) 4. Herlocker, J.L., Konstan, J.A., Borchers, A., Riedl, J.: An Algorithmic Framework for Performing Collaborative Filtering. In: The 22nd Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR 1999), pp. 230–237. ACM Press, New York (1999) 5. Adomavicius, G., Tuzhilin, A.: Toward the Next Generation of Recommender Systems: A Survey of the State-of-the-art and Possible Extensions. IEEE Transactions on Knowledge and Data Engineering 17(6), 734–749 (2005) 6. Kwon, H.-J., Hong, K.-S.: Moment Similarity of Random Variables to Solve Cold-start Problems in Collaborative Filtering. In: Third International Symposium on Intelligent Information Technology Application, pp. 584–587. IEEE Press, New York (2009) 7. Ahn, H.J.: A New Similarity Measure for Collaborative Filtering to Alleviate the New User Cold-start Problem. Information Science 178(1), 37–51 (2008) 8. New, T.L., Foo, S.W., De Silva, L.C.: Speech Emotion Recognition using Hidden Markov Models. Speech Communication 41(4), 603–623 (2003) 9. Roh, Y.-W., Kim, D.-J., Lee, W.-S., Hong, K.-S.: Novel Acoustic Features for Speech Emotion Recognition. Science in China Series E: Technological Sciences 52(7), 1838– 1848 (2009) 10. De Silva, L.C., Miyasato, T., Nakatsu, R.: Facial Emotion Recognition Using Multi-modal Information. In: The International Conference on Information, Communications and Signal Processing ICICS 1997, pp. 397–401. IEEE Press, Los Alamitos (1997)
A Study of Vision Ergonomic of LED Display Signs on Different Environment Illuminance Jeih-Jang Liou, Li-Lun Huang, Chih-Fu Wu, Chih-Lung Yeh, and Yung-Hsiang Chen The Graguate Institude of Design Science, Tatung University
Abstract. The LED (light emitting diode, also referred to as LED) have already been used widely. However, despite the high visibility of LED with high brightness performance, it also leads to a glare problem, which generates a direct security issue in applying to traffics. Therefore, this research aimed to study how to make the LED display sign be more legible under high illuminative environments and to avoid the observers feeling dazzling glare under low illuminative environments. This research firstly studied the literatures to explore the drivers’ visual ergonomic as well as the optical properties of LED, and investigated the relatively existing norms for engineering vehicle LED display signs. Three variables were set in this study: three kinds of ambient illumination, four kinds of luminance contrast and two kinds of character form. In the first phase of the experiment, subjects observed LED display signs in both near and distant locations and filled out the SWN scale (Subjective Well-being under Neuroleptics), and in the second phase, subjects were then asked to moved forward and recorded their perceptions of comfort and glare to distance range. The findings demonstrated that, there was no variation in subjective evaluation to display signs with no backgrounds either in the near or distant locations, while to display signs with backgrounds, the subjects perceptions were the farther the distance, the clearer the legibility; higher ambient illumination could effectively reduce observers’ glare perception to LED display signs; display signs with backgrounds at the luminance contrast of 3:1 (L max = 3100, L min = 1033 cd / ) showed the lowest uncomfortable and glare level to observers. The two forms of character showed no significant variation in affecting observers in terms of the comfort and glare perception.
㎡
Keywords: LED display signs, engineering vehicles, legibility, ambient illumination, luminance contrast.
1 Introduction LED, a light source made from the semiconductor technology, has been widely used in IT products, communication electronics, display panel, traffic signal, and various instrumental displays. Owing to the price drop and improving product features, HighBrightness LED has gradually replaced the traditional LED, in addition, as the emerging markets have used High-Brightness LED directly, the traditional LED is now only partly utilized in signs, lighting, electronic equipments etc., and its market has been extremely shrinked. D. Harris (Ed.): Engin. Psychol. and Cog. Ergonomics, HCII 2011, LNAI 6781, pp. 53–62, 2011. © Springer-Verlag Berlin Heidelberg 2011
54
J.-J. Liou et al.
Currently, High-Brightness LEDs are mainly applied to mobile phones, displays, automotives, lightings, signal lights, and other areas. In viewing the application and future forecast, the DisplaySearch pointed out that the outdoor display has been ranked in the top two demands of High-Brightness LED. Therefore, the LED display immediately provides important and dynamic traffic information to people. According to current experience of LED displays application, in order to obtain an ideal display effect outdoor, the brightness has to be over 4000 cd/ [1]; However, in the night, such high brightness might make the observers feel discomfortable and glare. Take the arrow direction traffic light on the rear of national highway engineering vehicle for example, according to the statistics of National Freeway Bureau, there is about six constructional fatal accidents per year, in which three of them usually occurred during the operation of mobile engineering trucks. Although those accidents were not cause by engineering trucks, all seemed to be relevant to the insufficiency of warning signs on the vehicles and related constructional spots, which also showed a high rate of car accident on the highway constructional operation. Therefore, this research aimed to study how to make the brightness of LED display sign on engineering trucks to reach the most comfortable visual distance which allows drivers to have more time to do proper reactions and judgments, as well as avoid causing glare to drivers when they are driving at close range. Kazunori Munehiro et al. [2] study pointed out that the legible distance of LED guide light in the daytime is father than road marking, while in the mist the glare level will rise with the increased brightness of LED. As this is the situation, an appropriate change of brightness of LED is necessary. Therefore, installing LED on road traffic does improve legibility of drivers. The study of Uchida Kazuhiro et al [3] has indicated that the low legibility during the night is because of the high contrast brought out by too high brightness, while the text and background within a certain contrast can enhance the effect of visual recognition. On aspect of the application of LED display signs in each country, according to the Highway Construction Traffic Control Manual of Directorate General of Highways, MOTC [4], there are at least four levels of the brightness of LED display signs in which the darkest level must be the half of the brightest, and the intensity of each LED light must not be less than 2cd. .In the manual of the road construction traffic control and safety equipment of Macao Special Administrative Region Transport Bureau [5], it also mentions that the brightness of LED light on the engineering vehicle should be adjusted in accordance with the surrounding illumination. Japan only stipulates two levels of adjustment for the brightness of LED display signs, dividing into day and night luminance in accordance with the difference of color. From the various regulations for variable message sign and LED signal on engineering vehicles set by each county, it shows that currently there is no uniform provision to the brightness of LED application on engineering truck. Therefore, this research focused on the brightness contrast background between LED variable message sign and LED display panel, conducting the study of visibility under different illumination environments, to investigate how the luminance contrast of LED display sign reached the farthest visible distance under different illumination environments and without creating glare to drivers in close distance.
㎡
A Study of Vision Ergonomic of LED Display Signs on Different Environment
55
2 Research Methods and Experimental Design 2.1 Research Methods The purpose of this study was to explore: (1) The farthest and comfortable distance of the luminance contrast of LED display sign under different ambient illuminances. (2) The effect of the character form of LED display signs to subjects’ legibility. Experimental Variables (1) The luminance contrast (Lc) of LED sign This experiment used dimmer to control the different brightnesses of LED and LP9221 UNM6 Luminometer to measure the luminance. The luminance contrast between sign and background was divided into four levels, Lc1(L max = 6200 cd/ L min = 0 cd/ ) Lc2(L max=6200 cd/ Lmin = 3100 cd/ ) Lc3(L max = 6200 cd/ L min = 2066 cd/ ) Lc4(L max =3100 cd/ L min = 1033 cd/ ). (2) Ambient illumination This experiment was conducted at a built darkroom in the third floor of the Department of Industrial Design of Tatung University, and simulated the illumination environments around the clock by using three halogen lamps, in which one lamp was connected to dimmer to control the illumination. The variables of environment illumination were set for three classes, including the bright day (30000 Lux), the dark day (5000 Lux) and night (10Lux). LP 9221 S1 Illuminometer was used to measure the illuminance. (3) Character exhibition According to the highway construction traffic control manual of Directorate General of Highways, the main directional contents of advance warning arrow sign on engineering warning car are "← -" and "