SlideShare a Scribd company logo
1 of 37
Beyond Mouse and Keyboard:
Post-WIMP and Novel Forms of
Interaction
                                                             Jacques Chueke
                                                        London, UK, May 2011

                                                             George Buchanan
                                                                (1st Supervisor)
                                                Lecturer, Centre for HCI Design

                                                              Stephanie Wilson
                                                               (2nd Supervisor)
                                                Lecturer, Centre for HCI Design




Master in Design, PUC-Rio, RJ, Brazil
PhD Researcher at the Centre for HCI Design
School of Informatics, City University London                                1
Things are changing…




                       iPad: 1 year old growing among touch screens and print

                                                                                2
Things are changing…




                       iPad for cats???

                                          3
Overview
• The introduction of novel
  hardware for computing and
  gaming during the last decade
  is changing the way we control
  everyday devices.

• It provides NUI control methods,
  such as haptic (e.g. iPhone,
  iPad, MS Surface), gesture-
  based and voice (e.g.
  Nintendo Wii, Microsoft Xbox
  360 console gaming with Kinect
  sensor) and eye tracking
                                     MIT Media Lab: DepthJS – 2011
  interactions (e. g. Tobii P-10).

• One specific impact this has
  had is on the user’s control of
  such devices.
                                                           4
It’s not supposed to be like this…




                                 Gmail Motion, April 2011

                                                            5
It’s not supposed to be like this…




                                 Gmail Motion, April 2011


                                                            6
New Modes of Interaction
• PrimeSense / MS Kinect: Swim Browser




                        Prime sense browser competition winner: Stolarsky 'SwimBrowser’ – 2011
                                                                                                 7
New Modes of Interaction
• KinVi: Kinect Virtual Interface




                            Prime sense browser competition 2n place: Windows Control with gestures - 2011



                                                                                                             8
New Modes of Interaction
• eviGroup Paddle Pro




                        Front-facing webcam to track head movements for cursor control – 2011
                                                                                                9
New Modes of Interaction
• Hitachi: Hitachi's Gesture Remote Control TV Prototype




                                           CES 2009: Hitachi's Gesture Remote Control TV Prototype

                                                                                                     10
Problem Statement
• New command vocabularies
                                 NEW CONTROL METHODS (INPUT)
have emerged and users do
not know how to access or           NEW INTERFACES (OUTPUT)
activate them.

• This is a timely moment to
research how people make            OLD OUTPUT X NEW INPUT
sense of these technologies
                                    NEW OUTPUT X OLD INPUT
for CONTROL whilst some
Post-WIMP interfaces do not         NEW OUTPUT X NEW INPUT
display appropriate visual
cues for NUI interactions.

                               WIMP-GUI (DESKTOP) X NUI (PHYSICAL)
                                 POST-WIMP X MOUSE/KEYBOARD
                                   POST-WIMP X NUI (PHYSICAL)


                                                                11
Case 1: Microsoft Surface




                                 Media Player, Microsoft Surface. Nov, 2010
    POST-WIMP + NUI (PHYSICAL)                                           12
Case 2: Windows 8 Metro Dashboard
• The new Windows 8 with similar features as used in Windows Phone and Xbox 360 Dashboard.




      POST-WIMP + NUI (PHYSICAL) x                 Metro Dashboard: Windows 8 Start Screen. Feb, 2012
           MOUSE/KEYBOARD                                                                          13
Example: GNOME 3 - desktop environment for GNU/Linux
and UNIX-type operating systems.




                                   GNOME 3 hotcorners and responsive interface, 2012
    POST-WIMP + NUI (PHYSICAL) x
         MOUSE/KEYBOARD                                                           14
Case 3: Tobii P-10 Eye Tracker: Gaze for Control




                                             15
Case 3: Tobii P-10 Eye Tracker: Gaze for Control




                           Tobii P-10 at the SmartLab (UEL), Oct. 2010
                                                                         16
Case 3: Tobii P-10 Eye Tracker: Gaze for Control
• Assistive Technology: Tobii P-10 at the SmartLab (UEL).




                                                      Tobii P-10 equipment, Oct. 2010
      WIMP-GUI (DESKTOP) X NUI (PHYSICAL)                                           17
Case 3: Tobii P-10 Eye Tracker: Gaze for Control




                      Mouse configuration pop up for Windows Control, Tobii P-10
                                                                                   18
Case 4: Tobii LeNovo: Gaze for Control
• Tobii Lenovo / PCEye / Acuity




                                           Tobii LeNovo, Jun 2011
     WIMP-GUI (DESKTOP) X NUI (PHYSICAL)                            19
New Visual Cues/Feedback

                                                             UI affordances are
 Just-in-time chrome                                         shown on tap. Applying
  Can be triggered by                                        the principle of
 touch or proximity –                                        scaffolding will lead you
hover effect (pg. 153)                                       to far more successful
                                                             multi-touch and gesture
                                                             UI’s. (pg. 154)




 Tethers indicate
       that a size                                                 The marking
  constraint (MS                                                   menu system
     Surface) has                                                  teaches users to
been reached on                                                    make pen-based
   an item being                                                   gestures (pg. 150)
  scaled (pg. 91)


                         Wigdor, D. Wixton, D. Brave NUI World, 2011




                                                                            20
Methodology
An empirical study with Eye Tracking (Tobii x60) was conducted to test the
Perceptible Affordances of Drag and Drop interactions within a iGoogle
Personal Web Portal. This study served to create a protocol for analysis which
focuses on the very first 10 seconds of a participant scrutinizing the screen while
trying to respond specific questions.




        Participant, 54, Beginner Expertise                                         Participant, 22, Advanced Expertise

                                         Gazeplot Comparison: Beginner x Advanced                                         21
Quantitative Data Analysis: Gazeplot




    Resulting gaze plot from seven participants during the first 10 seconds after question 01.
                                                                                                 22
Quantitative Data Analysis: Heatmap




    Resulting heat map from seven participants during the first 10 seconds after question 01.
                                                                                                23
Quantitative Data Analysis: Gazeplot




    Resulting heat map from seven participants during the first 10 seconds after question 02.
                                                                                                24
Quantitative Data Analysis: Heatmap




    Resulting heat map from seven participants during the first 10 seconds after question 02.
                                                                                                25
Quantitative Data Analysis: Gazeplot




    Resulting heat map from seven participants during the first 10 seconds after question 03.
                                                                                                26
Quantitative Data Analysis: Heatmap




    Resulting heat map from seven participants during the first 10 seconds after question 03.
                                                                                                27
Quantitative Data Analysis: Gazeplot




    Resulting heat map from seven participants during the first 10 seconds after question 04.
                                                                                                28
Quantitative Data Analysis: Heatmap




    Resulting heat map from seven participants during the first 10 seconds after question 04.
                                                                                                29
Model 1: Perceptible Affordances
 • According to Nielsen (2008):
 "Affordance" means what you can
 do to an object. For example, a
 checkbox affords turning on and off,
 and a slider affords moving up or
 down.

 • "Perceived Affordances" are
 actions you understand just by
 looking at the object, before you
 start using it (or feeling it, if it's a
 physical device rather than an on-
 screen UI element).

 • In Gaver’s (1991) words, “…Perceptible Affordances are
 inter-referential: the attributes of the object relevant for
 action are available for perception. What is perceived is
 what is acted upon.”                                         30
Perceptible Affordances in Post-WIMP

                    Post-WIMP GUI [INTERFACE LAYER]
                                         OUTPUT



 - LESS SYMBOLIC
 - MORE INTUITIVE
 (USER)               PERCEPTIBLE AFFORDANCE
 - MORE REACTIVE
 (COMPUTER)




                                          INPUT

                    NUI [MODE OF INTERACTION LAYER]



                                                      31
Model 2: Norman’s Theory of Action
                     Execution Cycle


         Execution   Specification of          Formulation of
                        Actions                  Intention
                       Sequence




                       Interaction




        Perception   Interpretation               Evaluation


                     Evaluation Cycle


                             Preece et al (2009: 121) quoting Norman, (1986)

                                                                               32
Research Question 1: ACTIVATION
                    Post-WIMP GUI [INTERFACE LAYER]
 Evaluation Cycle                        OUTPUT




                      PERCEPTIBLE AFFORDANCE




  Execution Cycle                         INPUT

                    NUI [MODE OF INTERACTION LAYER]



                                                      33
Conclusions and Future Work
• By developing a methodology for an empirical study, which focuses on
observation prior to any interaction, we are willing to identify what elements
people will focus on NUI screens.


• I believe both Norman’s theory of Action and Perceptible Affordances theories
can be combined with the more recent Piaget’s theory of INRC and the
Scaffolding concept; and used on my protocol for analysis.


• A prototype with Post-WIMP characteristics and NUI mode of interaction will be
built in order to to understand how users visually scan such interfaces to obtain
the gist of its interactive potential.


• Quantitative (Eye Tracking) and Qualitative (Verbalizations) data will be
combined to produce conclusions about what kind of information can be obtained
with the protocol – and how can this data be adapted to indicate better design
interactions with NUI systems.

                                                                                 34
Thank you for your attention!




Jacques Chueke
Jacques.chueke.1@city.ac.uk
                                  35
Bibliography
 Beaudouin-Lafon, M. (November 2000). "Instrumental Interaction: An Interaction Model for Designing Post-WIMP
 User Interfaces". CHI '00: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. The
 Hague, The Netherlands: ACM Press. pp. 446–453. doi:10.1145/332040.332473. ISBN 1-58113-216-6.
 http://www.daimi.au.dk/CPnets/CPN2000/download/chi2000.pdf.

 Breeze, James. Eye Tracking: Best Way to Test Rich App Usability. UX Magazine, access on 25 November 2010.
 (http://www.uxmag.com/technology/eye-tracking-the-best-way-to-test-rich-app-usability)

 Buxton, W. (2001). Less is More (More or Less), in P. Denning (Ed.), The Invisible Future: The seamless
 integration of technology in everyday life. New York: McGraw Hill, 145–179 ITU Internet Reports 2005: The
 Internet of Things – Executive Summary.

 Dam, A. (February 1997). "POST-WIMP User Interfaces". Communications of the ACM (ACM Press) 40 (2): pp.
 63–67. doi:10.1145/253671.253708.

 Dourish, P. Where the Action Is: The Foundations of Embodied Interaction. A Bradford Book: The MIT Press, USA,
 2004.

 Ehmke & Wilson, 2007. Identifying Web Usability Problems from Eye-Tracking Data. Published by the British
 Computer Society. People and Computers XXI – HCI…but not the way we know it: Proceedings of HCI 2007.

 Gaver, W. Technology Affordances. Copyright 1991 ACM 0-89791-383-3/91/0004/0079.

 Gentner, D. and Nielsen, J. (April 1993). "The Anti-Mac Interface". Communications of the ACM (ACM Press) 39
 (8): pp. 70–82. http://www.useit.com/papers/anti-mac.html.

 Jacob, R. et al. (2008). "Reality-Based Interaction: A Framework for Post-WIMP Interfaces". CHI '08: Proceedings
 of the Twenty-Sixth Annual SIGCHI Conference on Human Factors in Computing Systems. Florence, Italy: ACM.
 pp. 201–210. doi:http://doi.acm.org.ezproxy.lib.ucf.edu/10.1145/1357054.1357089. ISBN 978-1-60558-011-1. 36
Bibliography
 McGrenere, J., Ho, W. (2000). Affordances: Clarifying and Evolving a Concept. Procs. of Graphic Interfaces 2000,
 Montreal, May 2000.

 McNaughton, J. Utilizing Emerging Multi-touch Table Designs. Technology Enhanced Learning Research Group -
 Durham University. TR-TEL-10-01.

 Nielsen, J. (April 1993). "Noncommand User Interfaces". Communications of the ACM (ACM Press) 36 (4): pp.
 83–99. doi:10.1145/255950.153582. http://www.useit.com/papers/noncommand.html.

 Norman, D. (1999). Affordance, Conventions and Design. In ACM Interactions, (May + June, 1999), 38-42.
 Picard, R. Affective Computing. The MIT Press, Cambridge, Massachusetts. London, England, 1998.

 PREECE, Jenny. SHARP, Helen. ROGERS, Yvonne. Interaction Design: Beyond Human-Computer Interaction
 [2nd edition]. John Wiley & Sons, Ltd. West Sussex, UK, 2009.

 Ramduny-Ellis, D.; Dix, A.; Hare, J.; Gill, S. Physicality: Towards a Less-GUI Interface (Preface). Procs. Third
 International Workshop on Physicality. Cambridge, England, 2009.

 Sorensen, M. Making a Case for Biological and Tangible Interfaces. Proceedings of the Third International
 Workshop on Physicality. Cambridge, England, 2009.

 Sternberg, R. Cognitive Psychology. Wadsworth, Cengage Learning. Belmont, CA, USA, 2009, 2006.

 Vyas, D., Chisalita, C. Veer, G. Affordance in Interaction. ECCE '06 Proceedings of the 13th Eurpoean conference
 on Cognitive ergonomics: trust and control in complex socio-technical systems. ACM New York, NY, USA ©2006
 ISBN: 978-3-906509-23-5

 WIGDOR, Deniel. WIXON, Dennis. Brave NUI World: designing natural user interfaces for touch and gesture. Morgan
 Kauffman Publishers, USA, 2011.                                                                                    37

More Related Content

What's hot

Fjord@ The Future of Broadcasting
Fjord@ The Future of BroadcastingFjord@ The Future of Broadcasting
Fjord@ The Future of BroadcastingFjord
 
INTERACTIVE WHITEBOARD SOLUTIONS
INTERACTIVE WHITEBOARD SOLUTIONSINTERACTIVE WHITEBOARD SOLUTIONS
INTERACTIVE WHITEBOARD SOLUTIONSHelder Lopes
 
Saiful hidayat strengthening cooperation in ict research & development best...
Saiful hidayat strengthening cooperation in ict research & development   best...Saiful hidayat strengthening cooperation in ict research & development   best...
Saiful hidayat strengthening cooperation in ict research & development best...Saiful Hidayat
 
Augmented Human 2018
Augmented Human 2018Augmented Human 2018
Augmented Human 2018Woontack Woo
 
Technology Insight Report Touch Technology
Technology Insight Report  Touch TechnologyTechnology Insight Report  Touch Technology
Technology Insight Report Touch TechnologyPrashant Nair
 
Pointing, Selecting & Direct Manipulation
Pointing, Selecting & Direct ManipulationPointing, Selecting & Direct Manipulation
Pointing, Selecting & Direct ManipulationJ S
 
Big M Conference - Future Mobile Innovations
Big M Conference - Future Mobile InnovationsBig M Conference - Future Mobile Innovations
Big M Conference - Future Mobile InnovationsPaul Golding
 
RIT (Rakuten Institute of Technology) presentation about UI/UX
RIT (Rakuten Institute of Technology) presentation about UI/UXRIT (Rakuten Institute of Technology) presentation about UI/UX
RIT (Rakuten Institute of Technology) presentation about UI/UXRakuten Group, Inc.
 

What's hot (10)

Design for the iPad
Design for the iPadDesign for the iPad
Design for the iPad
 
Fjord@ The Future of Broadcasting
Fjord@ The Future of BroadcastingFjord@ The Future of Broadcasting
Fjord@ The Future of Broadcasting
 
INTERACTIVE WHITEBOARD SOLUTIONS
INTERACTIVE WHITEBOARD SOLUTIONSINTERACTIVE WHITEBOARD SOLUTIONS
INTERACTIVE WHITEBOARD SOLUTIONS
 
Saiful hidayat strengthening cooperation in ict research & development best...
Saiful hidayat strengthening cooperation in ict research & development   best...Saiful hidayat strengthening cooperation in ict research & development   best...
Saiful hidayat strengthening cooperation in ict research & development best...
 
Augmented Human 2018
Augmented Human 2018Augmented Human 2018
Augmented Human 2018
 
Technology Insight Report Touch Technology
Technology Insight Report  Touch TechnologyTechnology Insight Report  Touch Technology
Technology Insight Report Touch Technology
 
Pointing, Selecting & Direct Manipulation
Pointing, Selecting & Direct ManipulationPointing, Selecting & Direct Manipulation
Pointing, Selecting & Direct Manipulation
 
Big M Conference - Future Mobile Innovations
Big M Conference - Future Mobile InnovationsBig M Conference - Future Mobile Innovations
Big M Conference - Future Mobile Innovations
 
RIT (Rakuten Institute of Technology) presentation about UI/UX
RIT (Rakuten Institute of Technology) presentation about UI/UXRIT (Rakuten Institute of Technology) presentation about UI/UX
RIT (Rakuten Institute of Technology) presentation about UI/UX
 
Portfolio of work
Portfolio of workPortfolio of work
Portfolio of work
 

Viewers also liked

JChueke HCID Open Day_apr2012_pt01
JChueke HCID Open Day_apr2012_pt01JChueke HCID Open Day_apr2012_pt01
JChueke HCID Open Day_apr2012_pt01Jacques Chueke
 
JChueke HCID Open Day_apr2012_pt02
JChueke HCID Open Day_apr2012_pt02JChueke HCID Open Day_apr2012_pt02
JChueke HCID Open Day_apr2012_pt02Jacques Chueke
 
The eye-gaze-communication-system-1.doc(updated)
The eye-gaze-communication-system-1.doc(updated)The eye-gaze-communication-system-1.doc(updated)
The eye-gaze-communication-system-1.doc(updated)NIRAJ KUMAR
 
Hawk Eye Technology - An Understanding
Hawk Eye Technology - An UnderstandingHawk Eye Technology - An Understanding
Hawk Eye Technology - An UnderstandingAbhinay Bandaru
 
Hawk eye technology By RKO
Hawk eye technology By RKOHawk eye technology By RKO
Hawk eye technology By RKORahul Prajapat
 
Hawk eye 1st-presentation
Hawk eye 1st-presentationHawk eye 1st-presentation
Hawk eye 1st-presentationBruno Borras
 
A project on wheelchair motion control using eye gaze and blinks
A project on wheelchair motion control using eye gaze and blinksA project on wheelchair motion control using eye gaze and blinks
A project on wheelchair motion control using eye gaze and blinkspooja mote
 
Hawk eye technology by Mohit Agarwal
Hawk eye technology by Mohit AgarwalHawk eye technology by Mohit Agarwal
Hawk eye technology by Mohit AgarwalMohit Agarwal
 
Hawk Eye Technology ppt
Hawk Eye Technology pptHawk Eye Technology ppt
Hawk Eye Technology pptsravya raju
 
Eye Movement based Human Computer Interaction Technique
Eye Movement based Human Computer Interaction TechniqueEye Movement based Human Computer Interaction Technique
Eye Movement based Human Computer Interaction TechniqueJobin George
 
Hawk eye technology
Hawk eye technologyHawk eye technology
Hawk eye technologyN.CH Karthik
 

Viewers also liked (19)

JChueke HCID Open Day_apr2012_pt01
JChueke HCID Open Day_apr2012_pt01JChueke HCID Open Day_apr2012_pt01
JChueke HCID Open Day_apr2012_pt01
 
JChueke HCID Open Day_apr2012_pt02
JChueke HCID Open Day_apr2012_pt02JChueke HCID Open Day_apr2012_pt02
JChueke HCID Open Day_apr2012_pt02
 
Hawk eye
Hawk eyeHawk eye
Hawk eye
 
Hawk eye ppt
Hawk eye pptHawk eye ppt
Hawk eye ppt
 
The eye-gaze-communication-system-1.doc(updated)
The eye-gaze-communication-system-1.doc(updated)The eye-gaze-communication-system-1.doc(updated)
The eye-gaze-communication-system-1.doc(updated)
 
Hawk Eye Technology - An Understanding
Hawk Eye Technology - An UnderstandingHawk Eye Technology - An Understanding
Hawk Eye Technology - An Understanding
 
Eye tracking
Eye trackingEye tracking
Eye tracking
 
Hawk-Eye
Hawk-EyeHawk-Eye
Hawk-Eye
 
Tobii Eye Tracking
Tobii Eye TrackingTobii Eye Tracking
Tobii Eye Tracking
 
(eye gaze)
(eye gaze)(eye gaze)
(eye gaze)
 
Hawk eye technology
Hawk   eye   technologyHawk   eye   technology
Hawk eye technology
 
Hawk_s_Eye
Hawk_s_EyeHawk_s_Eye
Hawk_s_Eye
 
Hawk eye technology By RKO
Hawk eye technology By RKOHawk eye technology By RKO
Hawk eye technology By RKO
 
Hawk eye 1st-presentation
Hawk eye 1st-presentationHawk eye 1st-presentation
Hawk eye 1st-presentation
 
A project on wheelchair motion control using eye gaze and blinks
A project on wheelchair motion control using eye gaze and blinksA project on wheelchair motion control using eye gaze and blinks
A project on wheelchair motion control using eye gaze and blinks
 
Hawk eye technology by Mohit Agarwal
Hawk eye technology by Mohit AgarwalHawk eye technology by Mohit Agarwal
Hawk eye technology by Mohit Agarwal
 
Hawk Eye Technology ppt
Hawk Eye Technology pptHawk Eye Technology ppt
Hawk Eye Technology ppt
 
Eye Movement based Human Computer Interaction Technique
Eye Movement based Human Computer Interaction TechniqueEye Movement based Human Computer Interaction Technique
Eye Movement based Human Computer Interaction Technique
 
Hawk eye technology
Hawk eye technologyHawk eye technology
Hawk eye technology
 

Similar to Beyond Mouse and Keyboard: Eye Tracking Study of Post-WIMP Interfaces

hcid2011 - Gesture Based Interfaces: Jacques chueke (HCID, City University L...
hcid2011 -  Gesture Based Interfaces: Jacques chueke (HCID, City University L...hcid2011 -  Gesture Based Interfaces: Jacques chueke (HCID, City University L...
hcid2011 - Gesture Based Interfaces: Jacques chueke (HCID, City University L...City University London
 
Future of interface design 2010
Future of interface design 2010Future of interface design 2010
Future of interface design 2010Pavel Růžička
 
Sketch2presentation
Sketch2presentationSketch2presentation
Sketch2presentationjin.fan
 
Touch Screen
Touch ScreenTouch Screen
Touch ScreenIonela
 
Touching More than 3 Dimensions Research Into Novel Interfaces – three dimen...
Touching More than 3 Dimensions Research Into Novel Interfaces – three dimen...Touching More than 3 Dimensions Research Into Novel Interfaces – three dimen...
Touching More than 3 Dimensions Research Into Novel Interfaces – three dimen...Joaquim Jorge
 
COSC 426 Lect. 8: AR Research Directions
COSC 426 Lect. 8: AR Research DirectionsCOSC 426 Lect. 8: AR Research Directions
COSC 426 Lect. 8: AR Research DirectionsMark Billinghurst
 
Users becoming-designers-begona pino
Users becoming-designers-begona pinoUsers becoming-designers-begona pino
Users becoming-designers-begona pinoBegoña Pino
 
Tangible User Interface Showcase
Tangible User Interface ShowcaseTangible User Interface Showcase
Tangible User Interface ShowcaseSimone Mora
 
ITCamp 2012 - Tim Huckaby - Keynote
ITCamp 2012 - Tim Huckaby - KeynoteITCamp 2012 - Tim Huckaby - Keynote
ITCamp 2012 - Tim Huckaby - KeynoteITCamp
 
The Laws of User Experience: Making it or breaking it with the UX Factor
The Laws of User Experience: Making it or breaking it with the UX FactorThe Laws of User Experience: Making it or breaking it with the UX Factor
The Laws of User Experience: Making it or breaking it with the UX FactorEffectiveUI
 
The Laws of User Experience: Making it or Breaking It with the UX Factor
The Laws of User Experience: Making it or Breaking It with the UX FactorThe Laws of User Experience: Making it or Breaking It with the UX Factor
The Laws of User Experience: Making it or Breaking It with the UX FactorEffective
 
Hand Gesture Interface 기술동향 및 전망
Hand Gesture Interface 기술동향 및 전망Hand Gesture Interface 기술동향 및 전망
Hand Gesture Interface 기술동향 및 전망Dongchul Kim
 
Going Mobile With Mobile 2.0 (V0.2)
Going Mobile With Mobile 2.0 (V0.2)Going Mobile With Mobile 2.0 (V0.2)
Going Mobile With Mobile 2.0 (V0.2)Paul Golding
 
BFA Digital Design Thesis Proposal Presentation DRAFT
BFA Digital Design Thesis Proposal Presentation DRAFTBFA Digital Design Thesis Proposal Presentation DRAFT
BFA Digital Design Thesis Proposal Presentation DRAFTSkye Sant
 
Design and Evaluation Case Study: Evaluating The Kinect Device In The Task of...
Design and Evaluation Case Study: Evaluating The Kinect Device In The Task of...Design and Evaluation Case Study: Evaluating The Kinect Device In The Task of...
Design and Evaluation Case Study: Evaluating The Kinect Device In The Task of...Waqas Tariq
 
Design and Evaluation Case Study: Evaluating The Kinect Device In The Task of...
Design and Evaluation Case Study: Evaluating The Kinect Device In The Task of...Design and Evaluation Case Study: Evaluating The Kinect Device In The Task of...
Design and Evaluation Case Study: Evaluating The Kinect Device In The Task of...Waqas Tariq
 
Design and Evaluation Case Study: Evaluating The Kinect Device In The Task of...
Design and Evaluation Case Study: Evaluating The Kinect Device In The Task of...Design and Evaluation Case Study: Evaluating The Kinect Device In The Task of...
Design and Evaluation Case Study: Evaluating The Kinect Device In The Task of...Waqas Tariq
 

Similar to Beyond Mouse and Keyboard: Eye Tracking Study of Post-WIMP Interfaces (20)

hcid2011 - Gesture Based Interfaces: Jacques chueke (HCID, City University L...
hcid2011 -  Gesture Based Interfaces: Jacques chueke (HCID, City University L...hcid2011 -  Gesture Based Interfaces: Jacques chueke (HCID, City University L...
hcid2011 - Gesture Based Interfaces: Jacques chueke (HCID, City University L...
 
Future of interface design 2010
Future of interface design 2010Future of interface design 2010
Future of interface design 2010
 
Sketch2presentation
Sketch2presentationSketch2presentation
Sketch2presentation
 
Touch Screen
Touch ScreenTouch Screen
Touch Screen
 
Touching More than 3 Dimensions Research Into Novel Interfaces – three dimen...
Touching More than 3 Dimensions Research Into Novel Interfaces – three dimen...Touching More than 3 Dimensions Research Into Novel Interfaces – three dimen...
Touching More than 3 Dimensions Research Into Novel Interfaces – three dimen...
 
COSC 426 Lect. 8: AR Research Directions
COSC 426 Lect. 8: AR Research DirectionsCOSC 426 Lect. 8: AR Research Directions
COSC 426 Lect. 8: AR Research Directions
 
NUI_jaydev
NUI_jaydevNUI_jaydev
NUI_jaydev
 
The Second Web
The Second WebThe Second Web
The Second Web
 
Users becoming-designers-begona pino
Users becoming-designers-begona pinoUsers becoming-designers-begona pino
Users becoming-designers-begona pino
 
Tangible User Interface Showcase
Tangible User Interface ShowcaseTangible User Interface Showcase
Tangible User Interface Showcase
 
14 585
14 58514 585
14 585
 
ITCamp 2012 - Tim Huckaby - Keynote
ITCamp 2012 - Tim Huckaby - KeynoteITCamp 2012 - Tim Huckaby - Keynote
ITCamp 2012 - Tim Huckaby - Keynote
 
The Laws of User Experience: Making it or breaking it with the UX Factor
The Laws of User Experience: Making it or breaking it with the UX FactorThe Laws of User Experience: Making it or breaking it with the UX Factor
The Laws of User Experience: Making it or breaking it with the UX Factor
 
The Laws of User Experience: Making it or Breaking It with the UX Factor
The Laws of User Experience: Making it or Breaking It with the UX FactorThe Laws of User Experience: Making it or Breaking It with the UX Factor
The Laws of User Experience: Making it or Breaking It with the UX Factor
 
Hand Gesture Interface 기술동향 및 전망
Hand Gesture Interface 기술동향 및 전망Hand Gesture Interface 기술동향 및 전망
Hand Gesture Interface 기술동향 및 전망
 
Going Mobile With Mobile 2.0 (V0.2)
Going Mobile With Mobile 2.0 (V0.2)Going Mobile With Mobile 2.0 (V0.2)
Going Mobile With Mobile 2.0 (V0.2)
 
BFA Digital Design Thesis Proposal Presentation DRAFT
BFA Digital Design Thesis Proposal Presentation DRAFTBFA Digital Design Thesis Proposal Presentation DRAFT
BFA Digital Design Thesis Proposal Presentation DRAFT
 
Design and Evaluation Case Study: Evaluating The Kinect Device In The Task of...
Design and Evaluation Case Study: Evaluating The Kinect Device In The Task of...Design and Evaluation Case Study: Evaluating The Kinect Device In The Task of...
Design and Evaluation Case Study: Evaluating The Kinect Device In The Task of...
 
Design and Evaluation Case Study: Evaluating The Kinect Device In The Task of...
Design and Evaluation Case Study: Evaluating The Kinect Device In The Task of...Design and Evaluation Case Study: Evaluating The Kinect Device In The Task of...
Design and Evaluation Case Study: Evaluating The Kinect Device In The Task of...
 
Design and Evaluation Case Study: Evaluating The Kinect Device In The Task of...
Design and Evaluation Case Study: Evaluating The Kinect Device In The Task of...Design and Evaluation Case Study: Evaluating The Kinect Device In The Task of...
Design and Evaluation Case Study: Evaluating The Kinect Device In The Task of...
 

Recently uploaded

Developer Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLDeveloper Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLScyllaDB
 
Vertex AI Gemini Prompt Engineering Tips
Vertex AI Gemini Prompt Engineering TipsVertex AI Gemini Prompt Engineering Tips
Vertex AI Gemini Prompt Engineering TipsMiki Katsuragi
 
Artificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptxArtificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptxhariprasad279825
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubKalema Edgar
 
My INSURER PTE LTD - Insurtech Innovation Award 2024
My INSURER PTE LTD - Insurtech Innovation Award 2024My INSURER PTE LTD - Insurtech Innovation Award 2024
My INSURER PTE LTD - Insurtech Innovation Award 2024The Digital Insurer
 
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024BookNet Canada
 
Training state-of-the-art general text embedding
Training state-of-the-art general text embeddingTraining state-of-the-art general text embedding
Training state-of-the-art general text embeddingZilliz
 
Search Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdfSearch Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdfRankYa
 
DevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache MavenDevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache MavenHervé Boutemy
 
Streamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupStreamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupFlorian Wilhelm
 
Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Enterprise Knowledge
 
Story boards and shot lists for my a level piece
Story boards and shot lists for my a level pieceStory boards and shot lists for my a level piece
Story boards and shot lists for my a level piececharlottematthew16
 
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024BookNet Canada
 
Dev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebDev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebUiPathCommunity
 
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)Mark Simos
 
Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 3652toLead Limited
 
Gen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfGen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfAddepto
 
Unraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfUnraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfAlex Barbosa Coqueiro
 
Scanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsScanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsRizwan Syed
 
Commit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyCommit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyAlfredo García Lavilla
 

Recently uploaded (20)

Developer Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLDeveloper Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQL
 
Vertex AI Gemini Prompt Engineering Tips
Vertex AI Gemini Prompt Engineering TipsVertex AI Gemini Prompt Engineering Tips
Vertex AI Gemini Prompt Engineering Tips
 
Artificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptxArtificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptx
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding Club
 
My INSURER PTE LTD - Insurtech Innovation Award 2024
My INSURER PTE LTD - Insurtech Innovation Award 2024My INSURER PTE LTD - Insurtech Innovation Award 2024
My INSURER PTE LTD - Insurtech Innovation Award 2024
 
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
 
Training state-of-the-art general text embedding
Training state-of-the-art general text embeddingTraining state-of-the-art general text embedding
Training state-of-the-art general text embedding
 
Search Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdfSearch Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdf
 
DevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache MavenDevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache Maven
 
Streamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupStreamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project Setup
 
Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024
 
Story boards and shot lists for my a level piece
Story boards and shot lists for my a level pieceStory boards and shot lists for my a level piece
Story boards and shot lists for my a level piece
 
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
 
Dev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebDev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio Web
 
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
 
Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365
 
Gen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfGen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdf
 
Unraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfUnraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdf
 
Scanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsScanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL Certs
 
Commit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyCommit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easy
 

Beyond Mouse and Keyboard: Eye Tracking Study of Post-WIMP Interfaces

  • 1. Beyond Mouse and Keyboard: Post-WIMP and Novel Forms of Interaction Jacques Chueke London, UK, May 2011 George Buchanan (1st Supervisor) Lecturer, Centre for HCI Design Stephanie Wilson (2nd Supervisor) Lecturer, Centre for HCI Design Master in Design, PUC-Rio, RJ, Brazil PhD Researcher at the Centre for HCI Design School of Informatics, City University London 1
  • 2. Things are changing… iPad: 1 year old growing among touch screens and print 2
  • 3. Things are changing… iPad for cats??? 3
  • 4. Overview • The introduction of novel hardware for computing and gaming during the last decade is changing the way we control everyday devices. • It provides NUI control methods, such as haptic (e.g. iPhone, iPad, MS Surface), gesture- based and voice (e.g. Nintendo Wii, Microsoft Xbox 360 console gaming with Kinect sensor) and eye tracking MIT Media Lab: DepthJS – 2011 interactions (e. g. Tobii P-10). • One specific impact this has had is on the user’s control of such devices. 4
  • 5. It’s not supposed to be like this… Gmail Motion, April 2011 5
  • 6. It’s not supposed to be like this… Gmail Motion, April 2011 6
  • 7. New Modes of Interaction • PrimeSense / MS Kinect: Swim Browser Prime sense browser competition winner: Stolarsky 'SwimBrowser’ – 2011 7
  • 8. New Modes of Interaction • KinVi: Kinect Virtual Interface Prime sense browser competition 2n place: Windows Control with gestures - 2011 8
  • 9. New Modes of Interaction • eviGroup Paddle Pro Front-facing webcam to track head movements for cursor control – 2011 9
  • 10. New Modes of Interaction • Hitachi: Hitachi's Gesture Remote Control TV Prototype CES 2009: Hitachi's Gesture Remote Control TV Prototype 10
  • 11. Problem Statement • New command vocabularies NEW CONTROL METHODS (INPUT) have emerged and users do not know how to access or NEW INTERFACES (OUTPUT) activate them. • This is a timely moment to research how people make OLD OUTPUT X NEW INPUT sense of these technologies NEW OUTPUT X OLD INPUT for CONTROL whilst some Post-WIMP interfaces do not NEW OUTPUT X NEW INPUT display appropriate visual cues for NUI interactions. WIMP-GUI (DESKTOP) X NUI (PHYSICAL) POST-WIMP X MOUSE/KEYBOARD POST-WIMP X NUI (PHYSICAL) 11
  • 12. Case 1: Microsoft Surface Media Player, Microsoft Surface. Nov, 2010 POST-WIMP + NUI (PHYSICAL) 12
  • 13. Case 2: Windows 8 Metro Dashboard • The new Windows 8 with similar features as used in Windows Phone and Xbox 360 Dashboard. POST-WIMP + NUI (PHYSICAL) x Metro Dashboard: Windows 8 Start Screen. Feb, 2012 MOUSE/KEYBOARD 13
  • 14. Example: GNOME 3 - desktop environment for GNU/Linux and UNIX-type operating systems. GNOME 3 hotcorners and responsive interface, 2012 POST-WIMP + NUI (PHYSICAL) x MOUSE/KEYBOARD 14
  • 15. Case 3: Tobii P-10 Eye Tracker: Gaze for Control 15
  • 16. Case 3: Tobii P-10 Eye Tracker: Gaze for Control Tobii P-10 at the SmartLab (UEL), Oct. 2010 16
  • 17. Case 3: Tobii P-10 Eye Tracker: Gaze for Control • Assistive Technology: Tobii P-10 at the SmartLab (UEL). Tobii P-10 equipment, Oct. 2010 WIMP-GUI (DESKTOP) X NUI (PHYSICAL) 17
  • 18. Case 3: Tobii P-10 Eye Tracker: Gaze for Control Mouse configuration pop up for Windows Control, Tobii P-10 18
  • 19. Case 4: Tobii LeNovo: Gaze for Control • Tobii Lenovo / PCEye / Acuity Tobii LeNovo, Jun 2011 WIMP-GUI (DESKTOP) X NUI (PHYSICAL) 19
  • 20. New Visual Cues/Feedback UI affordances are Just-in-time chrome shown on tap. Applying Can be triggered by the principle of touch or proximity – scaffolding will lead you hover effect (pg. 153) to far more successful multi-touch and gesture UI’s. (pg. 154) Tethers indicate that a size The marking constraint (MS menu system Surface) has teaches users to been reached on make pen-based an item being gestures (pg. 150) scaled (pg. 91) Wigdor, D. Wixton, D. Brave NUI World, 2011 20
  • 21. Methodology An empirical study with Eye Tracking (Tobii x60) was conducted to test the Perceptible Affordances of Drag and Drop interactions within a iGoogle Personal Web Portal. This study served to create a protocol for analysis which focuses on the very first 10 seconds of a participant scrutinizing the screen while trying to respond specific questions. Participant, 54, Beginner Expertise Participant, 22, Advanced Expertise Gazeplot Comparison: Beginner x Advanced 21
  • 22. Quantitative Data Analysis: Gazeplot Resulting gaze plot from seven participants during the first 10 seconds after question 01. 22
  • 23. Quantitative Data Analysis: Heatmap Resulting heat map from seven participants during the first 10 seconds after question 01. 23
  • 24. Quantitative Data Analysis: Gazeplot Resulting heat map from seven participants during the first 10 seconds after question 02. 24
  • 25. Quantitative Data Analysis: Heatmap Resulting heat map from seven participants during the first 10 seconds after question 02. 25
  • 26. Quantitative Data Analysis: Gazeplot Resulting heat map from seven participants during the first 10 seconds after question 03. 26
  • 27. Quantitative Data Analysis: Heatmap Resulting heat map from seven participants during the first 10 seconds after question 03. 27
  • 28. Quantitative Data Analysis: Gazeplot Resulting heat map from seven participants during the first 10 seconds after question 04. 28
  • 29. Quantitative Data Analysis: Heatmap Resulting heat map from seven participants during the first 10 seconds after question 04. 29
  • 30. Model 1: Perceptible Affordances • According to Nielsen (2008): "Affordance" means what you can do to an object. For example, a checkbox affords turning on and off, and a slider affords moving up or down. • "Perceived Affordances" are actions you understand just by looking at the object, before you start using it (or feeling it, if it's a physical device rather than an on- screen UI element). • In Gaver’s (1991) words, “…Perceptible Affordances are inter-referential: the attributes of the object relevant for action are available for perception. What is perceived is what is acted upon.” 30
  • 31. Perceptible Affordances in Post-WIMP Post-WIMP GUI [INTERFACE LAYER] OUTPUT - LESS SYMBOLIC - MORE INTUITIVE (USER) PERCEPTIBLE AFFORDANCE - MORE REACTIVE (COMPUTER) INPUT NUI [MODE OF INTERACTION LAYER] 31
  • 32. Model 2: Norman’s Theory of Action Execution Cycle Execution Specification of Formulation of Actions Intention Sequence Interaction Perception Interpretation Evaluation Evaluation Cycle Preece et al (2009: 121) quoting Norman, (1986) 32
  • 33. Research Question 1: ACTIVATION Post-WIMP GUI [INTERFACE LAYER] Evaluation Cycle OUTPUT PERCEPTIBLE AFFORDANCE Execution Cycle INPUT NUI [MODE OF INTERACTION LAYER] 33
  • 34. Conclusions and Future Work • By developing a methodology for an empirical study, which focuses on observation prior to any interaction, we are willing to identify what elements people will focus on NUI screens. • I believe both Norman’s theory of Action and Perceptible Affordances theories can be combined with the more recent Piaget’s theory of INRC and the Scaffolding concept; and used on my protocol for analysis. • A prototype with Post-WIMP characteristics and NUI mode of interaction will be built in order to to understand how users visually scan such interfaces to obtain the gist of its interactive potential. • Quantitative (Eye Tracking) and Qualitative (Verbalizations) data will be combined to produce conclusions about what kind of information can be obtained with the protocol – and how can this data be adapted to indicate better design interactions with NUI systems. 34
  • 35. Thank you for your attention! Jacques Chueke Jacques.chueke.1@city.ac.uk 35
  • 36. Bibliography Beaudouin-Lafon, M. (November 2000). "Instrumental Interaction: An Interaction Model for Designing Post-WIMP User Interfaces". CHI '00: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. The Hague, The Netherlands: ACM Press. pp. 446–453. doi:10.1145/332040.332473. ISBN 1-58113-216-6. http://www.daimi.au.dk/CPnets/CPN2000/download/chi2000.pdf. Breeze, James. Eye Tracking: Best Way to Test Rich App Usability. UX Magazine, access on 25 November 2010. (http://www.uxmag.com/technology/eye-tracking-the-best-way-to-test-rich-app-usability) Buxton, W. (2001). Less is More (More or Less), in P. Denning (Ed.), The Invisible Future: The seamless integration of technology in everyday life. New York: McGraw Hill, 145–179 ITU Internet Reports 2005: The Internet of Things – Executive Summary. Dam, A. (February 1997). "POST-WIMP User Interfaces". Communications of the ACM (ACM Press) 40 (2): pp. 63–67. doi:10.1145/253671.253708. Dourish, P. Where the Action Is: The Foundations of Embodied Interaction. A Bradford Book: The MIT Press, USA, 2004. Ehmke & Wilson, 2007. Identifying Web Usability Problems from Eye-Tracking Data. Published by the British Computer Society. People and Computers XXI – HCI…but not the way we know it: Proceedings of HCI 2007. Gaver, W. Technology Affordances. Copyright 1991 ACM 0-89791-383-3/91/0004/0079. Gentner, D. and Nielsen, J. (April 1993). "The Anti-Mac Interface". Communications of the ACM (ACM Press) 39 (8): pp. 70–82. http://www.useit.com/papers/anti-mac.html. Jacob, R. et al. (2008). "Reality-Based Interaction: A Framework for Post-WIMP Interfaces". CHI '08: Proceedings of the Twenty-Sixth Annual SIGCHI Conference on Human Factors in Computing Systems. Florence, Italy: ACM. pp. 201–210. doi:http://doi.acm.org.ezproxy.lib.ucf.edu/10.1145/1357054.1357089. ISBN 978-1-60558-011-1. 36
  • 37. Bibliography McGrenere, J., Ho, W. (2000). Affordances: Clarifying and Evolving a Concept. Procs. of Graphic Interfaces 2000, Montreal, May 2000. McNaughton, J. Utilizing Emerging Multi-touch Table Designs. Technology Enhanced Learning Research Group - Durham University. TR-TEL-10-01. Nielsen, J. (April 1993). "Noncommand User Interfaces". Communications of the ACM (ACM Press) 36 (4): pp. 83–99. doi:10.1145/255950.153582. http://www.useit.com/papers/noncommand.html. Norman, D. (1999). Affordance, Conventions and Design. In ACM Interactions, (May + June, 1999), 38-42. Picard, R. Affective Computing. The MIT Press, Cambridge, Massachusetts. London, England, 1998. PREECE, Jenny. SHARP, Helen. ROGERS, Yvonne. Interaction Design: Beyond Human-Computer Interaction [2nd edition]. John Wiley & Sons, Ltd. West Sussex, UK, 2009. Ramduny-Ellis, D.; Dix, A.; Hare, J.; Gill, S. Physicality: Towards a Less-GUI Interface (Preface). Procs. Third International Workshop on Physicality. Cambridge, England, 2009. Sorensen, M. Making a Case for Biological and Tangible Interfaces. Proceedings of the Third International Workshop on Physicality. Cambridge, England, 2009. Sternberg, R. Cognitive Psychology. Wadsworth, Cengage Learning. Belmont, CA, USA, 2009, 2006. Vyas, D., Chisalita, C. Veer, G. Affordance in Interaction. ECCE '06 Proceedings of the 13th Eurpoean conference on Cognitive ergonomics: trust and control in complex socio-technical systems. ACM New York, NY, USA ©2006 ISBN: 978-3-906509-23-5 WIGDOR, Deniel. WIXON, Dennis. Brave NUI World: designing natural user interfaces for touch and gesture. Morgan Kauffman Publishers, USA, 2011. 37

Editor's Notes

  1. Welcome to my session, entitled ‘Beyond Mouse and Keyboard: Post-WIMP and Novel Forms of Interaction’. My name is Jacques. I’m a PhD researcher at the Centre for HCI Design. I’m a teacher in Brazil, teaching Usability for web and software on Postgraduation degrees and graphic design at PUC-Rio university. I started my research on October 2010.I have the pleasure to be accompanied by My first supervisor, Dr. George Buchanan, who will talk at the end about the research being developed at the centre and the master in human centred-systems. This presentation is about the core subject of my PhD research at the Centre for HCI Design, City Uni.I’d like to thank Mr. Hillmore and the BCS board for this opportunity. It’s my second time here, first I was on June 2010, presenting my work in a Doc Consortium – which was a great experience and where great sharing of information took place. I’m sure the same will take place tonight. Shall we start?I’ll walk you through new developments within technologies for interaction.Wiil present 4 case studies where I could use the technology – and make a few comments about some interface issues I’ve spotted that could be improved.We’ll discuss a protocol for analysis I’m developing with my supervisors which evolves participants with different expertise utilizing NUI technologies and eye tracking technology.I’ll explain the theories I’m using and updating to better understand how people learn and adapt (or not) to this new technologies.In case you’re wondering what is Post-WIMP (stands for Window Icon Menu and Pointing Device): Defined by van Dam as interfaces “containing at least one interaction technique not dependent on classical 2D widgets such as menus and icons”. Ultimately it will involve all senses in parallel, natural language communication and multiple users. Communications of ACM, 1997
  2. From early ages we’re in contact with ground breaking technology. PARADIGMS of user interaction are being dissolved at daily basis.
  3. People are developing games for kittens. Indeedthings are changing.
  4. After almost thirty years of the desktop metaphor as the dominant visual interface with mouse and keyboard as input methods, traditional paradigms of user interaction are changing rapidly (Wigdor, 2011: 1-5). The introduction of novel hardware for computing and gaming during the last decade is changing the way we control everyday devices (Dam, 1997) because it provides, for instance, haptic (e.g. iPhone, iPad, MS Surface), gesture-based and voice (e.g. Nintendo Wii, Microsoft Xbox 360 console gaming with Kinect sensor) and eye tracking interactions (e. g. Tobii P-10). Noticeably, traditional control modes of interaction such as buttons, links, icons and tools, generally activated by a pointer, are not present and no longer hold this kind of interaction. Therefore, the challenge is to step outside the GUI paradigm and enable the user to use control technologies in a non-GUI/WIMP interaction, which is mostly physical. As already mentioned, these technologies, although bringing Post-WIMP interfaces (Beaudouin-Lafon, 2000), might not be displaying appropriate visual cues for physical interaction. As you can see on the video at the right hand-side, people from MIT Media Lab is changing the Kinect SDK in order to research gestural interactions. In case u never heard of the Kinect, is a camera that comes with MS X-Box 360 which has the initial purpose to allow gaming with controllers – it identifies one’s structure in a skeleton view.By natural I DO NOT mean what u gonna see:
  5. I gonna let you decide what’s wrong with this picture.This is a jest from Google on April's fool, last year.
  6. They were moking this frenzy about NUI and gestural interfaces that was emerging last year. Imagine the need to learn such vocabulary/language. Does this look natural???I leave this as a warning: careful with the new vocabularies you’re introducing and complex gestures one might need to learn to interact...
  7. The community is actively changing CONTROL methods. There’s a genuine interest on this.Again: plan properly the gestures u imagine as being “natural” – imagine this in a public space: “excuse me I have to check my email”.
  8. Gestural interaction within regular desktop Windows. It brings one of the issues I’m talking about - subject of this presentation. Very clever, indeed. Still a conventional desktop with natural interaction.I see a problem on this. I believe the interface should change, to better convey the message of interacting with gestures, voice, touch, eye-gaze.This is what I’m researching: how to inform properly about possible physical interactions available from the system?How to design better visual, audible, tactile cues to inform about NUI interactions?At the video you can see the guy standing up but now we have a set of lenses(available at Amazon) that allow users to sit in front of the kinect, interacting at close range.
  9. Moving fromKinect hacks and people fiddling with the SDK let’s move towards the industry/mainstream companies. Head tracking to control the mouse pointer with eviGroup Paddle Pro.Again regular Windows OS on the background.
  10. Again a new learning of gestures vocabularies to shift channels, control volume…
  11. Right hand-side I tried to decompose the problem into variables of INPUT and OUTPUT. This regards the current configuration of some of the technologies presented and some that I could experience in different workshops and technology that we have at the Interaction Lab. If you cross over OLD/NEW OUTPUTS X OLD/NEW INPUTS. INPUT: understand input as a user’s command.OUTPUT: understand output as what the system displays.Left aside OLD INPUTxOLD OUTPUT situation. Too much research on this already aiming to improve the GUI-WIMP desktop WYSIWYG.In some cases, NUI modes of interaction just co-exist as additional features on a traditional GUI, presenting hybrid solutions, which could hamper even more the user interaction, creating control problems with the system. Different technologies from specific manufacturers will be regarded in this presentation, in order to exemplify the subject of novel interactions and input methods. By all means the companies here quoted are to receive great admiration for their efforts on R&D of technologies for HCI. Comments made about specific shortcomings or design flaws (as I see it) regarding their modes of interaction do not aim to question the quality of their work. My intention is to point out specific issues that were observed and represent the core issues that this research tackles with. I’ve recorded 4 case studies which exemplify these configurations.
  12. Microsoft Surface from 2007 (10-12 K) that we have at the Interaction Lab. This is a playback from my first encounter with the native Media Player, where I was trying to create my own play list with the albums available.No cues, no warnings, no error messages, no tooltips. One thing that would be helpful are tooltips for the newcomer, the inexperienced user. When one finally becomes experienced after learning new languages, new icons andWays to interact with the system could turn off any aids such as tooltips, tutorials, etc. But the very first interaction is very important. Could make the difference for one to choose the technology or not. One without experienceMight never go back to this. Might never become a customer or a user.
  13. CONSUMER PREVIEW VERSION – TUI and mouse/keyboard. HIDDEN MENUS/INTERACTIONS.APP SWITCHER / CHARMS / CONTEXTUAL MENUStart screen – could press spacebar to unveil the login screen (does mouse click works?)Metro Dashboard – tilesThe problem of INVISIBLE MENUS. Enter app – how to exit? No home button? No exit icon. Have to make a swipe gesture from the very bezel, the canvas to unveil the CHARMS menu. We praise the effort. Jon mentioned he learned because he watched tutorials. I couldn’t do almost anything.Multi-task. Again same swipe gesture from the bezel – no VISUAL CUES before any interaction takes place. No hot corners – no response while interacting to inform what is there and how to do it. Very specific moves to rearrange apps/create different views.
  14. Hot corners, similar to Mac EXPOSÉ but with visual feedback. Right corners that are displayed when you activate a window – informing there’s a connection.
  15. I’mveryinterestedonAssistive Technology. Perhaps R&D in thisareamightbenefitnotonlypeoplewithspecialneedsbuttheentirecommunity. SimilarTowhathappenedwith HTML/XML sourcecode for web – accessibilityissues.Accessibility: the technology conveys opportunities for people with special needs to interface with digital devices (multi-sensory).MENTION  GReAT (Gesture Recognition in Aphasia Therapy) SAM MUCROSOFT WORK
  16. Did a workshop at SmartLab, UEL with Mick Donegan, specialist on AT with eye gaze for control. Tobii studio with special software with large buttons, icons, shining colors. But when it comes to Windows control it’s a different ball game.
  17. Not convinced the visual cues available for control interaction are the best here.
  18. TASK: OpenSoftw > Print Screen > Open Sotw > Paste > Save File > Select Folder > Tried to change file format (FAIL) > Close softwChange system setings > look downChange mouse settings > look left
  19. Meeting with Scott Hodgins (Director, Acuity ETS Ltd) and Sara Hyléen (TobiiCorporate Marketing Manager)1. Change active windows like Alt + Tab: Spacebar + Gaze to select active Windows. Then you release the spacebar to activate.2. Pointer jumps to any spot you’re looking at with a small movement on the track pad. You never loose your pointer from sight (could that be annoying?). Features were tested with users. Sara got used to the pointer jump and misses it a lot now.3. Zoom/pan large 3d images with mouse/track pad scroll – really handy for zooming in and out over wherever you’re looking at. In another prototype you can use head movements for zooming, spinning, etc. 4. Tobii Media Studio: selecting thumbnails with eye gaze. No significant visual cues for swiping images (left/right gaze) and looking at the bottom to bring back thumb menu.5. Presentation Browser: PPT. Spacebar + eye gaze to select/activate slides. 6. Text browser: vertical scrolling and gaze feedback. How am I planning to investigate this? Visual Perception task – Eye Tracking technology present at Interaction Lab, Centre for HCID, City Uni.
  20. I reiterate: In my research I am especially interested on the moment when users scan the screen of NUI systems for the very first time. I’ll investigate what happens cognitively, when a user comes across visual cues they’re not familiar with, in systems they’ve never used before. The visual cues that indicate the range of controls available through physical modes of interaction. How am I planning to investigate this? Visual Perception task – Eye Tracking technology present at Interaction Lab, Centre for HCID, City Uni. Perfect to investigate where people look at.Small sample, 7 people. D and D was spotted as a hidden/invisible interaction. – we want to adapt the protocol to Post-WIMP with NUI and test some of these technologies. In time a prototype will be created novel visual metaphor, a reactive interface which should be tested in order to verify if its PA efficiently convey available interactions. No verbalizations during the 10 seconds – could generate false data – people might have detained their gaze over a spot or a feature they were trying to explain.Quantitative data was obtained with the Tobii x60 eye-tracker (e.g. saccades plots and fixations times) and was compared with verbalizations (qualitative) in order to produce conclusions about how hidden interactions affected participants with different expertise. Explain what is a FIXATION (grows the more one attempts to a specific spot) and a SACCADE (path that connects fixations). The inexperienced participant COULD NOT SEE THE DIFFERENCE between a PWP and a regular Portal and could NOT SPOT D AND DROP interactions. Verbalizations confirmed.
  21. Mention they’re instructed to observe only after each question and then with RTA technique they explain their interpretation of the screen based on the question made.‘Q1: What is this website for?’Recurrence between FIXATIONS could indicate hardship to understand some user interface object. There’s a very interesting paper from my 2nd supervisor trying to find patterns on fixations and saccades with specific usability problems.I’ll not discuss the results in depth – no time for this – but experience is a key feature on this kind of interaction (NO VISUAL CUES, no proper PA to inform the newcomer about CONFIGURATION and D AND D features). People need to compareWith what they know to SPOT this kind of interaction. HAD INEXPERIENCED PARTICIPANTS BEEN INFORMED properly – with efficient VISUAL CUES that they could CHANGE the initial CONFIG and RE-ORDER object (widgets in a dashboard) the Experience would have been different? Would eyetracking data show this? I’m most certain it would. That’s why I’ll test this with my prototype later.
  22. ‘Q1: What is this website for?’
  23. ‘Q2: What can you do in this kind of website?’
  24. ‘Q2: What can you do in this kind of website?’
  25. ‘Q3: Do you think is possible to change your screen the way you like it?’I was conducting them towards my main question, actually. Even cueing them about this possibility. Inexperienced didn’t spot or suspected besides my warning.
  26. ‘Q3: Do you think is possible to change your screen the way you like it?’
  27. ‘Q4: Is it possible to move anything in there?’More fixations over the top bar, top part of widgets took place.Question were of great influence over people’s gaze. As Jacques Aumont: the introduction of an order affects how a person scrutinizes an image and disrupts expected trends. We’ve seen landing s over pictures which were accidental, is hard to avoid looking at big pictures.
  28. ‘Q4: Is it possible to move anything in there?’
  29. PERCEIVED AFFORDANCE x (2) CULTURAL CONSTRAINT = CONVENTION + SYMBOLIC COMMUNICATION (SYMBOLIC MEANING ARBRITARY – LEARNED CONVENTION)EXAMPLE OF PERCEPTIBLE AFFORDANCE: SLIDER/BUTTONPerceptible Affordances theory is used in HCI to better understand how to make a system usable and how to shape the functions that users anticipate a system may have. It teaches us that a tools within systems should be identifiable; its use should be obvious as well its intended effect. What are the control actions and what supposedly are the results? The theory of Perceptible Affordances resonates and complements the very different stages of Norman’s Theory of Action. In particular it relates to the evaluation cycle, where the user is still assessing and trying to make sense of a system. This moment plays a pivotal role on the following interaction between user and system. Fewer mistakes are made if the evaluation cycle is well supported by the interface design. When execution takes place, less activation without awareness of the forthcoming results is prone to happen: users will not be mislead so often. The concept of affordance has been used in HCI to solve problems related to the usability of designed systems. The concept was originally coined by Gibson (1986) and introduced to the HCI field by Norman (1988) and was further appropriated by Gaver (1991), Bærentsen & Trettvik (2002), amongst others. Vyas, D. (2006)The concept of an affordance was coined by the perceptual psychologist James J. Gibson(1979) in his seminal book The Ecological Approach to Visual Perception. The concept was introduced to the HCI community by Donald Norman in his book The Psychology of Everyday Things from 1988.Donald Arthur Norman (born December 25, 1935), is an academic in the field of cognitive science, design and usability engineering and a co-founder and consultant with the Nielsen Norman Group. He is the author of the book The Design of Everyday Things.Much of Norman's work involves the advocacy of user-centered design. His books all have the underlying purpose of furthering the field of design, from doors to computers.‘In today’s screen design sometimes the cursor shape changes to indicate the desired action (e.g., the change from arrow to handshape in a browser), but this is a convention, not an affordance. After all, the user can still click anywhere, whatever the shape of the cursor. Now if we locked the mouse button when the wrong cursor appeared, that would be a real affordance, although somewhat ponderous. The cursor shape is visual information: it is a learned convention. When you learn not to click unless you have the proper cursorform, you are following a cultural constraint. Norman (1999) "Affordance" means what you can do to an object. For example, a checkbox affords turning on and off, and a slider affords moving up or down. "Perceived affordances" are actions you understand just by looking at the object, before you start using it (or feeling it, if it's a physical device rather than an on-screen UI element). All of this is discussed in Don Norman's book The Design of Everyday Things. (a.k.a POET: Psychology of Everyday Things). Jakob Nielsen's Alertbox, February 19, 2008: Top-10 Application-Design Mistakeshttp://www.useit.com/alertbox/application-mistakes.htmlWe view the affordances of an artefact as the possibilities (for both: thinking and doing) that are signified by the users during their interaction with the artefact. Acknowledging the work of Baerentsen & Trettvik, we propose an interaction-centered view of affordance, which we call Affordance in Interaction. From this view, affordances of an artefact are not the properties of the artefact but a relationship that is socially and culturally constructed between the users and the artefact in the lived world. This view strongly suggests that affordance emerges during a user’s interaction with the environment. In addition, the affordance in interaction view focuses on the ‘active interpretations’ of the users interacting with the artefact. From this view, users are actively participating in the interaction with the artefact and continuously interpreting the situation and constructing and re-building meanings about the artefact. We suggest that affordances can be better understood as an interpretative relationship between users and the artefact. Vyas et al (2006) SEMIOTICS SIGN: Charles Morris (Sintatic – Semantic – Pragmatic) Peirce (Representamen – Object – Interpretant x Icon – Index – Symbol)
  30. I made a distinction for the purpose of better clarification on how Perceptible Aff operate in Post-WIMP. As a Bridge between the interface layer (visual, acoustic, haptic) + mode of interaction.The trick is to show what is really possible to be done rather than what is apparently possible.‘In today’s screen design sometimes the cursor shape changes to indicate the desired action (e.g., the change from arrow to handshape in a browser), but this is a convention, not an affordance. After all, the user can still click anywhere, whatever the shape of the cursor. Now if we locked the mouse button when the wrong cursor appeared, that would be a real affordance, although somewhat ponderous. The cursor shape is visual information: it is a learned convention. When you learn not to click unless you have the proper cursorform, you are following a cultural constraint. Norman (1999)PERCEIVED AFFORDANCE x (2) CULTURAL CONSTRAINT = CONVENTION + SYMBOLIC COMMUNICATION (SYMBOLIC MEANING ARBRITARY – LEARNED CONVENTION)EXAMPLE OF PERCEPTIBLE AFFORDANCE: SLIDER/BUTTON
  31. I consider the Evaluation Cycle paramount during user-interaction
  32. We propose a view that identifies some fraction of a user interface as based on the Post-WIMP theme (1) plus some other fraction that provides computer-only functionality (2) that is not realistic. As a design approach or metric, the goal would be to make the first category as large as possible and use the second only as necessary, highlighting the tradeoff explicitly. Jacob at al, 2008With Eye tracking technology I will be able to analyze how participants scrutinize the screen (yielding Quantitative data). I will be able to cross reference quantitative data with participant’s utterances – which were organized in classes, with general inductive approach for qualitative data analysis. A prototype with Post-WIMP characteristics and novel technologies for interaction will be built in order to elicit user exploration of new and visually unfamiliar digital interfaces to understand how users visually scan such interfaces to obtain the gist of its interactive potential. HCI theories and Cognitive Psychology will be used to better understand those issues. I believe both theories of Norman’s theory of Action and Perceptible Affordances can be adapted and updated to the research question and the identified problem. They could also be combined with more recent theories such as Piaget’s theory of INRC (Wigdor, 2011: 137-138): Identity, Negation, Reciprocal and Commutative and the Scaffolding concept, based on the seminal thinking of the famous psychologist Vygotsky (Lajoie, 2005: 541-557) By developing a methodology for an empirical study, which focuses on observation only prior to any interaction, we are willing to identify what elements people will focus on NUI screens. We will be able to extract from their comments a more consistent understanding about what they misunderstood and even disregarded – specifically visual cues for potential interaction with novel and unfamiliar interfaces. With Eye tracking I will be able to analyze how participants scrutinize the screen. I will be able to cross reference this raw quantitative data with participant’s utterances – which were organized in classes, with general inductive approach (Thomas, 2006) for qualitative data analysis. Quantitative and Qualitative data were then combined to produce conclusions about what kind of information can be obtained with the protocol – and how can it be later adapted to a NUI prototype or system. -----------------------------------------The way we perceive things is changing. We need to re-interpret the shift we’re living and review the language itself that would better convey the message of Post-WIMP/NUI and encompass the experience of the very interaction itself.The interface should change to encompass TUI and NUI, rather than just co-exist with addictive features in an already exceeded GUI. Research about the different feedback (multi-sensorial) should take place in order to encompass more effcientlythe possibilitities of interfacing with eyes, gestures, voice, touch, emotions and the very mind itself. GUI additions such as Natural User Interfaces, Microsoft’s Surface Computer, eye-tracking and other Haptic interfaces are not transforming the underlying problems created with the GUI.Sorensen (2009)