Bernardos et al., 2016 - Google Patents
A comparison of head pose and deictic pointing interaction methods for smart environmentsBernardos et al., 2016
- Document ID
- 2750559148121214031
- Author
- Bernardos A
- Gómez D
- Casar J
- Publication year
- Publication venue
- International Journal of Human-Computer Interaction
External Links
Snippet
Pointing is one of the more meaningful gestures in human expression. In this article, deictic and head pose pointing methods are evaluated as interaction means, with the objective of comparing their efficiency and delivered comfort, usability, and user experience to apply …
- 230000003993 interaction 0 title abstract description 220
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/0354—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of 2D relative movements between the device, or an operating part thereof, and a plane or surface, e.g. 2D mice, trackballs, pens or pucks
- G06F3/03543—Mice or pucks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/0354—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of 2D relative movements between the device, or an operating part thereof, and a plane or surface, e.g. 2D mice, trackballs, pens or pucks
- G06F3/03549—Trackballs
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/038—Control and interface arrangements therefor, e.g. drivers or device-embedded control circuitry
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/04815—Interaction with three-dimensional environments, e.g. control of viewpoint to navigate in the environment
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/0304—Detection arrangements using opto-electronic means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for programme control, e.g. control unit
- G06F9/06—Arrangements for programme control, e.g. control unit using stored programme, i.e. using internal store of processing equipment to receive and retain programme
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/00362—Recognising human body or animal bodies, e.g. vehicle occupant, pedestrian; Recognising body parts, e.g. hand
- G06K9/00375—Recognition of hand or arm, e.g. static hand biometric or posture recognition
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Bernardos et al. | A comparison of head pose and deictic pointing interaction methods for smart environments | |
LaViola Jr et al. | 3D user interfaces: theory and practice | |
Besançon et al. | Mouse, tactile, and tangible input for 3D manipulation | |
Carmigniani et al. | Augmented reality technologies, systems and applications | |
Feuchtner et al. | Extending the body for interaction with reality | |
Plopski et al. | The eye in extended reality: A survey on gaze interaction and eye tracking in head-worn extended reality | |
Velloso et al. | Motion correlation: Selecting objects by matching their movement | |
Yan et al. | Eyes-free target acquisition in interaction space around the body for virtual reality | |
Spittle et al. | A review of interaction techniques for immersive environments | |
Chen et al. | User-defined gestures for gestural interaction: extending from hands to other body parts | |
US20220269333A1 (en) | User interfaces and device settings based on user identification | |
Shoemaker et al. | Body-centric interaction techniques for very large wall displays | |
Datcu et al. | On the usability and effectiveness of different interaction types in augmented reality | |
Badam et al. | Supporting visual exploration for multiple users in large display environments | |
Schwind et al. | Up to the finger tip: The effect of avatars on mid-air pointing accuracy in virtual reality | |
Aslan et al. | Mid-air authentication gestures: An exploration of authentication based on palm and finger motions | |
Pfeiffer et al. | EyeSee3D 2.0: Model-based real-time analysis of mobile eye-tracking in static and dynamic three-dimensional scenes | |
Schwind et al. | Understanding visual-haptic integration of avatar hands using a fitts' law task in virtual reality | |
Roth et al. | Hybrid avatar-agent technology–A conceptual step towards mediated “social” virtual reality and its respective challenges | |
De Paolis et al. | The effects of touchless interaction on usability and sense of presence in a virtual environment | |
Liebers et al. | Identifying users by their hand tracking data in augmented and virtual reality | |
Hochreiter et al. | Cognitive and touch performance effects of mismatched 3D physical and visual perceptions | |
Zhang et al. | Facilitating temporal synchronous target selection through user behavior modeling | |
Zhang et al. | A hybrid 2D–3D tangible interface combining a smartphone and controller for virtual reality | |
Fikkert et al. | Interacting with visualizations |