U.S. patent application number 16/008577 was filed with the patent office on 2018-12-20 for system for gaze interaction.
This patent application is currently assigned to Tobii AB. The applicant listed for this patent is Tobii AB. Invention is credited to Markus Cederlund, Robert Gavelin, Erland George-Svahn, Anders Kaplan, Anders Olsson, Marten Skogo, Anders Vennstrom.
Application Number | 20180364802 16/008577 |
Document ID | / |
Family ID | 53043388 |
Filed Date | 2018-12-20 |
United States Patent
Application |
20180364802 |
Kind Code |
A1 |
Cederlund; Markus ; et
al. |
December 20, 2018 |
SYSTEM FOR GAZE INTERACTION
Abstract
A control module for generating gesture based commands during
user interaction with an information presentation area is provided.
The control module is configured to acquire user input from a
touchpad and gaze data signals from a gaze tracking module; and
determine at least one user generated gesture based control command
based on a user removing contact of a finger of the user with the
touchpad; determine a gaze point area on the information
presentation area including the user's gaze point based on at least
the gaze data signals; and execute at least one user action
manipulating a view presented on the graphical information
presentation area based on the determined gaze point area and at
least one user generated gesture based control command, wherein the
user action is executed at said determined gaze point area.
Inventors: |
Cederlund; Markus;
(Sollentuna, SE) ; Gavelin; Robert; (Akerberga,
SE) ; Vennstrom; Anders; (Stockholm, SE) ;
Kaplan; Anders; (Uppsala, SE) ; Olsson; Anders;
(Stockholm, SE) ; Skogo; Marten; (Danderyd,
SE) ; George-Svahn; Erland; (Solna, SE) |
|
Applicant: |
Name |
City |
State |
Country |
Type |
Tobii AB |
Danderyd |
|
SE |
|
|
Assignee: |
Tobii AB
Danderyd
SE
|
Family ID: |
53043388 |
Appl. No.: |
16/008577 |
Filed: |
June 14, 2018 |
Related U.S. Patent Documents
|
|
|
|
|
|
Application
Number |
Filing Date |
Patent Number |
|
|
14601007 |
Jan 20, 2015 |
10025381 |
|
|
16008577 |
|
|
|
|
13646299 |
Oct 5, 2012 |
10013053 |
|
|
14601007 |
|
|
|
|
61583013 |
Jan 4, 2012 |
|
|
|
Current U.S.
Class: |
1/1 |
Current CPC
Class: |
G02B 2027/0187 20130101;
G06F 3/038 20130101; G06F 2203/0381 20130101; G02B 27/017 20130101;
G06F 3/0481 20130101; G02B 2027/0178 20130101; G06F 3/013 20130101;
G06F 3/014 20130101; G02B 2027/014 20130101; G06F 3/04883 20130101;
G02B 27/01 20130101; G06F 3/04845 20130101; G06F 3/03547 20130101;
G06F 3/017 20130101 |
International
Class: |
G06F 3/01 20060101
G06F003/01; G06F 3/038 20060101 G06F003/038; G06F 3/0488 20060101
G06F003/0488; G06F 3/0484 20060101 G06F003/0484; G06F 3/0481
20060101 G06F003/0481; G02B 27/01 20060101 G02B027/01; G06F 3/0354
20060101 G06F003/0354 |
Claims
1. A control module for generating gesture based commands during
user interaction with an information presentation area, wherein
said control module is configured to: acquire user input from a
touchpad and gaze data signals from a gaze tracking module; and
determine at least one user generated gesture based control command
based on a user removing contact of a finger of the user with the
touchpad; determine a gaze point area on said information
presentation area including the user's gaze point based on at least
the gaze data signals; and execute at least one user action
manipulating a view presented on said graphical information
presentation area based on said determined gaze point area and at
least one user generated gesture based control command, wherein
said user action is executed at said determined gaze point area.
Description
CROSS REFERENCE TO RELATED APPLICATION
[0001] This application is a Continuation of U.S. application Ser.
No. 14/601,007 filed on Jan. 20, 2015, which is a
Continuation-in-part application of U.S. application Ser. No.
13/646,299 filed on Oct. 5, 2012, which claims priority to, and the
benefit of U.S. Provisional Application No. 61/583,013 filed on
Jan. 4, 2012. The entire content of all of the above applications
are hereby incorporated by reference.
FIELD OF THE INVENTION
[0002] The invention generally relates to computer implemented
systems and methods for utilizing detection of eye movements in
connection with interactive graphical user interfaces. In
particular, the present invention relates to systems and methods
for assisting a user when interacting with a graphical user
interface by combining eye based input with gesture based input and
gesture based user commands.
BACKGROUND
[0003] Human computer interaction has been revolutionized by the
introduction of the graphical user interface (GUI). Thereby, an
efficient means was provided for presenting information to a user
with a bandwidth that immensely exceeded any prior channels. Over
the years the speed at which information can be presented has
increased further through color screens, enlarged displays,
intelligent graphical objects (e.g. pop-up windows), window tabs,
menus, toolbars, etc. During this time, however, the input devices
have remained essentially unchanged, i.e. the keyboard and the
pointing device (e.g. the mouse, track ball or touchpad). In recent
years, handwriting devices have been introduced (e.g. in the form
of a stylus or graphical pen). Nevertheless, while output bandwidth
has multiplied several times, the input bandwidth has been
substantially unchanged. Consequently, a severe asymmetry in the
communication bandwidth in the human computer interaction has
developed.
[0004] In order to decrease this bandwidth asymmetry as well as to
improve and facilitate the user interaction, various attempts have
been made to use eye-tracking for such purposes. By implementing an
eye tracking device in e.g. a laptop, the interaction possibilities
between the user and the different software applications run on the
computer can be significantly enhanced.
[0005] Hence, one interesting idea for improving and facilitating
the user interaction and for removing the bandwidth asymmetry is to
use eye gaze tracking instead or as a complement to mouse input.
Normally, the cursor is positioned on the display according to the
calculated point of gaze of the user. A number of different
techniques have been developed to select and activate a target
object in these systems. In one example, the system activates an
object upon detection that the user fixates his or her gaze at a
certain object for a certain period of time. Another approach is to
detect an activation of an object when the user's eye blinks.
[0006] However, there are problems associated with these solutions
using eye tracking. For example, the humans use their eye in
perceptive actions instead of controlling. Therefore, it may be
stressful to carefully use eye movements to interact with a
computer, for example, to activate and select an object presented
on the display of the computer. It may also be difficult to control
blinking or staring in order to interact with objects presented on
a display.
[0007] Thus, there is a need within the art for improved techniques
that enable user interaction with a computer provided with an eye
tracking device allowing the user to control, select and activate
objects and parts of objects presented on a display of the computer
using his or her eyes in a more intuitive and natural way.
Furthermore, there is also a need within the art for techniques
that in a more efficient way takes advantage the potential of using
eye tracking for improving and facilitating the user interaction
with a computer.
[0008] One such attempt is presented in US pat. appl. (publication
number 2005/0243054) to Beymer et al. in which a technology for
selecting and activating a target object using a combination of eye
gaze and key presses is disclosed. More specifically, a user looks
at a target object, for example, a button on a graphical user
interface and then presses a selection key of the keyboard. Once
the selection key is pressed, a most probable target is determined
using probability reasoning. The determined target object is then
highlighted and the user can select it by pressing the selection
key again. If the highlighted object is not the target object, the
user can select another target object using additional keys to
navigate to the intended target object.
[0009] However, this technology is limited to object selection and
activation based on a combination of eye gaze and two sequential
presses of one dedicated selection key.
[0010] In U.S. Pat. No. 6,204,828 to Amir et al., a computer-driven
system for aiding a user to positioning a cursor by integrating eye
gaze and manual operator input is disclosed. A gaze tracking
apparatus monitors the eye orientation of the user while the user
views a screen. Concurrently, the computer monitors an input
device, such as a mouse, for mechanical activation by the operator.
When the computer detects mechanical activation of the input
device, it determined an initial cursor display position within a
current gaze area. The cursor is then displayed on the screen at
the initial display position and thereafter the cursor is
positioned manually according to the user's handling of the input
device without regard to the gaze.
[0011] Consequently, there still remains a need within the art of
an improved technique that in a more efficient way takes advantage
of the potential in using eye tracking for improving and
facilitating the user interaction with a computer and in particular
user interaction with graphical user interfaces.
SUMMARY
[0012] An object of the present invention is to provide improved
methods, devices and systems for assisting a user when interacting
with a graphical user interface by combining gaze based input with
gesture based user commands.
[0013] Another object of the present invention is to provide
methods, devices and systems for user friendly and intuitive
interaction with graphical user interfaces.
[0014] A particular object of the present invention is to provide
systems, devices and methods that enable a user of a computer
system without a traditional touch-screen to interact with
graphical user interfaces in a touch-screen like manner using a
combination of gaze based input and gesture based user commands.
Furthermore, the present invention offers a solution for
touch-screen like interaction using gaze input and gesture based
input as a complement or an alternative to touch-screen
interactions with a computer device having a touch-screen, such as
for instance in situations where interaction with the regular
touch-screen is cumbersome or ergonomically challenging.
[0015] Another particular object of the present invention is to
provide systems, devices and methods for combined gaze and gesture
based interaction with graphical user interfaces to achieve a
touchscreen like environment in computer systems without a
traditional touchscreen or in computer systems having a touchscreen
arranged ergonomically unfavorable for the user or a touchscreen
arranged such that it is more comfortable for the user to use
gesture and gaze for the interaction than the touchscreen.
[0016] In the context of the present invention, the term "GUI"
(Graphical User Interface) refers to a graphics-based user
interface with pictures or images and words (including e.g. signs
and figures) on a display that incorporate, for example, movable
windows and icons.
[0017] Further, in the context of the present invention the terms
"object" or "object part" refer to an interactive graphical object
or GUI object such as a window, an icon, a button, a scroll bar, a
hyperlink, or non-interactive objects such as an image, text or a
word in a text that the user desires to select or activate.
[0018] In the context of the present invention, the term "touchpad"
(or the term "trackpad") refers to a surface sensor for detecting
the position and movement of one or multiple fingers and/or one or
multiple other objects intended for pointing, drawing or making
gestures, such as for instance a stylus.
[0019] These and other objects of the present invention are
achieved by means of a system having the features defined in the
independent claims. Embodiments of the invention are characterized
by the dependent claims.
[0020] According to an aspect of the present invention, there is
provided a control module for implementation in, for example, a
computer device or handheld device or a wireless transmit/receive
unit (WTRU) for handling and generating gesture based control
commands to execute user action based on these commands. The
control module is configured to acquire user input from input means
adapted to detect user generated gestures and gaze data signals
from a gaze tracking module and to determine at least one user
generated gesture based control command based on the user input.
Further, the control module is configured to determine a gaze point
area on the information presentation area including the user's gaze
point based on at least the gaze data signals and to execute at
least one user action manipulating a view presented on the
graphical information presentation area based on the determined
gaze point area and at least one user generated gesture based
control command, wherein the user action is executed with the
determined gaze point area as a starting point. The gaze point area
serving as a starting point may be an area at which the user
initially gazes at or a fine tuned area, i.e. an area that the user
has selected by tuning or correcting commands via, for example, the
input means, thereby correcting or tuning an initial gaze point
area to a selected area.
[0021] According to another aspect of the present invention, there
is provided a method for generating gesture based commands during
user interaction with an information presentation area, for
example, associated with or included in a computer device or
handheld device, or associated with or included in a wireless
transmit/receive unit (WTRU). The method comprises acquiring user
input corresponding to user generated gestures and gaze data
signals and determining at least one user generated gesture based
control command based on the user input. Further, a gaze point area
on the information presentation area including the user's gaze
point is determined based on at least the gaze data signals and at
least one user action manipulating a view presented on the
information presentation area is executed based on the determined
gaze point area and at least one user generated gesture based
control command, wherein the user action is executed with the
determined gaze point area as a starting point.
[0022] According to a further aspect of the present invention,
there is provided a handheld portable device provided with or
associated with an information presentation area and comprising
input means adapted to detect user generated gestures and a gaze
tracking module adapted to detect gaze data of a viewer of the
information presentation area. The handheld device further
comprises a control module configured to acquire user input from
the input means and gaze data signals from the gaze tracking module
and to determine at least one user generated gesture based control
command based on the user input. The control module is further
configured to determine a gaze point area on the information
presentation area including the user's gaze point based on at feast
the gaze data signals and to execute at least one user action
manipulating a view presented on the information presentation area
based on the determined gaze point area and at least one user
generated gesture based control command, wherein the user action is
executed with the determined gaze point area as a starting point.
In embodiments of the present invention, the handheld device may be
a cellular phone, a smartphone, an iPad or similar device, a
tablet, a phoblet/phablet, a laptop or similar device.
[0023] According to a further aspect of the present invention,
there is provided a wireless transmit/receive unit, WTRU,
associated with an information presentation area and comprising
input means adapted to detect user generated gestures and a gaze
tracking module adapted to detect gaze data of a viewer of the
information presentation area. The WTRU further comprises a control
module configured to acquire user input from the input means and
gaze data signals from the gaze tracking module and to determine at
least one user generated gesture based control command based on the
user input. The control module is further configured to determine a
gaze point area on the information presentation area including the
user's gaze point based on at least the gaze data signals and to
execute at least one user action manipulating a view presented on
the information presentation area based on the determined gaze
point area and at least one user generated gesture based control
command, wherein the user action is executed with the determined
gaze point area as a starting point.
[0024] The term "wireless transmit/receive unit (WTRU)" include but
is not limited to a user equipment (UE), a mobile station, a fixed
or mobile subscriber unit, a cellular telephone, a smartphone, a
personal digital assistant (PDA), a computer, or any other type of
device capable of operating in a wireless environment such as a
wireless local area network (WLAN) or wireless mobile communication
system (e.g. a third generation (3G) global system for mobile
communication and systems for mobile communication including long
term evolution (LTE) cells).
[0025] According to another aspect of the present invention, there
is provided a system for user interaction with an information
presentation area. The system comprises input means adapted to
detect user generated gestures and a gaze tracking module adapted
to detect gaze data of a viewer of the information presentation
area. Further, the system includes a control module configured to
acquire user input from the input means and gaze data signals from
the gaze tracking module and to determine at least one user
generated gesture based control command based on the user input.
The control module is further configured to determine a gaze point
area on the information presentation area where the user's gaze
point is located based on at least the gaze data signals and to
execute at least one user action manipulating a view presented on
the graphical information presentation area based on the determined
gaze point area and at least one user generated gesture based
control command, wherein the user action is executed with the
determined gaze point area as a starting point.
[0026] According to yet another aspect of the present invention,
there is provided a computer device associated with an information
presentation area. The computer device comprises input means
adapted to detect user generated gestures and a gaze tracking
module adapted to detect gaze data of a viewer of the information
presentation area. The computer device further comprises a control
module configured to acquire user input from input means adapted to
detect user generated gestures and gaze data signals from a gaze
tracking module and to determine at best one user generated gesture
based control command based on the user input. Moreover, the
control module is configured to determine a gaze point area on the
information presentation area including the user's gaze point based
on at least the gaze data signals and to execute at `east one user
action manipulating a view presented on the information
presentation area based on the determined gaze point area and at
least one user generated gesture based control command, wherein the
user action's executed with the determined gaze point area as a
starting point.
[0027] According to embodiments of the present invention, the
computer device may, for example, be any one from the group of a
personal computer, computer workstation, mainframe computer, a
processor or device in a vehicle, or a handheld device such as a
cell phone, smartphone or similar device, portable music player
(such as e.g. an iPod), laptop computers, computer games,
electronic books, an iPAD or similar device, a Tablet, a
Phoblet/Phablet.
[0028] According to embodiments of the present invention, the input
means is configured to detect user gestures by a hand or a finger
(or fingers), for example, relative a keyboard or an information
presentation area using, for example, an optical measurement
technique or capacitive measurement technique.
[0029] According to an aspect of the present invention, there is
provided a system for user interaction with a wearable head mounted
information presentation area. The system comprises input means
configured as a gyro ring adapted to detect user generated gestures
and adapted to wirelessly communicate with a control module also
communicatively connected to the information presentation area as
well as a gaze tracking module adapted to detect gaze data of a
viewer of the information presentation area. A control module
configured to: acquire user input from the input means and gaze
data signals from the gaze tracking module; determine at least one
user generated gesture based control command based on the user
input; determine a gaze point area on the information presentation
area including the user's gaze point based on at least the gaze
data signals; and execute at least one user action manipulating a
view presented on the graphical information presentation area based
on the determined gaze point area and at least one user generated
gesture based control command, wherein the user action is executed
with the determined gaze point area as a starting point.
[0030] According to a further aspect of the present invention,
there is provided a system for user interaction with an information
presentation area. The system comprises input means adapted to
detect user generated gestures, wherein the input means comprising
at least one touchpad arranged on a steering device of a vehicle or
adapted to be integrated in a steering device of a vehicle.
Further, the system comprises a gaze tracking module adapted to
detect gaze data of a viewer of the information presentation area
and a control module configured to: acquire user input from the
input means and gaze data signals from the gaze tracking module;
determine at least one user generated gesture based control command
based on the user input; determine a gaze point area on the
information presentation area including the users gaze point based
on at least the gaze data signals; and execute at least one user
action manipulating a view presented on the graphical information
presentation area based on the determined gaze point area and at
least one user generated gesture based control command, wherein the
user action is executed with the determined gaze point area as a
starting point.
[0031] According to embodiments of the present invention, the input
means includes a touchpad configured to enable a user to generate
gesture based control commands. The gesture based commands can for
example be generated by moving at least one finger over a surface
of the touchpad or touching a surface of the touchpad with, for
example, the finger.
[0032] According to embodiments of the present invention, a
dedicated part or area of the touchpad surface is configured to
receive gesture based control commands.
[0033] According to embodiments of the present invention, at least
a first dedicated part or area of the touchpad surface is
configured to receive a first set of gesture based control commands
and at least a second part or area of the touchpad surface is
configured to receive a second set of gesture based control
commands. For example, the touchpad may be configured to receive
gestures such as scrolling or zooming at a dedicated area or
part.
[0034] In embodiments of the present invention, the control module
is configured to determine at least one gesture based control
command based on multiple simultaneous user input via the input
means. Further, a gaze point area on the information presentation
area where the user's gaze point is located is determined based on
the gaze data signals and at least one user action manipulating a
view presented on the graphical information presentation area is
executed based on the determined gaze point area and the at least
one gesture based control command, wherein the user action is
executed with the determined gaze point area as a starting
point.
[0035] According to embodiments of the present invention, an input
module is configured to interpret signals representing at least one
user generated gesture to provide at least one gesture based
control command reflecting a user's gesture. According to
embodiments of the present invention, the input module is arranged
in the control module.
[0036] In embodiments of the present invention, the input module is
configured to interpret the signals representing the at least one
user generated gesture using gaze input signals and/or a
predetermined set of possible gesture based control commands, each
possible control command corresponding to a particular user gesture
relative the input means.
[0037] According to embodiments of the present invention, at least
one object is presented on the graphical information presentation
area, the object representing at least one graphical user interface
component and configured to be manipulated based on the
user-generated gesture based control commands, wherein the control
module is configured to determine if the gaze point of the user is
on an object or in an area surrounding that object based on the
gaze data signals. Further, the control module may be configured to
determine if the gaze point of the user has been on an object or in
an area surrounding that object at a predetermined point in time
based on the gaze data signals. For example, the control module may
be configured to determine if the gaze point of the user was on an
object or the area surrounding that object 0.1 seconds ago.
[0038] User activation of the object is enabled if the user's gaze
point is on or within an area surrounding that object synchronized
with a user generated activation command resulting from user input
via the input means, wherein the activated object can be
manipulated by user generated commands resulting from user input
via the input means. User activation of the object may also be
enabled if the user's gaze point was on or within an area
surrounding that object at the predetermined period of time
synchronized with a user generated activation command resulting
from user input via the input means, wherein the activated object
can be manipulated by user generated commands resulting from user
input via the input means.
[0039] According to embodiments of the present invention, when the
user touches the touchpad, the location of the initial gaze point
is indicated by a visual feedback, such as a crosshairs or similar
sign. The user may adjust this initial location by moving the
finger on the touchpad. Then, the user may, in a touchscreen like
manner, interact with the information presentation area using
different gestures. The strength of the visual feedback, e.g. the
strength of the light of a crosshairs, may be dependent on where
the user's gaze is located on the information presentation area.
For example, if a dragging operation to pan a window is initiated
at the gaze point, the visual feedback may initially be discrete.
When the dragging operation has been maintained for a period, the
visual feedback can be strengthened to indicate for the user where
the dragging operation is performed at the moment.
[0040] In the embodiments including a touchpad, the gestures are
finger movements relative the touchpad and each gesture is
associated with or corresponds to a particular gesture based
control command resulting in a user action. Below, a non-exhaustive
number of examples of user actions that can be executed using a
combination of gestures and gaze are discussed: [0041] By gazing,
for example, at an object presented on the information presentation
area and by, in connection to this, pressing down and holding a
finger on the touchpad during a predetermined period of time, a
visual feedback related to that object is presented. For example,
by pressing down and holding the finger on the touchpad during a
first period of time, the object may be highlighted and, by
continue to hold the finger on the touchpad for a second period of
time, an information box presenting information regarding the
object may be displayed. [0042] By gazing, for example, at an
object presented on the information presentation area and by in
connection to this tapping on the touchpad using a finger, a
primary action can be initiated. For example, an application can be
opened and started by gazing at an icon representing the
application and tapping on the touchpad using a finger. According
to a further embodiment of the present invention after enacting a
primary action by tapping on a touchpad a second tap may be
performed which fine adjusts the effect of the first tap. By way of
example, consider when a user is utilizing a video game whereby the
user gazes at a target and taps to shoot the target. The user may
observe that the first tap caused the shot to pass to the right of
the target, the user may then tap slightly to the left of their
original tap in the subsequent tap on the touchpad while
maintaining their gaze. The system will then provide the next shot
slightly to the left of the previous shot. This is also of use when
selecting an item for opening or highlighting, such as opening an
application. The user may observe that their original tap has
caused an action to be performed not at their desired target. This
may be due to an inaccuracy in the gaze tracking or an offset
caused by the gaze tracker. A subsequent tap on the touchpad may
then be used to adjust the gaze location and thus the subject of
any action caused by tapping the touchpad. [0043] By gazing, for
example, at an object presented on the information presentation
area and by, in connection to this, lifting a finger (or fingers)
that have been in contact with the touchpad, a primary action can
be initiated. For example, an application can be opened and started
by gazing at an icon representing the application and lifting a
finger (or fingers) that have been in contact with the touchpad.
[0044] The user may slide or drag the view presented by the
information presentation area by gazing at the information
presentation area and by, in connection to this, sliding his or her
finger over the touchpad. The dragging is then initiated at the
gaze point of the user. A similar action to slide an object over
the information presentation area can be achieved by gazing at the
object and by, in connection to this, sliding the finger over the
touchpad. Both of these objectives may instead be implemented in a
way where two fingers are required to do the swipe, or one finger
is used for swiping while another finger holds down a button.
[0045] According to an embodiment of the present invention, by
gazing, for example, at an object presented on the information
presentation area and by, in connection to this, pressing down and
holding at least one finger or thumb on the touchpad or
touchscreen, the object may be selected. By moving the at least one
finger across the touchpad or touchscreen, the object may be moved
around or resized within the information presentation area. The new
position or size of the object may be confirmed by removing the at
least one finger from the touchpad or touchscreen, placing an
additional finger on the touchpad or touchscreen, maintaining gaze
at a location on the information presentation area or saying a word
or phrase. [0046] One example of this embodiment is moving windows
or other such information containing objects on the information
presentation area. In a preferred embodiment, two, three or four
fingers are placed on a touchpad while gazing at a window or the
like. The window subsequently becomes the active window and moving
the fingers across the touchpad facilitates movement of the window
across the information display area. Releasing contact of all the
fingers with the touchpad will reposition the window within the
information presentation area. However releasing contact of less
than all the fingers with the touchpad will allow for a secondary
action to be performed, such as a pinch motion to enable resizing
of the window. Furthermore it is possible to reposition the window
such that it may jump directly to a gaze position upon release of
at least on finger on the touchpad. [0047] As a further
improvement, repositioning the window adjacent an edge of the
information display area will cause the window to link with the
edge of the information display area, otherwise sometimes referred
to as "docking." For example, repositioning the window such that it
is adjacent the top edge of the information display area will cause
the window to align its top most edge with the top most edge of the
information display area and may cause the window to occupy
substantially the entirety of the information display area. [0048]
As a further improvement, repositioning the window in a
predetermined area of the information display such as the edge may
perform an action associated with the window, for example closing
the window, minimizing the window, maximizing the window and the
like. In some embodiments, it is preferable to provide a visual
indication on the information presentation area to allow a user to
easily see where to position the window so as to perform a desired
act. The act may be an act associated with the window such as a
close, maximize or minimize act or it may be associated with a
program running inside the window such as a file open, file close,
new document or the like action. [0049] Furthermore, once the user
has selected a window in a manner as previously described, areas at
which a user may gaze to perform an action such as close the
window, minimize the window or maximize the window may be
highlighted on the information presentation area. The user may
select an area by dwelling their gaze within the area, voice,
movement on a touchpad, enacting contact with a touchpad and
releasing contact with a touchpad. [0050] In a further improvement,
a user need not reposition the window by moving their at least on
finger across the touchpad or touchscreen; the user may reposition
the window by gazing at the desired position and maintaining their
gaze for a predetermined period, i.e., a dwell, or by saying a
predetermined word or phrase such as "move here." It is further
desirable to provide a visual indicator as to where the window will
be repositioned, if the user completes a movement or resize action.
[0051] As a further improvement, when the user places at least one
finger on the touchpad the device may determine that the user's
gaze is adjacent an edge of a window or other such information
containing object. Instead of enacting a mode whereby the window is
moved by the user moving their at least one finger on the touchpad,
the device may enact one of the following modes: [0052] If the gaze
is located adjacent a vertical edge of the window, a mode which
allows resizing of the horizontal size of the window by moving the
at least one finger on the touchpad is enacted. [0053] If the gaze
is located adjacent a horizontal edge of the window, a mode which
allows resizing of the vertical size of the window by moving the at
least one finger on the touchpad is enacted. [0054] If the gaze is
located adjacent the bottom right hand corner of the window, a mode
which allows resizing of the window in both the horizontal and
vertical directions by moving the at least one finger on the
touchpad is enacted. [0055] As a further embodiment, where a
touchpad or touchscreen is capable of detecting pressure, contact
with the touchpad or touchscreen that is sufficiently hard will
cause a window to become active and able to be moved. When the
contact with the touchpad or touchscreen is decreased in pressure,
the location of the window will be fixed. [0056] According to an
embodiment of the present invention, placing at least one finger on
a touchpad and moving in a set direction for example upwards or
downwards a list of programs currently running on a device to be
displayed on an information display area, the list comprising
summaries or representations of each program. Moving the at least
one finger across the touchpad will cycle a selector through the
list of currently running programs, alternatively a summary may be
selected by positioning a user's gaze over it. Releasing contact
with the touchpad will cause the program associated with the last
selected summary to be made active and displayed on the information
display area. Moving the at least one finger in an opposite
direction to that which caused the list of programs to be displayed
will cause the list to cease being displayed. The display and
cessation of display may occur gradually to provide a fading in or
out effect. [0057] The user may select an object for further
actions by gazing at the object and by, in connection to this,
swiping his or her finger downwards on the touchpad. [0058] By
gazing at an object or object part presented on the information
presentation area and by, in connection to this, pinching with two
of his or hers finger, it is possible to zoom that object or object
part. The same function can be implemented also on a touchpad only
able to sense single touch by having for instance the thumb push a
button or keyboard key and the finger moving on the touchpad away
from, or towards, the button or keyboard key. [0059] By gazing at
an object or object part presented on the information presentation
area and by, in connection to this, rotating with two of his or
hers finger, it is possible to rotate that object or object part.
Similarly, when using a touchpad only able to sense single touch
the thumb can press a button while a finger moves on the touchpad
in a curve at a constant distance from the button to rotate an
object. Other manipulations of an object or object part of interest
are also possible in addition to rotation, for example after gazing
at an object a user may contact a touchpad and perform a movement
to hold and drag an object. For example this may be used in a video
game such as a card game or a game where an object is the subject
of a slingshot. The object such as the card may be gazed at and
then a press on a touchpad would cause that card to be selected or
held, moving on the touchpad would cause the card to be moved until
the touchpad was released. [0060] By gazing at an edge of the
information presentation area and sliding the finger over the
touchpad in the direction that would have been towards the center
of the information presentation area if the gesture had been done
at the gaze position, a menu or other window hidden during normal
use, such as a help menu, can be presented or displayed. That is, a
hidden menu or other window can be displayed or presented if the
user gazes at, for example, the left edge of the information
presentation area and swipes his or her finger over the touchpad in
the right direction. [0061] By gazing at a slider control, for
example a volume control, the finger can be moved up/down (or
left/right for a horizontal control) on the touch pad, on a
predefined area of a touch screen or above a keyboard to adjust the
value of the slider control. [0062] By gazing at a checkbox control
while doing a "check-gesture" (such as a "V") on the touchpad, the
checkbox can be checked or unchecked. [0063] By gazing at a
zoomable object or object part presented on the information
presentation area and while pressing hard on a pressure sensitive
touchpad with one finger (e.g. one of the thumbs), it is possible
to zoom in or out on said object using the gaze point as the zoom
center point, where each hard press toggles between different zoom
levels. [0064] By gazing at an object or object part where several
options are available, for example "copy" or "rename", the
different options can be displayed on different sides of the object
after a preset focusing dwell time has passed or after appropriate
user input has been provided. The touchpad or a predefined area of
a touch screen is thereafter used to choose action. For example,
slide left to copy and slide right to rename. [0065] Other gestures
may be performed to interact with an object at a gaze area,
examples of such gestures include: [0066] performing a swiping
gesture to dismiss or move away the object, [0067] performing a
slicing gesture to slice or cut the object, [0068] performing a
rotation gesture to adjust a value associated with the object,
[0069] performing a rotation gesture to perform an `undo` action
associated with the object, perform a user defined gesture to open
a menu, map or other such information selection tool, performing a
scratch gesture (e.g. a rapid back and forth movement) to highlight
an object or text or otherwise interact with the object, performing
a drawing of an object to cause an act to be performed, for example
drawing an X or O in a game of tic tac toe, or drawing a
representation of a magic rune or the like to cast a spell in a
video game.
[0070] According to another embodiment of the present invention,
the gaze tracking module and the user input means are implemented
in a touchscreen provided device such as an iPad or similar device.
The touchscreen functions both as information presentation area and
input device for input of user gestures. A control module is
included in the touchscreen provided device and is configured to
determine a gaze point area on the information presentation area,
i.e. the touchscreen, where the user's gaze point is located based
on the gaze data signals and to execute at least one user action
manipulating a view presented on the touchscreen based on the
determined gaze point area and at least one user generated gesture
based control command, wherein the user action is executed with the
determined gaze point area as a starting point. The user gestures
are inputted via the touchscreen. According to this embodiment, the
user gestures, or finger movements on the touchscreen, are relative
to the gaze point, which entails a more user friendly and ergonomic
use of touchscreen provided devices. For example, the user may hold
the device with both hands and interact with graphical user
interfaces on the touchscreen using the gaze and movement of the
thumbs, where all user actions and activations have the gaze point
of the user as starting point.
[0071] As mentioned, the gesture and gaze initiated actions
discussed above are only exemplary and there are a large number of
further gestures in combination with gaze point resulting in an
action that are conceivable. Below, some further examples are
described: [0072] Selection of an object or object part can be made
by gazing at that object or object part and pressing a finger (e.g.
a thumb), fine tuning by moving the finger and releasing the
pressure applied by the finger to select that object or object
part; [0073] Selection of an object or object part can be made by
gazing at that object or object part, pressing a finger (e.g. a
thumb), fine tuning by moving the finger, using another finger
(e.g. the other thumb) to tap for selecting that object or object
part. In addition, a double tap may be used for a "double click
action" and a quick downward movement may be used for a "right
click". [0074] By gazing at a zoomable object or object part
presented on the information presentation area while moving a
finger (e.g. one of the thumbs) in a circular motion, it is
possible to zoom in or out of said object using the gaze point as
the zoom center point, where a clockwise motion performs a "zoom
in" command and a counterclockwise motion performs a "zoom out"
command or vice versa. [0075] By gazing at a zoomable object or
object part presented on the information presentation area and in
connection to this holding one finger (e.g. one of the thumbs)
still while moving another finger (e.g. the other thumb) upwards or
downwards, it is possible to zoom in or out of said object using
the gaze point as the zoom center point, where an upwards motion
performs a "zoom in" command and a downwards motion performs a
"zoom out" command or vice versa. [0076] By gazing at a zoomable
object or object part presented on the information presentation
area while double-tapping on the touch screen with one finger (e.g.
one of the thumbs), it is possible to zoom in or out of said object
using the gaze point as the zoom center point, where each
double-tap toggles between different zoom levels. [0077] By gazing
at a zoomable object or object part presented on the information
presentation area while sliding two fingers (e.g. the two thumbs)
simultaneously in opposite horizontal directions, it is possible to
zoom that object or object part. [0078] By gazing at a zoomable
object and in connection to this holding a finger (e.g. one thumb)
still on the touchscreen while moving another finger (e.g. the
other thumb) in a circular motion, it is possible to zoom that
object or object part. [0079] By gazing at an object or object part
presented on the information presentation area and in connection to
this holding a finger (e.g. one of the thumbs) still on the
touchscreen while sliding another finger (e.g. the other thumb), it
is possible to slide or drag the view presented by the information
presentation area. [0080] By gazing at an object or object part
presented on the information presentation area and in connection to
this holding a finger (e.g. one of the thumbs) still on the
touchscreen while sliding another finger (e.g. the other thumb), it
is possible to slide or drag the view presented by the information
presentation area. [0081] By gazing at an object or object part
presented on the information presentation area and while tapping or
double-tapping with a finger (e.g. one of the thumbs), an automatic
panning function can be activated so that the presentation area is
continuously slided from one of the edges of the screen towards the
center while the gaze point is near the edge of the information
presentation area, until a second user input is received. [0082] By
gazing at an object or object part presented on the information
presentation area and while tapping or double-tapping with a finger
(e.g. one of the thumbs), the presentation area is instantly slided
according to the gaze point (e.g. the gaze point is used to
indicate the center of where the information presentation area
should be slided). [0083] By gazing at a rotatable object or object
part presented on the information presentation area while sliding
two fingers (e.g. the two thumbs) simultaneously in opposite
vertical directions, it is possible to rotate that object or object
part.
[0084] Before the two-finger gesture is performed, one of the
fingers can be used to fine-tune the point of action. For example,
a user feedback symbol like a "virtual finger" can be shown on the
gaze point when the user touches the touchscreen. The first finger
can be used to slide around to adjust the point of action relative
to the original point. When the user touches the screen with the
second finger, the point of action is fixed and the second finger
is used for "clicking" on the point of action or for performing
two-finger gestures like the rotate, drag and zoom examples
above.
[0085] According to another embodiment of the current invention,
the gaze tracking module and the user input means are implemented
in a portable device such as an iPad, ultrabook tablet or similar
device. However, instead of performing the gestures with the thumbs
on the presentation area, one or two separate touchpads are placed
on the back side of the device to allow two-finger gestures with
other fingers than the thumb.
[0086] According to another embodiment of the current invention,
the gaze tracking module and the user input means are implemented
in a vehicle. The information presentation area may be a heads-up
display or an infotainment screen. The input means may be one or
two separate touch pads on the backside (for use with the index
finger/s) or on the front side (for use with the thumb/s) of the
steering wheel.
[0087] According to another embodiment of the current invention,
the gaze tracking module and the information presentation area are
implemented in a wearable head mounted display that may be designed
to look as a pair of glasses (such as the solution described in
U.S. Pat. No. 8,235,529). The user input means may include a gyro
and be adapted to be worn on a wrist, hand or at least one finger.
For example the input means may be a ring with a wireless
connection to the glasses (or to a processing unit such as a smart
phone that is communicatively connected to the glasses) and a gyro
that detects small movements of the finger where the ring is worn.
The detected movements representing gesture data may then
wirelessly be communicated to the glasses where gaze is detected
and gesture based control commands based on the gesture data from
the input means is used to identify and execute user action.
[0088] Normally, in most applications, the touchpad is
significantly smaller than the information presentation area, which
entails that in certain situations the touchpad may impose
limitations on the possible user actions. For example, it may be
desired to drag or move an object over the entire information
presentation area while the user's movement of a finger or fingers
is limited by the smaller touchpad area. Therefore, in embodiments
of the present invention, a touchscreen like session can be
maintained despite that the user has removed the finger or fingers
from the touchpad if, for example, a specific or dedicated button
or keyboard key is held down or pressed. Thereby, it is possible
for the user to perform actions requiring multiple touches on the
touchpad. For example, an object can be moved or dragged across the
entire information presentation area by means of multiple dragging
movements on the touchpad.
[0089] In other embodiments of the present invention, a dragging
movement on the information presentation area or other user action
is continued after the finger or fingers has reached an edge of the
touchpad in the same direction as the initial direction of the
finger or fingers. The continued movement or other actions may be
continued until an interruption command is delivered, which may be,
for example, a pressing down of a keyboard key or button, a tap on
the touchpad or when the finger or fingers is removed from the
touchpad.
[0090] In further embodiments of the present invention, the speed
of the dragging movement or other action is increased or
accelerated when the user's finger or fingers approaches the edge
of the touchpad. The speed may be decreased if the fingers or
finger is moved in an opposite direction.
[0091] In embodiments of the present invention, the action, e.g. a
dragging movement of an object, can be accelerated based on gaze
position. For example, by gazing at an object, initiating a
dragging operation of that object in a desired direction and
thereafter gazing at a desired end position for that object, the
speed of the object movement will be higher the longer the distance
between the initial position of the object and the desired end
position is.
[0092] In other embodiments of the present invention voice commands
may be used to choose what action to perform on the object
currently being gazed at and then a gesture is required to fulfill
the action. For instance a voice command such as the word "move"
may allow the user to move the object currently being gazed at by
moving a finger over the touchpad or touchscreen. Another action to
perform may be to delete an object. In this case the word "delete"
may allow deletion of the object currently being gazed at, but
additionally a gesture, such as swiping downwards is required to
actually delete the object. Thus, the object to act on is chosen by
gazing at it, the specific action to perform is chosen by a voice
command and the movement to perform or the confirmation is done by
a gesture.
[0093] According to an embodiment of the present invention, a
touchpad may amplify an action being performed by a character or
object on the information presentation area. For example, while
gazing at an area on the information presentation area to perform
an act, contact with the touchpad may amplify the act. One example
of this is the case of a video game or the like where a character
may lean around a corner, and the lean may be initiated by an input
such as a mouse action, key press, head movement or gaze. While the
lean is being performed, performing a gesture on the touchpad may
amplify or extend the lean. In this manner the user may initiate an
act based on their gaze and then utilize a gesture on a touchpad to
amplify or extend that act.
[0094] According to a further embodiment of the present invention,
a touchpad may be mapped to a view of the information presentation
area. The touchpad may then be used to perform an action at the
mapped portion of the information presentation area. The mapping
may be 1:1 or in some instances may be enhanced so as to provide a
more amplified version of the action; for example, the mapping may
be 2:1 or 3:1. By way of example, this embodiment may be used in
the following manner: a user gazes at an area of the information
presentation area; a magnification around the gaze area may then be
displayed; this may be automatically displayed or displayed upon
the user performing some act. The magnified area may then be mapped
by the device to the touchpad, such that pressing a location on the
touchpad performs an action at the corresponding location in the
magnified area. For example, in the case of a video game, a
magnified area of a map or the like may be shown and tapping on the
touchpad may cause an action to happen at that location on the map.
For example, pressing a location on the touchpad approximately 75%
across and 25% down the touchpad would cause the action associated
with pressing to be performed 75% across and 25% down the magnified
area.
[0095] According to a further embodiment of the present invention,
the sensitivity of a pointer is adjusted based on a detected gaze
position. By way of example, if a gaze area is detected adjacent an
edge of a screen, then the speed of a pointer will proceed at a
faster pace than if the gaze area is detected in the center of the
screen. The speed or sensitivity of a pointer may also be effected
by the object at which a user is gazing; for example, if a user is
gazing at sand in a game, then the pointer may move at a slower
speed.
[0096] In a further embodiment of the present invention, a touch
sensitive surface is provided remote from an information display
area. The touch sensitive surface may be provided on a controller
or the like, such as a controller provided by Sony for the
PlayStation.RTM. 4 gaming console. Furthermore, a touch sensitive
tablet or the like may be provided remote from the information
display area. In this manner a user may daze at the information
display area and have their gaze detected of any previously or
otherwise described touchscreen or touchpad enabled
interactions.
[0097] Further objects and advantages of the present invention will
be discussed below by means of exemplifying embodiments.
[0098] These and other features, aspects and advantages of the
invention will be more fully understood when considered with
respect to the following detailed description, appended claims and
accompanying drawings.
BRIEF DESCRIPTION OF THE DRAWINGS
[0099] The drawings are not necessarily drawn to scale and
illustrate generally, by way of example, but no way of limitation,
various embodiments of the present invention. Thus, exemplifying
embodiments of the invention are illustrated by way of example and
not by way of limitation in the figures of the accompanying
drawings in which like references indicate similar elements. It
should be noted that references to "an" or "one" embodiment in this
discussion are not necessarily to the same embodiment, and such
references mean at least one.
[0100] FIG. 1 shows an overview picture of a user controlling a
computer apparatus in which the present invention is
implemented;
[0101] FIG. 2 is a block diagram illustrating an embodiment of an
arrangement in accordance with the present invention;
[0102] FIG. 3 is a block diagram illustrating another embodiment of
an arrangement in accordance with the present invention;
[0103] FIG. 4 illustrates an exemplary gesture resulting in a user
generated gesture based control command in accordance with the
present invention;
[0104] FIG. 5 illustrates another exemplary gesture resulting in a
user generated gesture based control command in accordance with the
present invention;
[0105] FIG. 6 illustrates a further exemplary gesture resulting in
a user generated gesture based control command in accordance with
the present invention;
[0106] FIG. 7 illustrates yet another exemplary gesture resulting
in a user generated gesture based control command in accordance
with the present invention;
[0107] FIG. 8 illustrates a further exemplary gesture resulting in
a user generated gesture based control command in accordance with
the present invention;
[0108] FIG. 9 illustrates another exemplary gesture resulting in a
user generated gesture based control command in accordance with the
present invention;
[0109] FIG. 10 illustrates yet another exemplary gesture resulting
in a user generated gesture based control command in accordance
with the present invention;
[0110] FIG. 11a shows an overview picture of a touchscreen provided
device in which a further embodiment of the present invention is
implemented;
[0111] FIG. 11b shows an overview picture of a device provided with
touchpads on a backside in which a further embodiment of the
present invention is implemented;
[0112] FIG. 12 is a block diagram illustrating the embodiment in
accordance with the present invention shown in FIG. 11a;
[0113] FIG. 13a is a schematic view of a control module according
to an embodiment of the present invention;
[0114] FIG. 13b is a schematic view of a control module according
to another embodiment of the present invention;
[0115] FIG. 13c is a schematic view of a control module according
to another embodiment of the present invention;
[0116] FIG. 14 is a schematic view of a wireless transmit/receive
unit, WTRU, according to an embodiment of the present
invention;
[0117] FIG. 15a is a schematic view of an embodiment of a computer
device or handheld device in accordance with an embodiment of the
present invention;
[0118] FIG. 15b is a schematic view of another embodiment of a
computer device or handheld device in accordance with the present
invention;
[0119] FIG. 16 is a schematic flow chart illustrating steps of an
embodiment of a method in accordance with an embodiment of the
present invention;
[0120] FIG. 17 is a schematic flow chart illustrating steps of
another embodiment of a method in accordance with the present
invention; and
[0121] FIG. 18 is a schematic flow chart illustrating steps of a
further embodiment of a method in accordance with an embodiment of
the present invention;
[0122] FIG. 19 is a schematic flow chart illustrating steps of
another embodiment of a method in accordance with an embodiment of
the present invention;
[0123] FIG. 20 is a block diagram illustrating a further embodiment
of an arrangement in accordance with the present invention;
[0124] FIG. 21 is a schematic illustration of yet another
implementation of the present invention;
[0125] FIG. 22 is a schematic illustration of a further
implementation of the present invention; and
[0126] FIG. 23 is a schematic illustration of an implementation of
the present invention.
DETAILED DESCRIPTION OF THE INVENTION
[0127] As used herein, the term "module" refers to an application
specific integrated circuit (ASIC), an electronic circuit, a
processor (shared, dedicated, or group) and memory that execute one
or more software programs, a combinational logic circuit, or other
suitable components that provide the described functionality. The
term "module" further refers to a specific form of software
necessary to practice the methods described herein and particularly
the functions described in connection with each specific "module".
It is believed that the particular form of software will be
determined primarily by the particular system architecture employed
in the system and by the particular methodologies employed by the
system according to the present invention.
[0128] The following is a description of exemplifying embodiments
in accordance with the present invention. This description is not
to be taken in limiting sense, but is made merely for the purposes
of describing the general principles of the invention. It is to be
understood that other embodiments may be utilized and structural
and logical changes may be made without departing from the scope of
the present invention.
[0129] With reference first to FIGS. 1, 2, 3 and 20, embodiments of
a computer system according to the present invention will be
described. FIG. 1 shows an embodiment of a computer system with
integrated gaze and manual control according to the present
invention. The user 110 is able to control the computer system 10
at least partly based on an eye-tracking signal D.sub.EYE, which
described the user's point of regard x, y on an information
presentation area or display 20 and based on user generated
gestures, i.e. a movement of at least one body part of the user can
be detected, generating gesture based control commands via user
input means 50 such as a touchpad 51.
[0130] In the context of the present invention, as mentioned above,
the term "touchpad" (or the term "trackpad") refers to a pointing
device featuring a tactile sensor, a specialized surface that can
translate the motion and position of a user's fingers to a relative
position on a screen (information presentation area). Touchpads are
a common feature of laptop computers, and are also used as a
substitute for a mouse where desk space is scarce. Because they
vary in size, they can also be found on personal digital assistants
(PDAs) and some portable media players. Wireless touchpads are also
available as detached accessories. Touchpads operate in one of
several ways, including capacitive sensing and conductance sensing.
The most common technology used today entails sensing the
capacitive virtual ground effect of a finger, or the capacitance
between sensors. While touchpads, like touchscreens, are able to
sense absolute position, resolution is limited by their size. For
common use as a pointer device, the dragging motion of a finger is
translated into a finer, relative motion of the cursor on the
screen, analogous to the handling of a mouse that is lifted and put
back on a surface. Hardware buttons equivalent to a standard
mouse's left and right buttons are positioned below, above, or
beside the touchpad. Netbooks sometimes employ the last as a way to
save space. Some touchpads and associated device driver software
may interpret tapping the pad as a click, and a tap followed by a
continuous pointing motion (a "click-and-a-half") can indicate
dragging. Tactile touchpads allow for clicking and dragging by
incorporating button functionality into the surface of the touchpad
itself. To select, one presses down on the touchpad instead of a
physical button. To drag, instead performing the "click-and-a-half"
technique, one presses down while on the object, drags without
releasing pressure and lets go when done. Touchpad drivers can also
allow the use of multiple fingers to facilitate the other mouse
buttons (commonly two-finger tapping for the center button). Some
touchpads have "hotspots", locations on the touchpad used for
functionality beyond a mouse. For example, on certain touchpads,
moving the finger along an edge of the touch pad will act as a
scroll wheel, controlling the scrollbar and scrolling the window
that has the focus vertically or horizontally. Apple uses
two-finger dragging for scrolling on their trackpads. Also, some
touchpad drivers support tap zones, regions where a tap will
execute a function, for example, pausing a media player or
launching an application. All of these functions are implemented in
the touchpad device driver software, and can be disabled. Touchpads
are primarily used in self-contained portable laptop computers and
do not require a flat surface near the machine. The touchpad is
close to the keyboard, and only very short finger movements are
required to move the cursor across the display screen; while
advantageous, this also makes it possible for a user's thumb to
move the mouse cursor accidentally while typing. Touchpad
functionality is available for desktop computers in keyboards with
built-in touchpads.
[0131] Examples of touchpads include one-dimensional touchpads used
as the primary control interface for menu navigation on
second-generation and later iPod Classic portable music players,
where they are referred to as "click wheels", since they only sense
motion along one axis, which is wrapped around like a wheel. In
another implementation of touchpads, the second-generation
Microsoft Zunc product line (the Zune 80/120 and Zune 4/8) uses
touch for the Zune Pad. Apple's PowerBook 500 series was its first
laptop to carry such a device, which Apple refers to as a
"trackpad". Apple's more recent laptops feature trackpads that can
sense up to five fingers simultaneously, providing more options for
input, such as the ability to bring up the context menu by tapping
two fingers. In late 2008 Apple's revisions of the MacBook and
MacBook Pro incorporated a "Tactile Touchpad" design with button
functionality incorporated into the tracking surface.
[0132] The present invention provides a solution enabling a user of
a computer system without a traditional touchscreen to interact
with graphical user interfaces in a touchscreen like manner using a
combination of gaze based input and gesture based user commands.
Furthermore, the present invention offers a solution for
touchscreen like interaction using gaze input and gesture based
input as a complement or an alternative to touchscreen interactions
with a computer device having a touchscreen.
[0133] The display 20 may hence be any type of known computer
screen or monitor, as well as combinations of two or more separate
displays. For example, the display 20 may constitute a regular
computer screen, a stereoscopic screen, a heads-up display (HUD) in
a vehicle, or at least one head-mounted display (HMD).
[0134] The computer 30 may, for example, be any one from the group
of a personal computer, computer workstation, mainframe computer, a
processor in a vehicle, or a handheld device such as a cell phone,
portable music player (such as e.g. an iPod), laptop computers,
computer games, electronic books and similar other devices. The
present invention may also be implemented in "intelligent
environment" where, for example, objects presented on multiple
displays can be selected and activated.
[0135] In order to produce the gaze tracking signal D.sub.EYE, a
gaze tracker unit 40 is included in the display 20, or is
associated with the display 20. A suitable gaze tracker is
described in the U.S. Pat. No. 7,572,008, titled "Method and
Installation for detecting and following an eye and the gaze
direction thereof", by the same applicant, which hereby is
incorporated in its entirety.
[0136] The software program or software implemented instructions
associated with the gaze tracking module 40 may be included within
the gaze tracking module 40. The specific example shown in FIGS. 2,
3 and 20 illustrates the associated software implemented in a gaze
tracking module, which may be included solely in the computer 30,
in the gaze tracking module 40, or in a combination of the two,
depending on the particular application.
[0137] The computer system 10 comprises a computer device 30, a
gaze tracking module 40, a display 20, a control module 36, 36' and
user input means 50, 50' as shown in FIGS. 2, 3 and 20. The
computer device 30 comprises several other components in addition
to those illustrated in FIGS. 2 and 20 but these components are
omitted from FIGS. 2, 3 and 20 in illustrative purposes.
[0138] The user input means 50, 50' comprises elements that are
sensitive to pressure, physical contact, gestures, or other manual
control by the user, for example, a touchpad 51. Further, the input
device means 50, 50' may also include a computer keyboard, a mouse,
a "track ball", or any other device, for example, an IR-sensor,
voice activated input means, or a detection device of body gestures
or proximity based input can be used. However, in the specific
embodiments shown in FIGS. 2, 3 and 20, a touchpad 51 is included
in the user input device 50, 50'.
[0139] An input module 32, which may be a software module included
solely in a control module 36' or in the user input means 50 or as
a module separate from the control module and the input means 50',
is configured to receive signals from the touchpad 51 reflecting a
user's gestures. Further, the input module 32 is also adapted to
interpret the received signals and provide, based on the
interpreted signals, gesture based control commands, for example, a
tap command to activate an object, a swipe command or a slide
command.
[0140] If the input module 32 is included in the input means 50,
gesture based control commands are provided to the control module
36, see FIG. 2. In embodiments of the present invention, the
control module 36' includes the input module 32 based on gesture
data from the user input means 50', see FIG. 3.
[0141] The control module 36, 36' is further configured to acquire
gaze data signals from the gaze tracking module 40. Further, the
control module 36, 36' is configured to determine a gaze point area
120 on the information presentation area 20 where the user's gaze
point is located based on the gaze data signals. The gaze point
area 120 is preferably, as illustrated in FIG. 1, a local area
around a gaze point of the user.
[0142] Moreover, the control module 36, 36' is configured to
execute at least one user action manipulating a view presented on
the graphical information presentation area 20 based on the
determined gaze point area and the at least one user generated
gesture based control command, wherein the user action is executed
with the determined gaze point area as a starting point. The
control module 36, 36' may be integrated in the computer device 30
or may be associated or coupled to the computer device 30.
[0143] Hence, the present invention allows a user to interact with
a computer device 30 in touchscreen like manner, e.g. manipulate
objects presented on the information presentation area 20, using
gaze and gestures, e.g. by moving at least one finger on a touchpad
51.
[0144] Preferably, when the user touches the touchpad 51, the
location of the initial gaze point is indicated by a visual
feedback, such as a crosshairs or similar sign. This initial
location can be adjusted by moving the finger on the touchpad 51.
Thereafter, the user can, in a touchscreen like manner, interact
with the information presentation area 20 using different gestures
and the gaze. In the embodiment including a touchpad, the gestures
are finger movements relative the touchpad 51 and each gesture is
associated with or corresponds to particular gesture based user
command resulting in a user action.
[0145] Below, a non-exhaustive number of examples of user actions
that can be executed using a combination of gestures and gaze will
be discussed with regard to FIG. 4-10: [0146] By gazing, for
example, at an object presented on the information presentation
area 20 and by in connection to this, touching the touchpad or
pressing down and holding a finger 60 (see FIG. 4) on the touchpad
51 during a period of y ms, that object is highlighted. If the
finger 60 is held down during a second period of z ms, an
information box may be displayed presenting information regarding
that object. In FIG. 4, this gesture is illustrated in relation to
a touchpad 51. [0147] By gazing, for example, at an object
presented on the information presentation area 20 and by in
connection to this tapping on the touchpad 51 using a finger 71, a
primary action can be initiated. For example, an application can be
opened and started by gazing at an icon representing the
application and tapping on the touchpad 51 using a finger. In FIG.
5, this gesture is illustrated in relation to a touchpad 51.
According to a further embodiment of the present invention after
enacting a primary action by tapping on the touchpad 51 a second
tap may be performed which fine adjusts the effect of the first
tap. By way of example, consider when a user is utilizing a video
game whereby the user gazes at a target and taps to shoot the
target. The user may observe that the first tap caused the shot to
pass to the right of the target, the user may then tap slightly to
the left of their original tap in the subsequent tap on the
touchpad 51 while maintaining their gaze. The system will then
provide the next shot slightly to the left of the previous shot.
This is also of use when selecting an item for opening or
highlighting, such as opening an application. The user may observe
that their original tap has caused an action to be performed not at
their desired target. This may be due to an inaccuracy in the gaze
tracking or an offset caused by the gaze tracker. A subsequent tap
on the touchpad 51 may then be used to adjust the gaze location and
thus the subject of any action caused by tapping the touchpad.
[0148] The user may slide or drag the view presented by the
information presentation area 20 by gazing somewhere on the
information presentation area 20 and by, in connection to this,
sliding his or her finger 81 over the touchpad 51. A similar action
to slide an object over the information presentation area 20 can be
achieved by gazing at the object and by, in connection to this,
sliding the finger 81 over the touchpad 51. This gesture is
illustrated in FIG. 6 in relation to the touchpad 51. Of course,
this gesture can be executed by means of more than one finger, for
example, by using two fingers. [0149] According to an embodiment of
the present invention, by gazing, for example, at an object
presented on the information presentation area 20 and by, in
connection to this, pressing down and holding at least one finger
or thumb on the touchpad 51 or touchscreen, the object may be
selected. By moving the at least one finger across the touchpad 51
or touchscreen, the object may be moved around or resized within
the information presentation area 20. The new position or size of
the object may be confirmed by removing the at least one finger
from the touchpad 51 or touchscreen, placing an additional finger
on the touchpad 51 or touchscreen, maintaining gaze at a location
on the information presentation area 20 or saying a word or phrase.
[0150] One example of this embodiment is moving windows or other
such information containing objects on the information presentation
area 20. In a preferred embodiment, two, three or four fingers are
placed on a touchpad 51 while gazing at a window or the like. The
window subsequently becomes the active window and moving the
fingers across the touchpad 51 facilitates movement of the window
across the information display area. Releasing contact of all the
fingers with the touchpad 51 will reposition the window within the
information presentation area 20. However releasing contact of less
than all the fingers with the touchpad 51 will allow for a
secondary action to be performed, such as a pinch motion to enable
resizing of the window. Furthermore it is possible to reposition
the window such that it may jump directly to a gaze position upon
release of at least on finger on the touchpad 51. [0151] As a
further improvement, repositioning the window adjacent an edge of
the information display area will cause the window to link with the
edge of the information display area, otherwise sometimes referred
to as "docking." For example, repositioning the window such that it
is adjacent the top edge of the information display area will cause
the window to align its top most edge with the top most edge of the
information display area and may cause the window to occupy
substantially the entirety of the information display area. [0152]
As a further improvement, repositioning the window in a
predetermined area of the information display such as the edge may
perform an action associated with the window, for example closing
the window, minimizing the window, maximizing the window and the
like. In some embodiments, it is preferable to provide a visual
indication on the information presentation area 20 to allow a user
to easily see where to position the window so as to perform a
desired act. The act may be an act associated with the window such
as a close, maximize or minimize act or it may be associated with a
program running inside the window such as a file open, file close,
new document or the like action. [0153] Furthermore, once the user
has selected a window in a manner as previously described, areas at
which a user may gaze to perform an action such as close the
window, minimize the window or maximize the window may be
highlighted on the information presentation area 20. The user may
select an area by dwelling their gaze within the area, voice,
movement on a touchpad 51, enacting contact with a touchpad 51 and
releasing contact with a touchpad 51. [0154] In a further
improvement, a user need not reposition the window by moving their
at least on finger across the touchpad 51 or touchscreen; the user
may reposition the window by gazing at the desired position and
maintaining their gaze for a predetermined period, i.e., a dwell,
or by saying a predetermined word or phrase such as "move here." It
is further desirable to provide a visual indicator as to where the
window will be repositioned, if the user completes a movement or
resize action. [0155] As a further improvement, when the user
places at least one finger on the touchpad 51 the device may
determine that the user's gaze is adjacent an edge of a window or
other such information containing object. Instead of enacting a
mode whereby the window is moved by the user moving their at least
one finger on the touchpad 51, the device may enact one of the
following modes: [0156] If the gaze is located adjacent a vertical
edge of the window, a mode which allows resizing of the horizontal
size of the window by moving the at least one finger on the
touchpad 51 is enacted. [0157] If the gaze is located adjacent a
horizontal edge of the window, a mode which allows resizing of the
vertical size of the window by moving the at least one finger on
the touchpad 51 is enacted. [0158] If the gaze is located adjacent
the bottom right hand corner of the window, a mode which allows
resizing of the window in both the horizontal and vertical
directions by moving the at least one finger on the touchpad 51 is
enacted. [0159] As a further embodiment, where a touchpad 51 or
touchscreen is capable of detecting pressure, contact with the
touchpad 51 or touchscreen that is sufficiently hard will cause a
window to become active and able to be moved. When the contact with
the touchpad 51 or touchscreen is decreased in pressure, the
location of the window will be fixed. [0160] According to an
embodiment of the present invention, placing at least one finger on
a touchpad 51 and moving in a set direction for example upwards or
downwards a list of programs currently running on a device to be
displayed on an information display area, the list comprising
summaries or representations of each program. Moving the at least
one finger across the touchpad 51 will cycle a selector through the
list of currently running programs, alternatively a summary may be
selected by positioning a user's gaze over it. Releasing contact
with the touchpad 51 will cause the program associated with the
last selected summary to be made active and displayed on the
information display area. Moving the at least one finger in an
opposite direction to that which caused the list of programs to be
displayed will cause the list to cease being displayed. The display
and cessation of display may occur gradually to provide a fading in
or out effect. [0161] The user may select an object for further
actions by gazing at the object and by, in connection to this,
swiping his or her finger 91 on the touchpad 51 in a specific
direction. This gesture is illustrated in FIG. 7 in relation to the
touchpad 51. Of course, this gesture can be executed by means of
more than one finger, for example, by using two fingers. [0162] By
gazing at an object or object part presented on the information
presentation area 20 and by, in connection to this, pinching with
two of his or hers finger 101 and 102, it is possible to zoom out
that object or object part. This gesture is illustrated in FIG. 8
in relation to the touchpad 51. Similarly, by gazing at an object
or object part presented on the information presentation area 20
and by, in connection to this, moving the fingers 101 and 102
apart, it is possible to expand or zoom in that object or object
part. [0163] By gazing at an object or object part presented on the
information presentation area 20 and by, in connection to this,
rotating with two of his or hers finger 111 and 112, it is possible
to rotate that object or object part. This gesture is illustrated
in FIG. 9 in relation to the touchpad 51. Other manipulations of an
object or object part of interest are also possible in addition to
rotation, for example after gazing at an object a user may contact
the touchpad 51 and perform a movement to hold and drag an object.
For example this may be used in a video game such as a card game or
a game where an object is the subject of a slingshot. The object
such as the card may be gazed at and then a press on the touchpad
51 would cause that card to be selected or held, moving on the
touchpad would cause the card to be moved until the touchpad was
released. [0164] By gazing at an edge or frame part of the
information presentation area 20 or at an area in proximity to the
edge or frame and, in connection to this, sliding his or her finger
or fingers 124 on the touchpad 51 in a direction which if performed
at the point of gaze would have been from the edge towards a center
of the information presentation area a menu may come in from the
edge. [0165] By gazing at a slider control, for example a volume
control, the finger can be moved up/down (or left/right for a
horizontal control) to adjust the value of the slider control. With
appropriate input means this gesture can be detected on a touchpad,
on a touch screen or in air without physically touching the input
means. [0166] By gazing at a checkbox control while doing a
"check-gesture" (such as a "V") on the touchpad, the checkbox can
be checked or unchecked. With appropriate input means this gesture
can be detected on a touchpad, on a touch screen or in air without
physically touching the input means. [0167] By gazing at an object
or object part where several options are available, for example
"copy" or "rename", the different options can be displayed on
different sides of the object after a preset focusing dwell time
has passed or after appropriate user input has been provided.
Thereafter a gesture is done to choose action. For example, swipe
left to copy and swipe right to rename. With appropriate input
means this gesture can be detected on a touchpad, on a touch screen
or in air without physically touching the input means. [0168] Other
gestures may be performed to interact with an object at a gaze
area, examples of such gestures include: [0169] performing a
swiping gesture to dismiss or move away the object, [0170]
performing a slicing gesture to slice or cut the object, [0171]
performing a rotation gesture to adjust a value associated with the
object, [0172] performing a rotation gesture to perform an `undo`
action associated with the object, [0173] perform a user defined
gesture to open a menu, map or other such information selection
tool, [0174] performing a scratch gesture (e.g. a rapid back and
forth movement) to highlight an object or text or otherwise
interact with the object, [0175] performing a drawing of an object
to cause an act to be performed, for example drawing an X or O in a
game of tic tac toe, or drawing a representation of a magic rune or
the like to cast a spell in a video game.
[0176] By pressing the finger harder on the touchpad, i.e.
increasing the pressure of a finger touching the touchpad, a
sliding mode can be initiated. For example, by gazing at an object,
touching the touchpad, increasing the pressure on the touchpad and
moving the finger or finger over the touchscreen, the object can be
moved or dragged over the information presentation area. When the
user removes the finger from the touchpad 51, the touchscreen like
session is finished.
[0177] The user may thereafter start a new touchscreen like session
by gazing at the information presentation area 20 and placing the
finger on the touchpad 51.
[0178] As mentioned, the gesture and gaze initiated actions
discussed above are only exemplary and there are a large number of
further gestures in combination with gaze point resulting in an
action that are conceivable. With appropriate input means many of
these gestures can be detected on a touchpad, on a predefined area
of a touch screen, in air without physically touching the input
means, or by an input means worn on a finger or a hand of the user.
Below, some further examples are described: [0179] Selection of an
object or object part can be made by gazing at that object or
object part and pressing a finger (e.g. a thumb), fine tuning by
moving the finger and releasing the pressure applied by the finger
to select that object or object part; [0180] Selection of an object
or object part can be made by gazing at that object or object part,
pressing a finger (e.g. a thumb), fine tuning by moving the finger,
using another finger (e.g. the other thumb) to tap for selecting
that object or object part. In addition, a double tap may be used
for a "double click action" and a quick downward movement may be
used for a "right click". [0181] By gazing at a zoomable object or
object part presented on the information presentation area while
moving a finger (e.g. one of the thumbs) in a circular motion, it
is possible to zoom in or out of the said object using the gaze
point as the zoom center point, where a clockwise motion performs a
"zoom in" command and a counterclockwise motion performs a "zoom
out" command or vice versa. [0182] By gazing at a zoomable object
or object part presented on the information presentation area and
in connection to this holding one finger (e.g. one of the thumbs)
still while moving another finger (e.g. the other thumb) upwards
and downwards, it is possible to zoom in or out of the said object
using the gaze point as the zoom center point, where an upwards
motion performs a "zoom in" command and a downwards motion performs
a "zoom out" command or vice versa. [0183] By gazing at a zoomable
object or object part presented on the information presentation
area and while pressing hard on a pressure-sensitive touchpad with
one finger (e.g. one of the thumbs), it is possible to zoom in or
out on the said object using the gaze point as the zoom center
point, where each hard press toggles between different zoom levels.
[0184] By gazing at a zoomable object or object part presented on
the information presentation area while double-tapping on a
touchpad with one finger (e.g. one of the thumbs), it is possible
to zoom in or out of the said object using the gaze point as the
zoom center point, where each double-tap toggles between different
zoom levels. [0185] By gazing at a zoomable object or object part
presented on the information presentation area while sliding two
fingers (e.g. the two thumbs) simultaneously in opposite horizontal
directions, it is possible to zoom that object or object part.
[0186] By gazing at a zoomable object and in connection to this
holding finger (e.g. one thumb) still on the touchscreen while
moving another finger (e.g. the other thumb) in a circular motion,
it is possible to zoom that object or object part. [0187] By gazing
at an object or object part presented on the information
presentation area and in connection to this holding a finger (e.g.
one of the thumbs) still on the touchscreen while sliding another
finger (e.g. the other thumb), it is possible to slide or drag the
view presented by the information presentation area. [0188] By
gazing at an object or object part presented on the information
presentation area and in connection to this holding a finger (e.g.
one of the thumbs) still on the touchscreen while sliding another
finger (e.g. the other thumb), it is possible to slide or drag the
view presented by the information presentation area. [0189] By
gazing at an object or object part presented on the information
presentation area and while tapping or double-tapping with a finger
(e.g. one of the thumbs), an automatic panning function can be
activated so that the presentation area is continuously slided from
one of the edges of the screen towards the center while the gaze
point is near the edge of the information presentation area, until
a second user input is received. [0190] By gazing at an object or
object part presented on the information presentation area and
while tapping or double-tapping with a finger (e.g. one of the
thumbs), the presentation area is instantly slided according to the
gaze point (e.g. the gaze point is used to indicate the center of
where the information presentation area should be slided). [0191]
By gazing at a rotatable object or object part presented on the
information presentation area while sliding two fingers (e.g. the
two thumbs) simultaneously in opposite vertical directions, it is
possible to rotate that object or object part.
[0192] Before the two-finger gesture is performed, one of the
fingers can be used to fine-tune the point of action. For example,
a user feedback symbol like a "virtual finger" can be shown on the
gaze point when the user touches the touchscreen. The first finger
can be used to slide around to adjust the point of action relative
to the original point. When user touches the screen with the second
finger, the point of action is fixed and the second finger is used
for "clicking" on the point of action or for performing two-finger
gestures like the rotate, drag and zoom examples above.
[0193] In embodiments of the present invention, the touchscreen
like session can be maintained despite that the user has removed
the finger or fingers from the touchpad if, for example, a specific
or dedicated button or keyboard key is held down or pressed.
Thereby, it is possible for the user to perform actions requiring
multiple touches on the touchpad. For example, an object can be
moved or dragged across the entire information presentation area by
means of multiple dragging movements on the touchpad.
[0194] With reference now to FIGS. 11a, 11b and 12, further
embodiments of the present invention will be discussed. FIG. 11a
shows a further embodiment of a system with integrated gaze and
manual control according to the present invention. This embodiment
of the system is implemented in a device 100 with a touchscreen 151
such as an iPad or similar device. The user is able to control the
device 100 at least partly based on gaze tracking signals which
describes the user's point of regard x, y on the touchscreen 151
and based on user generated gestures, i.e. a movement of at least
one body part of the user can be detected, generating gesture based
control commands via user input means 150 including the touchscreen
151.
[0195] The present invention provides a solution enabling a user of
a device 100 with a touchscreen 151 to interact with a graphical
user interfaces using gaze as direct input and gesture based user
commands as relative input. Thereby, it is possible, for example,
to hold the device 100 with both hands and interact with a
graphical user interface 180 presented on the touchscreen with gaze
and the thumbs 161 and 162 as shown in FIG. 11a.
[0196] In an alternative embodiment, one or more touchpads 168 can
be arranged on the backside of the device 100', i.e. on the side of
the device on which the user normally do not look at during use.
This embodiment is illustrated in FIG. 11b. Thereby, a user is
allowed to control the device at least partly based on gaze
tracking signals which describes the user's point of regard x, y on
the information presentation area and based on user generated
gestures, i.e. a movement of at least one finger on the one or more
touchpads 1168 on the backside of the device 100', generating
gesture based control commands interpreted by the control module.
In order to produce the gaze tracking signal, a gaze tracking
module 140 is included in the device 100, 100'. A suitable gaze
tracker is described in the U.S. Pat. No. 7,572,008, titled "Method
and Installation for detecting and following an eye and the gaze
direction thereof", by the same applicant, which hereby is
incorporated in its entirety.
[0197] The software program or software implemented instructions
associated with the gaze tracking module 140 may be included within
the gaze tracking module 140.
[0198] The device 100 comprises a gaze tracking module 140, user
input means 150 including the touchscreen 151 and an input module
132, and a control module 136 as shown in FIG. 12. The device 100
comprises several other components in addition to those illustrated
in FIG. 12 but these components are omitted from FIG. 12 in
illustrative purposes.
[0199] The input module 132, which may be a software module
included solely in a control module or in the user input means 150,
is configured to receive signals from the touchscreen 151
reflecting a user's gestures. Further, the input module 132 is also
adapted to interpret the received signals and provide, based on the
interpreted signals, gesture based control commands, for example, a
tap command to activate an object, a swipe command or a slide
command.
[0200] The control module 136 is configured to acquire gaze data
signals from the gaze tracking module 140 and gesture based control
commands from the input module 132. Further, the control module 136
is configured to determine a gaze point area 180 on the information
presentation area, i.e. the touchscreen 151, where the user's gaze
point is located based on the gaze data signals. The gaze point
area 180 is preferably, as illustrated in FIG. 1, a local area
around a gaze point of the user.
[0201] Moreover, the control module 136 is configured to execute at
least one user action manipulating a view presented on the
touchscreen 151 based on the determined gaze point area and the at
least one user generated gesture based control command, wherein the
user action is executed with the determined gaze point area as a
starting point. All user actions described in the context of this
application may also be executed with this embodiment of the
present invention.
[0202] In a possible further embodiment, when the user touches the
touchscreen 151, the location of the initial gaze point is
indicated by a visual feedback, such as a crosshairs or similar
sign. This initial location can be adjusted by moving the finger on
the touchscreen 151, for example, using a thumb 161 or 162.
Thereafter, the user can interact with the touchscreen 151 using
different gestures and the gaze, where the gaze is the direct
indicator of the user's interest and the gestures are relative to
the touchscreen 151. In the embodiment including a touchscreen, the
gestures are finger movements relative the touchscreen 151 and each
gesture is associated with or corresponds to particular gesture
based user command resulting in a user action.
[0203] With reference now to FIGS. 13a, 13b and 13c, control
modules for generating gesture based commands during user
interaction with an information presentation area 201, for example,
associated with a WTRU (described below with reference to FIG. 14),
or a computer device or handheld portable device (described below
with reference to FIG. 15a or 15b), or in a vehicle (described
below with reference to FIG. 21), or in a wearable head mounted
display (described below with reference to FIG. 22) will be
described. Parts or modules described above will not be described
in detail again in connection to this embodiment.
[0204] According to an embodiment of the present invention shown in
FIG. 13a, the control module 200 is configured to acquire user
input from input means 205, for example, included in a device in
which the control module may be arranged in, adapted to detect user
generated gestures. For this purpose, the control module 200 may
include an input module 232 comprising a data acquisition module
210 configured to translate the gesture data from the input means
205 into an input signal. The input means 205 may include elements
that are sensitive to pressure, physical contact, gestures, or
other manual control by the user, for example, a touchpad. Further,
the input means 205 may also include a computer keyboard, a mouse,
a "track ball", or any other device, for example, an IR-sensor,
voice activated input means, or a detection device of body gestures
or proximity based input can be used.
[0205] Further, the input module 232 is configured to determine at
least one user generated gesture based control command based on the
input signal. For this purpose, the input module 232 further
comprises a gesture determining module 220 communicating with the
data acquisition module 210. The gesture determining module 220 may
also communicate with the gaze data analyzing module 240. The
gesture determining module 220 may be configured to check whether
the input signal corresponds to a predefined or predetermined
relative gesture and optionally use gaze input signals to interpret
the input signal. For example, the control module 200 may comprise
a gesture storage unit (not shown) storing a library or list of
predefined gestures, each predefined gesture corresponding to a
specific input signal. Thus, the gesture determining module 220 is
adapted to interpret the received signals and provide, based on the
interpreted signals, gesture based control commands, for example, a
tap command to activate an object, a swipe command or a slide
command.
[0206] A gaze data analyzing module 240 is configured to determine
a gaze point area on the information presentation area 201
including the user's gaze point based on at least the gaze data
signals from the gaze tracking module 235. The information
presentation area 201 may be a display of any type of known
computer screen or monitor, as well as combinations of two or more
separate displays, which will depend on the specific device or
system in which the control module is implemented in. For example,
the display 201 may constitute a regular computer screen, a
stereoscopic screen, a heads-up display (HUD) in a vehicle, or at
least one head-mounted display (HMD). Then, a processing module 250
may be configured to execute at least one user action manipulating
a view presented on the information presentation area 201 based on
the determined gaze point area and at least one user generated
gesture based control command, wherein the user action is executed
with the determined gaze point area as a starting point. Hence, the
user is able to control a device or system at least partly based on
an eye-tracking signal which described the user's point of regard
x, y on the information presentation area or display 201 and based
on user generated gestures, i.e. a movement of at least one body
part of the user can be detected, generating gesture based control
commands via user input means 205 such as a touchpad.
[0207] According to another embodiment a control module according
to the present invention shown in FIG. 13b, the control module 260
is configured to acquire gesture based control commands from an
input module 232'. The input module 232' may comprise a gesture
determining module and a data acquisition module as described above
with reference to FIG. 13a. A gaze data analyzing module 240 is
configured to determine a gaze point area on the information
presentation area 201 including the user's gaze point based on at
least the gaze data signals received from the gaze tracking module
235. The information presentation area 201 may be a display of any
type of known computer screen or monitor, as well as combinations
of two or more separate displays, which will depend on the specific
device or system in which the control module is implemented in. For
example, the display 201 may constitute a regular computer screen,
a stereoscopic screen, a heads-up display (HUD) in a vehicle, or at
least one head-mounted display (HMD). A processing module 250 may
be configured to execute at least one user action manipulating a
view presented on the information presentation area 201 based on
the determined gaze point area and at least one user generated
gesture based control command, wherein the user action is executed
with the determined gaze point area as a starting point. Hence, the
user is able to control a device or system at least partly based on
an eye-tracking signal which described the user's point of regard
x, y on the information presentation area or display 201 and based
on user generated gestures, i.e. a movement of at least one body
part of the user can be detected, generating gesture based control
commands via user input means 205 such as a touchpad.
[0208] With reference to FIG. 13c, a further embodiment of a
control module according to the present invention will be
discussed. The input module 232'' is distributed such that the data
acquisition module 210 is provided outside the control module 280
and the gesture determining module 220 is provided in the control
module 280. A gaze data analyzing module 240 is configured to
determine a gaze point area on the information presentation area
201 including the user's gaze point based on at least the gaze data
signals received from the gaze tracking module 235. The information
presentation area 201 may be a display of any type of known
computer screen or monitor, as well as combinations of two or more
separate displays, which will depend on the specific device or
system in which the control module is implemented in. For example,
the display 201 may constitute a regular computer screen, a
stereoscopic screen, a heads-up display (HUD) in a vehicle, or at
least one head-mounted display (HMD). A processing module 250 may
be configured to execute at least one user action manipulating a
view presented on the information presentation area 201 based on
the determined gaze point area and at least one user generated
gesture based control command, wherein the user action is executed
with the determined gaze point area as a starting point. Hence, the
user is able to control a device or system at least partly based on
an eye-tracking signal which described the user's point of regard
x, y on the information presentation area or display 201 and based
on user generated gestures, i.e. a movement of at least one body
part of the user can be detected, generating gesture based control
commands via user input means 205 such as a touchpad.
[0209] With reference to FIG. 14, a wireless transmit/receive unit
(WTRU) such as a cellular telephone or a smartphone, in accordance
with the present invention will be described. Parts or modules
described above will not be described in detail again. Further,
only parts or modules related to the present invention will be
described below. Accordingly, the WTRU includes a large number of
additional parts, units and modules that are not described herein
such as antennas and transmit/receive units. The wireless
transmit/receive unit (WTRU) 300 is associated with an information
presentation area 301 and further comprises input means 305,
including e.g. an input module as has been described above, adapted
to detect user generated gestures and a gaze tracking module 325
adapted to detect gaze data of a viewer of the information
presentation area 301. The WTRU further comprises a control module
200, 260 or 280 as described above with reference to FIGS. 13a, 13b
and 13c. The user is able to control the WTRU at least partly based
on an eye-tracking signal which describes the user's point of
regard x, y on the information presentation area or display 301 and
based on user generated gestures, i.e. a movement of at least one
body part of the user can be detected, generating gesture based
control commands via user input means 305 such as a touchpad. All
user actions described in the context of this application may also
be executed with this embodiment of the present invention.
[0210] With reference to FIGS. 15a and 15b, a computer device or
handheld portable device in accordance with the present invention
will be described. Parts or modules described above will not be
described in detail again. Further, only parts or modules related
to the present invention will be described below. Accordingly, the
device includes a large number of additional parts, units and
modules that are not described herein such as memory units (e.g.
RAM/ROM), or processing units. The computer device or handheld
portable device 400 may, for example, be any one from the group of
a personal computer, computer workstation, mainframe computer, a
processor or device in a vehicle, or a handheld device such as a
cell phone, smartphone or similar device, portable music player
(such as e.g. an iPod), laptop computers, computer games,
electronic books, an iPAD or similar device, a Tablet, a
Phoblet/Phablet.
[0211] The computer device or handheld device 400a is connectable
to an information presentation area 401a (e.g. an external display
or a heads-up display (HUD), or at least one head-mounted display
(HMD)), as shown in FIG. 15a, or the computer device or handheld
device 400b includes an information presentation area 401b, as
shown in FIG. 15b, such as a regular computer screen, a
stereoscopic screen, a heads-up display (HUD), or at least one
head-mounted display (HMD). Furthermore, the computer device or
handheld device 400a, 400b comprises input means 405 adapted to
detect user generated gestures and a gaze tracking module 435
adapted to detect gaze data of a viewer of the information
presentation area 401. Moreover, the computer device or handheld
device 400a, 400b comprises a control module 200, 260, or 280 as
described above with reference to FIG. 13a, 13b or 13c. The user is
able to control the computer device or handheld device 400a, 400b
at least partly based on an eye-tracking signal which described the
user's point of regard x, y on the information presentation area or
display 401 and based on user generated gestures, i.e. a movement
of at least one body part of the user can be detected, generating
gesture based control commands via user input means 405 such as a
touchpad. All user actions described in the context of this
application may also be executed with this embodiment of the
present invention.
[0212] With reference now to FIG. 16-19, example embodiments of
methods according to the present invention will be described. The
method embodiments described in connection with FIGS. 16-19 are
implemented in an environment where certain steps are performed in
a device; e.g. a WTRU described above with reference to FIG. 14, or
a computer device or handheld device described above with reference
to FIG. 15a or 15b and certain steps are performed in a control
module, e.g. a control module as described above with reference to
FIGS. 13a, 13b and 13c. As the skilled person realizes, the methods
described herein can also be implemented in other environments, as.
for example, in a system as described above with reference to FIGS.
2, 3 and 20 or in implementations illustrated in FIGS. 21-23.
Similar or like steps performed in the different embodiments will
be denoted with the same reference numeral hereinafter.
[0213] With reference first to FIG. 16, the device is waiting for
user input in step S500. In step S510, the user touches a touch
sensitive area on the device (e.g. input means as described above)
with one or more fingers of each hand. This step is not a part of
the method according to embodiments of the invention. There are a
large number of conceivable gestures that the user can use to
control actions of the device, and a non-exhaustive number of such
gestures have been described above. At step S520, the gesture data,
i.e. the user input, is translated into an input signal. At step
S530, it is checked whether the input signal corresponds to a
predefined or predetermined relative gesture. If not, the procedure
returns back to step S500. On the other hand, if yes (i.e. the
input signal corresponds to a predefined gesture), a gesture based
control command is Generated at step S570. At step S540, the user
looks at a screen or an information presentation area and at step
S550 the user's gaze is detected at the information presentation
area. The step S540 is not a part of the method according to
embodiments of the present invention. In step S560, a gaze point
area including a user's point of gaze on the screen or information
presentation area At step S580, an action corresponding to the
relative gesture at the user's point of gaze is performed based on
the gesture based control command and the determined gaze point at
the information presentation area.
[0214] With reference to FIG. 17, the device is waiting for user
input in step S500. In step S590, the user makes a gesture with one
or more fingers and/or at least one hand in front of the
information presentation area (which gesture is interpreted by
input means as described above). The step S590 is not a part of the
method according to embodiments of the present invention. There are
a large number of conceivable gestures that the user can use to
control actions of the device, and a non-exhaustive number of such
gestures have been described above. At step S520, the gesture data,
i.e. the user input, is translated into an input signal. At step
S530, it is checked whether the input signal corresponds to a
predefined or predetermined relative gesture. If not, the procedure
returns back to step S500. On the other hand, if yes (i.e. the
input signal corresponds to a predefined gesture), a gesture based
control command is generated at step S570. At step S540, the user
looks at a screen or an information presentation area and at step
S550 the user's gaze is detected at the information presentation
area. As mentioned above, the step S540 is not a part of the method
according to embodiments of the present invention. In step S560, a
gaze point area including a user's point of gaze on the screen or
information presentation area. At step S580, an action
corresponding to the relative gesture at the user's point of gaze
is performed based on the gesture based control command and the
determined gaze point at the information presentation area.
[0215] With reference to FIG. 18, the device is waiting for user
input in step S500. In step S592, the user generates input by
touching touchpad or predefined area of touch-screen. The step S592
is not a part of the method according to embodiments of the present
invention. There are a large number of conceivable gestures that
the user can use to control actions of the device, and a
non-exhaustive number of such gestures have been described above.
At step S520, the gesture data, i.e. the user input, is translated
into an input signal. At step S530, it is checked whether the input
signal corresponds to a predefined or predetermined relative
gesture. If not, the procedure returns back to step S500. On the
other hand, if yes (i.e. the input signal corresponds to a
predefined gesture), a gesture based control command is generated
at step S570. At step S540, the user looks at a screen or an
information presentation area and at step S550 the user's gaze is
detected at the information presentation area. The step S540 is not
a part of the method according to embodiments of the present
invention. In step S560, a gaze point area including a user's point
of gaze on the screen or information presentation area is
determined. At step S580, an action corresponding to the relative
gesture at the user's point of gaze is performed based on the
gesture based control command and the determined gaze point at the
information presentation area.
[0216] With reference to FIG. 19, the device is waiting for user
input in step S500. In step S594, the user generates input by
making a gesture with one or more of his or hers fingers and/or at
least one hand. The step S594 is not a part of the method according
to embodiments of the present invention. There are a large number
of conceivable gestures that the user can use to control actions of
the device, and a non-exhaustive number of such gestures have been
described above. At step S520, the gesture data, i.e. the user
input, is translated into an input signal. At step S530, it is
checked whether the input signal corresponds to a predefined or
predetermined relative gesture. If not, the procedure returns back
to step S500. On the other hand, if yes (i.e. the input signal
corresponds to a predefined gesture), a gesture based control
command is generated at step S570. At step S540, the user looks at
a screen or an information presentation area and at step S550 the
user's gaze is detected at the information presentation area. The
step S540 is not a part of the method according to embodiments of
the present invention. In step S560, a gaze point area including a
user's point of gaze on the screen or information presentation area
is determined. At step S580, an action corresponding to the
relative gesture at the user's point of gaze is performed based on
the gesture based control command and the determined gaze point at
the information presentation area.
[0217] With reference to FIG. 21, a further implementation of the
present invention will be discussed. A gaze tracking module (not
shown) and a user input means 900 are implemented in a vehicle (not
shown). The information presentation area (not shown) may be a
heads-up display or an infotainment screen. The input means 900 may
be one or two separate touch pads on the backside (for use with the
index finger/s) or on the front side (for use with the thumb/s) of
the steering wheel 910 of the vehicle. A control module 950 is
arranged in a processing unit configured to be inserted into a
vehicle or a central processing unit of the vehicle. Preferably,
the control module is a control module as described with reference
to FIGS. 13a-13c.
[0218] With reference to FIG. 22, another implementation of the
present invention will be discussed. A gaze tracking module (not
shown) and an information presentation area (not shown) are
implemented in a wearable head mounted display 1000 that may be
designed to look as a pair of glasses. One such solution is
described in U.S. Pat. No. 8,235,529. The user input means 1010 may
include a gyro and be adapted to be worn by the user 1020 on a
wrist, hand or at least one finger. For example, the input means
1010 may be a ring with a wireless connection to the glasses and a
gyro that detects small movements of the finger where the ring is
worn. The detected movements representing gesture data may then
wirelessly be communicated to the glasses where gaze is detected
and gesture based control commands based on the gesture data from
the input means is used to identify and execute user action.
Preferably, a control module as described with reference to FIG.
13a-13c is used with this implementation.
[0219] With reference to FIG. 23, an implementation of the present
invention will be discussed. In this implementation, the user 1120
is able to control a computer device 1100 at least partly based on
an eye-tracking signal which describes the user's point of regard
x, y on an information presentation area 1140 and based on user
generated gestures, i.e. a movement of at least one body part of
the user can be detected, generating gesture based control commands
via user input means 1150. In this embodiment, the user 1120 can
generate the gesture based control commands by performing gestures
above or relative the keyboard of the computer device 1100. The
input means 1140 detects the gestures, for example, using an
optical measurement technique or capacitive measurement technique.
Preferably, a control module as described with reference to FIG.
13a-13c is used with this implementation and may be arranged in the
computer device 1100. The computer device 1100 may, for example, be
any one from the group of a personal computer, computer
workstation, mainframe computer, or a handheld device such as a
cell phone, portable music player (such as e.g. an iPod), laptop
computers, computer games, electronic books and similar other
devices. The present invention may also be implemented in an
"intelligent environment" where, for example, objects presented on
multiple displays can be selected and activated. In order to
produce the gaze tracking signals, a gaze tracker unit (not shown)
is included in the computer device 1100, or is associated with the
information presentation area 1140. A suitable gaze tracker is
described in the U.S. Pat. No. 7,572,008, titled "Method and
Installation for detecting and following an eye and the gaze
direction thereof", by the same applicant, which hereby is
incorporated in its entirety. While this specification contains a
number of specific embodiments, these should not be construed as
limitation to the scope of the present invention or of what may be
claimed, but rather as descriptions of features specific to
exemplary implementations of the present invention. Certain
features that are described in this specification in the context of
separate implementations can also be implemented in combinations in
a single implementation. Conversely, various features that are
described in the context of a single implementation can also be
implemented in multiple implementations separately or in any
suitable sub-combination. Moreover, although feature may be
described above as acting in certain combinations or even initially
claimed as such, one or more features from a claimed combination
can in some cases be excised from the combination, and the claimed
combination may be directed to a sub-combination or variation of a
sub-combination.
[0220] According to an embodiment of the present invention, a
touchpad may amplify an action being performed by a character or
object on the information presentation area. For example, while
gazing at an area on the information presentation area to perform
an act, contact with the touchpad may amplify the act. One example
of this is the case of a video game or the like where a character
may lean around a corner, and the lean may be initiated by an input
such as a mouse action, key press, head movement or gaze. While the
lean is being performed, performing a gesture on the touchpad may
amplify or extend the lean. In this manner the user may initiate an
act based on their gaze and then utilize a gesture on a touchpad to
amplify or extend that act.
[0221] According to a further embodiment of the present invention,
a touchpad may be mapped to a view of the information presentation
area. The touchpad may then be used to perform an action at the
mapped portion of the information presentation area. The mapping
may be 1:1 or in some instances may be enhanced so as to provide a
more amplified version of the action; for example, the mapping may
be 2:1 or 3:1. By way of example, this embodiment may be used in
the following manner: a user gazes at an area of the information
presentation area; a magnification around the gaze area may then be
displayed; this may be automatically displayed or displayed upon
the user performing some act. The magnified area may then be mapped
by the device to the touchpad, such that pressing a location on the
touchpad performs an action at the corresponding location in the
magnified area. For example, in the case of a video game, a
magnified area of a map or the like may be shown and tapping on the
touchpad may cause an action to happen at that location on the map.
For example, pressing a location on the touchpad approximately 75%
across and 25% down the touchpad would cause the action associated
with pressing to be performed 75% across and 25% down the magnified
area.
[0222] According to a further embodiment of the present invention,
the sensitivity of a pointer is adjusted based on a detected gaze
position. By way of example, if a gaze area is detected adjacent an
edge of a screen, then the speed of a pointer will proceed at a
faster pace than if the gaze area is detected in the center of the
screen. The speed or sensitivity of a pointer may also be effected
by the object at which a user is gazing; for example, if a user is
gazing at sand in a game, then the pointer may move at a slower
speed.
[0223] In a further embodiment of the present invention, a touch
sensitive surface is provided remote from an information display
area. The touch sensitive surface may be provided on a controller
or the like, such as a controller provided by Sony for the
PlayStation.RTM. 4 gaming console. Furthermore, a touch sensitive
tablet or the like may be provided remote from the information
display area. In this manner a user may gaze at the information
display area and have their gaze detected of any previously or
otherwise described touchscreen or touchpad enabled
interactions.
[0224] Similarly, while operations are depicted in the drawings in
a particular order, this should not be understood as require such
operations be performed in the particular order shown or in
sequential order, or that all illustrated operation be performed to
achieve desirable results. In certain circumstances, multitasking
and parallel processing may be advantageous. Moreover, the
separation of various system components in the implementation
described above should not be understood as requiring such
separation in all implementations, and it should be understood that
the described program components and systems can generally be
integrated together in a single software product or packaged into
multiple software products.
* * * * *