U.S. patent application number 11/172239 was filed with the patent office on 2007-01-04 for methods and apparatuses for monitoring attention of a user during a collaboration session.
Invention is credited to Jitendra Chawla, David Knight, Manish Kumar, Edward Wong.
Application Number | 20070005752 11/172239 |
Document ID | / |
Family ID | 37591075 |
Filed Date | 2007-01-04 |
United States Patent
Application |
20070005752 |
Kind Code |
A1 |
Chawla; Jitendra ; et
al. |
January 4, 2007 |
Methods and apparatuses for monitoring attention of a user during a
collaboration session
Abstract
In one embodiment, the methods and apparatuses detect devices
participating in a collaboration session; monitor the devices for
an attribute corresponding to an attention level of each of the
devices; and determine whether each device is active based on the
attention level.
Inventors: |
Chawla; Jitendra; (Union
City, CA) ; Knight; David; (Belmont, CA) ;
Wong; Edward; (Fremont, CA) ; Kumar; Manish;
(Sunnyvale, CA) |
Correspondence
Address: |
BLAKELY SOKOLOFF TAYLOR & ZAFMAN
12400 WILSHIRE BOULEVARD
SEVENTH FLOOR
LOS ANGELES
CA
90025-1030
US
|
Family ID: |
37591075 |
Appl. No.: |
11/172239 |
Filed: |
June 29, 2005 |
Current U.S.
Class: |
709/224 |
Current CPC
Class: |
H04N 7/15 20130101; H04L
67/22 20130101; H04L 12/1831 20130101; G06Q 10/103 20130101; H04N
21/44218 20130101 |
Class at
Publication: |
709/224 |
International
Class: |
G06F 15/173 20060101
G06F015/173 |
Claims
1. A method comprising: detecting devices participating in a
collaboration session; monitoring the devices for an attribute
corresponding to an attention level of each of the devices; and
determining whether each device is active based on the attention
level.
2. The method according to claim 1 further comprising storing the
attention level and corresponding attribute for each device with a
record associated with each device.
3. The method according to claim 1 further comprising presenting
content during the collaboration session.
4. The method according to claim 1 wherein the attribute is a
location of content on each device during the collaboration session
wherein the content is utilized during the collaboration
session.
5. The method according to claim 1 wherein the attribute is a
frequency of interfacing with each device during the collaboration
session.
6. The method according to claim 1 wherein the attribute is a
duration of interfacing with each device during the collaboration
session.
7. The method according to claim 1 further comprising comparing the
attention level with a predetermined threshold to determine whether
each device is active.
8. The method according to claim 1 further comprising scoring a
plurality of attributes.
9. The method according to claim 8 further comprising generating
the attention level from scoring the plurality of attributes.
10. A system, comprising: a storage module configured to store a
record containing an attention attribute associated with a device
participating in a collaboration session; a content detection
module configured to detect a location of content utilized by the
device wherein the content is utilized during the collaboration
session and the location of the content contributes to the
attention attribute; and a scoring module configured to generate an
attention level for the device based on the attention
attribute.
11. The system according to claim 10 wherein storage module is
configured to store the attention level for the device.
12. The system according to claim 10 further comprising an
interface detection module configured to detect an input from the
device during the collaboration session and contributes to the
attention attribute.
13. The system according to claim 12 wherein the input from the
device is initiated from a pointing device.
14. The system according to claim 12 wherein the input from the
device is an audio signal.
15. The system according to claim 10 further comprising a duration
module configured to detect a duration of the content in a
particular location.
16. The system according to claim 10 wherein the location of the
content refers to a particular location on a display of the
device.
17. The system according to claim 10 wherein the location of the
content refers to a priority order of the content relative to other
content displayed on the device.
18. A system comprising: means for detecting devices participating
in a collaboration session; means for monitoring the devices for an
attribute corresponding to an attention level of each of the
devices; and means for determining whether each device is active
based on the attention level.
19. The system according to claim 18 wherein the means for
determining includes a scoring module configured to weigh a
plurality of attributes corresponding to the device.
20. The system according to claim 18 wherein the means for
determining includes a comparing module configured to compare the
attention level with a predetermined threshold.
Description
RELATED APPLICATION
[0001] The present invention is related to, and claims the benefit
of U.S. patent application Ser. No. 10/946,562, filed on Sep. 20,
2004, entitled, "Attention Indicator For Use In A Web Conference,"
by Jitendra Chawla.
FIELD OF INVENTION
[0002] The present invention relates generally to monitoring
attention of a user and, more particularly, to monitoring attention
of a user during a collaboration session.
BACKGROUND
[0003] There has been an increased use in collaboration sessions
that are Internet or web-based to communicate with employees,
vendors, and clients. During these collaboration sessions,
information is typically exchanged between multiple participants.
This exchanged information may include audio, graphical, and/or
textual information. Often times the ongoing effectiveness of the
collaboration session depends on the participants paying attention
in the collaboration session. There are many reasons as to why the
participant fails to pay attention in the collaboration session.
Several examples include exchanged information that is poorly
designed, participants falling behind during the collaboration
session, lack of interest on the part of the participant, and
technical difficulties that prevent the participant from following
the collaboration session.
SUMMARY
[0004] In one embodiment, the methods and apparatuses detect
devices participating in a collaboration session; monitor the
devices for an attribute corresponding to an attention level of
each of the devices; and determine whether each device is active
based on the attention level.
BRIEF DESCRIPTION OF THE DRAWINGS
[0005] The accompanying drawings, which are incorporated in and
constitute a part of this specification, illustrate and explain one
embodiment of the methods and apparatuses for monitoring attention
of a user during a collaboration session. In the drawings,
[0006] FIG. 1 is a diagram illustrating an environment within which
the methods and apparatuses for monitoring attention of a user
during a collaboration session are implemented;
[0007] FIG. 2 is a simplified block diagram illustrating one
embodiment in which the methods and apparatuses for monitoring
attention of a user during a collaboration session are
implemented;
[0008] FIG. 3 is a simplified block diagram illustrating a system,
consistent with one embodiment of the methods and apparatuses
monitoring attention of a user during a collaboration session;
[0009] FIG. 4 is an exemplary record for use with the methods and
apparatuses for monitoring attention of a user during a
collaboration session;
[0010] FIG. 5 is a flow diagram consistent with one embodiment of
the methods and apparatuses for monitoring attention of a user
during a collaboration session;
[0011] FIG. 6 is a flow diagram consistent with one embodiment of
the methods and apparatuses for monitoring attention of a user
during a collaboration session;
[0012] FIG. 7 is a flow diagram consistent with one embodiment of
the methods and apparatuses for monitoring attention of a user
during a collaboration session;
[0013] FIG. 8 is a flow diagram consistent with one embodiment of
the methods and apparatuses for monitoring attention of a user
during a collaboration session;
[0014] FIG. 9 illustrates an exemplary screen shot showing one
embodiment of the methods and apparatuses for monitoring attention
of a user during a collaboration session; and
[0015] FIG. 10 illustrates an exemplary screen shot showing one
embodiment of the methods and apparatuses for monitoring attention
of a user during a collaboration session.
DETAILED DESCRIPTION
[0016] The following detailed description of the methods and
apparatuses for monitoring attention of a user during a
collaboration session refers to the accompanying drawings. The
detailed description is not intended to limit the methods and
apparatuses for monitoring attention of a user during a
collaboration session. Instead, the scope of the methods and
apparatuses for monitoring attention of a user during a
collaboration session is defined by the appended claims and
equivalents. Those skilled in the art will recognize that many
other implementations are possible, consistent with the present
invention.
[0017] References to a "device" include a device utilized by a user
such as a desktop computer, a portable computer, a personal digital
assistant, a video phone, a landline telephone, a cellular
telephone, and a device capable of receiving/transmitting an
electronic signal.
[0018] References to a desktop are directed to an entire portion of
a display area of a corresponding device.
[0019] References to a collaboration session include a plurality of
devices that are configured to view content submitted by one of the
devices.
[0020] References to a participant device include devices that are
participating in the collaboration session.
[0021] References to a presenter device include a device that is
participant and shares content shared with other participants.
[0022] References to an attendee device include a device that is a
participant and receives content shared by another participant
device. The attendees are capable of view content that is offered
by the presenter device. In some instances, the attendee devices
are capable of modifying the content shared by the presenter
device.
[0023] In one embodiment, the methods and apparatuses for
monitoring attention of a user during a collaboration session
detects the attention of an attendee device during a collaboration
session. In one embodiment, different attributes of the attendee
device that reflect the attention of the user of the attendee
device are detected, recorded, and analyzed to determine whether
the user of the attendee device is paying attention to the content
associated with the collaboration session. In one embodiment, by
indicating the attention of the user of the attendee device, the
attendee device may be prompted to pay attention. Further, the
presenter device may be alerted to the attention status of the
attendee devices. Additionally, the attention status for each of
the attendee devices may be correlated to the content presented
during the collaboration session at predetermined times such that
the attention of the attendee devices may be determined at any
point during the collaboration session.
[0024] FIG. 1 is a diagram illustrating an environment within which
the methods and apparatuses for monitoring attention of a user
during a collaboration session are implemented. The environment
includes an electronic device 110 (e.g., a computing platform
configured to act as a client device, such as a computer, a
personal digital assistant, and the like), a user interface 115, a
network 120 (e.g., a local area network, a home network, the
Internet), and a server 130 (e.g., a computing platform configured
to act as a server).
[0025] In one embodiment, one or more user interface 115 components
are made integral with the electronic device 110 (e.g., keypad and
video display screen input and output interfaces in the same
housing such as a personal digital assistant. In other embodiments,
one or more user interface 115 components (e.g., a keyboard, a
pointing device such as a mouse, a trackball, etc.), a microphone,
a speaker, a display, a camera are physically separate from, and
are conventionally coupled to, electronic device 110. In one
embodiment, the user utilizes interface 115 to access and control
content and applications stored in electronic device 110, server
130, or a remote storage device (not shown) coupled via network
120.
[0026] In accordance with the invention, embodiments of monitoring
attention of a user during a collaboration session below are
executed by an electronic processor in electronic device 110, in
server 130, or by processors in electronic device 110 and in server
130 acting together. Server 130 is illustrated in FIG. 1 as being a
single computing platform, but in other instances are two or more
interconnected computing platforms that act as a server.
[0027] FIG. 2 is a simplified diagram illustrating an exemplary
architecture in which the methods and apparatuses for monitoring
attention of a user during a collaboration session are implemented.
The exemplary architecture includes a plurality of electronic
devices 202, a server device 210, and a network 201 connecting
electronic devices 202 to server 210 and each electronic device 202
to each other. The plurality of electronic devices 202 are each
configured to include a computer-readable medium 209, such as
random access memory, coupled to an electronic processor 208.
Processor 208 executes program instructions stored in the
computer-readable medium 209. In one embodiment, a unique user
operates each electronic device 202 via an interface 115 as
described with reference to FIG. 1.
[0028] The server device 130 includes a processor 211 coupled to a
computer-readable medium 212. In one embodiment, the server device
130 is coupled to one or more additional external or internal
devices, such as, without limitation, a secondary data storage
element, such as database 240.
[0029] In one instance, processors 208 and 211 are manufactured by
Intel Corporation, of Santa Clara, Calif. In other instances, other
microprocessors are used.
[0030] In one embodiment, the plurality of client devices 202 and
the server 210 include instructions for a customized application
for monitoring attention of a user during a collaboration session.
In one embodiment, the plurality of computer-readable media 209 and
212 contain, in part, the customized application. Additionally, the
plurality of client devices 202 and the server 210 are configured
to receive and transmit electronic messages for use with the
customized application. Similarly, the network 210 is configured to
transmit electronic messages for use with the customized
application.
[0031] One or more user applications are stored in media 209, in
media 212, or a single user application is stored in part in one
media 209 and in part in media 212. In one instance, a stored user
application, regardless of storage location, is made customizable
based on monitoring attention of a user during a collaboration
session as determined using embodiments described below.
[0032] FIG. 3 illustrates one embodiment of a system 300. In one
embodiment, the system 300 is embodied within the server 130. In
another embodiment, the system 300 is embodied within the
electronic device 110. In yet another embodiment, the system 300 is
embodied within both the electronic device 110 and the server
130.
[0033] In one embodiment, the system 300 includes a content
detection module 310, an interface detection module 320, a storage
module 330, an interface module 340, a control module 350, a
duration module 360, a scoring module 370, and a record module
380.
[0034] In one embodiment, the control module 350 communicates with
the content detection module 310, the interface detection module
320, the storage module 330, the interface module 340, the duration
module 360, the scoring module 370, and the record module 380. In
one embodiment, the control module 350 coordinates tasks, requests,
and communications between the content detection module 310, the
interface detection module 320, the storage module 330, the
interface module 340, the duration module 360, the scoring module
370, and the record module 380.
[0035] In one embodiment, the content detection module 310 detects
content that is utilized by the user in connection with the device.
In one embodiment, the content is utilized in connection with
multiple devices within a collaboration session between multiple
parties. For example, the collaboration session may include a data
conference or a video conference through a network, a phone line,
and/or the Internet.
[0036] In one embodiment, the content is a document utilized within
a collaboration session. In another embodiment, the content is
audio visual media that is utilized within a collaboration
session.
[0037] In one embodiment, the content detection module 310 detects
the location of the content within each device. For example, the
content detection module 310 is capable of detecting whether the
content is currently being displayed on the device. Further, when
there are multiple pieces of content on a particular device, the
content detection module 310 is capable of detecting whether the
content that is the subject of the collaboration session is
currently the priority content and shown on the display relative to
the other pieces of content. For example, the subject of the
collaboration session is currently on top and shown on the display
relative to other content.
[0038] In another embodiment, the content detection module 310
detects the size of the content that is being displayed on the
device. For example, the content detection module 310 is capable of
determining the size of the window that displays the content being
utilized within the collaboration session.
[0039] In yet another embodiment, the content detection module 310
detects the displayed location of the content that is the subject
of the collaboration session. For example, centrally positioned
locations on a display device may be considered better locations
for a user to view the content that is the subject of the
collaboration session. Similarly, locations that are on the
periphery of the display device may be considered less
desirable.
[0040] In one embodiment, the interface detection module 320
monitors input from various interface devices connected to devices
that are participating in the collaboration session. These various
interface devices include a keyboard, a pointing device, a
microphone, a telephone, a video camera, and the like.
[0041] In one embodiment, the interface detection module 320
detects when the cursor is moved within the device participating in
the collaboration session. In another embodiment, the interface
detection module 320 monitors the voice transmissions originating
from the device participating in the collaboration session. In yet
another embodiment, the interface detection module 320 detects any
activity by the device participating in the collaboration
session.
[0042] For example, in one embodiment, the presence of input from
the device participating in a collaboration session indicates that
the device is being actively utilized and that the user may be
viewing the device.
[0043] In one embodiment, the storage module 330 stores a record
including a list of attributes associated with the each device
participating in a collaboration session. An exemplary list of
attributes is shown in a record 400 within FIG. 4.
[0044] In one embodiment, the interface module 340 receives a
signal from one of the electronic devices 110. In one embodiment,
the electronic devices 110 are participating in a collaboration
session. In another embodiment, the interface module 340 delivers a
signal to one of the electronic devices 110.
[0045] In one embodiment, the duration module 360 monitors the
duration of the interaction of the devices participating in the
collaboration session. For example, if one of the devices has the
content located in a prominent location, then the duration module
360 monitors the length of time that the content is shown in the
prominent location. In another example, the duration module 360
also monitors the duration of voice transmissions and/or use of
other interface devices.
[0046] In another embodiment, the duration module 360 also monitors
the frequency of interaction of the devices participating in the
collaboration session. If the device participating in the
collaboration session is frequently being utilized during the
duration of the collaboration session, there is a higher chance
that a user of the device is viewing the device during the
collaboration session.
[0047] In one embodiment, the scoring module 370 receives
information from the content detection module 310, the interface
detection module 320, and the duration module 360. In one
embodiment, the scoring module 370 determines the level of
participation of the user of each device participating in the
collaboration session is considered active. In one embodiment, if
the level of participation exceeds an arbitrary threshold, then the
device is considered "active". Similarly, if the level of
participation is below an arbitrary threshold, then the device is
considered "inactive".
[0048] In one embodiment, the level of participation among each
device is utilized to determine the level of participation for all
participants within the collaboration session.
[0049] In one embodiment, the record module 380 stores the
collaboration session. In one embodiment, the record module 380
stores the content that is presented during the collaboration
session. In another embodiment, the record module 380 stores the
annotations and comments produced by the participants of the
collaboration session. In yet another embodiment, the record module
380 the activity levels of each of the devices participating in the
collaboration session.
[0050] In another embodiment, the record module 380 stores the
participation values for each device participating in the
collaboration session and corresponding content that is the subject
of the collaboration session. In one embodiment, the participation
values for each device are stored for a particular moment that
corresponds with the same moment during the collaboration
session.
[0051] The system 300 in FIG. 3 is shown for exemplary purposes and
is merely one embodiment of the methods and apparatuses for
monitoring attention of a user during a collaboration session.
Additional modules may be added to the system 300 without departing
from the scope of the methods and apparatuses for monitoring
attention of a user during a collaboration session. Similarly,
modules may be combined or deleted without departing from the scope
of the methods and apparatuses for monitoring attention of a user
during a collaboration session.
[0052] FIG. 4 illustrates an exemplary record 400 for use with the
methods and apparatuses for monitoring attention of a user during a
collaboration session. In one embodiment, the record 400
illustrates an exemplary record associated with the participation
of a particular device during a collaboration session.
[0053] In one embodiment, there are multiple records such that each
record 400 is associated with a particular device. Further, each
device corresponds with multiple records wherein each record 400
for a particular device corresponds to a particular collaboration
session.
[0054] In one embodiment, the record 400 includes a content
activity field 410, an interface activity field 420, a duration
field 430, a scoring field 440, and an aggregated scoring field
450. In one embodiment, the record 400 resides within the storage
module 330. In one embodiment, the record 400 describes a
attributes detected through the system 300.
[0055] In one embodiment, the content activity field 410 includes
information related to the content activity associated with the
particular device during the collaboration session. In one
embodiment, the content activity field 410 indicates a time stamp
when the content activity is detected through the content detection
module 310. For example, a time stamp that describes when the
content that is the subject of the collaboration session is pushed
to the background is recorded and stored within the content
activity field 410.
[0056] In one embodiment, the interface activity field 420 includes
information related to the interface activity associated with the
particular device during the collaboration session. In one
embodiment, the content activity field 420 indicates a time stamp
when the interface activity is detected through the content
detection module 320. For example, a time stamp that describes when
the cursor is utilized by a user of the particular device during
the collaboration session is recorded and stored within the
interface activity field 420.
[0057] In one embodiment, the duration field 430 includes the
length of time that corresponds to activity attributed from the
content and/or the interface device associated with the particular
device during the collaboration session. In one embodiment, the
duration field 430 receives information through the duration module
360.
[0058] In one embodiment, the scoring field 440 includes a score
that represents a rating of both the duration and the activity
level of the particular device during the collaboration session. In
one embodiment, the scoring field 440 receives information through
the scoring module 370.
[0059] In one embodiment, the aggregated scoring field 450 includes
a score that represents an overall rating of all the devices during
the collaboration session. In one embodiment, the overall rating is
based on both the duration and the activity level of all the
devices during the collaboration session. In one embodiment, the
aggregated scoring field 450 receives information through the
scoring module 370.
[0060] The flow diagrams as depicted in FIGS. 5, 6, 7 and 8 are one
embodiment of the methods and apparatuses for monitoring attention
of a user during a collaboration session. The blocks within the
flow diagrams can be performed in a different sequence without
departing from the spirit of the methods and apparatuses for
monitoring attention of a user during a collaboration session.
Further, blocks can be deleted, added, or combined without
departing from the spirit of the methods and apparatuses for
monitoring attention of a user during a collaboration session.
[0061] The flow diagram in FIG. 5 illustrates capturing activities
occurring on a device that is participating on a collaboration
session according to one embodiment of the invention.
[0062] In Block 510, the content is detected. In one embodiment,
the content is one of the items utilized within the collaboration
session. In one embodiment, the content is detected through the
content detection module 310. For example, the manner in which the
content is displayed on the device that is participating in the
collaboration session is detected. The manner in which the content
is displayed includes the size of the content being displayed, the
priority of the content (whether the content is displayed on the
top-most window), and/or the position of the content within the
display area.
[0063] In Block 520, the interface is detected. In one embodiment,
the interface input is detected from one of the devices
participating in the collaboration session. In one embodiment, the
interface input is detected through the interface detection module
320. For example, movement of the cursor or selection of items by
devices during the collaboration session is detected.
[0064] In Block 530, if interface input is detected, then the input
activity is recorded within the Block 540. If no interface input is
detected, then the absence of input activity is recorded within the
Block 550.
[0065] In Block 530, if a change in the display of the content is
detected, then the content activity is recorded within the Block
540. If no change in the display of the content is detected, then
the absence of content activity is recorded within the Block
550.
[0066] In Block 560, a time stamp is applied to either the
recordation of either activity or absence of activity. In one
embodiment, the time stamp is accorded a real time of day such as
1:30 PM. In another embodiment, the time stamp is accorded a time
relative to the collaboration session such as 2 minutes 15 seconds
into the collaboration session. In either embodiment, the
granularity of the time stamp may be adjustable. For example, the
time stamp can occur every minute, every second, or every fraction
of a second depending on the collaboration session.
[0067] In one embodiment, the activity information that arises from
the content detection module 310 and the interface detection module
320 are recorded within a record 400 of the storage module 330.
[0068] The flow diagram in FIG. 6 illustrates determining levels of
activity from devices participating in a collaboration session
according to one embodiment of the invention.
[0069] In Block 610, activity information is received. In one
embodiment, the activity information is recorded within FIG. 5.
[0070] In Block 620, the activity information associated with a
device participating in the collaboration session is compared
against a predetermined threshold. In one embodiment, the
predetermined threshold includes parameters from the content
detection module 310, the interface detection module 320, and the
duration module 360. In another embodiment, the predetermined
threshold is measured in terms of the parameter corresponding with
the scoring module 370.
[0071] In one embodiment, the predetermined threshold is assigned
to differentiate between devices that show indication that they are
active during the collaboration session versus devices that show
indication that they are inactive during the collaboration
session.
[0072] In one embodiment the predetermined threshold is based on a
cumulative score over the period of the collaboration session. In
another embodiment, the predetermine threshold is based on a
periodic score that is reset at particular intervals.
[0073] In Block 630, if the activity from a particular device
participating in the collaboration session exceeds the
predetermined threshold, then the activity associated with the
particular device is considered active and recorded in the Block
640.
[0074] In Block 630, if the activity from a particular device
participating in the collaboration session falls below the
predetermined threshold, then the activity associated with the
particular device is considered inactive and recorded in the Block
650.
[0075] In Block 660, the attendee device that falls below the
predetermined threshold is notified. In one embodiment, the
notification includes one of a buzzer sound from the attendee
device, a visual indication on the attendee device, and the content
that is the subject of the collaboration session being placed in
the forefront. In another embodiment, the attendee device is
notified by a pop-up window notifying the user of the attendee
device that additional attention is requested.
[0076] In another embodiment, the presenter device is also notified
when the attendee device falls below the predetermined
threshold.
[0077] The flow diagram in FIG. 7 illustrates displaying the status
of the devices participating in a collaboration session according
to one embodiment of the invention.
[0078] In Block 710, current activity status of each device
participating in the collaboration session is received. In one
embodiment, the current activity status is received during the
collaboration session. In one embodiment, the current activity
status of a device may be either active or inactive. In one
embodiment, the determination of whether the device is either
active or inactive depends on factors such as content detection,
interface detection, and duration of the content and/or interface
detection. An exemplary determination of the current activity
status is found in FIG. 6.
[0079] In Block 720, the current activity status of each device
participating in the collaboration session is displayed to one of
the devices participating in the collaboration session. In one
embodiment, the device that receives the displayed current activity
status is the host or originator of the collaboration session. In
another embodiment, the device that receives the displayed current
activity status is the current presenter of the collaboration
session; the presenter of the collaboration session may change
multiple times during the collaboration session.
[0080] An exemplary screen shot illustrating a display of the
current activity status for devices participating in the
collaboration session is shown in FIG. 10.
[0081] In Block 730, notification regarding the current activity
status is received by one of the devices participating in the
collaboration session. In one embodiment, the notification
indicates that a predetermined number of devices within the
collaboration session that are inactive. In another embodiment, the
notification indicates that a predetermined percentage of devices
within the collaboration session that are inactive.
[0082] In one embodiment, the notification is presented to the
device associated with current presenter in the collaboration
session. In another embodiment, the notification is presented to
the device associated with host or originator of the collaboration
session.
[0083] In one instance, when a particular device associated with
the current presenter receives a notification, then the current
presenter is able to modify the presentation to other devices with
the hope of increasing the number of devices that are actively
participating in the collaboration session.
[0084] The flow diagram in FIG. 8 illustrates displaying the status
of the devices participating in a collaboration session according
to one embodiment of the invention.
[0085] In Block 810, the collaboration session is recorded. In one
embodiment, the comments made by various devices participating in
the collaboration session as well as the content presented during
the collaboration session is recorded. In one embodiment, a time is
recorded simultaneously with the collaboration session. For
example, a time stamp is recorded periodically as the collaboration
session is also recorded. As an exemplary scenario, the time stamp
periodically occurs every second. In one embodiment, the time stamp
refers to the actual time such as 1:25:33 PM. In another
embodiment, the time stamp is relative to the start time of the
collaboration session.
[0086] In Block 820, the activity status and the time stamp
corresponding to the activity status is recorded. In one
embodiment, the activity status of each device participating in the
collaboration session is recorded. An exemplary determination of
the current activity status is found in FIG. 6.
[0087] In one embodiment, the activity status may change over the
course of the collaboration session. The activity status for each
device may be recorded at a predetermined interval. The
predetermined interval may range for a fraction of a second to
multiple seconds.
[0088] In Block 830, the activity status for each device at a given
time is correlated to the collaboration session. In one embodiment,
the time stamps corresponding to the activity status are matched
with time stamps corresponding to the recorded collaboration
session. For example, the activity status for device #3 at time t1
is matched to the content presented at the collaboration session at
the same time t1.
[0089] In Block 840, the activity status and content presented at
the collaboration session are displayed.
[0090] An exemplary screen shot illustrating the display of the
current activity status for devices participating in the
collaboration session and the content presented at the
collaboration session are shown in FIG. 9.
[0091] FIG. 9 illustrates an exemplary screen shot 900 that shows
one embodiment of the methods and apparatuses for monitoring
attention of a user during a collaboration session. In one
embodiment, the screen shot 900 includes a visual display area 910,
a time line 920, and an activity display 930.
[0092] In one embodiment, the visual display area 910 displays
content utilized during the collaboration session. In one
embodiment, the content is textual, graphical, audio, and/or video.
Further, the content may be annotations, conversations, and content
presented during the collaboration session.
[0093] In one embodiment, the time line 920 graphically displays a
timing diagram that shows the duration of at least a portion of the
collaboration session. In one embodiment, different times are shown
on the time line 920 such as t0, t1, and t2. For example, to may
represent the beginning of the collaboration session, and t2 may
represent the termination of the collaboration session. In one
embodiment, a marker 921 shows a current time location of the
collaboration session. For example, the marker 921 associates the
content displayed within the graphical display area 910 with a
location of the collaboration session.
[0094] In one embodiment, the activity display 930 graphically
illustrates the percentage of devices that are actively
participating in the collaboration session. In this example,
gradations of 100%, 50%, and 0% are shown. However, any number of
intervals may be utilized without departing from the spirit or
scope of the invention. In one embodiment, at times t0, t1, and t2,
the activity display 930 highlights those participation percentages
as milestone time periods. In one embodiment, the marker 921 shows
the percentage of device participation at a particular time.
[0095] In another embodiment, the activity display 930 graphically
illustrates the actual number of devices that are actively
participating in the collaboration session.
[0096] FIG. 10 illustrates an exemplary screen shot 1000 that shows
one embodiment of the methods and apparatuses for monitoring
attention of a user during a collaboration session. In one
embodiment, the screen shot 1000 includes a graphical display area
1010, a time line 1020, and an activity display 1030.
[0097] In one embodiment, the visual display area 1010 displays
content utilized during the collaboration session. In one
embodiment, the content is textual, graphical, audio, and/or video.
Further, the content may be annotations, conversations, and content
presented during the collaboration session.
[0098] In one embodiment, the time line 1020 graphically displays a
timing diagram that shows the duration of at least a portion of the
collaboration session. In one embodiment, different times are shown
on the time line 1020 such as t0, t1, and t2. For example, to may
represent the beginning of the collaboration session, and t2 may
represent the termination of the collaboration session. In one
embodiment, a marker 1021 shows a current time location of the
collaboration session. For example, the marker 1021 associates the
content displayed within the visual display area 1010 with a
location of the collaboration session.
[0099] In one embodiment, the activity display 1030 graphically
identifies devices that are actively participating in the
collaboration session. In this example, pictorial representations
1031 of each device are shown. In one embodiment, a different color
of the pictorial representations 1031 represents whether the device
is actively participating or inactive. In another embodiment, a
highlighted pictorial representation 1031 represents that the
device is actively participating in the collaboration session.
However, any number of pictorial and/or textual representations may
be utilized without departing from the spirit or scope of the
invention.
[0100] In one embodiment, at times t0, t1, and t2, the activity
display 1030 highlights those devices that are actively
participating as milestone time periods. In one embodiment, the
marker 1021 shows the devices that are actively participating at a
particular time.
[0101] In another embodiment, the activity display 930 graphically
illustrates the actual number of devices that are actively
participating in the collaboration session.
[0102] The foregoing descriptions of specific embodiments of the
invention have been presented for purposes of illustration and
description. The invention may be applied to a variety of other
applications.
[0103] They are not intended to be exhaustive or to limit the
invention to the precise embodiments disclosed, and naturally many
modifications and variations are possible in light of the above
teaching. The embodiments were chosen and described in order to
explain the principles of the invention and its practical
application, to thereby enable others skilled in the art to best
utilize the invention and various embodiments with various
modifications as are suited to the particular use contemplated. It
is intended that the scope of the invention be defined by the
Claims appended hereto and their equivalents.
* * * * *