U.S. patent application number 14/753183 was filed with the patent office on 2015-10-22 for system and method in a television receiver for providing user-selection of objects in a television program.
This patent application is currently assigned to Broadcom Corporation. The applicant listed for this patent is Broadcom Corporation. Invention is credited to Jeyhan Karaoguz, Nambirajan Seshadri.
Application Number | 20150304721 14/753183 |
Document ID | / |
Family ID | 43730008 |
Filed Date | 2015-10-22 |
United States Patent
Application |
20150304721 |
Kind Code |
A1 |
Karaoguz; Jeyhan ; et
al. |
October 22, 2015 |
System And Method In A Television Receiver For Providing
User-Selection Of Objects In A Television Program
Abstract
A system and method in a television receiver (e.g., a television
receiver external to a television) for providing user-selection of
objects in a television program, substantially as shown in and/or
described in connection with at least one of the figures, as set
forth more completely in the claims.
Inventors: |
Karaoguz; Jeyhan; (Irvine,
CA) ; Seshadri; Nambirajan; (Irvine, CA) |
|
Applicant: |
Name |
City |
State |
Country |
Type |
Broadcom Corporation |
Irvine |
CA |
US |
|
|
Assignee: |
Broadcom Corporation
|
Family ID: |
43730008 |
Appl. No.: |
14/753183 |
Filed: |
June 29, 2015 |
Related U.S. Patent Documents
|
|
|
|
|
|
Application
Number |
Filing Date |
Patent Number |
|
|
12850866 |
Aug 5, 2010 |
9098128 |
|
|
14753183 |
|
|
|
|
61242234 |
Sep 14, 2009 |
|
|
|
Current U.S.
Class: |
725/37 |
Current CPC
Class: |
H04N 21/42206 20130101;
H04N 21/482 20130101; H04N 21/41265 20200801; H04N 21/436 20130101;
H04N 21/44008 20130101; H04N 21/8545 20130101; H04N 9/8205
20130101; H04N 21/2668 20130101; H04N 21/858 20130101; G06F 3/0346
20130101; H04N 21/4826 20130101; H04N 21/2408 20130101; G06F 3/0308
20130101; H04N 21/42222 20130101; H04N 21/472 20130101; G06F 3/0304
20130101; H04N 21/4524 20130101; H04N 21/812 20130101; H04N
21/234318 20130101; H04N 21/42209 20130101; H04N 21/4828 20130101;
H04N 5/445 20130101; H04N 21/8133 20130101; H04N 21/438 20130101;
G06F 3/0325 20130101; G06F 3/0386 20130101; H04N 21/4725 20130101;
H04N 21/25841 20130101; H04N 21/47815 20130101; H04N 5/76 20130101;
H04N 21/4782 20130101; G06F 3/0428 20130101; H04N 21/4722 20130101;
H04N 21/4728 20130101; H04N 21/845 20130101; H04N 21/4334 20130101;
H04N 21/8173 20130101; G06F 3/0412 20130101; H04N 21/42204
20130101; H04N 21/4325 20130101; H04N 21/8126 20130101; H04N
21/4622 20130101; H04N 21/47805 20130101; H04N 21/23892
20130101 |
International
Class: |
H04N 21/4725 20060101
H04N021/4725; H04N 21/44 20060101 H04N021/44 |
Claims
1. A method for identifying a user-selected object in a video
stream, the method comprising: in a display device: communicating
the video stream for presentation on a screen, the video stream
comprising user-selectable objects; determining an on-screen
pointing location; identifying a set of user-selectable objects in
the video stream, at least in part, on the on-screen pointing
location; determining, from a set of types of user-selectable
objects, object types of the set of user selectable objects; and
selecting the user-selectable object based on an object type of the
user-selectable object.
2. The method of claim 1, wherein the object type is identified as
favorite object type of a user.
3. The method of claim 1, wherein an information presentation
functionality of the display device is determined based on the
object type.
4. The method of claim 1, wherein a communication functionality of
the display device is determined based on the object type.
5. The method of claim 1, wherein a business transaction
functionality of the display device is determined based on the
object type.
6. The method of claim 1, wherein an user interaction functionality
of the display device is determined based on the object type.
7. The method of claim 1, comprising receiving information
identifying selectable objects in the video stream.
8. The method of claim 7, wherein receiving information identifying
selectable objects in the video stream comprises receiving said
information embedded in a same data stream as a stream
communicating the video stream.
9. The method of claim 7, wherein receiving information identifying
selectable objects in the video stream comprises receiving said
information in a data stream communicated in parallel with a stream
communicating the video stream.
10. The method of claim 7, wherein the information identifying
selectable objects in the video stream is communicated with the
video stream without being requested.
11. The method of claim 7, wherein the information identifying
selectable objects in the video stream comprises timing information
associated with the selectable objects.
12. The method of claim 7, wherein the information identifying
selectable objects in the video stream comprises information
defining respective regions of the video stream associated with
respective selectable objects in the video stream.
13. The method of claim 12, wherein the information defining
respective regions of the video stream associated with the
respective selectable objects comprises information describing
movement of the respective regions.
14. The method of claim 1, comprising communicating information
indicating the user-selectable object to a remote control
device.
15. A receiver comprising: at least one processor operable to, at
least: communicate video stream for presentation on a screen, the
video stream comprising user-selectable objects; determine an
on-screen pointing location; identify a set of user-selectable
objects in the video stream, at least in part, on the on-screen
pointing location; and select the user-selectable object from the
set of user-selectable objects based on a unique identifier of the
user-selectable object, where the identifier is associated with a
group of user-selectable objects.
16. The receiver of claim 15, wherein the at least one module is
operable to receive information identifying selectable objects in
the video stream.
17. The receiver of claim 16, where the information identifying
selectable objects in the video stream is communicated with the
video stream without being requested by the receiver.
18. The receiver of claim 16, where the information identifying
selectable objects in the video stream comprises timing information
associated with the selectable objects.
19. The receiver of claim 16, where the information identifying
selectable objects in the video stream comprises information
defining respective regions of the video stream associated with
respective selectable objects in the video stream.
20. A method for identifying a user-selected object in a video
stream, the method comprising: in a display device: communicating
the video stream for presentation on a screen, the video stream
comprising user-selectable objects; determining an on-screen
pointing location; identifying a set of user-selectable objects in
the video stream, at least in part, on the on-screen pointing
location; determining whether the set of user-selectable objects
belongs to a group of related user-selectable objects; and
selecting the user-selectable object based on whether the
user-selectable object belongs to the group of related
user-selectable objects.
Description
CROSS-REFERENCE TO RELATED APPLICATIONS/INCORPORATION BY
REFERENCE
[0001] This patent application is a continuation application of
non-provisional application Ser. No. 12/850,866, filed Aug. 5,
2010, and titled "SYSTEM AND METHOD IN A TELEVISION RECEIVER FOR
PROVIDING USER-SELECTION OF OBJECTS IN A TELEVISION PROGRAM", and
is related to and claims priority from provisional patent
application Ser. No. 61/242,234 filed Sep. 14, 2009, and titled
"TELEVISION SYSTEM," the contents of which are hereby incorporated
herein by reference in their entirety. This patent application is
also related to U.S. patent application Ser. No. 12/774,380, filed
May 5, 2010, titled "SYSTEM AND METHOD IN A TELEVISION FOR
PROVIDING USER-SELECTION OF OBJECTS IN A TELEVISION PROGRAM"; U.S.
patent application Ser. No. 12/850,832, filed Aug. 5, 2010, titled
"SYSTEM AND METHOD IN A DISTRIBUTED SYSTEM FOR PROVIDING
USER-SELECTION OF OBJECTS IN A TELEVISION PROGRAM"; U.S. patent
application Ser. No. 12/850,911, filed Aug. 5, 2010, titled "SYSTEM
AND METHOD IN A TELEVISION CONTROLLER FOR PROVIDING USER-SELECTION
OF OBJECTS IN A TELEVISION PROGRAM"; U.S. patent application Ser.
No. 12/850,945, filed Aug. 5, 2010, titled "SYSTEM AND METHOD IN A
TELEVISION CONTROLLER FOR PROVIDING USER-SELECTION OF OBJECTS IN A
TELEVISION PROGRAM"; U.S. patent application Ser. No. 12/851,036,
filed Aug. 5, 2010, titled "SYSTEM AND METHOD IN A TELEVISION
SYSTEM FOR PROVIDING FOR USER-SELECTION OF AN OBJECT IN A
TELEVISION PROGRAM"; U.S. patent application Ser. No. 12/851,075,
filed Aug. 5, 2010, titled "SYSTEM AND METHOD IN A PARALLEL
TELEVISION SYSTEM FOR PROVIDING USER-SELECTION OF OBJECTS IN A
TELEVISION PROGRAM". The contents of each of the above-mentioned
applications are hereby incorporated herein by reference in their
entirety.
FEDERALLY SPONSORED RESEARCH OR DEVELOPMENT
[0002] [Not Applicable]
SEQUENCE LISTING
[0003] [Not Applicable]
MICROFICHE/COPYRIGHT REFERENCE
[0004] [Not Applicable]
BACKGROUND OF THE INVENTION
[0005] Present television systems are incapable of providing for
and/or conveniently providing for user-selection of objects in a
television program. Further limitations and disadvantages of
conventional and traditional approaches will become apparent to one
of skill in the art, through comparison of such systems with the
present invention as set forth in the remainder of the present
application with reference to the drawings.
BRIEF SUMMARY OF THE INVENTION
[0006] Various aspects of the present invention provide a system
and method in a television receiver (e.g., external to a
television) for providing for user selection of objects in a
television program, substantially as shown in and/or described in
connection with at least one of the figures, as set forth more
completely in the claims. These and other advantages, aspects and
novel features of the present invention, as well as details of
illustrative aspects thereof, will be more fully understood from
the following description and drawings.
BRIEF DESCRIPTION OF SEVERAL VIEWS OF THE DRAWINGS
[0007] FIG. 1 is a diagram illustrating an exemplary television
system, in accordance with various aspects of the present
invention.
[0008] FIG. 2 is a flow diagram illustrating an exemplary method
for providing user-selection of objects in television programming,
in accordance with various aspects of the present invention.
[0009] FIG. 3 is a flow diagram illustrating an exemplary method
for providing user-selection of objects in television programming,
in accordance with various aspects of the present invention.
[0010] FIG. 4 is a diagram illustrating an exemplary television
receiver, in accordance with various aspects of the present
invention.
[0011] FIG. 5 is a diagram illustrating exemplary modules and/or
sub-modules for a television receiver, in accordance with various
aspects of the present invention.
DETAILED DESCRIPTION OF VARIOUS ASPECTS OF THE INVENTION
[0012] The following discussion will refer to various communication
modules, components or circuits. Such modules, components or
circuits may generally comprise hardware and/or a combination of
hardware and software (e.g., including firmware). Such modules may
also, for example, comprise a computer readable medium (e.g., a
non-transitory medium) comprising instructions (e.g., software
instructions) that, when executed by a processor, cause the
processor to perform various functional aspects of the present
invention. Accordingly, the scope of various aspects of the present
invention should not be limited by characteristics of particular
hardware and/or software implementations of a module, component or
circuit unless explicitly claimed as such. For example and without
limitation, various aspects of the present invention may be
implemented by one or more processors (e.g., a microprocessor,
digital signal processor, baseband processor, microcontroller,
etc.) executing software instructions (e.g., stored in volatile
and/or non-volatile memory). Also for example, various aspects of
the present invention may be implemented by an application-specific
integrated circuit ("ASIC") and/or other hardware components.
[0013] Additionally, the following discussion will refer to various
television system modules (e.g., television receiver modules). It
should be noted that the following discussion of such various
modules is segmented into such modules for the sake of illustrative
clarity. However, in actual implementation, the boundaries between
various modules may be blurred. For example, any or all of the
functional modules discussed herein may share various hardware
and/or software components. For example, any or all of the
functional modules discussed herein may be implemented wholly or
in-part by a shared processor executing software instructions.
Additionally, various software sub-modules that may be executed by
one or more processors may be shared between various software
modules. Accordingly, the scope of various aspects of the present
invention should not be limited by arbitrary boundaries between
various hardware and/or software components, unless explicitly
claimed.
[0014] The following discussion may also refer to communication
networks and various aspects thereof. For the following discussion,
a communication network is generally the communication
infrastructure through which a communication device (e.g., a
portable communication device, television, television control
device, television provider, television programming provider,
television receiver, video recording device, etc.) may communicate
with other systems. For example and without limitation, a
communication network may comprise a cable and/or satellite
television communication network, a cellular communication network,
a wireless metropolitan area network (WMAN), a wireless local area
network (WLAN), a wireless personal area network (WPAN), any home
or premises communication network, etc. A particular communication
network may, for example, generally have a corresponding
communication protocol according to which a communication device
may communicate with the communication network. Unless so claimed,
the scope of various aspects of the present invention should not be
limited by characteristics of a particular type of communication
network.
[0015] The following discussion will at times refer to an on-screen
pointing location. Such a pointing location refers to a location on
the television screen to which a user (either directly or with a
pointing device) is pointing. Such a pointing location is to be
distinguished from other types of on-screen location
identification, such as, for example, using arrow keys and/or a
mouse to move a cursor or to traverse blocks (e.g., on an on-screen
program guide) without pointing. Various aspects of the present
invention, while referring to on-screen pointing location, are also
readily extensible to such other forms of on-screen location
identification.
[0016] Additionally, the following discussion will at times refer
to television programming. Such television programming generally
includes various types of television programming (e.g., television
programs, news programs, sports programs, music television, movies,
television series programs and/or associated advertisements,
educational programs, live or recorded television programming,
broadcast/multicast/unicast television programming, etc.). Such
television programming may, for example, comprise real-time
television broadcast programming (or multicast or unicast
television programming) and/or user-stored television programming
that is stored in a user device (e.g., a VCR, PVR, etc.). Such
television programming video content is to be distinguished from
other non-programming video content that may be displayed on a
television screen (e.g., an electronic program guide, user
interface menu, a television set-up menu, a typical web page, a
document, a graphical video game, etc.). Various aspects of the
present invention may, for example in a television receiver (e.g.,
a receiver external to a television), comprise receiving television
programming, communicating the received television programming to a
television for presentation to a user on the screen of the
television, determining an on-screen pointing location pointed to
by the user of the television, and identifying a user-selected
object in the presented television programming.
[0017] Also, the following discussion will at times refer to
user-selectable objects in television programming. Such
user-selectable objects includes both animate (i.e., living) and
inanimate (i.e., non-living) objects, both still and moving. Such
objects may, for example, comprise characteristics of any of a
variety of objects present in television programming. Such objects
may, for example and without limitation, comprise inanimate
objects, such as consumer good objects (e.g., clothing,
automobiles, shoes, jewelry, furniture, food, beverages,
appliances, electronics, toys, artwork, cosmetics, recreational
vehicles, sports equipment, safety equipment, computer equipment,
communication devices, books, etc.), premises objects (e.g.,
business locations, stores, hotels, signs, doors, buildings,
landmarks, historical sites, entertainment venues, hospitals,
government buildings, etc.), objects related to services (e.g.,
objects related to transportation, objects related to emergency
services, objects related to general government services, objects
related to entertainment services, objects related to food and/or
drink services, etc.), objects related to location (e.g., parks,
landmarks, streets, signs, road signs, etc.), etc. Such objects
may, for example, comprise animate objects, such as people (e.g.,
actors/actresses, athletes, musicians, salespeople, commentators,
reports, analysts, hosts/hostesses, entertainers, etc.), animals
(e.g., pets, zoo animals, wild animals, etc.) and plants (e.g.,
flowers, trees, shrubs, fruits, vegetables, cacti, etc.).
[0018] Turning first to FIG. 1, such figure is a diagram
illustrating a non-limiting exemplary television system 100 in
accordance with various aspects of the present invention. The
exemplary system 100 includes a television provider 110. The
television provider 110 may, for example, comprise a television
network company, a cable company, a movie-providing company, a news
company, an educational institution, etc. The television provider
110 may, for example, be an original source of television
programming (or related information). Also for example, the
television provider 110 may be a communication company that
provides programming distribution services (e.g., a cable
television company, a satellite television company, a
telecommunication company, a data network provider, etc.). The
television provider 110 may, for example, provide television
programming and non-programming information and/or video content.
The television provider 110 may, for example, provide information
related to a television program (e.g., information describing or
otherwise related to selectable objects in television programming,
etc.).
[0019] The exemplary television system 100 may also include a third
party program information provider 120. Such a provider 120 may,
for example, provide information related to a television program.
Such information may, for example, comprise information describing
selectable objects in programming, program guide information,
etc.
[0020] The exemplary television system 100 may include one or more
communication networks (e.g., the communication network(s) 130).
The exemplary communication network 130 may comprise
characteristics of any of a variety of types of communication
networks over which television programming and/or information
related to television programming (e.g., information identifying
and/or describing and/or otherwise related to user-selectable
objects in television programming) may be communicated. For example
and without limitation, the communication network 130 may comprise
characteristics of any one or more of: a cable television network,
a satellite television network, a telecommunication network, the
Internet, a local area network (LAN), a personal area network
(PAN), a metropolitan area network (MAN), any of a variety of
different types of home networks, etc.
[0021] The exemplary television system 100 may include a first
television 140. Such a first television 140 may, for example,
comprise networking capability enabling such television 140 to
communicate directly with the communication network 130. For
example, the first television 140 may comprise one or more embedded
television receivers or transceivers (e.g., a cable television
receiver, satellite television transceiver, Internet modem, etc.).
Also for example, the first television 140 may comprise one or more
recording devices (e.g., for recording and/or playing back video
content, television programming, etc.).
[0022] The exemplary television system 100 may include a first
television controller 160. Such a first television controller 160
may, for example, operate to (e.g., which may include "operate when
enabled to") control operation of the first television 140. The
first television controller 160 may comprise characteristics of any
of a variety of television controlling devices. For example and
without limitation, the first television controller 160 may
comprise characteristics of a dedicated television control device,
a universal remote control, a cellular telephone or personal
computing device with television control capability, etc.
[0023] The first television controller 160 (or television control
device) may, for example, transmit signals directly to the first
television 140 to control operation of the first television 140.
The first television controller 160 may also, for example, operate
to transmit signals (e.g., via the communication network 130) to
the television provider 110 to control television programming (or
related information) being provided to the first television 140, or
to conduct other transactions (e.g., business transactions,
etc.).
[0024] As will be discussed in more detail later, the first
television controller 160 may operate to communicate screen
pointing information with the first television 140 and/or other
devices. Also, as will be discussed in more detail later, various
aspects of the present invention include a user pointing to a
location on a television screen (e.g., pointing to an animate or
inanimate object presented in television programming). In such a
scenario, the user may perform such pointing in any of a variety of
manners. One of such exemplary manners includes pointing with a
television control device. The first television controller 160
provides a non-limiting example of a device that a user may utilize
to point to an on-screen location.
[0025] The exemplary television system 100 may also include a
television receiver 151. The television receiver 151 may, for
example, operate to (e.g., which may include "operate when enabled
to") provide a communication link between a television and/or
television controller and a communication network and/or
information provider. For example, the television receiver 151 may
operate to provide a communication link between the second
television 141 and the communication network 130, or between the
second television 141 and the television provider 110 (and/or third
party program information provider 120) via the communication
network 130.
[0026] The television receiver 151 may comprise characteristics of
any of a variety of types of television receivers. For example and
without limitation, the television receiver 151 may comprise
characteristics of a cable television receiver, a satellite
television receiver, etc. Also for example, the television receiver
151 may comprise a data communication network modem for data
network communications (e.g., with the Internet, a LAN, PAN, MAN,
telecommunication network, etc.). The television receiver 151 may
also, for example, comprise recording capability (e.g., programming
recording and playback, etc.). The television receiver 151 may, for
example, be a stand-alone entity or may, for example, be integrated
into any of a variety of other devices (e.g., integrated into a
DVR, DVD and/or Blueray player, integrated into a game box,
integrated into a personal computer, integrated into a router
and/or gateway, etc.). The television receiver 151 may, for
example, operate to (which includes "operate when enabled to")
perform any or all of the functionality discussed herein.
[0027] The exemplary television system 100 may include a second
television controller 161. Such a second television controller 161
may, for example, operate to (e.g., which may include "operate when
enabled to") control operation of the second television 141 and the
television receiver 151. The second television controller 161 may
comprise characteristics of any of a variety of television
controlling devices. For example and without limitation, the second
television controller 161 may comprise characteristics of a
dedicated television control device, a dedicated television
receiver control device, a universal remote control, a cellular
telephone or personal computing device with television control
capability, etc.
[0028] The second television controller 161 may, for example,
operate to transmit signals directly to the second television 141
to control operation of the second television 141. The second
television controller 161 may, for example, operate to transmit
signals directly to the television receiver 151 to control
operation of the television receiver 151. The second television
controller 161 may additionally, for example, operate to transmit
signals (e.g., via the television receiver 151 and the
communication network 130) to the television provider 110 to
control television programming (or related information) being
provided to the television receiver 151, or to conduct other
transactions (e.g., business transactions, etc.). The second
television controller 161 may further, for example, operate to
receive signals from the second television 141 and/or television
receiver 151. Such signals may, for example, comprise signals
communicating television programming, information identifying
and/or describing user-selectable objects in television programming
and/or any of a variety of other information to the second
television controller 161.
[0029] As will be discussed in more detail later, various aspects
of the present invention include a user pointing to a location on a
television screen (e.g., pointing to an animate or inanimate object
presented in television programming). In such a scenario, the user
may perform such pointing in any of a variety of manners. One of
such exemplary manners includes pointing with a television control
device. The second television controller 161 provides one
non-limiting example of a device that a user may utilize to point
to an on-screen location.
[0030] The exemplary television system 100 was provided to provide
a non-limiting illustrative foundation for discussion of various
aspects of the present invention. Thus, the scope of various
aspects of the present invention should not be limited by any
characteristics of the exemplary television system 100 unless
explicitly claimed.
[0031] FIG. 2 is a flow diagram illustrating an exemplary method
200 for providing user-selection of objects in television
programming, in accordance with various aspects of the present
invention. Any or all aspects of the exemplary method 200 may, for
example, be implemented in a television receiver (e.g., external to
a television), for example the television receiver 151 shown in
FIG. 1 and discussed previously.
[0032] The exemplary method 200 may, for example, begin executing
at step 205. The exemplary method 200 may begin executing in
response to any of a variety of causes and/or conditions. For
example, the exemplary method 200 may begin executing in response
to a user command to begin, in response to user selection of a
television program (or corresponding television channel) that
includes user selectable objects, upon television receiver reset
and/or power-up, in response to a user input indicating a desire to
provide object selection capability to the user, in response to
identification of a user and/or user equipment for which object
selection capability is to be provided, in response to user payment
of a fee, etc.
[0033] The exemplary method 200 may, for example at step 210,
comprise receiving television programming. Many non-limiting
examples of such television programming were provided above (e.g.,
television programming with user-selectable objects). Step 210 may
comprise receiving the television programming from any of a variety
of sources. For example and without limitation, step 210 may
comprise receiving the television programming from a television
broadcasting company, from a movie streaming company, from a user
(or consumer) video recording device (e.g., internal and/or
external to the television receiver), from an Internet television
programming provider, etc.
[0034] Step 210 may also comprise receiving the television
programming via any of a variety of types of communication
networks. Such networks may, for example, comprise a wireless
television network (e.g., terrestrial and/or satellite) and/or
cable television network. Such networks may, for example, comprise
any of variety of data communication networks (e.g., the Internet,
a local area network, a personal area network, a metropolitan area
network, etc.).
[0035] In general, step 210 may comprise receiving television
programming. The scope of various aspects of the present invention
should not be limited by characteristics of any particular
television programming, television programming source, television
programming network or manner of receiving television programming
unless explicitly claimed.
[0036] The exemplary method 200 may, at step 220, comprise
communicating television programming to a television (e.g., for
presentation to a user on a screen of the television). Step 220
may, for example, comprise communicating the television programming
received at step 210 to a television in any of a variety of
manners. For example, step 220 may comprise communicating raw video
data and/or decoded video data in one or more data signals, one or
more television/display driver signals (e.g., composite and/or
component video signals), etc. Also for example, step 220 may
comprise communicating the television programming over any of a
variety of communication media (e.g., wired, tethered optical,
wireless RF, non-tethered optical, etc.). Further for example, step
220 may comprise communicating the television programming to
another video presentation device external to the television.
[0037] The communicated television programming may, for example,
comprise user-selectable objects in the television programming.
Many non-limiting examples of such user-selectable objects were
presented above. In general, such user-selectable objects may, for
example, comprise animate and/or inanimate objects in television
programming that a user may select (e.g., using a pointing device
or other user interface by which a user may specify a screen
location). Information of such user-selectable objects may be
received by a television receiver and retained for processing by
the television receiver and/or communicated to the television as
part of the television programming or separate from the television
programming (e.g., via a data interface separate from a television
programming interface).
[0038] The exemplary method 200 may, at step 230, comprise
determining an on-screen pointing location pointed to by a user of
the television receiver and/or television. Step 230 may comprise
determining an on-screen pointing location in any of a variety of
manners, non-limiting examples of which will now be provided.
Various non-limiting examples of on-screen pointing location
determining are provided in U.S. Provisional Application No.
61/242,234, which is hereby incorporated herein by reference in its
entirety. An on-screen pointing location may, for example, be
expressed in a screen-centric coordinate system (e.g., x-y pixel
coordinates), a screen independent coordinate system (e.g., based
on location within a moving image, where such location is generic
to all television screens), a world coordinate and/or universal
coordinate system, a video frame-based coordinate system, etc.
[0039] Step 230 may, for example, comprise the television receiver
analyzing sensor information (e.g., associated with sensors
on-board and/or off-board the television and/or television
receiver) to determine user on-screen pointing location. Step 230
may also, for example, comprise the television receiver receiving
information describing the on-screen pointing location from a
device external to the television receiver (e.g., the television, a
television controller, a television network device, a user pointing
device, etc.).
[0040] Step 230 may, for example, comprise identifying a timestamp
temporally identifying the instance of a determined on-screen
pointing location. Such timestamp may, for example, be obtained by
a clock, timestamp embedded in a video stream, timestamp embedded
in a stream including object information, timestamp associated with
a signal transmitted from a user pointing device, etc.
Determination of such a timestamp may, for example, be based on
user command (e.g., a user indicating that a selection has
occurred) or automatically without a direct indication from the
user that a selection has occurred (e.g., the system determining
that the user has pointed to an object for at least a particular
amount of time), etc. Such timestamp may be utilized, for example,
for determining selection of a moving, changing and/or temporally
transient object in the television programming.
[0041] In general, step 230 may comprise determining an on-screen
pointing location pointed to by a user of the television receiver
and/or television. Accordingly, the scope of various aspects of the
present invention should not be limited by characteristics of any
particular type of on-screen pointing location or any particular
manner of determining such an on-screen pointing location unless
explicitly claimed.
[0042] The exemplary method 200 may, at step 240, comprise
identifying a user-selectable object in the television programming
(e.g., the television programming communicated at step 220) at
which the user is pointing based, at least in part, on the
determined on-screen pointing location (e.g., as determined at step
230). Step 240 may comprise performing such identifying in any of a
variety of manners, non-limiting examples of which will now be
presented.
[0043] For example, step 240 may comprise determining the on-screen
location and/or dimensions of one or more user-selectable objects
(e.g., or associated selection region) in the television
programming, and identifying a user-selected object by analyzing
the respective on-screen locations of the one or more
user-selectable objects and the determined on-screen pointing
location (e.g., at a particular time instance and/or particular
timeframe) to determine the television programming object selected
by the user. For example, such on-screen location may, for example,
comprise the on-screen location of one or more points, areas and/or
volumes associated with respective locations of user-selectable
objects.
[0044] In such an exemplary scenario or any scenario involving
determining on-screen location of one or more objects in the
presented television programming, step 240 may, for example,
comprise determining such on-screen object location in any of a
variety of manners. For example, step 240 may comprise receiving
information identifying and/or describing the user-selectable
objects in the television program.
[0045] For example, step 240 may comprise receiving information
identifying and/or describing such user-selectable objects from the
same source as the received television programming. For example,
step 240 may comprise receiving such information embedded in a same
data stream as a stream communicating the television programming to
the television receiver (e.g., embedded in the received television
program data). For example, a television stream protocol may
comprise specialized elements (and/or the utilization of unassigned
elements) that include information about selectable objects (e.g.,
object identity, shape, location, size, coloration, movement
characteristics, timing, appearance time window, etc.).
[0046] Also for example, step 240 may comprise receiving the
information identifying and/or describing such user-selectable
objects in a data stream communicated in parallel with a stream
communicating the presented television programming. In such a
scenario, the television program stream and object information
stream may be received (e.g., by the television receiver) from a
same source over a same television programming communication
channel. Additionally for example, step 240 may comprise receiving
such information from a same source but over a communication
channel different from the channel over which the television
programming is received and/or over a type of communication channel
different from a television programming communication channel.
[0047] Further for example, step 240 may comprise receiving the
information identifying and/or describing such user-selectable
objects from a source (e.g., a third party information provider, a
television network source, etc.) different from the source from
which the television programming is received. In such an exemplary
scenario, step 240 may comprise receiving such information via
different respective communication networks or via one or more same
communication networks. Also for example, step 240 may comprise
receiving such information over a different communication medium
than that over which the television programming is received.
[0048] Step 240 may, for example, comprise receiving the
information identifying and/or describing such user-selectable
objects in a data stream, where such information is always
transmitted in the data stream (e.g., whether or not requested by a
user and/or other system). Alternatively for example, step 240 may
comprise receiving such information, where such information is
communicated (e.g., to the television) only when requested (e.g.,
only when requested by the television receiver, by a television
controller, by a television, by a user electronic device, by the
user, etc.).
[0049] Step 240 may, for example, comprise receiving the
information identifying and/or describing such user-selectable
objects in real-time (i.e., as the television programming is
received). Also for example, step 240 may comprise receiving such
information from a source of user-stored television programming.
For example, such information may be stored with stored television
programming in a user storage device (e.g., in a same data file, in
separate but related files, etc.). In such an exemplary
implementation, such information may be received from the user's
television programming storage device in time synchronization with
television programming.
[0050] As mentioned above, the information identifying and/or
describing user selectable objects in television programming may
comprise timing information associated with such selectable
objects. For example, movement of a selectable object may be
expressed as a function of time. Also for example, appearance of a
selectable object in television programming may be associated with
a time window during which such object appears. As will be
discussed in more detail below, timing associated with a user
on-screen pointing (or object selection) event may be synchronized
to the timing of selectable object location in a television program
to determine whether a user pointed to (or selected) a particular
object at a particular location at a particular time.
[0051] The information identifying and/or describing user
selectable objects in television programming may comprise
information defining respective regions of the presented television
programming that are associated with respective user-selectable
objects in the presented television programming. For example, such
information may comprise information describing respective
geometric shapes (e.g., 2-D and/or 3-D geometric constructs)
associated with respective user-selectable objects. For example, a
circle, oval, square, rectangle, pentagon or any polygon may be
associated with a user-selectable object. User-selection of one of
such geometric shapes (e.g., a determined on-screen pointing
location within the boundaries of such geometric shape(s) at the
relevant point in time) may indicate user selection of the
respective object.
[0052] As discussed above, the object description (whether
geometric or not) may comprise information (e.g., temporal
information) describing movement, appearance, size changing, shape
changing, etc. associated with the user-selectable object. For
example, in an exemplary scenario where a plurality of geometric
shapes are utilized to describe a user-selectable object (and/or a
user-selectable area associated with such object), the description
of the user-selectable object may comprise information describing
the manner in which each of the respective objects move in the
programming (e.g., as a function of time, as a function of video
frame number, etc.) and/or information describing the manner in
which dimensions of each of the respective objects change.
[0053] Once on-screen pointing location and object location (e.g.,
at a particular point in time) are known, such information may be
processed to determine a user-selectable object to which a user is
pointing. As mentioned above, such processing may comprise
determining a respective region of the television screen and/or a
television programming frame associated with a respective
user-selectable object, where the respective region correlates to
an on-screen pointing location pointed to by the user (e.g., at a
particular point in time or during a particular timeframe).
[0054] During performing such processing, step 240 may comprise
low-pass filtering the determined on-screen pointing location
(e.g., as determined at step 230) to compensate for unintended
movement of pointing location (e.g., due to unsteady or unstable
user pointing), thus increasing reliability of object selection
determination. For example, successful user-selection of a
television programming object may require a user to point to an
on-screen object for a particular amount of time (or for a
particular number of frames).
[0055] Note that such low-pass filtering may also be performed at
step 230. Such filtering may, for example, comprise filtering over
a particular period of time, over a particular number of on-screen
pointing direction determinations, over a particular number of
television programming frames, etc. Such filtering may comprise
averaging a plurality of on-screen pointing direction
determinations, utilizing a finite impulse response filtering
scheme, etc.
[0056] Step 240 may also, for example, comprise identifying a
user-selected object by identifying a most likely object selected
by the user. For example, step 240 may comprise identifying a
closest user-selectable object to the on-screen pointing location
determined at step 230. Also for example, step 240 may comprise
determining a most likely user-selected object based, at least in
part, on respective popularity of user selectable objects (e.g.,
favoring the most often selected inanimate and/or animate objects).
Additionally for example, step 240 may comprise identifying a most
likely user-selected object based, at least in part, on monetary
considerations (e.g., placing a higher likelihood on
user-selectable objects associated with relatively higher paying
advertisers, placing a higher likelihood on user-selectable objects
associated with a per-selection based advertising fee, etc.).
Further for example, step 240 may comprise identifying a most
likely user-selected programming object based on history of
selection from a particular user (e.g., favoring types of objects
most often selected by a particular user). Also for example, step
240 may comprise identifying a most likely user-selected
programming object based on object newness (e.g., a new object is
likely to garner more interest than an object that has been shown
for a relatively long period of time). Additionally, for example,
step 240 may comprise identifying a most likely user-selected
television programming object based on object size.
[0057] A selected object may be identified with a unique identifier
(e.g., unique to the object, unique to a group of related objects,
etc.). Step 240 may thus comprise determining the identifier
associated with the determined user-selected object. Such
identifier may then be utilized in the performance of any of a
variety of further functions, non-limiting examples of which will
be presented below. For example, the information identifying and/or
describing a user-selectable object may comprise information
describing functions associated with such object (e.g., information
presentation functionality, communication functionality, business
transaction functionality, user interaction functionality,
etc.).
[0058] The steps of the exemplary method 200 (or aspects thereof)
may, for example, be performed (e.g., by a television receiver) in
real-time. In such manner, the user may have relatively expeditious
access to functionality associated with the user-selected object.
Alternatively for example, the exemplary method 200 (or aspects
thereof) may be performed off-line in a manner in which
functionality associated with the user-selected object is provided
to the user at a later time (e.g., after presentation of the
television program, upon the user pausing presentation of the
television program, upon the user logging into the user's computer
system, upon the user accessing email, etc.).
[0059] As mentioned above, any or all of the steps of the exemplary
method 200 may be performed for user selection of an object in
television programming as the programming is broadcast in real-time
and/or may be performed for user selection of an object in
television programming that has been recorded on a user (or home)
television programming recorder (e.g., a personal video recorder
(PVR), video cassette recorder (VCR), etc.) and is currently being
communicated to the television for presentation to the user (e.g.,
at step 220) in a time-shifted manner. For example, a user may
record a broadcast television program on a PVR for later viewing,
view such recorded programming at a later time, and while viewing
such time-shifted television programming at a later time, select
user-selectable objects in such programming.
[0060] Similarly, any or all of the steps of the exemplary method
200 may be performed for user selection of an object in television
programming that has been provided to the user (or stored by the
user) on a physical storage medium (e.g., on a digital versatile
disc (DVD), video cassette recorder tape, non-volatile memory
device, etc.). For example, a user may purchase a set of DVDs
including all episodes of a season of a television series, view
each of such episodes at the convenience of the user, and while
viewing such episodes, select user-selectable objects in such
programming. In such a scenario, a system (e.g., a television
receiver) implementing the method 200 may receive the television
programming and/or object information from a television program
record and playback device (e.g., on-board and/or off-board a
television receiver).
[0061] In an exemplary scenario where on-screen pointing location
at a particular point in time is utilized to determine object
selection, any of a variety of time references may be utilized. For
example, synchronization of on-screen pointing location and
user-selectable object location (e.g., on-screen and/or in-frame
location) may be based on a presentation timestamp (PTS) and/or a
decoding timestamp (DTS), or the like, which may be encoded in a
broadcast and/or recorded television program or determined as such
program is being communicated to a television or presented to a
user. In such a scenario, so long as the object location and
pointing determination are based on a common and/or synchronized
time reference, the identification of a pointed-to object may be
performed accurately.
[0062] As mentioned previously, object information identifying
and/or describing user-selectable objects may be received encoded
in a video program stream or may be received in a separate stream
(and/or channel). In a scenario where television programming
information is stored (either short term or long term), the object
information may also be stored (e.g., with the stored programming
information in a same data file, in a separate but related data
file, etc.). Such storage may, for example, be housed in a same
housing as a television receiver implementing the method and/or in
a separate housing. In such a manner, when the user determines to
view a time-shifted program, the object information is accessible
to the television receiver.
[0063] Alternatively, such information identifying and/or
describing user-selectable objects in television programming may be
requested from a third party when such information is needed. In
yet another scenario, for example, where video information may be
decoded separately from the television receiver (e.g., in a
separate set top box (STB), cable and/or satellite television
receiver, PVR, personal computer, etc.) and communicated to the
television receiver, such object information may also be received
by such separate device and provided to the television receiver
(e.g., in an information channel separate from a television
programming information channel).
[0064] Note that although a portion of the previous discussion
concerned analyzing on-screen pointing location and on-screen
object location to identify a user-selected object, such analysis
may also be similarly performed by analyzing on-frame pointing
location and on-frame object location. In other words, such
analysis may comprise performing any of a variety of coordinate
transformations to perform such analysis in any of a variety of
different respective coordinate domains.
[0065] In general, step 240 may comprise identifying a
user-selectable object in the presented television programming at
which the user is pointing (e.g., based, at least in part, on the
determined on-screen pointing location (e.g., as determined at step
230)). Accordingly, the scope of various aspects of the present
invention should not arbitrarily be limited by any particular
manner of performing such identifying unless explicitly
claimed.
[0066] The exemplary method 200 may, for example at step 295,
comprise performing continued operations. Step 295 may comprise
performing any of a variety of continued operations, non-limiting
examples of such continued operation(s) will be presented below.
For example, step 295 may comprise returning execution flow to any
of the previously discussed method steps. For example, step 295 may
comprise returning execution flow of the exemplary method 200 to
step 230 for determining additional on-screen pointing locations
and corresponding user-selected objects in the television
programming.
[0067] Also for example, step 295 may comprise generating a signal
comprising characteristics (e.g., non-data characteristics, analog
characteristics, digital characteristics, data, etc.) adapted to
cause the television to generate a user output indicating the
identified user-selectable object (e.g., as identified at step
240). For example, such a generated signal may comprise
characteristics adapted to cause the television to overlay a
graphical feature coinciding with the identified user-selectable
object on the television programming as the television programming
is presented to the user. For example, as discussed above, a
user-selectable object (and/or the user-selectable portion of a
user-selectable object) may be defined by one or more geometric
shapes. In such an exemplary scenario, step 295 may comprise
generating a signal comprising characteristics adapted to cause the
television to highlight such geometric shapes (or the borders
thereof) when step 240 determines that the user has selected a
user-selectable object associated with such geometric shapes. Also
for example, step 295 may comprise generating a signal comprising
characteristics adapted to cause the television to present an
outline of the identified object on the television screen,
temporarily brighten or alter the color of the identified object,
temporarily display a message on the screen as an indication of the
identified object, etc. Step 295 may also, for example, comprise
outputting an audio indication that a user-selected object has been
identified and/or generating a signal comprising characteristics
that cause another device (e.g., a television) to output such an
audio indication.
[0068] Additionally for example, step 295 may comprise
communicating information indicating the identified user-selectable
object to a device other than a television external to the
television receiver (e.g., a user device at the same premises as
the television receiver and/or a device communicatively coupled to
the television receiver via a communication network). For example,
step 295 may comprise communicating such information to a
television remote control device (e.g., in a scenario where the
television remote control device may provide the user an indication
of the identified user-selectable object). In such an exemplary
scenario, the television remote control device may comprise a video
screen on which the television program may be displayed, and the
identified user-selectable object may then be graphically indicated
on such video screen (e.g., instead of being identified on a screen
of the television and/or in addition to being identified on the
screen of the television).
[0069] Further for example, step 295 may comprise processing
information of an identified user-selected object (e.g., as
determined at step 240) to determine an action to perform with
regard to such selected object. Various non-limiting examples of
such actions are provided in U.S. Provisional Application No.
61/242,234, which is hereby incorporated herein by reference in its
entirety.
[0070] In general, step 295 may comprise performing continued
operations (e.g., performing additional operations corresponding to
a user-selected television programming object, repeating various
method steps for additional user-selected objects, etc.).
Accordingly, the scope of various aspects of the present invention
should not be limited by characteristics of any particular type of
continued processing unless explicitly claimed.
[0071] Turning next to FIG. 3, such figure is a flow diagram
illustrating an exemplary method 300 for providing user-selection
of objects in television programming, in accordance with various
aspects of the present invention. The exemplary method 300 may, for
example, share any or all characteristics with the exemplary method
200 illustrated in FIG. 2 and discussed previously. Any or all
aspects of the exemplary method 300 may, for example, be
implemented in a television receiver (e.g., external to a
television), for example the television receiver 151 shown in FIG.
1 and discussed previously.
[0072] The exemplary method 300 may, for example, begin executing
at step 305. The exemplary method 300 may begin executing in
response to any of a variety of causes or conditions. Step 305 may,
for example, share any or all characteristics with step 205 of the
exemplary method 200 illustrated in FIG. 2 and discussed
previously.
[0073] The exemplary method 300 may, for example at step 310,
comprise receiving television programming. Step 310 may, for
example, share any or all characteristics with step 210 of the
exemplary method 200 illustrated in FIG. 2 and discussed
previously.
[0074] For example, step 310 may comprise, for example at sub-step
312, receiving a television program broadcast as such program is
broadcast in real-time. Alternatively for example, step 310 may
comprise, for example at sub-step 314, receiving a previously
broadcast program from a user recording device (e.g., a PVR, VCR,
etc.) in a time-shifted manner.
[0075] The exemplary method 300 may, for example at step 320,
comprise communicating television programming (e.g., as received at
step 310) to a television (e.g., for presentation to a user on a
screen of the television). Step 320 may, for example, share any or
all characteristics with step 220 of the exemplary method 200
illustrated in FIG. 2 and discussed previously.
[0076] For example, step 320 may comprise, for example at sub-step
322, communicating the received television programming to a
television (e.g., of a television communicatively coupled to a
television receiver implementing the exemplary method 300 or a
portion thereof). Alternatively for example, step 320 may comprise,
for example at sub-step 324, communicating received television
programming to another user device (e.g., different from a
television) for presentation to the user (e.g., to a display device
different from the television, to a television remote control
device with a display, to a user's handheld computer, etc.).
[0077] The exemplary method 300 may, for example at step 330,
comprise determining on-screen pointing location pointed to by a
user of the television and/or television receiver. Step 330 may,
for example, share any or all characteristics with step 230 of the
exemplary method 200 illustrated in FIG. 2 and discussed
previously.
[0078] For example, step 330 may comprise, for example at sub-step
332, analyzing sensor information (e.g., associated with sensors
on-board and/or off-board the television and/or television
receiver) to determine user on-screen pointing location.
Alternatively for example, step 330 may comprise, for example at
sub-step 334, the television receiver receiving information
describing the on-screen pointing location from a device external
to the television receiver (e.g., the television, a television
controller, a television network device, etc.).
[0079] The exemplary method 300 may, for example at step 340,
comprise identifying a user-selectable object in the television
programming (e.g., the television programming communicated at step
320) at which the user is pointing based, at least in part, on the
determined on-screen pointing location (e.g., as determined at step
330). Step 340 may, for example, share any or all characteristics
with step 240 of the exemplary method 200 illustrated in FIG. 2 and
discussed previously.
[0080] For example, step 340 may comprise, for example at sub-step
342, determining the on-screen location of one or more
user-selectable objects in the television programming.
Alternatively for example, step 330 may comprise, for example at
sub-step 344, identifying a user-selected object by analyzing the
respective on-screen locations of the one or more user-selectable
objects and the determined on-screen pointing location (e.g., at a
particular time instance and/or particular timeframe) to determine
the object selected by the user.
[0081] The exemplary method 300 may, for example at step 395,
comprise performing continued operations. Step 395 may, for
example, share any or all characteristics with step 295 of the
exemplary method 200 illustrated in FIG. 2 and discussed
previously.
[0082] Turning next to FIG. 4, such figure is a diagram
illustrating an exemplary television receiver 400, in accordance
with various aspects of the present invention. The exemplary
television receiver 400 may, for example, share any or all
characteristics with the exemplary television receiver 151
illustrated in FIG. 1 and discussed previously. Also, the exemplary
television receiver 400 (e.g., various modules thereof) may operate
to perform any or all of the functionality discussed previously
with regard to the exemplary methods 200 and 300 illustrated in
FIGS. 2-3 and discussed previously.
[0083] The exemplary television receiver 400 includes a first
communication interface module 410. The first communication
interface module 410 may, for example, operate to communicate over
any of a variety of communication media and utilizing any of a
variety of communication protocols. For example, though the first
communication interface module 410 is illustrated coupled to a
wireless RF antenna via a wireless port 412, the wireless medium is
merely illustrative and non-limiting. The first communication
interface module 410 may, for example, operate to communicate with
one or more communication networks (e.g., cable television
networks, satellite television networks, telecommunication
networks, the Internet, local area networks, personal area
networks, metropolitan area networks, etc.) via which television
video content (e.g., television programming) and/or other data is
communicated. Also for example, the first communication interface
module 410 may operate to communicate with local sources of
television video content (e.g., video recorders, receivers, gaming
devices, etc.). Additionally, for example, the first communication
interface module 410 may operate to communicate with a television
controller (e.g., directly or via one or more intermediate
communication networks).
[0084] The exemplary television receiver 400 includes a second
communication interface module 420. The second communication
interface module 420 may, for example, operate to communicate over
any of a variety of communication media and utilizing any of a
variety of communication protocols. For example, the second
communication interface module 420 may communicate via a wireless
RF communication port 422 and antenna, or may communicate via a
non-tethered optical communication port 424 (e.g., utilizing laser
diodes, photodiodes, etc.). Also for example, the second
communication interface module 420 may communicate via a tethered
optical communication port 426 (e.g., utilizing a fiber optic
cable), or may communicate via a wired communication port 428
(e.g., utilizing coaxial cable, twisted pair, HDMI cable, Ethernet
cable, any of a variety of wired component and/or composite video
connections, etc.). The second communication interface module 420
may, for example, operate to communicate with one or more
communication networks (e.g., cable television networks, satellite
television networks, telecommunication networks, the Internet,
local area networks, personal area networks, metropolitan area
networks, etc.) via which television video content (e.g.,
television programming) and/or other data is communicated. Also for
example, the second communication interface module 420 may operate
to communicate with local sources of television video content
(e.g., video recorders, receivers, gaming devices, etc.).
Additionally, for example, the second communication interface
module 420 may operate to communicate with a television controller
(e.g., directly or via one or more intervening communication
networks).
[0085] The exemplary television receiver 400 may also comprise
additional communication interface modules, which are not
illustrated. Such additional communication interface modules may,
for example, share any or all aspects with the first 410 and second
420 communication interface modules discussed above.
[0086] The exemplary television receiver 400 may also comprise a
communication module 430. The communication module 430 may, for
example, operate to control and/or coordinate operation of the
first communication interface module 410 and the second
communication interface module 420 (and/or additional communication
interface modules as needed). The communication module 430 may, for
example, provide a convenient communication interface by which
other components of the television receiver 400 may utilize the
first 410 and second 420 communication interface modules.
Additionally, for example, in an exemplary scenario where a
plurality of communication interface modules are sharing a medium
and/or network, the communication module 430 may coordinate
communications to reduce collisions and/or other interference
between the communication interface modules.
[0087] The exemplary television receiver 400 may additionally
comprise one or more user interface modules 440. The user interface
module 440 may generally operate to provide user interface
functionality to a user of the television receiver 400. For
example, and without limitation, the user interface module 440 may
operate to provide for user control of any or all standard
television receiver commands (e.g., channel control, on/off, input
selection, programming, recording/playback control, etc.). The user
interface module 440 may, for example, operate and/or respond to
user commands utilizing user interface features disposed on the
television receiver 400 (e.g., buttons, etc.) and may also utilize
the communication module 430 (and/or first 410 and second 420
communication interface modules) to communicate with a television
controller (e.g., a dedicated television remote control, a
universal remote control, a cellular telephone, personal computing
device, gaming controller, etc.). Further for example, the user
interface module 440 may utilize the communication module 430
(and/or first 410 and/or second 420 communication interface
modules) to communicate with a television to utilize the user
interface features of the television (e.g., to interface with a
user of the television receiver 400).
[0088] The user interface module 440 may also comprise one or more
sensor modules that operate to interface with and/or control
operation of any of a variety of sensors that may be utilized to
ascertain an on-screen pointing location. For example and without
limitation, the user interface module 440 (or sensor module(s)
thereof) may operate to receive signals associated with respective
sensors (e.g., raw or processed signals directly from the sensors,
through intermediate devices, via the communication interface
modules 410, 420, etc.). Also for example, in scenarios in which
such sensors are active sensors (as opposed to purely passive
sensors), the user interface module 440 (or sensor module(s)
thereof) may operate to control the transmission of signals (e.g.,
RF signals, optical signals, acoustic signals, etc.) from such
sensors. Additionally, the user interface module 440 may perform
any of a variety of video output functions (e.g., in an exemplary
scenario where the television receiver 400 has an onboard display,
in an exemplary scenario where the television receiver 400 is
utilizing a display of another device, etc.).
[0089] The exemplary television receiver 400 may comprise one or
more processors 450. The processor 450 may, for example, comprise a
general purpose processor, digital signal processor,
application-specific processor, microcontroller, microprocessor,
etc. For example, the processor 450 may operate in accordance with
software (or firmware) instructions. As mentioned previously, any
or all functionality discussed herein may be performed by a
processor executing instructions. For example, though various
modules are illustrated as separate blocks or modules in FIG. 4,
such illustrative modules, or a portion thereof, may be implemented
by the processor 450.
[0090] The exemplary television receiver 400 may comprise one or
more memories 460. As discussed above, various aspects may be
performed by one or more processors executing instructions. Such
instructions may, for example, be stored in the one or more
memories 460. Such memory 460 may, for example, comprise
characteristics of any of a variety of types of memory. For example
and without limitation, such memory 460 may comprise one or more
memory chips (e.g., ROM, RAM, EPROM, EEPROM, flash memory,
one-time-programmable OTP memory, etc.), hard drive memory, CD
memory, DVD memory, etc.
[0091] The exemplary television receiver 400 may comprise one or
more modules 452 that operate to perform and/or manage the receipt
and/or communication of television programming. For example, such
one or more modules 452 may operate to utilize the communication
module 430 (e.g., and at least one of the communication interface
modules 410, 420) to receive and/or communicate television
programming. For example, such one or more modules 452 may operate
to perform step 210 of the exemplary method 200 discussed
previously and/or step 310 of the exemplary method 300 discussed
previously.
[0092] Also for example, such one or more modules 452 may operate
to utilize the communication module 430 (and/or first 410 and/or
second 420 communication interface modules) to communicate
television programming to a television for presentation to a user.
Additionally for example, such one or more modules 452 may operate
to utilize the communication module 430 (e.g., and at least one of
the communication interface modules 410, 420) to communicate
television programming or other video output information to one or
more devices communicatively coupled to the television receiver 400
(e.g., devices other than a television). For example, such one or
more modules 452 may operate to perform step 220 of the exemplary
method 200 discussed previously and/or step 320 of the exemplary
method 300 discussed previously.
[0093] The exemplary television receiver 400 may comprise one or
more on-screen pointing location determination module(s) 454. Such
on-screen pointing location determination module(s) 454 may, for
example, operate to determine an on-screen pointing location
pointed to by a user of the television receiver and/or a television
communicatively coupled thereto. Such module(s) 454 may, for
example, operate to perform step 230 of the exemplary method 200
and/or step 330 of the exemplary method 300 discussed previously.
For example, the module(s) 454 may operate to analyze sensor
information to determine an on-screen pointing location. Also for
example, the module(s) 454 may operate to receive on-screen
pointing location information from a device (e.g., a television, a
television controller, a television network device, a user pointing
device, etc.) external to the television receiver 400 (e.g.,
utilizing the communication module 430).
[0094] The exemplary television receiver 400 may comprise one or
more user-selected object identification modules 456. Such
module(s) 456 may, for example, operate to identify a
user-selectable object in television programming at which a user of
the television receiver 400 and/or a television communicatively
coupled thereto is pointing. For example, such module(s) 456 may
operate to identify such user-selected object based, at least in
part, on on-screen pointing location determined by the on-screen
pointing location determination module(s) 454. Such module(s) 456
may, for example, operate to perform step 240 of the exemplary
method 200 and/or step 340 of the exemplary method 300 discussed
previously. For example, the module(s) 456 may operate to determine
the on-screen location of one or more user-selectable objects in
the television programming, and identify a user-selected object by
analyzing the respective on-screen locations of the one or more
user-selectable objects and the determined on-screen pointing
location or in-frame pointing location (e.g., at a particular time
instance and/or particular timeframe) to determine the object
selected by the user.
[0095] Though not illustrated, the exemplary television receiver
400 may, for example, comprise one or more modules that operate to
perform any or all of the continued processing discussed previously
with regard to step 295 of the exemplary method 200 and step 395 of
the exemplary method 300, discussed previously. Such modules (e.g.,
as with the one or more modules 452, 454 and 456) may be performed
by the processor(s) 450 executing instructions stored in the memory
460.
[0096] Turning next to FIG. 5, such figure is a diagram
illustrating exemplary modules and/or sub-modules for a television
receiver 500, in accordance with various aspects of the present
invention. The exemplary television receiver 500 may share any or
all aspects with any of the television receivers 151 and 400
discussed herein and illustrated in FIGS. 1 and 4. For example, the
exemplary television receiver 500 (or various modules thereof) may
operate to perform any or all functionality discussed herein with
regard to the exemplary method 200 illustrated in FIG. 2 and the
exemplary method 300 illustrated in FIG. 3. As with the exemplary
television receiver 400, the components of the exemplary television
receiver 500 may be disposed in a single television receiver (e.g.,
a cable television receiver, a satellite television receiver, an
Internet television receiver, etc.).
[0097] For example, the television receiver 500 comprises a
processor 530. Such a processor 530 may, for example, share any or
all characteristics with the processor 450 discussed with regard to
FIG. 4. Also for example, the television receiver 500 comprises a
memory 540. Such memory 540 may, for example, share any or all
characteristics with the memory 460 discussed with regard to FIG.
4.
[0098] Also for example, the television receiver 500 may comprise
any of a variety of user interface module(s) 550. Such user
interface module(s) 550 may, for example, share any or all
characteristics with the user interface module(s) 440 discussed
previously with regard to FIG. 4. For example and without
limitation, the user interface module(s) 550 may comprise: a
display device, a camera (for still or moving picture acquisition),
a speaker, an earphone (e.g., wired or wireless), a microphone, a
video screen (e.g., a touch screen), a vibrating mechanism, a
keypad, and/or any of a variety of other user interface devices
(e.g., a mouse, a trackball, a touch pad, touch screen, light pen,
game controlling device, etc.).
[0099] The exemplary television receiver 500 may also, for example,
comprise any of a variety of communication modules (505, 506, and
510). Such communication module(s) may, for example, share any or
all characteristics with the communication interface module(s) 410,
420 discussed previously with regard to FIG. 4. For example and
without limitation, the communication interface module(s) 510 may
comprise: a Bluetooth interface module; an IEEE 802.11, 802.15,
802.16 and/or 802.20 module; any of a variety of cellular
telecommunication interface modules (e.g., GSM/GPRS/EDGE,
CDMA/CDMA2000/1x-EV-DO, WCDMA/HSDPA/HSUPA, TDMA/PDC, WiMAX, etc.);
any of a variety of position-related communication interface
modules (e.g., GPS, A-GPS, etc.); any of a variety of
wired/tethered communication interface modules (e.g., USB, Fire
Wire, RS-232, HDMI, Ethernet, wireline and/or cable modem, etc.);
any of a variety of communication interface modules related to
communicating with external memory devices; etc. The exemplary
television receiver 500 is also illustrated as comprising various
wired 506 and/or wireless 505 front-end modules that may, for
example, be included in the communication interface modules and/or
utilized thereby.
[0100] The exemplary television receiver 500 may also comprise any
of a variety of signal processing module(s) 590. Such signal
processing module(s) 590 may share any or all characteristics with
modules of the exemplary television receiver 400 that perform
signal processing. Such signal processing module(s) 590 may, for
example, be utilized to assist in processing various types of
information discussed previously (e.g., with regard to sensor
processing, position determination, video processing, image
processing, audio processing, general user interface information
data processing, etc.). For example and without limitation, the
signal processing module(s) 590 may comprise: video/graphics
processing modules (e.g. MPEG-2, MPEG-4, H.263, H.264, JPEG, TIFF,
3-D, 2-D, MDDI, etc.); audio processing modules (e.g., MP3, AAC,
MIDI, QCELP, AMR, CMX, etc.); and/or tactile processing modules
(e.g., keypad I/O, touch screen processing, motor control,
etc.).
[0101] In summary, various aspects of the present invention provide
a system and method in a television receiver (e.g., external to a
television) for providing user-selection of objects in a television
program. While the invention has been described with reference to
certain aspects and embodiments, it will be understood by those
skilled in the art that various changes may be made and equivalents
may be substituted without departing from the scope of the
invention. In addition, many modifications may be made to adapt a
particular situation or material to the teachings of the invention
without departing from its scope. Therefore, it is intended that
the invention not be limited to the particular embodiment
disclosed, but that the invention will include all embodiments
falling within the scope of the appended claims.
* * * * *