U.S. patent application number 15/350218 was filed with the patent office on 2017-04-06 for system and method for providing event spectators with audio/video signals pertaining to remote events.
The applicant listed for this patent is immersion Entertainment, LLC. Invention is credited to Geoffrey L. Anderson, Tazwell L. Anderson, JR., Robert J. Kniskern, Todd H. Steele, Mark A. Wood.
Application Number | 20170099510 15/350218 |
Document ID | / |
Family ID | 36250839 |
Filed Date | 2017-04-06 |
United States Patent
Application |
20170099510 |
Kind Code |
A1 |
Anderson, JR.; Tazwell L. ;
et al. |
April 6, 2017 |
SYSTEM AND METHOD FOR PROVIDING EVENT SPECTATORS WITH AUDIO/VIDEO
SIGNALS PERTAINING TO REMOTE EVENTS
Abstract
A system and method for providing content for access by a
portable device at an event are provided. The system includes an
interface to receive signals from a plurality of sources. The
plurality of sources are located at an event and remote from the
event. The signals define event content and non-event content. The
system further includes a processor for processing the received
signals to generate combined signals based upon the received
signals from the plurality of sources. The system also includes a
transmitter for transmitting to a plurality of portable devices at
the event the combined signals. The combined signals are configured
to allow a user to select content using the portable device
receiving the combined signals.
Inventors: |
Anderson, JR.; Tazwell L.;
(Atlanta, GA) ; Anderson; Geoffrey L.; (Atlanta,
GA) ; Wood; Mark A.; (Boca Raton, FL) ;
Kniskern; Robert J.; (Fort Wayne, IN) ; Steele; Todd
H.; (Fort Wayne, IN) |
|
Applicant: |
Name |
City |
State |
Country |
Type |
immersion Entertainment, LLC |
Atlanta |
GA |
US |
|
|
Family ID: |
36250839 |
Appl. No.: |
15/350218 |
Filed: |
November 14, 2016 |
Related U.S. Patent Documents
|
|
|
|
|
|
Application
Number |
Filing Date |
Patent Number |
|
|
14593895 |
Jan 9, 2015 |
|
|
|
15350218 |
|
|
|
|
13076013 |
Mar 30, 2011 |
8725064 |
|
|
14593895 |
|
|
|
|
12557725 |
Sep 11, 2009 |
7929903 |
|
|
13076013 |
|
|
|
|
11070353 |
Mar 2, 2005 |
7593687 |
|
|
12557725 |
|
|
|
|
10680612 |
Oct 7, 2003 |
7725073 |
|
|
11070353 |
|
|
|
|
Current U.S.
Class: |
1/1 |
Current CPC
Class: |
H04N 21/4627 20130101;
H04N 21/41407 20130101; H04N 21/6125 20130101; H04L 65/80 20130101;
H04N 21/2665 20130101; H04N 21/43637 20130101; H04N 21/6118
20130101; H04N 21/6143 20130101; H04N 21/2365 20130101; H04N
21/2541 20130101; H04N 21/252 20130101; H04L 29/06027 20130101;
H04N 21/2223 20130101; H04N 7/181 20130101; H04L 65/4076 20130101;
H04N 21/6137 20130101; H04H 20/61 20130101; H04N 21/4347 20130101;
H04L 65/605 20130101; H04N 21/25841 20130101 |
International
Class: |
H04N 21/254 20060101
H04N021/254; H04N 21/4363 20060101 H04N021/4363; H04N 21/61
20060101 H04N021/61; H04N 21/4627 20060101 H04N021/4627; H04N
21/414 20060101 H04N021/414 |
Claims
1-64. (canceled)
65. A system for providing entertainment event content to a
portable device, the system configured to communicate with at least
one facility that receives video signals defining event content
associated with multiple entertainment events of a similar type,
the system comprising: multiple antenna located to service
different coverage areas, the multiple antenna including a first
antenna having a first coverage area where a first user is located;
a processor configured to determine a criteria regarding the
multiple entertainment events, the multiple entertainment events
including a remote entertainment event that is remote from the
first user and the first coverage area; the processor configured to
designate the first antenna to receive a first video signal
associated with the remote entertainment event indicated by the
criteria; and a router capable of routing the first video signal,
associated with the remote entertainment event, to the first
antenna; and the first antenna configured to wirelessly transmit
the first video signal to at least a first portable device within a
corresponding first coverage area, the first video signal
configured to allow the user to view the event content on the first
portable device; wherein the first antenna is configured to service
a coverage area at a first entertainment event and the first
antenna is configured to support active two-way communication with
the first portable device through which the user actively makes a
request for the first video signal from the system.
66. (canceled)
66. The system of claim 65, wherein the router routes first video
event content to at least a first antenna and second video event
content to at least a second antenna, the first and second antenna
defining WiFi nodes at a first entertainment event.
67. The system of claim 65, wherein one of the corresponding
coverage areas include a stadium where a first entertainment event
occurs.
68. The system of claim 65, wherein the first antenna is located
proximate to a stadium where a local entertainment event is
occurring and a second antenna is located remote from the local
entertainment event, the first and second antenna transmitting
local event content.
69. The system of claim 65, wherein sources comprise at least one
of over air television, event cameras, AM/FM radio, satellite,
VCR/DVD, cable television, satellite television, and Internet.
70. The system of claim 65, wherein the processor is configured to
designate antenna of a cellular network to receive the video
signal.
71. The system of claim 65, further comprising a modulator
configured to modulate multiple video signals on different
channels.
72. The system of claim 65, wherein the first portable device
represents a cellular phone and the first antenna communicates
through WiFi with the cellular phone.
73. The system of claim 65, wherein the video signal includes
real-time video event content from a live entertainment event and
recorded video event content from a prior entertainment event,
where the live and prior entertainment events are a same type of
event.
74. The system of claim 65, where the first antenna defines a WiFi
node with the corresponding coverage area at a first entertainment
event, the video signal including the event content associated with
the first entertainment event.
75. A method for routing entertainment event content to a portable
device from a system that receives video signals defining event
content associated with multiple entertainment events of a similar
type, the method comprising: connecting to multiple antenna located
to service different coverage areas, the multiple antenna including
a first antenna having a first coverage area where a first user is
located; determining a criteria regarding the multiple
entertainment events, the multiple entertainment events including a
remote entertainment event that is remote from the first user and
the first coverage area; designating the first antenna to receive a
first video signal associated with the remote entertainment event
indicated by the criteria; routing the first video signal,
associated with the remote entertainment event, to the first
antenna; wirelessly transmitting, from the first antenna, the video
signal to at least a first portable device within the coverage area
of the first antenna; and configuring the video signal to allow a
user of the first portable device to view the event content on the
first portable device; wherein the first antenna is configured to
service a coverage area at a first entertainment event and the
first antenna is configured to support active two-way communication
with the first portable device through which the user actively
makes a request for the video signal from the system.
76. The method of claim 75, wherein the first video signal
represents real-time video event content from the remote
entertainment event.
77. The method of claim 75, wherein the antenna are located to
service corresponding coverage areas within a stadium where a first
entertainment event occurs.
78. The method of claim 75, further comprising receiving a user
request for user specified event content and routing the video
signal to the user in responses to the user specified event
content.
79. The method of claim 75, further comprising wirelessly
transmitting a first video signal to a first portable device in
response to a user request.
80. The method of claim 75, wherein at least a portion of the
antenna are configured to service coverage areas remote from the
entertainment event and are configured to wirelessly transmit the
video signal to cellular telephones while at locations other than
at the entertainment event.
81. The method of claim 75, wherein at least one antenna is
configured to service a coverage area at a first entertainment
event and the at least one antenna is configured to support active
two-way communication with the first portable device through which
the user actively makes a request for the video signal from the
system.
82. The method of claim 75, wherein the video signal includes video
event content from a first video source at a live entertainment
event and audio event content from a first audio source at the live
entertainment event.
83. The method of claim 75, wherein the antenna and first portable
device communicate through WiFi communications.
84. The method of claim 75, wherein the first antenna defines a
WiFi node with the corresponding coverage area at a first
entertainment event, the video signal including the event content
associated with the first entertainment event.
85. The system of claim 75, wherein the first video signal
represents real-time video event content from the remote
entertainment event.
Description
CROSS REFERENCE TO RELATED APPLICATIONS
[0001] This application is a continuation and claims priority to
copending Non-Provisional U.S. patent application assigned Ser. No.
12/557,725 filed Sep. 11, 2009, which is a continuation and claims
priority to copending Non-Provisional U.S. patent application
assigned Ser. No. 11/070,353 filed Mar. 2, 2005, now U.S. Pat. No.
7,593,687, which is a continuation-in-part and claims priority to
copending Non-Provisional U.S. patent application assigned Ser. No.
10/680,612 and filed on Oct. 7, 2003, now U.S. Pat. No. 7,725,073
all entitled "SYSTEM AND METHOD FOR PROVIDING EVENT SPECTATORS WITH
AUDIO/VIDEO SIGNALS PERTAINING TO REMOTE EVENTS" and all of which
are hereby incorporated by reference in their entirety.
BACKGROUND OF THE INVENTION
[0002] This invention relates generally to video and audio signal
processing techniques and, more particularly, to methods and
systems for receiving video and audio signals from a plurality of
sources and for providing a user with combinations of these signals
from which to select.
[0003] Spectators attending a stadium for viewing an event, such as
a sporting event, for example, are often interested in events
simultaneously occurring at other locations, such as, other
stadiums. As an example, a spectator attending a football game at a
stadium may be interested in other football games at other
stadiums, particularly at other stadiums within the same general
region, for example, hearing commentary or viewing images of other
games.
[0004] Radio stations may broadcast AM and/or FM audio signals of
stadium events. Such audio signals normally define the comments of
one or more commentators who watch a stadium event and comment on
the event as it is occurring such that listeners of the audio
broadcast may receive real-time information pertaining to the
stadium event. For sporting events, the commentary provided by such
commentators is often referred to as a "play-by-play"
commentary.
[0005] Moreover, a spectator at one stadium may utilize a radio to
receive at least one audio event signal from a radio station, which
may be any audio signal that provides information (e.g., real-time
and/or prerecorded information) pertaining to the event. The radio
broadcast allows a spectator to hear information pertaining to an
event occurring at another stadium. Unfortunately, an audio event
signal pertaining to a stadium event of interest to a spectator is
not always broadcast by a radio station using AM or FM signals,
thereby preventing spectators outside the stadium from receiving
the broadcast. Further, in situations where an audio event signal
of interest to a spectator is broadcast in such a manner that it
may be received, the spectator may be attending a stadium event out
of range of the audio event signal of interest. Thus, a spectator
at one stadium is not always able to receive an audio event signal
providing detailed audio information about an event occurring at
another stadium. A spectator likewise may be unable to receive a
video event signal providing detailed video information about an
event occurring at another stadium.
[0006] Further, spectators may be interested in viewing or
accessing other content while at an event. For example, a spectator
may want to check the news during half-time of a game. However,
known systems do not allow spectators to selectively access
different content from different sources while at an event.
BRIEF DESCRIPTION OF THE INVENTION
[0007] In one exemplary embodiment, a system for providing content
for access by a portable device at an event includes an interface
to receive signals from a plurality of sources. The plurality of
sources are located at an event and remote from the event. The
signals define event content and non-event content. The system
further includes a processor for processing the received signals to
generate combined signals based upon the received signals from the
plurality of sources. The system also includes a transmitter for
transmitting to a plurality of portable devices at the event the
combined signals. The combined signals are configured to allow a
user to select content using the portable device receiving the
combined signals.
[0008] In another exemplary embodiment, a method for providing
content for access by a portable device at an event includes
receiving at an event a plurality of signals from a plurality of
sources. The plurality of signals define event content and
non-event content. The method further includes determining received
signals to combine to define at least one combined signal. The
method also includes broadcasting the at least one combined signal
at the event, with the at least one combined signal configured to
be received by at least one portable device.
[0009] In yet another exemplary embodiment, a method for providing
content for access by a user at an event includes receiving via a
portable device event content and non-event content defined by at
least one signal combined from a plurality of signals from a
plurality of sources. The method further includes outputting at
least one of video and audio content based upon at the at least one
signal.
BRIEF DESCRIPTION OF THE DRAWINGS
[0010] FIG. 1 is a block diagram of a communication system in
accordance with one exemplary embodiment of the present
invention.
[0011] FIG. 2 is a block diagram of a central facility system in
accordance with one exemplary embodiment of the present
invention.
[0012] FIG. 3 is a block diagram of a broadcast system in
accordance with one exemplary embodiment of the present
invention.
[0013] FIG. 4 is a flowchart of exemplary processes utilized by
various embodiments of a communication system of the present
invention.
[0014] FIG. 5 is a block diagram of a local broadcast system in
accordance with one exemplary embodiment of the present
invention.
[0015] FIG. 6 is a flowchart of an exemplary process for accessing
and broadcasting content from a plurality of sources at an event in
accordance with an exemplary embodiment of the present
invention.
[0016] FIG. 7 is a block diagram of one exemplary embodiment of a
video/audio receiving system constructed according to the
principles of the present invention;
[0017] FIG. 8 is a block diagram of one exemplary embodiment of an
image display system of FIG. 1;
[0018] FIG. 9 is a block diagram of another exemplary embodiment of
a video/audio receiving system constructed according to the
principles of the present invention.
DETAILED DESCRIPTION OF THE INVENTION
[0019] Exemplary embodiments of methods and systems providing
spectators at an event with a plurality of audio and video event
signals pertaining to other remote events and other content, and
from which a spectator may select one or more of the audio and
video event signals are described below in detail. A technical
effect of the methods and systems described herein include at least
one of generating a plurality of different audio and/or video
signals at one or more remote locations (e.g., generating multiple
audio signal events at different stadiums, with the audio signal
events defining information, such as play-by-play commentary
pertaining to events at the different stadiums), transmitting the
plurality of audio and/or video signals to a central facility,
wherein the audio and/or video signals are selectively transmitted
to broadcast systems that each service one or more events (e.g.,
one or more stadiums), and allowing spectators to select from the
audio and/or video signals transmitted via the broadcast
systems.
[0020] It should be noted that when reference is made herein to
audio and/or video event signals, this refers to any audio or video
signal generated and/or transmitted to and/or from an event, such
as, for example, audio and video signals that provide information,
including, but not limited to, real-time and/or prerecorded
information, pertaining to an event or other content from different
sources. It should also be noted that when reference is made herein
to non-event or non-event related audio and/or video signals, this
refers to any live, delayed, or archived and recorded audio and/or
video signals such as for example, live audio and/or video signals
being generated from another event, rebroadcast audio and/or video
signals from an another event, any other live or recorded
audio/video content, such as, for example, content relating to
sporting events, entertainment events (e.g. movies, cartoons,
advertisements, statistics and data) and or other third party
content (e.g., news).
[0021] In the exemplary embodiments, the methods generally provide
spectators at an event with a selection of audio and/or video event
signals pertaining to other remote events that may be of interest
to the spectators. In particular, a plurality of different audio
and/or video event signals are generated at various locations
(e.g., remote locations). For example, multiple audio event signals
may be generated at different stadiums, and each audio event signal
may define information, such as a play-by-play commentary
pertaining to the event occurring at the stadium from which the
audio event signal is generated or from another stadium.
[0022] Each of the audio and/or video event signals may be
transmitted to a central facility, and various communication
techniques may be employed to communicate the audio and/or video
event signals to the central facility. In an exemplary embodiment,
audio and/or video event signals are communicated to a central
facility via a communication network (e.g., an existing or
specifically designed network), such as the publicly switched
telephone network (PSTN). In order to increase performance, the
audio and/or video event signals may be transmitted over a high
speed communication link, such as, for example, a T1 communication
line. It should be noted that the communication links utilized to
transmit the audio and/or video event signals to the central
facility are not limited to specific types of links or physical
land lines, but may include, for example, terrestrial links and/or
satellite links.
[0023] At the central facility, the audio and/or video event
signals may be selectively coalesced and transmitted to a plurality
of broadcast systems that each service one or more events and/or
stadiums. For example, a determination may be made that spectators
at a particular stadium are likely to be interested in particular
ones of the audio event signals. In an exemplary embodiment, such
audio event signals are then transmitted to the broadcast system
servicing the particular stadium. The broadcast system then may
broadcast the audio event signal to the spectators being serviced
by the broadcast system. For example, a determination may be made
that spectators at a first stadium for viewing a Southeastern
Conference (SEC) football game are interested in other SEC football
games occurring at other stadiums. In such a case, audio event
signals defining the play-by-play commentary of the other SEC
football games may be transmitted, by the central facility, to the
broadcast system servicing the first stadium. The broadcast system
may then broadcast these audio event signals to the spectators at
the first stadium.
[0024] For other stadiums, different sets of audio event signals
may be provided. For example, for a stadium where a BIG 10 football
game is occurring, the central facility may transmit to the
broadcast system servicing that stadium, audio event signals
pertaining to other BIG 10 football games. These audio event
signals may then be broadcast to the spectators of the stadium via
the broadcast system. Further, and as another example, for a
stadium where a Major League Baseball (MLB) game is occurring,
audio event signals pertaining to other MLB baseball games may be
transmitted to the broadcast system servicing the stadium. Various
combinations of audio event signals thereby may be provided to
spectators at various stadiums and are not limited to the
combinations described herein. It should be noted that video event
signals likewise may be communicated as described herein.
[0025] Transmission of a set of audio and/or video event signals to
a particular broadcast system may occur via various transmission
methodologies. In an exemplary embodiment, the audio and/or video
event signals are communicated to the broadcast systems via a
communication network (e.g., an existing or specifically designed
network), such as, for example, the PSTN. To provide increased
performance, the audio and/or video event signals may be
transmitted over high speed communication links such as, for
example, a T1 communication line. It should be noted that the
communication links utilized to transmit the audio and/or video
event signals to the broadcast systems are not limited to specific
types of links or physical land lines, but may include, for
example, terrestrial links and/or satellite links. Furthermore,
audio and/or video event signals may be generated as or may be
converted into digital signals in order to enable better and/or
faster transmission to and/or from a broadcast system.
[0026] In an exemplary embodiment, when a broadcast system receives
audio and/or video event signals from the central facility, the
broadcast system may communicate such signals to spectators at the
stadium serviced by the broadcast system. The communication of such
signals may be wireless and/or non-wireless. For example, the
broadcast system may modulate each of the audio and/or video event
signals on a carrier signal of a different frequency and wirelessly
transmit the modulated signals to the stadium spectators. The
stadium spectators may be provided with portable receiver units
that demodulate and filter the audio and/or video event signals
transmitted by the broadcast system. Specifically, using a receiver
unit, a spectator may select one or more of the audio and/or video
event signals received by the receiver unit, and the receiver unit
may filter out all of the other audio and/or video event signals,
thereby recovering the selected audio and/or video event signal or
signals. The recovered audio and/or video event signal or signals
may then be converted into sound via one or more speakers of the
receiver unit or images via a display of the receiver unit.
[0027] It should be noted that a broadcast system and, in
particular, an antenna of the broadcast system may be located at or
close to the stadium serviced by it. In such an embodiment, the
audio and/or video event signals transmitted by the broadcast
system can be transmitted over a short distance (e.g., less than
approximately one-half of a mile), thereby limiting various
restrictions imposed by regulation agencies, such as the Federal
Communication Commission (FCC). Various frequency ranges may be
employed, for example, each broadcast system in an exemplary
embodiment may transmit audio and/or video event signals within the
ultra-high frequency (UHF) range, microwave range (e.g., two to ten
gigahertz range), etc.
[0028] FIGS. 1 through 3 illustrate exemplary embodiments of a
communication system 10 that will be described in more detail
below. For illustrative purposes, the communication system 10 will
be described as enabling spectators attending NCAA football games
at stadiums to select and hear a play-by-play commentary and/or see
video images of various football games. However, it should be noted
that, in other embodiments, the communication system 10 may enable
spectators to listen to and view other types of information
pertaining to other types of events (e.g., football, auto racing,
golf, horse racing, etc.) located at different venues (e.g., race
track, open road race course, golf course, etc.) or other types of
information (e.g., advertisements).
[0029] FIG. 1 is a block diagram of a communication system 10 of an
exemplary embodiment of the present invention. The communication
system includes a central facility 12 for receiving audio and/or
video signals (e.g., audio and/or video feeds and audio and/or
video event signals) from a plurality of sources (e.g., local event
and remote sources). The central facility 12 combines various audio
and/or video signals as described herein for transmission to one or
more broadcast systems 32. The broadcast systems 32 will be
described in more detail with reference to FIGS. 5 and 6. In an
exemplary embodiment, a separate broadcast system is provided in
connection with or corresponding to each of a plurality of events
or, for example, a plurality of stadiums. The broadcast systems 32
transmit a plurality of audio and/or video signals to receiver
units 65, which allow users attending events, for example, at
various stadiums to select audio and or video signals from one or
more sources. The audio and video signals define content for audio
output or display with the receiver units 65. The sources may be
located at the event at which the spectator is attending and
providing audio and video signals pertaining to that event, or may
be located remote from the event attended by the spectator, and
provide audio and video signals pertaining to the remote event.
Further, the sources may provide audio and video signals pertaining
to other content, such as, for example, advertising.
[0030] As shown in FIG. 2, audio and/or video signals, for example,
play-by-play feeds defining play-by-play commentaries of various
football games are transmitted to the central facility 12. These
play-by-play feeds may be transmitted via satellite, T1, and/or
other types of communication links. Further, the play-by-play feeds
may be transmitted as digital or analog signals. In an exemplary
embodiment, the play-by-play feeds are generated as digital signals
or converted into digital signals before being transmitted to the
central facility 12.
[0031] The central facility 12, in an exemplary embodiment,
includes a feed mixer 15 and a data router 18. The feed mixer 15
combines the audio and/or video feeds and transmits the audio
and/or video feeds to the data router 18. The feed mixer 15
combines the play-by-play feeds such that, for example,
play-by-play feeds that are likely to be of interest to a group of
spectators or users are combined together. For example, the
play-by-play feeds may be mixed such that play-by-play feeds
associated with games or teams within or from the same region are
combined (e.g., multiplexed) together.
[0032] For example, stadiums 13 participating in (e.g., serviced
by) the communication system 10 may be categorized based on their
regional location. As shown in FIG. 2, the stadiums 13 may be
categorized into a plurality of regions 14, such as first through
fifth regions (e.g., a Southeast Region, a Southern Region, a
Midwest Region, a Central Region, and a Western Region). Each
play-by-play feed from the same region may be combined by the feed
mixer 15. Therefore, the feed mixer 15 may output combined signals
in which each combined signal comprises play-by-play feeds from the
same region. For example, one of the combined signals may comprise
play-by-play feeds from each stadium within the Southern Region
that is hosting a football game. Such a combined signal may
comprise play-by-play feeds from Neyland Stadium, Bryant-Denny
Stadium, Griffin Stadium, and Memorial Stadium (Clemson).
[0033] The mixing of play-by-play feeds as described herein may be
controlled by mixing logic 25. In one exemplary embodiment, the
mixing logic 25 (e.g., user interface such as a computer) receives
inputs from a user and mixes the play-by-play feeds based on these
inputs, thereby allowing a user to control which feeds are mixed
together. In another exemplary embodiment, user preferences (e.g.,
viewing preferences) or predetermined criteria (e.g., closest
games) are used control which feeds are mixed together. Further,
another content insertion control logic 27 may be configured to
insert, for example, advertisements and/or commercials into the
feeds mixed by the feed mixer 15. It should be noted that the
mixing logic 25 and the other content insertion control logic 27
may each be implemented in hardware, software, or any combination
thereof.
[0034] The data router 18 may be configured to transmit each of the
combined signals to, for example, corresponding events or ones of
the stadiums. For example, a combined signal including feeds likely
to be of interest to spectators at Neyland Stadium may be
transmitted to Neyland stadium. Because spectators attending a
sporting event at a stadium are likely to be interested in similar
sporting events occurring within the same approximate region, a
combined signal having one or more play-by-play feeds from a
particular region may be transmitted to the stadiums within the
same region. For example, a combined signal including play-by-play
feeds from football games within the Southern Region may be
transmitted to one or more stadiums within this same region. As a
result, the play-by-play feeds defining the combined signal
received by a particular stadium are likely to be of interest to
the spectators at the particular stadium. In the exemplary
embodiment shown in FIG. 2, the combined signals are communicated
to the stadiums over a telecommunication network using high speed
communication links, such as T1 connections. However, other types
of communication techniques may be employed in other embodiments.
Further, it should be noted that various other methodologies for
determining which stadiums are to receive which feeds may be
employed in other exemplary embodiments as desired or needed and as
described herein.
[0035] Referring now to FIG. 3, and for example, for each event or
stadium, an exemplary embodiment of a broadcast system 32 receives
the combined signal transmitted to it by the central facility 12
(shown in FIG. 2). A demultiplexer 35, controlled by control logic
37, demultiplexer the combined signal to thereby recover, for
example, each play-by-play feed within the combined signal. A
digital-to-analog (D/A) converter 41 converts each recovered
play-by-play feed into analog form and provides each of the analog
feeds to a modulator 45. Each modulator 45 modulates its received
play-by-play feed onto a carrier signal of a different frequency,
and a signal combiner 48 combines the modulated signals, which are
then amplified by a power amplifier 51, before being wirelessly
transmitted by an antenna 52. It should be noted that if the
combined signal received by the broadcast system 32 of a particular
stadium or associated with a particular event does not include, for
example, a feed pertaining to the event occurring at the same
stadium, a local play-by-play feed 49 generated at the stadium may
be included in the feeds modulated by the modulators 45.
[0036] In one exemplary embodiment, one or more spectators, for
example, at the stadium, is provided with a receiver unit 65
capable of receiving and demodulating the signals transmitted by
the broadcast system 32. The selected audio and/or video signals,
such as play-by-play feeds are then output by the receiver unit 65
(e.g., output as audio from a speaker or video from a display),
thereby reproducing, for example, the play-by-play commentary or
video images defined by the selected feed(s). For example, assume
that a spectator at Neyland Stadium, while watching a Tennessee
game occurring at Neyland Stadium, is interested in the football
game occurring at Griffin Stadium. In such an example, the
spectator may tune his receiver unit 65 (e.g., select a channel)
such that the receiver unit 65 filters out all play-by-play feeds
except the feed originally generated at Griffin stadium. As a
result, the receiver unit 65 outputs the play-by-play commentary of
the football game occurring at Griffin Stadium. Moreover, the
spectator may listen to this commentary while watching the
Tennessee football game or another game on the receiver unit
65.
[0037] In an exemplary embodiment, the receiver unit 65 is portable
and allows the user to provide inputs for selecting at least one or
more of the audio and/or video signals, such as, for example,
play-by-play feeds and/or video signals of interest to the user. In
another exemplary embodiment as shown in FIG. 3, a receiver unit 65
includes a receiver and converter component 70 that receives and
converts each audio and/or video signal received from broadcast
system 32 via antenna 52 and antenna 72. The receiver unit 65 also
includes a tuner 74 for tuning the audio and/or video feed provided
on different frequencies such that feeds on different frequencies
can be output from the receiver unit 65.
[0038] Additional components for processing the different types of
feeds (e.g., audio and video signals) may be provided, such as, for
example, an FM receiver component 76 for receiving and processing
frequency-modulated signals. The receiver unit 65 also may include
user controls for controlling operation of the receiver unit 65
including a selector (not shown) for selecting between different
feeds and sources, a display (not shown) for displaying event
images from received video signals, a power on/off component 78
(e.g., switch) for powering on and off the receiver unit 65, and a
volume up/down component 80 for controlling the volume output of
the receiver unit 65. The receiver unit 65 also includes a power
supply 82 (e.g., removable battery, rechargeable battery, etc.) for
powering the receiver unit 65. It should be noted that although the
receiver unit 65 has been described with particular component
parts, it is not so limited, and additional or different component
parts may be provided as desired or required (e.g., storage
component for storing audio and/or video feeds).
[0039] It also should be noted that it is not necessary for the
broadcast system 32 used to transmit signals, for example, to
spectators at a particular stadium, to be located at or close to
the particular stadium. However, locating the broadcast system 32
at or close to the stadium may be desired or needed (e.g., results
in less strenuous FCC restrictions). Additionally, it is not
necessary for the central facility 12 to selectively determine
which of the audio and/or video signals (e.g., play-by-play feeds)
are to be provided, for example, to the different stadiums. For
example, the central facility 12 may be configured to provide each
audio and/or vide signal to each of the stadiums regardless of each
stadium's regional location. Further, it should be noted that video
event signals in addition to or instead of audio event signals may
be processed by the communication system 10 described above. For
example, in addition to or instead of generating and providing a
play-by-play audio signal of an event, a camera may generate a
play-by-play video signal of the event. This video signal then may
be transmitted to the central facility 12, which mixes the video
signal with, for example, other audio and/or video signals from
other events. After being transmitted to a broadcast system 32, the
broadcast system 32 may modulate and transmit the video signal to,
for example, spectators at the stadium, similar to the manner in
which the audio signals are modulated and transmitted to spectators
at a stadium, as described above. The receiver unit 65 may include
a video display for displaying the video signal when such signal is
selected by the spectator's input. An exemplary system for
distributing the audio and/or video signals is described in U.S.
Pat. No. 6,578,203, the entire disclosure of which is hereby
incorporated by reference herein. For example, the system described
therein may be employed to provide to spectators at a stadium the
audio and/or video signals received from the central facility
12.
[0040] Moreover, the spectator may view a play-by-play video signal
from an event at a remote stadium and may simultaneously listen to
a play-by-play audio signal from the same event. Alternatively, the
spectator may view a play-by-play video signal from an event at a
remote stadium and may simultaneously listen to a play-by-play
audio signal from another event at another remote stadium or from
the stadium at which the spectator is present. Furthermore, it is
not necessary for all of the signals processed by the communication
system 10 to pertain to an event occurring at a stadium. For
example, a spectator may view a play-by-play video signal
pertaining to an event at a remote stadium while listening to an
audio signal defining music unrelated to any of the events at any
of the stadiums or a video signal defining programming unrelated to
any of the events at any of the stadiums (e.g., advertisement or
commercial).
[0041] FIG. 4 is a flowchart 100 illustrating exemplary processes
utilized by the communication system 10. The technical effect of
the communication system 10 is achieved by first receiving a
plurality of audio and video signals from one or more sources
(e.g., audio and video signals from a plurality of stadiums) at
102. A determination is then made at 104 as to which audio and
video signals to combine. This may include, for example,
categorizing the audio and video signals based upon the geographic
location of the source of the signals (e.g., signals from stadiums
within a particular geographic region or relating to teams in that
region) and/or combining the signals based upon predetermined
criteria (e.g., combine signals based upon user preferences). It
should be noted that the audio and video signals may be combined
based upon other criteria, determinations or factors, such as, for
example, time of day (e.g., events occurring at the same time), the
type of event (e.g., all sports events, all football games, all
baseball games, etc.) and/or variety (e.g., combine signals to
provide diverse programming). A determination is also made at 106
as to other content to be included as part of the signals to be
combined (e.g., advertisements or commercials).
[0042] A determination is then made at 108 as to which broadcast
systems 32 (shown in FIG. 1) the signals to be combined are to be
transmitted. This also may include a determination of the type of
communication link to use to transmit the combined signals. Based
upon the determinations at 104, 106, 108 and any other user inputs
(e.g., used specified content), combined signals are transmitted to
the determined broadcast systems 32 at 110.
[0043] The combined signals are received by the broadcast systems
32 and processed (e.g., demultiplexed and modulated in analog form)
at 112, for transmission (e.g., local transmission) to a one or
more receiver units 65 (shown in FIG. 3), for example, within and
in close proximity to a stadium. This processing may also include
modulating the signals onto different frequencies. The processed
signals are then broadcast at 114 to the receiver units 65 within
the broadcasting range of the broadcast systems 32. For example,
the signals may be broadcast to receiver units 65 within a stadium
or within one hundred feet of the stadium. A user having a receiver
unit 65 may then select the specific signals, for example, defining
different feeds to output from the receiver unit 65.
[0044] In operation, a user with a receiver unit 65 at an event
(e.g., football game) may select from different audio and video
feeds that are broadcast at the event. For example, a user may
select audio and/or video signals defining content from another
event (e.g., another football game), video feeds of different
angles or portions of the event being attended and/or other content
such advertising and/or commercials. The audio and/or video content
may be accessed real-time or stored for later hearing and/or
viewing (e.g., to review a football play or listen to specific
commentary). The selectable audio and video signals may be
provided, for example, on predetermined channels selectable by a
user. Further, the sources of these audio and video signals may be
located at the event, remote from the event, or accessed from
stored or archived audio and video signals. Additionally, the
communication system 10 may be configured such that the receiver
unit 65 may receive audio and video signals at locations other than
at an event (e.g., at a stadium event), such as, for example, at a
home in connection with satellite television broadcasting. Further,
conditional access may be provided such that, for example, only
authorized receiver units 65 (e.g., receiver units 65 having
authorized serial numbers) can receive the audio and video signals
(e.g., encrypted audio and video signals).
[0045] It should be noted that various embodiments of the present
invention also allow users with, for example, portable devices, and
more particularly, wireless portable devices to receive video and
audio content from a plurality of sources and to select at the
portable wireless device the content that the user desires or needs
to access. For example, a user may select from programming from
different sources provided on a plurality of selectable channels.
Using the various embodiments of the present invention, a
transmission or broadcast system may provide a plurality of
combined audio/video signals at an event for access by a user with
a wireless portable device.
[0046] Thus, spectators at an event may be provided with a
selection of audio and/or video event signals pertaining to other
remote events and other non-event related content that may be of
interest to the spectators. In particular, a plurality of different
audio and/or video event signals are received from various
locations and sources (e.g., remote locations, cable TV channels,
Satellite TV channels, Satellite or Digital radio, VCR or DVD
players). These different signals may be provided, for example, by
the communication system 10 via one or more broadcast systems 32.
For example, multiple audio event signals may be received from
different events being provided via a Satellite Radio, and each
audio event signal may define information, such as a play-by-play
commentary pertaining to the event occurring at the stadium from
which the audio event signal is generated or from another
stadium.
[0047] Each of the audio and/or video event signals and each of the
non-event remote signals may be received, monitored and then
selected for broadcast by the broadcast system 32 at an event based
on the relevance of the content, or by the particular desires and
preferences of the event spectators that will be using portable
receiving devices. For example, a determination may be made that
spectators at a particular stadium are likely to be interested in
particular ones of the audio event signals. In an exemplary
embodiment, such audio event signals are then transmitted to the
broadcast system 32 servicing the particular stadium. The broadcast
system 32 then may broadcast the audio event signal to the
spectators being serviced by the broadcast system 32. For example,
a determination may be made that spectators at a first stadium for
viewing a Southeastern Conference (SEC) football game are
interested in other SEC football games occurring at other stadiums.
In such a case, audio event signals defining the play-by-play
commentary of the other SEC football games may be transmitted to
the broadcast system 32 servicing the first stadium. The broadcast
system 32 may then broadcast these audio event signals to the
spectators at the first stadium.
[0048] For other stadiums, different sets of audio event signals
may be received, monitored and selected for broadcast at the event.
For example, for a stadium where a BIG 10 football game is
occurring, a Satellite Radio system could be monitored for audio
event signals pertaining to other BIG 10 football games. These
audio event signals may then be selected for broadcast to the
spectators of the stadium via the broadcast system 32. Further, and
as another example, for a stadium where a Major League Baseball
(MLB) game is occurring, audio and/or video event signals
pertaining to other MLB baseball games being provided by a Cable TV
system may be monitored and selected for broadcast by the broadcast
system 32 servicing the stadium. Various combinations of audio
and/or video event signals thereby may be provided to spectators at
various stadiums and are not limited to the combinations described
herein.
[0049] In addition to providing a user with live, non-event audio
and/or video content from another sporting or entertainment event,
other non-event related content can be monitored via a plurality of
broadcast sources that are connected to the broadcast system 32 at
the event. These other broadcast sources (e.g. Cable TV, Satellite
TV, Satellite or Digital Radio, Local Broadcast TV or Radio) can be
monitored for relevant non-event related content (e.g. news,
weather, sports tickers and scores, movies, cartoons,
advertisements, statistics and data) that can then be selected for
broadcast at a specific event by the broadcast system 32 located
at, for example, a stadium.
[0050] In an exemplary embodiment, when the broadcast system 32
receives audio and/or video event signals from various input
sources, the broadcast system 32 may communicate such signals to
spectators at the stadium serviced by the broadcast system. The
communication of such signals may be wireless and/or non-wireless.
For example, the broadcast system 32 may modulate each of the audio
and/or video event signals on a carrier signal of a different
frequency and wirelessly transmit the modulated signals to the
stadium spectators. The stadium spectators may be provided with
portable receiver units that demodulate and filter the audio and/or
video event signals transmitted by the broadcast system 32.
Specifically, using a receiver unit, a spectator may select one or
more of the audio and/or video event signals received by the
receiver unit, and the receiver unit may filter out all of the
other audio and/or video event signals, thereby recovering the
selected audio and/or video event signal or signals. The recovered
audio and/or video event signal or signals may then be converted
into sound via one or more speakers of the receiver unit or images
via a display of the receiver unit.
[0051] It should be noted that a broadcast system 32 and, in
particular, an antenna of the broadcast system may be located at or
close to the stadium serviced by it. In such an embodiment, the
audio and/or video event signals transmitted by the broadcast
system 32 can be transmitted over a short distance (e.g., less than
approximately one-half of a mile), thereby limiting various
restrictions imposed by regulation agencies, such as the Federal
Communication Commission (FCC). Various frequency ranges may be
employed, for example, each broadcast system 32 in an exemplary
embodiment may transmit audio and/or video event signals within the
ultra-high frequency (UHF) range, microwave range (e.g., two to ten
gigahertz range), etc.
[0052] FIG. 5 illustrates an exemplary embodiment showing the
various input sources for both event and non-event audio and/or
video signals that are monitored and selected for broadcast by the
broadcast system located at the event, also referred to herein as a
local broadcast system. For illustrative purposes, the broadcast
system will be described as enabling spectators attending NCAA
football games at stadiums to select and hear a play-by-play
commentary and/or see video images of various football games, as
well as access other non-event content. However, it should be noted
that, in other embodiments, the broadcast system may enable
spectators to listen to and view other types of information
pertaining to other types of events (e.g., football, auto racing,
golf, horse racing, etc.) located at different venues (e.g., race
track, open road race course, golf course, etc.) or other types of
information (e.g., news or stock quotes).
[0053] FIG. 5 is a block diagram of a local broadcast system 32 of
an exemplary embodiment of the present invention. The broadcast
system 32 generally includes means for receiving audio and/or video
signals (e.g., audio and/or video feeds and audio and/or video
event signals, non-event content, recorded or archived content)
from a plurality of sources 152 (e.g., local event and remote
sources). The broadcast system 32 then may combine various audio
and/or video signals as described herein for broadcast of a
plurality of audio and/or video signals to receiver units 180 or
receiver units 65 (shown in FIG. 3), which allow users attending
events, for example, at various stadiums to select audio and or
video signals from one or more sources. The audio and video signals
define content for audio output or display with the receiver units.
The sources may be located at the event at which the spectator is
attending and providing audio and video signals pertaining to that
event, or may be located remote from the event attended by the
spectator, and provide audio and video signals pertaining to the
remote event. Further, the sources may provide audio and video
signals pertaining to other content, such as, for example,
non-event content such as television programming.
[0054] In one exemplary embodiment, one or more spectators, for
example, at the stadium, is provided with the receiver unit 180
capable of receiving and demodulating the signals transmitted by
the broadcast system 32. The selected audio and/or video signals,
such as play-by-play feeds, advertising, television programming,
etc. are then output by the receiver unit 180 (e.g., output as
audio from a speaker or video from a display), thereby reproducing,
for example, the play-by-play commentary or video images defined by
the selected feed(s). For example, assume that a spectator at
Neyland Stadium, while watching a Tennessee game occurring at
Neyland Stadium, is interested in the football game occurring at
Griffin Stadium. In such an example, the spectator may tune his
receiver unit 180 (e.g., select a channel) such that the receiver
unit filters out all play-by-play feeds except the feed originally
generated at Griffin stadium. As a result, the receiver unit
outputs the play-by-play commentary of the football game occurring
at Griffin Stadium. Moreover, the spectator may listen to this
commentary while watching the Tennessee football game or another
game on the receiver unit 180. As another example, the spectator
may tune the receiver unit 180 to receive a television feed or to
access Internet content.
[0055] It also should be noted that it is not necessary for the
system used to transmit signals, for example, to spectators at a
particular stadium, to be located at or close to the particular
stadium. However, locating the broadcast system at or close to the
stadium may be desired or needed (e.g., results in less strenuous
FCC restrictions). Further, it should be noted that any type of
video and/or audio event and/or non-event signals may be processed
by the broadcast system 32. For example, in addition to or instead
of generating and providing a play-by-play audio signal of an
event, a camera may generate a play-by-play video signal of the
event or a satellite system may generate satellite television
programming. This video content then may be routed to the broadcast
system 32 at the event, which mixes the video signal with, for
example, other audio and/or video signals from other events or
non-event content or signal.
[0056] Specifically, the broadcast system 32 receives content in
the form of audio and video signals from a plurality of sources
152. In one exemplary embodiment, the system may be a mobile
broadcast system (e.g., within a broadcast van) that includes an
interface 154, shown as an audio/video patch panel for receiving
signals from each of the various sources. It should be noted that
broadcast system 32 may be a fixed system or partially fixed (e.g.,
some of the components fixed at a stadium and other components
within a mobile unit). A plurality of multiplexers 156 receive the
signals from the interface 154. The multiplexers 156 combine the
signals as described in more detail herein. Each of the
multiplexers 156 outputs the combined signals to a monitoring
system 158 and a quality control system 160. The monitoring system
158 may include a plurality of displays 162 (e.g., monitors with a
broadcast van) for viewing, for example, by a program controller
operating the broadcast system 32 for monitoring the content. The
quality control system 160 may include a plurality of signal
analyzers 164 or other components for measuring signal quality and
maintaining a level of quality assurance/quality control (QA/QC).
For example, one or more oscilloscopes, spectrum analyzers and/or
video monitors may be provided for use onsite to ensure that the
quality and strength of each multiplexed signal is within or
satisfies a predetermined or predefined specification or level
prior to be being combined and transmitted.
[0057] The outputs of the multiplexers 156, and in particular the
multiplexed signals are provided to a plurality of frequency
modulators 166 for modulating the plurality of multiplexed signals
on different frequencies. The modulated signals are then provided
to a combiner 168 for combining the signals and transmission via a
transmitter 170 using, for example, an antenna 172. In various
embodiments, the multiplexers 156, frequency modulators 166 and
combiner 168 together may define, or be embodied in, one or more
processors.
[0058] Thus, the broadcast system 32, after receiving a plurality
of signals from a plurality of sources, modulates and transmits
selected signals to, for example, spectators at the stadium having
receiver units 180.
[0059] FIG. 6 is a flowchart 200 illustrating exemplary processes
utilized by the broadcast system 32. A plurality of audio and video
signals are received from a plurality of sources at 202. This
includes receiving audio and video signals defining content from,
for example, both local and remote sources, such as Cable TV,
Satellite TV, Satellite or Digital Radio, Over the Air TV or Radio,
VCR or DVD Players, and generally defining event content and
non-event content. A determination is then made at 204 as to which
audio and video signals to use (i.e., event related content). This
may include using signal selection criteria for an event at 206 to
make a determination of which signals to use. This criteria may
include, for example, spectator or regional preferences,
information relating to the geographic location of the source of
the signals (e.g., signals from stadiums within a particular
geographic region or relating to teams in that region) and/or other
predetermined criteria (e.g., type of game being played). It should
be noted that the audio and video signals may be selected based
upon other criteria, determinations or factors, such as, for
example, time of day (e.g., events occurring at the same time), the
type of event (e.g., all sports events, all football games, all
baseball games, etc.) and/or variety (e.g., select signals to
provide diverse programming). At 208 additional content (i.e.,
non-event related content) available for use is identified and at
210 selection criteria for determining which additional content to
include is used to determine additional content to be included as
part of the signals to be broadcast (e.g., news from an over the
air source) at 212.
[0060] At 214 a transmission frequency is assigned to each of the
signals and at 216 the signals are multiplexed. The multiplexed
signals are combined at 218 and the combined signals are broadcast
at the event at 220. The combined signals are received at 222 by
portable devices at the event and at 224 a user may select
particular content to be displayed and/or output as sound by the
portable device.
[0061] Additionally, the various embodiments described herein are
not limited for use in connection with particular systems or for
receiving signals from specific sources, but may receive signals
from other sources, such as, for example, the system for
distributing audio and/or video signals described in U.S. Pat. No.
6,578,203, the entire disclosure of which is hereby incorporated by
reference herein.
[0062] FIG. 7 illustrates an exemplary embodiment of a video/audio
receiving system 712 constructed according to the principles of the
present invention. At least one video signal 714 and at least one
audio signal 715 are received by a receiver 716 via a signal
interface 718. Each of the video signals 714 defines, for example,
a view of an event, such as from different perspectives. For
example, the video signals 714 may be generated by different video
cameras located at different locations at an event (e.g.,
positioned at different locations around a stadium or racetrack, at
various holes at a golf tournament, close-up to or on the stage at
a concert). Furthermore, each of the audio signals 715 defines
different sounds associated with an event. For example, at least
one of the audio signals 715 may be generated from a microphone
located close to the sideline of a game or in one of the helmets of
one of the players of the game (e.g., the helmet of a football
player) such that the audio signal defines sounds from the
participants in the game, or may be generated from a microphone in
a pit area of a racetrack with the audio signal defining sounds
from the pit crew. Alternatively, at least one of the audio signals
715 may define the comments of television commentators, and at
least one of the audio signals may define the comments of radio
commentators. Further, and for example, the video signals 714 may
define live television broadcasts of the event.
[0063] It should be noted that the video and audio signals 714 and
715 may be received from a plurality of different sources (e.g.,
local broadcast, closed circuit broadcast at the event, cable
television, satellite broadcast and the Internet) and define
content related to the event being attended or another event. It
should also be noted that the video and audio signals 714 and 715
are not limited to providing images and sounds of one event or the
event being attended. Event content defined by the video and audio
signals 714 and 715 and/or other signals (e.g., data signals) may
include, but is not limited to, audio/video from other events,
public television broadcasts, cable television broadcasts,
satellite broadcasts, Internet data, such as, for example, emails
or news, and interactive media or data, such as, for example,
trivia contests or other games.
[0064] In particular, at least one of the audio and one of the
video signals may be transmitted as a single combined signal from
an audio/video system such as described in U.S. Pat. No. 6,578,203
entitled "Audio/Video Signal Distribution System for Head Mounted
Displays," the entire disclosure of which is hereby incorporated
herein by reference. Additionally, one or more of the video and/or
audio signals may be wireless, in which case, the signal interface
718 may include one or more antennas for receiving the wireless
signals. However, various other types of signal interfaces 718 are
contemplated. For example, the signal interface 718 may be a cable
or other type of wired or signal transmission apparatus. Any type
of wireless and/or non-wireless technique may be used to transmit
signals to receiver 716 via the signal interface 718.
[0065] Some of the video and audio signals 714 and 715 may be
unmodulated when transmitted to the receiver 716 through the signal
interface 718 and, therefore, do not need to be demodulated by the
system. However, some of the video signals 714 and/or audio signals
715 may be modulated when received by the receiver 716, and,
therefore, may need to be demodulated by the system 712. For
example, at least one of the audio signals 715 defining the
comments of radio commentators may be modulated as a radio signal
for transmission to radios located at or away from the event, and
at least one of the video signals 714 may be modulated as a
television signal for transmission to televisions located at or
away from the event. Therefore, as shown in FIG. 7, the a
video/audio receiving system 712 preferably includes a demodulator
720 configured to demodulate modulated video signals 714 and/or
audio signals 715 received by the receiver 716 through the signal
interface 718.
[0066] Once demodulated, if necessary, the video and audio signals
714 and 715 are processed by signal processing logic 722, which
selects and conditions the signals 714 and 715. Specifically, the
signal processing logic 722 selects, based on inputs from the user
as described herein, one or more of the video signals 714 and one
or more of the audio signals 715. It should be noted that the
signal processing logic 722 may be implemented via hardware,
software, or a combination thereof. Further, the signal processing
logic 722 may include one or more filters for filtering out
unselected signals 714 and 715. After selecting one of the video
and audio signals 714 and 7107, the signal processing logic 722
conditions the selected video signals 714 such that they are
compatible with an image display system 730, and the signal
processing logic 722 conditions the selected audio signals 715 such
that they are compatible with speakers 734. The signal processing
logic 722 then transmits the conditioned audio signals 715 to the
speakers 734, which converts the conditioned audio signals 715 into
sound. The signal processing logic 722 also transmits the
conditioned video signals 714 to the image display system 730,
which displays the image defined by the conditioned video signals
714 according to techniques known in the art. It should be noted
that the processing performed by the signal processing logic 722
may be provided as described in U.S. Pat. No. 6,578,203.
[0067] Further, an input device 724, which may include, for
example, one or more buttons, knobs, dials, or other types of
switches, may be used to provide the inputs for the processing
performed by the signal processing logic 722. It should be noted
that these exemplary input devices 724 may be interchanged,
modified or replaced with other input devices as desired or needed.
By controlling the components of the input device 724, the user may
control various aspects of the processing performed by the signal
processing logic 722, including which video signals 714 are
selected for viewing, as well as which audio signals 715 are heard
and the volume of the audio signals 715.
[0068] FIG. 8 illustrates an exemplary embodiment of an image
display system 730 constructed according to the principles of the
present invention. Specifically, a processed video signal 714 is
displayed onto a Liquid Crystal Display (LCD) 734. The LCD 734 may
be lit from the back via a backlight 736, with the light shining
through the LCD 734, creating an image on the other side of the LCD
734. On the opposite side of the LCD 734 from the backlight 736, a
distance from the LCD 734, is a half-silvered mirror 738. The
half-silvered mirror 738 is set at an approximately forty-five
degree angle from the LCD 734. The image reflects off the
half-silvered mirror 738 onto a separate curved mirror 740 set a
distance away from the half-silvered mirror 738. The curved mirror
740 magnifies the image. The magnified image reflects off of the
curved mirror 740, back to the half-silvered mirror 738. The
magnified image passes through the half-silvered mirror 738 to a
lens 742 located on the opposite side of the half-silvered mirror
738 from the curved mirror 740. The magnified image passes through
the lens 7133, which focuses the magnified image.
[0069] Other embodiments of a video/audio receiving system having
additional or different components and performing additional or
different functions are contemplated (e.g., enhanced viewing
capabilities using binocular functions or video/audio storage
capabilities). Specifically, in another exemplary embodiment, a
video/audio receiving system 712' as shown in FIG. 9 includes a
front end tuner/receiver 760 provided for receiving a signal (e.g.,
modulated RF signal from an antenna within a receiving device)
containing video signals 714 and/or audio signals 7107, or a
combination thereof. A processor, such as, for example, a digital
processor 762 processes the received signal to provide video
signals 714 defining images for display via a viewing system 764.
The digital processor 762 may process the received signals to
provide audio signals 715 defining audio for output by the handheld
device via an audio system 763 (e.g., output using speakers or to
headphone connected to an audio jack). In one embodiment, the
video/audio receiving system 712' includes a memory 766 for storing
video or audio content as described herein. A power supply 768 is
also provided for powering the video/audio receiving system 712',
and specifically the digital processor 762 and memory 766. The
video/audio receiving system 712' also includes an optics system
770 for capturing images of an event, which are then processed by
the digital processor 762 for display on the viewing system 764 or
storage within the memory 766.
[0070] In particular, in one embodiment, the front end
tuner/receiver 760 includes a digital video receiver/demodulator
(i.e. tuner) that enables the video/audio receiving system 712' to
receive both digital video and audio signals transmitted, for
example, over standardized television, Wireless Fidelity (WiFi), or
other RF frequency bands. It should be noted that the received
broadcast signal provides live and recorded video and audio
content, and may include processor serial number specific enabling
codes to indicate whether a particular video/audio receiving system
712' is permitted to receive and display the broadcast signal (i.e.
conditional access). Thus, conditional access allowing for both
rental of devices containing the video/audio receiving system 712'
and/or pay per view functionality when devices are owned by a user
44 may be provided.
[0071] The digital video/audio output of the front end
tuner/receiver 760 is provided to the digital processor 762,
wherein the received signals are processed (e.g., conditioned) for
display on the viewing system 764 or for storing in the memory 766
for later access and display. The front end tuner/receiver 760 is
configured to receive transmissions having different transmission
requirements, such as, for example, from 8 Virtual Side Band (8VSB)
commercial television broadcasts, Coded Orthogonal Frequency
Division Multiplex (COFDM) commercial television broadcasts and/or
locally transmitted event content, such as provided using the
system described in U.S. Pat. No. 6,578,203. Further, the front end
tuner/receiver 760 also provides received audio signals to the
digital processor for processing and outputting processed digital
audio outputs for listening by a user or for storage.
[0072] The digital processor 762 is configured for processing video
and audio signals 714 and 715 from the front end tuner/receiver
760. Further, the memory 766 and the optics system 770 are
configured such that processed video and audio signals 714 and
7107, which may include, for example, live view, real and recorded
video and stored video and digital images, may be viewed using the
viewing system 764 (e.g., via an LCD). The digital processor 762
interfaces directly with both the front end tuner/receiver 760 and
the optics system 770 such that a user, via hardware and/or
software controlled using a user input 767, can select the desired
viewing or audio input. The user input 767, may include, for
example, one or more buttons, knobs, dials, or other types of
switches. It should be noted that these exemplary user inputs 767
may be interchanged, modified or replaced with other user inputs as
desired or needed.
[0073] Additionally, the output of the digital processor 762, for
example in the form of still images or continuous video, may be
stored in the memory 766. The stored images/video may then be
available, for example, for future viewing using the viewing system
764, or downloading to a printer or computer for further
processing.
[0074] User control of the video/audio receiving system 712' to
control the operation of the digital processor 762 may be provided
by a user input 767 (e.g., a standard NSEW four position button)
provided as part of a handheld device. The user input 767, such as,
for example, a multi-function button select system allows the user
to select the mode of operation (e.g., broadcast video, binocular,
digital camera with various telephoto settings, record and
playback), as well as other features specific to a particular mode.
For example, this may include telephoto options, video record time,
start, stop, and rewind; image store (e.g., take a picture); store
a continuous view (e.g., camcorder recording), etc. Additionally,
the user input buttons may be used to control other functions, such
as, for example, volume and video channel selection.
[0075] In one exemplary embodiment, the optics system 770 includes
two fixed focus lenses each providing a signal to a charge coupled
device (CCD). The CCD converts the focused optical signal into a
digital signal that is processed by the digital processor 762 for
display using the viewing system 764. In operation, the two fixed
focus lenses enable, for example, a wide field view and a telephoto
view, depending on the selection made by a user via the user input
767. For example, the optical zoom allows for a higher resolution
zoom capability than an electronic zoom, in which a portion of the
signal received by the CCD is expanded or "blown-up" to provide
zoom capability. Thus, by including two lens/CCD subsystems, both
optical and electronic zoom capabilities may be provided that
allows for different settings (e.g., wide field (optical),
telephoto 1 (digital from the wide field lens), telephoto 2
(optical telephoto), and telephoto 3 (digital from the optical
telephoto lens)).
[0076] Viewing system 764 receives processed signals from the
digital processor 762 or processed signals stored in the memory
766, and using "near-to-the-eye" optics, provides a user with an
image (e.g., video image) of the processed signals. Using known
displays and associated optics, a video image is provided such that
a user appears to be viewing an image that is much larger than
actually displayed. It should be noted that the viewing system 764
displays the output of the digital processor 762 based upon any of
the video/audio/optical input sources.
[0077] The memory 766 may be provided using permanent memory,
removable memory (e.g., DRAM), or a combination of both, such that
a user may store single images and/or continuous video. The stored
images and/or continuous video may be, for example, reviewed or
replayed at the event to ensure that the contents stored is what is
desired or needed by the user or to allow a user to view part of
the event again (e.g., view a close call in a football game). In
one embodiment, removable memory may be provided, such as, for
example, a memory stick/cartridge that may be removed by a user
after use at the event. Other interfaces may also be provided to
access the images and/or continuous videos stored in the memory
766, such as a USB connector allowing for the downloading of the
stored memory (e.g., captured video) to a computer.
[0078] In one exemplary embodiment, the power supply 768 includes a
rechargeable battery, such as a rechargeable Li Ion or Li Polymer
battery that may be permanent or removable from a device for
recharging. The power supply 768 may also include a recharge outlet
for recharging a battery while still in the device using a standard
AC/DC power converter. The power supply 768 may also include a
smaller replaceable battery (e.g. NiCad battery), that provides
constant power to the memory 766 to ensure that a user's settings
are not lost when main battery power falls below a predetermined
operating level.
[0079] In the various embodiments of the present invention, the
video/audio receiving systems 712 and 712' are embodied within
portable devices. Using handheld devices for viewing video signals
714, a user's viewing experience may be enhanced. For example, when
using the handheld devices, a field view of the game from a camera
located on another side of the stadium may be selected by a user,
thereby allowing the user 44 to see a similar view as spectators
located in that portion of the stadium. Further, in some
embodiments, because the handheld devices may limit the user's
peripheral view of the environment around him or her, the user 44
focuses on the view provided by the handheld devices. In these
embodiments, when the user 44 desires to view the event (e.g.,
game) directly, the user may quickly lower the handheld device so
that the user's view of the game is not obstructed. It should be
noted that the handheld devices may enhance a user's experience at
any event, such as, for example, any sporting event or other event
where a user 44 is unable to view the entire event (e.g., unable to
see the entire racetrack).
[0080] Furthermore, because the handheld devices are handheld, they
are easily portable, and the user 44 may carry the handheld devices
with him or her, and choose where he or she would like to view the
images produced by the handheld devices. For example, the user 44
may walk throughout a stadium with the handheld device in hand
while intermittently viewing the images and hearing the sounds
produced by the video/audio receiving system 712 or 712'. Further,
by manipulating user input 767, the user 44 may control which video
signals 714 are displayed and which audio signals 715 are produced
by the video/audio receiving systems 712 or 712'. Accordingly, the
handheld devices provide the user 44 more flexibility to observe
and listen to an event, such as a sporting event, and results in a
more enjoyable experience.
[0081] Different types of materials (e.g., part molded and part
flexible material), casings or housings for the handheld devices
may be employed to implement the various embodiments of the present
invention.
[0082] Thus, a user 44 may be provided with different video and
audio content associated with an event at which the user is
attending, another event of interest, or content of interest to the
viewer (e.g., business news, cartoons, etc.). Further, a user 44
may select different options for viewing the event (e.g., binocular
viewing) or store some or all of the event content (e.g., store
still images of the event).
[0083] It is not necessary for the user 44 to keep the handheld
device within the confines of the event (e.g., within the stadium).
For example, the video and audio signals 714 and 715 may be
transmitted via satellites and/or communication networks to various
locations around the world, and the user 44 may select the view he
or she prefers the most from a remote location capable of receiving
a video signal 714 and/or audio signal 715.
[0084] The handheld device also may be retrieved from the user 44
after the user 44 is finished viewing the event so that the
handheld device can be provided to another spectator for another
event, for example, at another stadium. Each user 44 may be charged
a usage fee for use of the handheld device, or alternatively, a
user may purchase the handheld device and pay a monthly
subscription fee for use of the device. In some of the various
embodiments, payment of the fee may be required before the user 44
is provided with the handheld device. In other embodiments, the
handheld device may receive information, via video and audio
signals 714 and 7107, or otherwise, indicating whether the handheld
device is authorized to produce sounds and images defined by the
signals (e.g., authorized serial number transmitted to the
video/audio system 712 or 712'). In this embodiment, the handheld
device is configured to produce images and sound only when
authorized, with authorization information (e.g., authorization
code unique to a handheld device) transmitted from an audio/video
transmitter using an audio/video system such as described in U.S.
Pat. No. 6,578,203, to the handheld device, only when the user 44
of the handheld device has provided proper payment.
[0085] It should also be noted that various modifications and
changes may be made to the various embodiments of the present
invention. For example, the signal interface 718 and front end
tuner/receiver 760 may be constructed using wireless fidelity WiFi
hardware and software for receiving transmission of content
provided on different bands (e.g., 2.4 GHz, 5.8 GHz, or 10 GHz),
instead of or in addition to a UHF TV frequency band (e.g., 400
MHz-800 MHz). Thus, the handheld device may operate and receive
content via lower UHF frequency bands or higher WiFi bands as
desired or needed.
[0086] Further, use of WiFi also allows for passive and active
two-way communication. For example, in operation using passive
communication, automatic "device to device" communication is
transparent to a user 44. This communication may include, for
example, communication of conditional access codes, collection of
statistical data regarding viewing habits, etc. With respect to
using active communication, interactive communication may be
provided in which the user 44 actively makes requests for
information, purchases, etc., which requests are conveyed to the
system for further action. This also may include, for example,
accessing the Internet or email. Thus, different types of data
embodied in different signals, instead of or in addition to the
video and audio signals 714 and 7107, may be implemented within the
various embodiments of the present invention (e.g., transmitted and
received by the handheld device).
[0087] Further, it should be noted that using the video/audio
system 712 or 712' in connection with the handheld devices allows
for operation of a venue based transmission system in which signals
from a production facility not located at the event may be
provided. For example, the signals, such as the video and audio
signals 714 and 7107, may be available via the Internet or
satellite with a transmission system operated and monitored
remotely from the production facility. Further, and for example, at
least one of the video signals 714 and one of the audio signals 715
may be transmitted as a single combined signal from an audio/video
system such as described in U.S. Pat. No. 6,578,203 and that is
provided at a production facility not located at the event.
Thereafter, transmission is provided via multiple WiFi nodes at the
event. The production facility may receive its content via
satellite download.
[0088] Further, the portable device, including the portable
receiver unit 65 or 180, may be any device having a receiver
configured to receive signals from the various embodiments of the
present invention described herein. For example, the wireless
portable device may be any suitable device capable of displaying
images or outputting sounds based on signals transmitted from the
system 50. These devices include, but are not limited to, laptop
computers, electronic tablet devices, cellular telephones, personal
digital assistants (PDAs) and/or other handheld portable devices
having wireless capabilities. Further examples of these devices,
include, but are not limited to, the device described in co-pending
U.S. application entitled "Electronic Handheld Audio/Video Receiver
and Listening/Viewing Device" filed on Apr. 18, 2001 and having
Ser. No. 09/837,128, the entire disclosure of which is hereby
incorporated by reference herein; the device described in
co-pending U.S. application "Audio/Video System and Method
Utilizing a Head Mounted Apparatus with Noise Attenuation" filed on
Aug. 31, 1999 and having Ser. No. 09/386,613, the entire disclosure
of which is hereby incorporated by reference herein; and the device
described in co-pending U.S. application entitled "Electronic
Handheld Audio/Video Receiver and Listening/Viewing Device" filed
on Jul. 30, 2003 and having Ser. No. 10/630,069, the entire
disclosure of which is hereby incorporated by reference herein.
[0089] Thus, in operation, various embodiments of the present
invention allow a user at an event to select and view and/or listen
to event content or non-event content from a plurality of sources.
The plurality of sources may be located locally at the event or
remote from the event.
[0090] It should be noted that the component parts of the various
embodiments may be modified as desired or needed. For example, the
interface may be modified to accommodate signals from additional or
different sources.
[0091] While the invention has been described in terms of various
specific embodiments, those skilled in the art will recognize that
the invention can be practiced with modification.
* * * * *