U.S. patent application number 17/153173 was filed with the patent office on 2021-05-13 for method and apparatus for managing advertisement content and personal content.
This patent application is currently assigned to AT&T Intellectual Property I, L.P.. The applicant listed for this patent is AT&T Intellectual Property I, L.P.. Invention is credited to Andrea Basso, Tara Hines, Aleksey Ivanov, Jeffrey Mikan, Nadia Morris.
Application Number | 20210142375 17/153173 |
Document ID | / |
Family ID | 1000005347321 |
Filed Date | 2021-05-13 |
United States Patent
Application |
20210142375 |
Kind Code |
A1 |
Ivanov; Aleksey ; et
al. |
May 13, 2021 |
METHOD AND APPARATUS FOR MANAGING ADVERTISEMENT CONTENT AND
PERSONAL CONTENT
Abstract
A system that incorporates teachings of the present disclosure
may include, for example, monitoring access by communication
devices of second users to personal content associated with a first
user, selecting advertisement content based on the monitoring of
the access, and providing the advertisement content for
presentation with the personal content when the personal content is
accessed by a communication device of a third user. Other
embodiments are disclosed.
Inventors: |
Ivanov; Aleksey;
(Middletown, NJ) ; Basso; Andrea; (Torino, IT)
; Hines; Tara; (New York, NY) ; Mikan;
Jeffrey; (ATLANTA, GA) ; Morris; Nadia;
(Houston, TX) |
|
Applicant: |
Name |
City |
State |
Country |
Type |
AT&T Intellectual Property I, L.P. |
Atlanta |
GA |
US |
|
|
Assignee: |
AT&T Intellectual Property I,
L.P.
Atlanta
GA
|
Family ID: |
1000005347321 |
Appl. No.: |
17/153173 |
Filed: |
January 20, 2021 |
Related U.S. Patent Documents
|
|
|
|
|
|
Application
Number |
Filing Date |
Patent Number |
|
|
15714492 |
Sep 25, 2017 |
10929900 |
|
|
17153173 |
|
|
|
|
13207872 |
Aug 11, 2011 |
9799061 |
|
|
15714492 |
|
|
|
|
Current U.S.
Class: |
1/1 |
Current CPC
Class: |
G06Q 30/06 20130101 |
International
Class: |
G06Q 30/06 20060101
G06Q030/06 |
Claims
1. A method comprising: generating, by a system comprising a
processor, personal content associated with a first user, the
personal content being generated based on sensory information
received by the system from a first mobile communication device,
wherein the first mobile communication device determines
information about a sensory environment, wherein the first mobile
communication device broadcasts a wireless notice signal requesting
the sensory information, wherein the broadcasting the wireless
notice signal comprises transmitting the wireless notice signal
according to a signal strength, wherein the signal strength is
selected by the first mobile communication device based on a size
of the sensory environment so that the wireless notice signal is
received by a group mobile communication devices that are in
proximity to the sensory environment, wherein the sensory
information is received by the first mobile communication device
from a group of mobile communication devices responsive to the
wireless notice signal, wherein the generating is based on an
identification of an object captured in one or more images
associated with the sensory environment of the first mobile
communication device, the first mobile communication device being
associated with a first user; adjusting, by the system, the object
in the personal content, wherein the adjusting includes enabling
user input to select the object when the personal content is
presented; monitoring, by the system, access by communication
devices of second users to the personal content; tracking, by the
system, portions of the personal content that are accessed by the
second users accessing the personal content to identify particular
portions of the personal content; evaluating, by the system, a
follower base of the first user to determine an exposure of the
personal content, the follower base comprising users accessing the
personal content; determining, by the system, a value for
advertising in the personal content based on the monitoring of the
access, the particular portions, and the evaluating of the follower
base, the value being based on a geographical area across which the
second users who access the personal content are spread, the value
being relatively lower in a first case that the geographic area is
a first size and the value being relatively higher in a second case
that the geographic area is a second size that is larger than the
first size, the determining resulting in a determined value; and
providing, by the system, an offer to an entity to include
advertisement content with the personal content, the offer being
based on the determined value.
2. The method of claim 1, wherein, the identification of the object
is based on image recognition applied by the system to the one or
more images.
3. The method of claim 2, further comprising analyzing, by the
first mobile communication device, the information about the
sensory environment to identify obstructions in the sensory
environment and selecting the strength of the wireless notice
signal by selecting a stronger strength of the wireless notice
signal based on the analyzing.
4. The method of claim 3, further comprising: identifying the
entity based on an association with the object.
5. The method of claim 4, further comprising: adjusting the
personal content to provide a link, in accordance with a selection
of the object by a user input, to a website associated with the
entity to purchase the object.
6. The method of claim 1, further comprising: providing the
advertisement content for presentation with the personal content
responsive to the personal content being accessed by communication
devices of third users; determining a number of times a selectable
portion of the advertisement content is accessed by the
communication devices of the third users; and responsive to the
number of times the selectable portion of the advertisement content
is accessed by the communication devices of the third users,
adjusting the advertisement content.
7. The method of claim 1, wherein the monitoring of the access
comprises determining a number of times that the personal content
is accessed by the communication devices of the second users, and
wherein the adjusting the object further comprises providing an
indicator in the one or more images to indicate that the object is
selectable.
8. The method of claim 7, wherein the advertisement content is
adjusted based on the number of times that the personal content is
accessed by the communication devices of the second users.
9. The method of claim 1, further comprising: determining a sale
associated with the advertisement content; and calculating revenue
for the first user based on the determining of the sale.
10. The method of claim 1, further comprising: providing the
advertisement content for presentation with the personal content
responsive to the personal content being accessed by communication
devices of third users; and calculating revenue based on a sale
associated with the advertisement content, for a service provider
that provides the access to the personal content for the
communication devices of the third users.
11. The method of claim 10, further comprising: receiving a user
selection associated with the advertisement content, wherein a
determination of the sale is associated with the user
selection.
12. A device comprising: a processing system including a processor;
and a memory that stores executable instructions that, when
executed by the processing system, facilitate performance of
operations, the operations comprising: receiving information about
a sensory environment of a first mobile communication device;
receiving sensory information from a group of mobile communication
devices including a first mobile communication device associated
with a first user, the sensory information including one or more
images associated with the sensory environment of the first mobile
communication device; generating personal content associated with
the first user, the personal content being generated based on an
identification of an object captured in one or more images
associated with the sensory environment of the first mobile
communication device; adjusting the object in the personal content,
wherein the adjusting includes enabling user input to select the
object when the personal content is presented; monitoring access by
a plurality of communication devices of a respective plurality of
second users to the personal content; tracking portions of the
personal content that are accessed by the plurality of second users
accessing the personal content to identify particular portions of
the personal content; determining a follower base of the first user
to identify an exposure of the personal content, the determining
based on users accessing the personal content; establishing a value
for advertising in the personal content, the value being
established based on a spread rate of the personal content
according to a set period, the monitoring of the access, the
particular portions, and the determining of the follower base, the
value being based on a geographical area across which the second
users who access the personal content are spread, the establishing
resulting in an established value; and providing an offer to an
entity to include advertisement content with the personal content,
the offer being based on the established value.
13. The device of claim 12, wherein the identification of the
object is based on image recognition applied by the processing
system to the one or more images.
14. The device of claim 12, wherein the generating the personal
content is further based on the sensory information from the group
of mobile communication devices.
15. The device of claim 14, wherein the one or more images
comprises a plurality of images, and wherein the sensory
information comprises the plurality of images.
16. The device of claim 15, wherein the operations further
comprise: identifying the entity based on an association with the
object; and adjusting the personal content to provide a link, in
accordance with a selection of the object by a user input, to a
website associated with the entity to purchase the object.
17. A non-transitory machine-readable storage medium comprising
executable instructions that, when executed by a system including a
processor, facilitate performance of operations, the operations
comprising: receiving information about a sensory environment of a
first mobile communication device; receiving one or more images
associated with the sensory environment of the first mobile
communication device; generating personal content associated with a
first user, the personal content being generated based on an
identification of an object captured in the one or more images
associated with the sensory environment of the first mobile
communication device, the first mobile communication device being
associated with the first user; adjusting the object in the
personal content to enabling user input to select the object when
the personal content is presented; monitoring access to the
personal content by a plurality of communication devices associated
with a respective plurality of second users; identifying particular
portions of the personal content by tracking portions of the
personal content that are accessed by the plurality of second users
accessing the personal content; evaluating a follower base of users
accessing the personal content of the first user to determine an
exposure of the personal content; determining a value for
advertising in the personal content, the value being determined
based on the monitoring access, the particular portions that are
identified, and the evaluating of the follower base, the value
being based on a geographical area across which the second users
who access the personal content are spread, the determining
resulting in a determined value; and providing an offer to an
entity, the offer comprising including advertisement content with
the personal content, and the offer being based on the determined
value.
18. The non-transitory machine-readable storage medium of claim 17,
wherein the operations further comprise: providing the
advertisement content, wherein the advertisement content is
provided for presentation with the personal content responsive to
the personal content being accessed by communication devices of
third users, and wherein the advertisement content comprises a
selectable portion; determining a number of times the selectable
portion of the advertisement content is accessed by the
communication devices of the third users; and responsive to the
number of times the selectable portion of the advertisement content
is accessed by the communication devices of the third users,
adjusting the advertisement content.
19. The non-transitory machine-readable storage medium of claim 17,
wherein the value is relatively lower in a first case wherein the
monitoring access comprises determining a number of times that the
personal content is accessed by the plurality of communication
devices associated with the plurality of second users, and wherein
the advertisement content is selected based on the number of times
that the personal content is accessed by the plurality of
communication devices of the plurality associated with second
users.
20. The non-transitory machine-readable storage medium of claim
C17, further comprising determining a spread rate of the personal
content according to a time period in which a threshold number of
target users accesses the personal content, the target users being
in one of a particular category of users, a particular location of
users, or a combination thereof, wherein the determining the value
is further based on the spread rate.
Description
CROSS REFERENCE TO RELATED APPLICATIONS
[0001] This application is a continuation of U.S. patent
application Ser. No. 15/714,492 filed on Sep. 25, 2017, pending,
which is a Division of U.S. patent application Ser. No. 13/207,872
(now U.S. Pat. No. 9,799,061), filed on Aug. 11, 2011. The contents
of each of the foregoing are hereby incorporated by reference into
this application as if set forth herein in full.
[0002] This application is related to U.S. application Ser. No.
13/193,019 entitled "METHOD AND APPARATUS FOR GENERATING MEDIA
CONTENT" filed on Jul. 28, 2011 by Ivanov et al., Attorney Docket
No. 2011-0179_7785-0889; U.S. application Ser. No. 13/193,166
entitled "METHOD AND APPARATUS FOR GENERATING MEDIA CONTENT" filed
on Jul. 28, 2011 by Ivanov et al., Attorney Docket No.
2011-0180_7785-0881; and U.S. application Ser. No. 13/195,335
entitled "METHOD AND APPARATUS FOR MANAGING PERSONAL CONTENT" filed
on Aug. 1, 2011 by Ivanov et al., Attorney Docket No.
2011-0181_7785-0882. The disclosure of each of these applications
is hereby incorporated by reference herein in their entirety.
FIELD OF THE DISCLOSURE
[0003] The present disclosure relates generally to personal content
and more specifically to managing advertisement content and the
personal content.
BACKGROUND
[0004] Media content is generally experienced by consumers via
devices such as computers, televisions, radios, and mobile
electronics. Media content is frequently delivered by service
providers, who send media content, such as television programming,
radio, and video, directly to consumers for enjoyment at their
physical location. Service providers also often provide access to
the internet and to internet-based media content, such as websites,
games, social networking, and virtual-reality applications.
Consumers often desire to customize media content and share that
content with others.
BRIEF DESCRIPTION OF THE DRAWINGS
[0005] FIGS. 1-2 depict illustrative embodiments of communication
systems that provide media services;
[0006] FIG. 3 depicts an illustrative embodiment of a web portal
for interacting with the communication systems of FIGS. 1-2;
[0007] FIG. 4 depicts an illustrative embodiment of a communication
device utilized in the communication systems of FIGS. 1-2;
[0008] FIGS. 5-6 depict illustrative embodiments of communication
systems that provide media services;
[0009] FIG. 7 depicts an illustrative embodiment of a method
operating in portions of the systems described in FIGS. 1-6;
[0010] FIG. 8 depicts an illustrative embodiment of a communication
system that provides media services;
[0011] FIG. 9 depicts an illustrative embodiment of a method
operating in portions of the systems and methods described in FIGS.
1-8;
[0012] FIG. 10 depicts an illustrative embodiment of a
communication system that provides media services;
[0013] FIG. 11 depicts an illustrative embodiment of a method
operating in portions of the systems and methods described in FIGS.
1-10;
[0014] FIG. 12 depicts an illustrative embodiment of a
communication system provides media services;
[0015] FIG. 13 depicts an illustrative embodiment of a method
operating in portions of the systems and methods described in FIGS.
1-12; and
[0016] FIG. 14 is a diagrammatic representation of a machine in the
form of a computer system within which a set of instructions, when
executed, may cause the machine to perform any one or more of the
methods discussed herein.
DETAILED DESCRIPTION
[0017] The present disclosure describes, among other things,
illustrative embodiments for generating media content utilizing a
mobile communication device, where the media content is associated
with an environment. The mobile communication device can retrieve
sensory information associated with the environment from a sensor
device coupled with the mobile communication device, including
still or moving images of the environment, as well as from other
devices. One or more of the exemplary embodiments can customize
advertising for the personal content based on a number of factors,
including demographics of the personal content, popularity of the
personal content and so forth. The advertisement content can be
overlayed on images of the personal content and/or can be objects
presented in the images that are made selectable by user input.
Revenue sharing based on the advertising can be performed for the
users that generate the personal content and/or the service
provider that provides access to the personal content. In one
embodiment, image recognition can be performed on objects captured
in the images and an identification of merchants associated with
the objects can further be performed.
[0018] Personal content generated from the sensory information can
be analyzed based on one or both of image and audio recognition to
identify individual(s) presented in the personal content. The
personal content or a limited portion thereof can be shared with
the individual(s), such as for the individual(s) to generate
further personal content that incorporates the shared portion. The
identification analysis can be performed using a number of
different techniques, such as image samples and/or audio samples
selected via user input and/or samples provided along with the
sensory information received from another mobile device. The
identification analysis can be a centralized process, such as
performed by a remote server and/or can be a localized process,
such as being performed by a set top box, personal computer or
other local processor. The sensory information can be utilized to
generate a plurality of personal content, where each of the
personal content includes at least some images captured at
different perspectives. The different personal content can be
selectively accessed by different parties, such as based on a
relationship between the requesting party and the user that
generated the personal content.
[0019] The mobile communication device can broadcast a notice to
obtain other sensory information associated with the environment
from one or more other mobile communication devices and one or more
fixed communication devices that have captured the other sensory
information associated with the environment. In one example, the
devices can exchange sensory information so that each device can
generate its own media content. The devices can negotiate with each
other for access to the sensory information. In one or more
embodiments, the devices can anonymously provide access to the
sensory information. In one or more embodiments, the media content
can be generated into three dimensional (3D) media content, such as
based on images captured by different devices from different
perspectives. The access to the sensory information can be shared
so that a first device receives sensory information via a second
device where the sensory information was captured by a third device
and where the first and third devices do not need to communicate
with each other to share the sensory information. The media content
can be generated by the mobile communication device and/or can be
generated by another device, such as a remote server which receives
the sensory information from the mobile communication device. Other
embodiments are contemplated by the present disclosure.
[0020] One embodiment of the present disclosure includes a server
that includes a memory and a processor coupled to the memory. The
processor can be operable to monitor access by communication
devices of second users to personal content associated with a first
user, where the personal content is generated based on sensory
information obtained by at least a group of mobile communication
devices in response to a broadcast by a first mobile communication
device of a wireless signal representing a notice to obtain the
sensory information, The sensory information can be associated with
at least one environment of the first mobile communication device.
The sensory information can include images of the at least one
environment. The first mobile communication device can be
associated with the first user. The processor can be operable to
determine demographic information representative of the personal
content, select advertisement content based on the monitoring of
the access and the determination of the demographic information,
and provide the advertisement content for presentation with the
personal content when the personal content is accessed by a
communication device of a third user.
[0021] One embodiment of the present disclosure can be a method
that includes monitoring access by communication devices of second
users to personal content associated with a first user. The
personal content can be generated based on sensory information
obtained by at least a group of mobile communication devices in
response to a broadcast by a first mobile communication device of a
wireless signal representing a notice to obtain the sensory
information. The sensory information can be associated with at
least one environment of the first mobile communication device. The
sensory information can include images of the at least one
environment. The first mobile communication device can be
associated with the first user. The method can also include
determining a value for advertising in the personal content based
on the monitoring of the access and providing an offer to an entity
to include advertisement content with the personal content. The
offer can be based on the determined value. The monitoring,
determining and providing steps can be performed by a processor
device.
[0022] One embodiment of the present disclosure includes a
non-transitory computer-readable storage medium comprising computer
instructions for monitoring access by communication devices of
second users to personal content associated with a first user. The
personal content can be generated based on sensory information
obtained by at least a group of mobile communication devices in
response to a broadcast by a first mobile communication device of a
wireless signal representing a notice to obtain the sensory
information. The sensory information can be associated with at
least one environment of the first mobile communication device. The
sensory information can include images of the at least one
environment. The first mobile communication device is associated
with the first user. The computer instructions can enable selecting
advertisement content based on the monitoring of the access and
providing the advertisement content for presentation with the
personal content when the personal content is accessed by a
communication device of a third user.
[0023] FIG. 1 depicts an illustrative embodiment of a first
communication system 100 for delivering media content. The
communication system 100 can generate media content related to an
environment 101, such as a concert, a sporting event, a user
walking through a shopping mall, a user playing at a park, and so
forth. The communication system 100 can generate the media content
by capturing different sensory information from different
communication devices that are in proximity to the environment 101.
The sensory information can include still or moving images, and can
include other types of sensory information such as temperature,
weather, lighting, color, audio, location information, time, season
of the year, aroma associated with the environment, and so forth.
The communication system 100 can allow mobile devices to
selectively exchange sensory information, including in a daisy
chain fashion where devices indirectly share sensory information
via other devices without direct communication, so that the mobile
devices can enhance their access to sensory information of the
environment 101.
[0024] The system 100 can allow advertisement content to be
customized for personal content, such as based on demographics of
the personal content, popularity (e.g., number of times accessed by
others), and so forth. The advertisement content can be utilized in
revenue sharing with the user that generates the personal content
and/or with the service provider that provides others access to the
personal content. Image recognition can be utilized to convert or
otherwise adjust objects that were captured in the personal content
into selectable objects that allow for purchase of the object, such
as via a link to a website selling the object. Other advertising
schemes can also be used, including selectable icons overlayed on
the images, banners, audio advertisement and so forth. Personal
content can be evaluated, such as based on demographics and the
number of followers of a particular user who generated the personal
content. The evaluation can then be used in making offers to
entities for advertising in the personal content.
[0025] The system 100 can provide for tagging personal content that
includes target individuals and selectively sharing all or portions
of personal content that includes the target individuals. The
target individuals can be identified via recognition analysis, such
as image and/or audio recognition. Samples for performing the
recognition analysis can be provided by a user, such as selecting a
sample in one personal content, and/or can be retrieved from a
communication device that is a source of a portion of the sensory
information. The system 100 can allow for generating different
versions of personal content from the sensory information, such as
from different perspectives, different people speaking and so
forth. Sharing of the personal content can be based on a number of
factors, including a relationship between the requestor and the
user that generated the content.
[0026] The communication system 100 can represent an Internet
Protocol Television (IPTV) media system. The IPTV media system can
include a super head-end office (SHO) 110 with at least one super
headend office server (SHS) 111 which receives media content from
satellite and/or terrestrial communication systems. In the present
context, media content can represent in whole or in part, for
example, audio content, moving image content such as 2D or 3D
videos, video games, virtual reality content, still image content,
and combinations thereof. The SHS server 111 can forward packets
associated with the media content to one or more video head-end
servers (VHS) 114 via a network of video head-end offices (VHO) 112
according to a common multicast communication protocol.
[0027] The VHS 114 can distribute multimedia content, including
broadcast content, via an access network 118 to commercial and/or
residential buildings 102 housing a gateway 104 (such as a
residential or commercial gateway). The access network 118 can
represent a group of digital subscriber line access multiplexers
(DSLAMs) located in a central office or a service area interface
that provide broadband services over fiber optical links or copper
twisted pairs 119 to buildings 102. The gateway 104 can use common
communication technology to distribute broadcast signals to media
processors 106 such as Set-Top Boxes (STBs) which in turn present
broadcast channels to media devices 108 such as computers or
television sets managed in some instances by a media controller 107
(such as an infrared or RF remote control).
[0028] The gateway 104, the media processors 106, and/or media
devices 108 can utilize tethered communication technologies (such
as coaxial, powerline or phone line wiring) or can operate over a
wireless access protocol such as Wireless Fidelity (WiFi),
Bluetooth, Zigbee, or other present or next generation local or
personal area wireless network technologies. By way of these
interfaces, unicast communications can also be invoked between the
media processors 106 and subsystems of the IPTV media system for
services such as video-on-demand (VoD), browsing an electronic
programming guide (EPG), or other infrastructure services.
[0029] A satellite broadcast television system 129 can be used also
in the media system of FIG. 1. The satellite broadcast television
system can be overlaid, operably coupled with, or replace the IPTV
system as another representative embodiment of communication system
100. In this embodiment, signals transmitted by a satellite 115
carrying media content can be received by a satellite dish receiver
131 coupled to the building 102. Modulated signals received by the
satellite dish receiver 131 can be transferred to the media
processors 106 for demodulating, decoding, encoding, and/or
distributing broadcast channels to the media devices 108. The media
processors 106 can be equipped with a broadband port to the ISP
network 132 to enable interactive services such as VoD and EPG as
described above.
[0030] In yet another embodiment, an analog or digital cable
broadcast distribution system such as cable TV system 133 can be
overlaid, operably coupled with, or replace the IPTV system and/or
the satellite TV system as another representative embodiment of
communication system 100. In this embodiment, the cable TV system
133 can also provide Internet, telephony, and interactive media
services.
[0031] The embodiments of the present disclosure can apply to other
present or next generation over-the-air and/or landline media
content services system.
[0032] Some of the network elements of the IPTV media system can be
coupled one or more computing devices 130, a portion of which can
operate as a web server for providing web portal services over an
Internet Service Provider (ISP) network 132 to wireline media
devices 108 or portable communication devices 116.
[0033] Multiple forms of media services can be offered to media
devices over landline technologies in communication system 100 such
as through the devices and/or techniques described above.
Additionally, media services can be offered to media devices by way
of a wireless access base station 117 operating according to common
wireless access protocols such as Global System for Mobile or GSM,
Code Division Multiple Access or CDMA, Time Division Multiple
Access or TDMA, Universal Mobile Telecommunications or UMTS, World
interoperability for Microwave or WiMAX, Software Defined Radio or
SDR, Long Term Evolution or LTE, and so on. Other present and next
generation wide area wireless network technologies are contemplated
by the present disclosure.
[0034] Communication system 100 can also provide for all or a
portion of the computing devices 130 to function as a sensory
information server (herein referred to as server 130). The server
130 can use common computing and communication technology to
perform function 162, which can include among things, coordinating
the gathering of sensory information from different mobile and
fixed communications devices in proximity to the environment
101.
[0035] The server 130 can be used to receive sensory data
representing characteristics of an environment of a source device,
such as one of the mobile communication devices 116. The server 130
can also be used to facilitate the gathering of the sensory data,
such as providing a platform for exchanging sensory data between
mobile communication devices and/or providing a platform for
negotiating obtaining media content, such as from a commercial
entity that is capturing images of an event occurring in the
environment (e.g., images of a baseball game).
[0036] In one or more embodiments, the server 130 can be used for
generating media content based on sensory information captured by
the group of communication devices (e.g., mobile and/or fixed
communication devices). For instance, the server 130 can generate
media content that shows various images of the environment,
including as a live scrap book of the user. The media content can
be generated by various devices, including the server 130, the
mobile device 116, the STB 106 or other other devices of the
communication system 100.
[0037] The media content can present the environment at various
viewing angles and/or at a selected viewing angle. In one
embodiment, the live scrap book can be presented at a viewing angle
designated by a user (such as in a user profile) and can further
allow access to other viewing angles when the user desires, such as
pausing the presentation of the live scrap book so that the other
viewing angles can be presented. The sensory information also
enables utilizing different sensory data from different devices to
further enhance the media content. For example, images can be
captured from a far distance to an event of the environment (e.g.,
a football game) by a first communication device to allow capturing
a view of all of the players in the football game while audio can
be captured from a close distance to the field of the football game
by a second communication device to allow hearing the event as if
the user were present on the field of the football game.
[0038] In one embodiment, portions of the sensory information can
be filtered based on quality. For instance, audio content can be
compared and selected, including utilizing only portions of a first
audio content and combining it with portions of a second audio
content, so that the audio associated with the environment has more
clarity. Other criteria can also be utilized for filtering portions
of the sensory information. In one example, the sensory information
can be filtered based on accuracy. For example, if the sensory
information is temperature then a temperature reading that is
believed to be inaccurate, such as based on a comparison with other
sensory information or based on a location of the temperature
sensor being too close to a heat source, can be ignored. As another
example, parental controls, user preferences, digital rights
management and so forth can be utilized in determining whether
captured sensory information should be filtered out of the
generation of the media content. For instance, images of portions
of events (e.g., actions occurring in an audience of the event in
the environment 101) that are inappropriate for children can be
filtered out of the media content. The filtering can be based on
user preferences, information stored in a user profile, monitored
consumption of a user and so forth.
[0039] In one or more embodiments, the server 130 can perform
recognition analysis on personal content to identify individuals
presented in the content. The individual and/or the portion of the
content presenting the individuals can be tagged or otherwise
flagged. The recognition analysis can be performed in a number of
different ways, including image and/or audio recognition, based on
samples selected by a user or otherwise obtained, such as with the
sensory information from a particular mobile communication device.
Selective access can be provided to communication devices
associated with flagged individuals. The personal content can be
obtained from various sources. For instance, the personal content
can be stored locally at a set top box and/or a DVR coupled with
the set top box. In one embodiment, the personal content can be
stored at a remote storage device, such as via cloud computing.
Combinations of remote and local storage techniques can also be
utilized, such as having various copies both stored locally and
stored in a cloud computing environment. The exemplary embodiments
can also include a combination of the server 130 and the media
processor 106 performing the recognition analysis.
[0040] The server 130 can access personal content of a user to
determine if the content can be provided with targeted
advertisement content. For instance, the personal content can be
stored in a database accessible to the server 130 and/or can be
stored locally, such as at a DVR coupled with a media processor 106
of the user. The server 130 can monitor access to the personal
content to determine various information, such as the identity of
users accessing the personal content, the popularity of the
content, whether the content is being re-distributed to others, and
so forth. In one embodiment, the server 130 can determine or rate
the following of the personal content and/or the following of the
user that generated the personal content. One or more of these
factors can be utilized in selecting advertising content to be
provided with the personal content and/or for making an offer to an
entity to advertise in the personal content.
[0041] In one embodiment, the server 130 can track the number of
times selectable advertisement content has been accessed in
personal content. If the access does not meet a threshold then the
advertisement content can be adjusted, such as based on maintaining
the same product but presenting it in a different type of
advertisement and/or switching the product that is being
advertised. In one embodiment, the personal content can be
evaluated as to a type of subject matter, such as based on image,
audio and/or text recognition, and demographic information can be
determined based on the subject matter. Other demographic
information associated with the personal content can also be
determined, such as via image recognition to determine age, gender
and so forth of individuals that have been captured in the images
of the personal content.
[0042] The media processors 106 and mobile communication devices
116 can be adapted with software functions 164 and 166,
respectively, to further utilize the services of the server 130.
For example, the media processors 106 and portable communication
devices 116 can be adapted to execute software functions 164 and
166, respectively, to collect and report sensory data to the server
130 representing characteristics of the environments of the media
processors 106 and mobile communication devices 116, respectively.
In one embodiment, the software functions 166 can include
retrieving sensory information from a sensor device coupled with
the mobile communication device 116. The sensor device can be a
camera, such as integrated with the mobile communication device 116
or in communication therewith via wireless and/or tethered
communication. Other types of sensor devices can also be utilized,
including audio recorders, temperature sensors, location detectors,
and so forth.
[0043] In one embodiment, the sensory information can be provided
to the media processor 106 for generating personal content,
including a plurality of different versions of the personal
content. The different versions can be based on criteria
established by the user, such as a version for family and a version
for friends. The different versions can highlight different people,
different audio, different objects and so forth. Other variations
of the versions can include different supplemental content being
added to the personal content, such as different background music
or different tags for tangential information related to the
environment. For example, a family version may include tags that
provide links or other information to similar family outings where
as a friend's version may provide links or other information to
particular friends and their personal information.
[0044] In one or more embodiments, the media processor 106 can
perform the advertisement analysis, recognition analysis and/or the
providing of selective analysis as described above.
[0045] FIG. 2 depicts an illustrative embodiment of a communication
system 200 employing IP Multimedia Subsystem (IMS) network
architecture to facilitate the combined services of
circuit-switched and packet-switched systems. Communication system
200 can be overlaid or operably coupled with communication system
100 as another representative embodiment of communication system
100.
[0046] The communication system 200 can generate media content
related to an environment in proximity to one or more of the mobile
communication devices of the system. The communication system 200
can generate the media content by capturing different sensory
information from different communication devices that are in
proximity to the environment. The sensory information can include
still or moving images, and can include other types of sensory data
such as temperature, weather, lighting, color, audio, location
information, time, season of the year, aroma associated with the
environment, and so forth. The communication system 200 allows
mobile devices to selectively exchange sensory information,
including in a daisy chain fashion where devices indirectly share
sensory information via other devices without direct communication,
so that the mobile devices can enhance their access to sensory
information of the environment. The communication system 200 also
allows users providing and/or receiving sensory information to
remain anonymous.
[0047] The communication system 200 enables monetization of
personal content via targeted advertising schemes. The monetization
can be performed by sharing revenue with a user that generates the
personal content and/or the service provider provides the access to
the personal content when a sale results from advertisement content
contained in the personal content. Various methodologies can be
utilized for targeting the advertisement, including using a natural
language engine to determine a subject matter(s) of the personal
content.
[0048] The communication system 200 allows for tagging personal
content that includes target individuals and selectively sharing
all or portions of personal content that includes the target
individuals. The target individuals can be identified via
recognition analysis, such as image and/or audio recognition.
Samples for performing the recognition analysis can be provided by
a user, such as selecting a sample in one personal content, and/or
can be retrieved from a communication device that is a source of a
portion of the sensory information.
[0049] Communication system 200 enables exchange of the sensory
information via packet-switched and circuit-switched communication
technologies.
[0050] Communication system 200 can comprise a Home Subscriber
Server (HSS) 240, a tElephone NUmber Mapping (ENUM) server 230, and
other common network elements of an IMS network 250. The IMS
network 250 can establish communications between IMS-compliant
communication devices (CDs) 201, 202, Public Switched Telephone
Network (PSTN) CDs 203, 205, and combinations thereof by way of a
Media Gateway Control Function (MGCF) 220 coupled to a PSTN network
260. The MGCF 220 need not be used when a communication session
involves IMS CD to IMS CD communications. A communication session
involving at least one PSTN CD may utilize the MGCF 220.
[0051] IMS CDs 201, 202 can register with the IMS network 250 by
contacting a Proxy Call Session Control Function (P-CSCF) which
communicates with an interrogating CSCF (I-CSCF), which in turn,
communicates with a Serving CSCF (S-CSCF) to register the CDs with
the HSS 240. To initiate a communication session between CDs, an
originating IMS CD 201 can submit a Session Initiation Protocol
(SIP INVITE) message to an originating P-CSCF 204 which
communicates with a corresponding originating S-CSCF 206. The
originating S-CSCF 206 can submit the SIP INVITE message to one or
more application servers (ASs) 217 that can provide a variety of
services to IMS subscribers.
[0052] For example, the application servers 217 can be used to
perform originating call feature treatment functions on the calling
party number received by originating S-CSCF 206 in the SIP INVITE
message. Originating treatment functions can include determining
whether the calling party number has international calling
services, call ID blocking, calling name blocking, 7-digit dialing,
and/or is requesting special telephony features (e.g., *72 forward
calls, *73 cancel call forwarding, *67 for caller ID blocking, and
so on). Based on initial filter criteria (iFCs) in a subscriber
profile associated with a CD, one or more application servers may
be invoked to provide various call originating feature
services.
[0053] Additionally, the originating S-CSCF 206 can submit queries
to the ENUM system 230 to translate an E.164 telephone number in
the SIP INVITE message to a SIP Uniform Resource Identifier (URI)
if the terminating communication device is IMS-compliant. The SIP
URI can be used by an Interrogating CSCF (I-CSCF) 207 to submit a
query to the HSS 240 to identify a terminating S-CSCF 214
associated with a terminating IMS CD such as reference 202. Once
identified, the I-CSCF 207 can submit the SIP INVITE message to the
terminating S-CSCF 214. The terminating S-CSCF 214 can then
identify a terminating P-CSCF 216 associated with the terminating
CD 202. The P-CSCF 216 may then signal the CD 202 to establish
Voice over Internet Protocol (VoIP) communication services, thereby
enabling the calling and called parties to engage in voice and/or
data communications. Based on the iFCs in the subscriber profile,
one or more application servers may be invoked to provide various
call terminating feature services, such as call forwarding, do not
disturb, music tones, simultaneous ringing,
[0054] stances the aforementioned communication process is
symmetrical. Accordingly, the terms "originating" and "terminating"
in FIG. 2 may be interchangeable. It is further noted that
communication system 200 can be adapted to support video
conferencing. In addition, communication system 200 can be adapted
to provide the IMS CDs 201, 202 with the multimedia and Internet
services of communication system 100 of FIG. 1, including the
sensory information gathered from various mobile device in
proximity to an event or environment. It is further contemplated
that the CDs of FIG. 2 can operate as wireline and/or wireless
devices. For example, the CDs of FIG. 2 can be communicatively
coupled to a cellular base station 117 such as shown in FIG. 1, a
femtocell (not shown), a WiFi router, a DECT base unit, or another
suitable wireless access unit to establish communications with the
IMS network 250 of FIG. 2.
[0055] If the terminating communication device is instead a PSTN CD
such as 203 or CD 205 (in instances where the cellular phone only
supports circuit-switched voice communications), the ENUM system
230 can respond with an unsuccessful address resolution which can
cause the originating S-CSCF 206 to forward the call to the MGCF
220 via a Breakout Gateway Control Function (BGCF) 219. The MGCF
220 can then initiate the call to the terminating PSTN CD over the
PSTN network 260 to enable the calling and called parties to engage
in voice and/or data communications.
[0056] It is further appreciated that the CDs of FIG. 2 can operate
as wireline or wireless devices. Although not shown, the CDs of
FIG. 2 can be communicatively coupled to a cellular base station
221, a femtocell, a WiFi router, a DECT base unit, or another
suitable wireless access unit to establish communications with the
IMS network 250 of FIG. 2. The cellular access base station 121 can
operate according to common wireless access protocols such as
Global System for Mobile (GSM), Code Division Multiple Access
(CDMA), Time Division Multiple Access (TDMA), Universal Mobile
Telecommunications (UMTS), World interoperability for Microwave
(WiMAX), Software Defined Radio (SDR), Long Term Evolution (LTE),
and so on. Other present and next generation wireless network
technologies are contemplated by the present disclosure.
Accordingly, multiple wireline and wireless communication
technologies are contemplated for the CDs of FIG. 2.
[0057] It is further contemplated that cellular phones supporting
LTE can support packet-switched voice and packet-switched data
communications and thus may operate as IMS-compliant mobile
devices. In this embodiment, the cellular base station 221 may
communicate directly with the IMS network 250.
[0058] Communication system 200 can include or otherwise be coupled
with server 130 of FIG. 1 for purposes similar to those described
above. It is further contemplated by the present disclosure that
the server 130 can perform the function 162 and thereby facilitate
gathering of sensory information and/or generating media content
based on gathered sensory information for subscribers associated
with CDs 201, 202, 203, and 205. CDs 201, 202, 203, and 205 can be
adapted with software to perform function 170 to utilize and
integrate with the media content generation and/or sensory
information gathering performed by the server 130. It is further
contemplated that the server 130 can be an integral part of the
application server(s) 217, where the application server(s) 217
performs function 172, which can be substantially similar to
function 162 and adapted to the operation of the IMS network
250.
[0059] FIG. 3 depicts an illustrative embodiment of a web portal
302 which can be hosted by server applications operating from the
computing devices 130 of the communication system 100 illustrated
in FIG. 1. The web portal 302 can be used for managing services of
communication systems 100-200. A web page of the web portal 302 can
be accessed by a Uniform Resource Locator (URL) with an Internet
browser such as Microsoft's Internet Explorer.TM., Mozilla's
Firefox.TM., Apple's Safari.TM., or Google's Chrome.TM. using an
Internet-capable communication device such as those described in
FIGS. 1-2. The web portal 302 can be configured, for example, to
access a media processor 106 and services managed thereby such as a
Digital Video Recorder (DVR), a Video on Demand (VoD) catalog, an
Electronic Programming Guide (EPG), or a personal catalog (such as
personal videos, pictures, audio recordings, etc.) stored at the
media processor 106. The web portal 302 can also be used for
provisioning IMS services described earlier, provisioning Internet
services, provisioning cellular phone services, and so on.
[0060] It is contemplated by the present disclosure that the web
portal 302 can further be utilized to manage and provision software
applications 162-172 of the devices of FIGS. 1-2, respectively, as
described earlier. In one embodiment, web portal 302 can be
utilized for selecting the notice signal that is broadcast by the
communication device to obtain sensory information from other
communication devices. As an example, a user can access the web
portal 302 via a web browser operating on a mobile communication
device 116. The web portal can be utilized for selecting a strength
of the wireless signal that is to be broadcast, where the selection
can be based on the characteristics of the environment. For
instance, the strength of the broadcast signal can be based on the
size of the environment so that the broadcast is being received by
other mobile communication devices that are in proximity to the
environment rather than mobile communication devices that are too
far from the environment to provide any useful sensory information.
While this embodiment describes use of the web portal to select the
signal strength, the exemplary embodiments can also select the
signal strength without use of the web portal, such as based on an
analysis made by the processor of the mobile communication device.
The analysis can be based on characteristics of the environment
that are detected or otherwise determined by the mobile
communication device, such as using GPS location information to
determine that a stronger signal is needed because the environment
includes numerous obstructions such as buildings. The analysis can
also be based on inputs of a user such as a type of environment
(e.g., a soccer stadium), size of environment, location of
environment, type of location (open areas rather than a city with
buildings obstructing the wireless broadcast), and so forth.
[0061] FIG. 4 depicts an exemplary embodiment of a communication
device 400. Communication device 400 can serve in whole or in part
as an illustrative embodiment of the devices depicted in FIGS. 1-2.
The communication device 400 can comprise a wireline and/or
wireless transceiver 402 (herein transceiver 402), a user interface
(UI) 404, a power supply 414, a location receiver 416, and a
controller 406 for managing operations thereof. The transceiver 402
can support short-range or long-range wireless access technologies
such as Bluetooth, WiFi, Digital Enhanced Cordless
Telecommunications (DECT), or cellular communication technologies,
just to mention a few. Cellular technologies can include, for
example, CDMA-1X, UMTS/HSDPA, GSM/GPRS, TDMA/EDGE, EV/DO, WiMAX,
SDR, LTE, as well as other next generation wireless communication
technologies as they arise. The transceiver 402 can also be adapted
to support circuit-switched wireline access technologies (such as
PSTN), packet-switched wireline access technologies (such as TCPIP,
VoIP, etc.), and combinations thereof.
[0062] The UI 404 can include a depressible or touch-sensitive
keypad 408 with a navigation mechanism such as a roller ball, a
joystick, a mouse, or a navigation disk for manipulating operations
of the communication device 400. The keypad 408 can be an integral
part of a housing assembly of the communication device 400 or an
independent device operably coupled thereto by a tethered wireline
interface (such as a USB cable) or a wireless interface supporting
for example Bluetooth. The keypad 408 can represent a numeric
keypad commonly used by phones, and/or a Qwerty keypad with
alphanumeric keys. The UI 404 can further include a display 410
such as monochrome or color LCD (Liquid Crystal Display), OLED
(Organic Light Emitting Diode) or other suitable display technology
for conveying images to an end user of the communication device
400. In an embodiment where the display 410 is touch-sensitive, a
portion or all of the keypad 408 can be presented by way of the
display 410 with navigation features.
[0063] The UI 404 can also include an audio system 412 that
utilizes common audio technology for conveying low volume audio
(such as audio heard only in the proximity of a human ear) and high
volume audio (such as speakerphone for hands free operation). The
audio system 412 can further include a microphone for receiving
audible signals of an end user. The audio system 412 can also be
used for voice recognition applications. The UI 404 can further
include an image sensor 413 such as a charged coupled device (CCD)
camera for capturing still or moving images.
[0064] The power supply 414 can utilize common power management
technologies such as replaceable and rechargeable batteries, supply
regulation technologies, and charging system technologies for
supplying energy to the components of the communication device 400
to facilitate long-range or short-range portable applications. The
location receiver 416 can utilize common location technology such
as a global positioning system (GPS) receiver capable of assisted
GPS for identifying a location of the communication device 400
based on signals generated by a constellation of GPS satellites,
thereby facilitating common location services such as
navigation.
[0065] The communication device 400 can use the transceiver 402 to
also determine a proximity to a cellular, WiFi, Bluetooth, or other
wireless access points by common sensing techniques such as
utilizing a received signal strength indicator (RSSI) and/or a
signal time of arrival (TOA) or time of flight (TOF). The
controller 406 can utilize computing technologies such as a
microprocessor, a digital signal processor (DSP), and/or a video
processor with associated storage memory such as Flash, ROM, RAM,
SRAM, DRAM or other storage technologies.
[0066] The communication device 400 can be coupled to a sensor
device 450 which captures sensory information associated with an
environment, such as environment 101 of FIG. 1. The sensory
information can be of various types and can be captured by
corresponding types of detection devices. For example, the sensory
information can be one or more of images, audio, temperature,
weather, lighting, color, location information, time, season of the
year, and aroma associated with the environment. The coupling of
the sensor device 450 with the communication device 400 can be of
various types, including: the sensor device being integrated in the
communication device, the sensor device being tethered or otherwise
connected via a hardwire to the communication device and/or the
sensor device being wirelessly in communication with the
communication device. The exemplary embodiments can also include
the image sensor 413, the audio system 412 and/or the location
receiver 416 functioning as the sensor device 450. The exemplary
embodiments also provide for the communication device 400 capturing
multiple types of sensory information. For example, the sensor
device 450 can be a lighting detector that provides lighting
information associated with the environment, along with images
captured by the image sensor 413 and audio captured by the audio
system 412.
[0067] The communication device 400 can be adapted to perform the
functions of the media processor 106, the media devices 108, or the
portable communication devices 116 of FIG. 1, as well as the IMS
CDs 201-202 and PSTN CDs 203-205 of FIG. 2. It will be appreciated
that the communication device 400 can also represent other common
devices that can operate in communication systems 100-200 of FIGS.
1-2 such as a gaming console and a media player.
[0068] The communication device 400 can store recognition
information associated with a user of the communication device,
such as images of the user and/or audio samples. For instance, the
images can be facial images, full body images, side images, back
images and so forth. The audio samples can be of various sentences
or phrases spoken by the user. The communication device 400 can
provide the recognition information to other communication devices
to enable recognition analysis to be performed on personal content
that is generated in association with the recipient communication
device. In one embodiment, the communication device 400 can provide
the recognition information along with sensory information captured
by sensor 450. In one embodiment, the communication device 400 can
provide the recognition information to a recipient communication
device in exchange for receipt of recognition information from the
recipient communication device. In one embodiment, the exchange of
the recognition information can be subject to a sharing agreement
that limits or otherwise precludes distribution of the recognition
information to other communication devices or other users.
[0069] Illustrative embodiments of methods that can operate in
portions of the communication device of FIG. 4 are described
below.
[0070] FIG. 5 depicts an illustrative embodiment of a system 500
for gathering sensory information associated with an environment
501 and generating media content based on the gathered sensory
information. System 500 can be overlaid or operably coupled to
communication systems 100-200 as another representative embodiment
of communication systems 100-200. System 500 can include a server
530 in communication with media processors 506 (such as set-top
boxes), communication device 516 (such as a laptop computer, a
server, etc.), and mobile wireless communication devices 512 and
514 (such as mobile phones, smart phones, etc.). The mobile
communication devices 512 and 514 can be communicatively coupled to
the server 530 by way of a network 550 (e.g., a mobility network)
coupled to the Internet or other communication means. The
communication device 516 can also be communicatively coupled to the
server 530 by way of the Internet or other communication means.
[0071] In this example, a first mobile communication device 512 can
be coupled with one or more sensor devices 513 for capturing
sensory information associated with the sensory environment 1. A
second mobile communication device 514 can be coupled with one or
more sensor devices 515 for capturing sensory information
associated with the sensory environment 2. A fixed communication
device 516 can be coupled with one or more sensor devices 517 for
capturing sensory information associated with the sensory
environment 3. The sensors 513, 515 and 517 can be in proximity to
the environment 501 in order to capture the sensory information of
the environment 501. As shown in FIG. 5, the sensory environments
1, 2 and 3 can overlap each other and encompass all or a portion of
the environment 501 to facilitate capturing sensory information
associated with the environment 501 that provides a full or more
encompassing perspective of the environment 501.
[0072] The sensors 513, 515 and 517 can be of various numbers and
can use various components, to capture various types of sensory
information including images, audio, temperature, weather,
lighting, color, location information, time, season of the year,
and aroma associated with the environment 501. While system 500
depicts three devices that are utilized for capturing sensory
information associated with environment 501, it should be
understood that any number of devices, including mobile and/or
fixed communication devices can be used for gathering sensory
information for generating media content.
[0073] In one embodiment, the mobile communication device 512 can
broadcast a wireless signal which is received by the mobile
communication device 514 and the fixed communication device 516.
The wireless signal can be a notice to obtain sensory information
associated with the environment 501. The notice can advise other
devices, including the mobile communication device 514 and the
fixed communication device 516, that mobile communication device
512 is utilizing sensor 513 to capture sensory information
associated with the environment 501, such as capturing still or
moving images of the environment 501. The notice can also request
sensory information associated with the environment 501 from other
devices, including the mobile communication device 514 and the
fixed communication device 516. The broadcast signal can be of
various types utilizing various communication protocols. For
example, the mobile communication device 512 can be a multi-mode
communication device and the wireless notice signal can be
broadcast using a number of different communication protocols
(simultaneously and/or serially) so that different devices
operating under different communication protocols can be put on
notice of the capture of the sensory information and/or the request
for sensory information. In one embodiment, the broadcast notice
signal can be a mid-range wireless signal (e.g., 100 meters).
[0074] The receipt of sensory information from other devices,
including the mobile communication device 514 and the fixed
communication device 516, can be performed in a number of different
ways. For instance, sensory information can be exchanged between
devices, such as the mobile communication device 512 transmitting
sensory information captured by sensors 513 to the mobile
communication device 514 while the mobile communication device 514
is transmitting sensory information captured by sensors 515 to the
mobile communication device 512. In one embodiment, the sharing or
otherwise providing of sensory information can be done based on an
opt-in procedure in response to broadcast of the notice. In one
embodiment, the providing of sensory information can be performed
while maintaining anonymity of users associated with the
communication devices, such as mobile communication devices 512 and
514 and fixed communication device 516.
[0075] In one embodiment, obtaining sensory information can be
negotiated. For example, the mobile communication device 512 can
transmit the notice which is received by the fixed communication
device 516. The fixed communication device 516 can agree to
transmit sensory information captured by sensor 517. The agreement
to send the sensory information captured by sensor 517 can be based
on a number of factors in the negotiation. For example, the user of
the mobile communication device 512 may have a subscription to
services provided by a service provider affiliated with the fixed
communication device 516. For instance, the environment 501 may be
a concert in a park where a service provider is operating a fixed
communication device 516 to capture video and audio of the concert
for broadcast over a channel of the service provider. The user of
the mobile communication device 512 may have a subscription for
media services provided over the service provider channel. The
fixed communication device 516 can provide the sensory information
captured by the sensor 517 to the mobile communication device 512,
which in this case is the video and audio of the concert in the
park. Media content can then be generated based on the video and
audio content, as well as other sensory information captured by
sensor 513 and/or received from mobile communication device 514
(which is captured by sensor 515). As an example, the media content
can be a presentation of the concert but from one or more viewing
angles that differ from the viewing angles utilized by the sensor
517 of the fixed communication device 516 when capturing the video
content. In this example, media content can be generated based on
the concert in the park, but where the media content differs, in
whole or in part, from that which is broadcast by the service
provider. The differences can be managed or otherwise determined by
the user of the mobile communication device 512, such as based on
user preferences, device capabilities and so forth.
[0076] In one embodiment, the exchange of sensory information can
be performed in a daisy chain fashion so that the sensory
information does not need to be received directly from a mobile
device that captured the sensory information, but rather can be
forwarded from a mobile device that received the sensory
information from another mobile device. In this example, the use of
the daisy chain techniques can enhance the ability to capture
varying perspectives of the environment 501.
[0077] System 500 allows for the generation of media content based
on the sensory information that has been captured with respect to
environment 501. For example, mobile communication device 512 can
gather sensory information which is used for presenting first media
content 538 at location 548. The first media content can be
generated by one or more of the mobile communication device 512,
the server 530 and the STB 506. The first media content can be
generated based on some or all of the sensory information captured
by one or more of sensors 513, 515 and 517. Mobile communication
device 514 can gather sensory information which is used for
presenting second media content 539 at location 549. The second
media content can be generated by one or more of the mobile
communication device 514, the server 530 and the STB 506. The
second media content can be generated based on some or all of the
sensory information captured by one or more of sensors 513, 515 and
517. Although the first and second media content 538 and 539 may be
similar because they are generated utilizing the same pool of
sensory information, there can be differences, including viewing
angles, portions of images or other sensory information presented,
and so forth. These differences can be selected based on user
preferences, including user profiles, monitored user behavior, and
so forth.
[0078] FIG. 6 depicts an illustrative embodiment of a system 600
for gathering sensory information associated with an environment
601 and generating media based on the gathered sensory information.
System 600 can be overlaid or operably coupled to communication
systems 100-200 and 500 as another representative embodiment of
communication systems 100-200 and 500. System 600 can include a
server 530 in communication with media processors 506 (such as
set-top boxes) and mobile wireless communication devices 512 and
612 (such as mobile phones, smart phones, etc.). The mobile
communication devices 512 and 612 can be communicatively coupled to
the server 530 by way of a network 550 (e.g., a mobility network)
coupled to the Internet or other communication means. Media content
can be presented at location 548 using the display device 508.
[0079] In this example, a first mobile communication device 512 can
be coupled with one or more sensor devices 513 for capturing
sensory information associated with the environment 601. The first
mobile communication device 512 can be in communication with second
mobile communication device 612A which are coupled with one or more
sensor devices 613A for capturing sensory information associated
with the environment 601. The second communication devices 612A can
be in communication with other second communication devices 612B
which are coupled with one or more sensor devices 613B or 613C for
capturing sensory information associated with the environment 601.
In this example, a daisy chain process is utilized to provide first
mobile communication device 512 with access to a greater pool of
sensory information. In this example, first mobile communication
device 512 can receive sensory information captured by sensors 613A
and 613B from second mobile communication devices 612A without
needing to communicate with any of other second mobile
communication devices 612B. As also shown by this example, access
to sensory information can be selectively controlled, including by
the mobile device that captured the sensory information and/or the
mobile device that is transmitting the sensory information. For
example, first mobile communication device 512 can receive via
second communication devices 612A, the sensory information captured
by sensors 613B without receiving the sensory information captured
by sensor 613C. In this example, the mobile communication device
612B that is coupled with sensor 613C can forward other sensory
information that it receives without sending the sensory
information that it retrieves from its own sensor 613C. Whether
sensory information is shared with other devices can be controlled
by various entities according to various procedures. For example, a
mobile device may agree to exchange sensory information with
another mobile device subject to the sensory information not being
further distributed to other mobile devices.
[0080] FIG. 7 depicts an illustrative method 700 that operates in
portions of the devices of FIGS. 1-6. Method 700 can begin with
step 702 in which a mobile communication device transmits a notice
to obtain sensory information associated with an environment, such
as images, audio and so forth. Based on the notice, other
communication devices, including fixed and/or mobile devices, can
opt-in or opt-out of providing and/or receiving sensory
information. In step 704, devices desiring to participate, can
exchange sensory information with the mobile communication device
and/or can provide sensory information to the mobile communication
device without receiving sensory information in return. The
exchange of sensory information can be performed with controls on
the distribution of the sensory information, such as permitting use
of the sensory information in media content generated by a
recipient device but prohibiting further distribution of the
sensory information to other devices.
[0081] In step 706, the mobile communication device can negotiate
to obtain other sensory information, such as requesting sensory
information from an entity operating a security camera in proximity
to the environment. The negotiation can be based on various
criteria, such as payment, exchange of sensory information,
establishing a relationship with an entity operating the other
communication device (such as becoming a member affiliated with the
entity) and so forth. In one embodiment, the negotiation can be
based on future or past exchanges of sensory information. For
instance, a communication device can agree to presently provide
sensory information to the mobile communication device if the
mobile communication device agrees to exchange sensory information
in the future (in the event that the two devices again cross paths
in a particular environment). In step 708, the negotiated sensory
information can be received from the other communication
devices.
[0082] The gathered sensory information can be filtered in step 710
by any number of devices, including the mobile communication
device, a remote server, an STB associated with the mobile
communication device and so forth. The filtering can be performed
based on various criteria, including user preferences, parental
controls, digital rights management (e.g., detecting that some of
the sensory information is subject to proprietary rights of
another), and so forth.
[0083] The media content can be generated in step 712 based on some
or all of the pool of sensory information that has been gathered.
In step 714, the presentation of the media content can be adjusted
using portions of the sensory information that provides a different
perspective of the environment. In one embodiment, editing of the
media content can be performed in a number of different ways, such
as piecing together a user's life from a default perspective (e.g.,
primarily front view) to create a live scrap book, while other
perspectives remain hidden where the user can launch the other
perspectives such as when pausing a segment or otherwise requesting
a different perspective. The perspectives can include a view from
afar (e.g., sensor from building 100 meters away), a view from
behind (bystander sensor captures view from behind), a view from
friends or family while in a group setting. In one embodiment, if
there is a stadium or concert hall recording (e.g., football,
baseball, soccer, concert, etc.), a user can request views
throughout the stadium based on the presence of other mobile
communication device throughout the stadium.
[0084] In one embodiment, the media content can be generated into
3D images from 2D images captured by the sensors. The remote
server, the mobile communication device, the STB and/or some other
device can utilize the different viewing angles provided by
different images of the sensory information to generate 3D content
and/or holographic content. In another embodiment, a depth map can
be utilized to generate the 3D content.
[0085] The media content can also be generated using other
information. For example, media content can be generated based on a
user watching his or her child's baseball game at a ball field.
Other media content, such as images of professional ball players,
and clips of professional announcers which correspond to action
occurring during the child's game can be spliced or otherwise
combined with the media content generated from the sensory
information. Additionally, media content previously generated by
the user and/or sensory information previously gathered that is
related to the child's baseball game can be spliced or otherwise
combined with the media content generated from the present sensory
information, such as adding images of the child from earlier games
when similar plays were made or images showing the age progression
of the child.
[0086] FIG. 8 depicts an illustrative embodiment of a system 800
for gathering sensory information associated with an environment
801 (having audio 802) and generating media content based on the
gathered sensory information. System 800 can be overlaid or
operably coupled to communication systems 100-200 and 500-600 as
another representative embodiment of communication systems 100-200
and 500-600. System 800 can include a server 530 in communication
with media device or processor 506 (such as set-top boxes) and
mobile wireless communication devices 512 and 514 (such as mobile
phones, smart phones, etc.), as well as fixed communication device
516. The mobile communication devices 512 and 514 and the fixed
communication device 516 can be communicatively coupled to the
server 530 by way of a network 550 (e.g., a mobility network)
coupled to the Internet or other communication means. Media content
can be presented at location 849 using the display device 508.
[0087] In this example, sensory information can be obtained from
the various communication devices in proximity to the environment
801, such as mobile devices 512 and 514 and fixed device 516. The
media device 506 can then generate a plurality of personal content,
shown as content 815, 820 and 825, from the sensory information. In
the exemplary embodiment, the media device 506, such as an STB, is
generating the plurality of content. However, it should be
understood that the media device 506 can be a number of different
devices (such as a personal computer, smart phone, etc.).
Additionally, other devices can be used for generating all or some
of the plurality of personal content (including portions of one or
more of the personal content), including the server 530 and/or
other media devices associated with other users.
[0088] The personal content 815, 820 and 825 can be different
versions of the content generated from portions of the sensory
information. The different versions of the personal content can be
based on criteria established or otherwise associated with the user
seeking to generate the personal content. For example, a user
profile can be used to piece together the personal content from the
sensory information. Monitored viewing behavior, as well as
previously generated personal content, can be used for generating
the plurality of different versions. The user can also define the
criteria for generating the personal content, such as target
individuals, sounds, objects, time and so forth to be included in
the personal content to the extent that the sensory information has
captured these criteria. In personal content 815, audio for one or
more individuals in the content can be included. The sensory
information can be analyzed to provide images of the individuals
without background images. In personal content 820, the perspective
of the sensory information (e.g., the viewing angle and zoom) can
be selected so that only select individuals are shown while the
audio is only heard without the audio of others in the environment.
This can be accomplished by reviewing the sensory information and
selecting sensory information captured by sensor devices that
provide the desired viewing angle and desired audio output.
Personal content 825 can provide a complete panoramic perspective
of the environment, with or without the accompanying audio.
[0089] The different versions of the personal content can be
accessed by different requestors based on a number of factors. In
one embodiment, a relationship between the user who generated the
personal content and the requestor can be determined and utilized
to provide or deny the access. For instance, relationships such as
immediate family, family, close friends, friends, acquaintances can
be used to determine if access can be granted. In one embodiment,
the user can customize the relationships including providing
monitorable definitions, such as an acquaintance becoming a friend
when it is determined that a threshold number of messages have been
exchanged between the user and the acquaintance. Different channels
can be provided by a service provider to view the different
versions of the content.
[0090] FIG. 9 depicts an illustrative method 900 that operates in
portions of the devices and methods of FIGS. 1-8. Method 900 can
begin with receiving sensory information at 902. The sensory
information can be received at a media device, such as an STB,
and/or can be received at another device that generates personal
content, such as remote server 130. The sensory information can be
associated with an environment and received from various devices
including a mobile device associated with the media device, mobile
device(s) not associated with the media device and/or fixed
communication devices. The sensory information can be received via
the related mobile device and/or directly from the other
communication devices.
[0091] At 904, criteria for generating personal content can be
obtained. For instance, the user can input criteria in a user
interface of the media device. As another example, the criteria can
be established based on a user profile and/or based on monitored
viewing behavior of the user and/or previous personal content that
was generated. In one embodiment, the criteria can be independent
of the preferences of the user, such as based on parental controls
or based on an automatic generation algorithm. At 906, a number of
different versions of the personal content can be generated based
on the criteria. The different versions can include the same or
different portions of the sensory information, such as images from
different viewing angles, different audio, and so forth. The
different versions can also be directed towards different target
audiences. For example, an immediate family version can be
generated that provides close-up images and audio of the user with
a limited amount of images and audio from other individuals. In one
embodiment, versions can be target to particular individuals. For
example, a version of the personal content may be generated that
includes all of the images and audio of the user and his friend,
while limiting images and audio of other individuals.
[0092] At 908, the media device or another component can analyze
the video and/or audio of each of the personal content for
objectionable material. For instance, the analysis can be automatic
without user interaction, based on image and speech recognition.
The analysis can be based on user interaction and/or the version of
the personal content that is being generated. For example, certain
images or words may be objectionable in the family version of the
personal content but may be acceptable the friend's version of the
content. At 910, one or more versions of the personal content can
be adjusted based on objectionable material, as well as based on
other criteria. In one embodiment, the audio generated from certain
individuals may be muted. In another embodiment, the images of
certain individuals may be obscured or otherwise blocked. The
muting and/or obscuring can be based on objectionable material,
user preferences, agreement negotiated with one or more suppliers
of the sensory information, and so forth. For instance, a second
mobile communication device may agree to provide sensory
information that includes images provided that images of the user
of the second mobile communication device are obscured in the
personal content.
[0093] At 912, a request for personal content can be received and a
determination of a relationship between the requestor and the
generator of the personal content can be made. The relationship can
be of many forms, including business and personal, and can be
customized or automatically defined. At 914, access to personal
content can be granted or denied based on a determined relationship
of a requestor. The access can be provided in a number of different
ways, such as via channels in an interactive television network
(e.g., IPTV), streaming between media devices, and so forth.
[0094] FIG. 10 depicts an illustrative embodiment of a system 1000
for analyzing and processing personal content that is generated
based at least in part on gathered sensory information associated
with an environment. System 1000 can be overlaid or operably
coupled to communication systems 100-200, 500-600 and 800 as
another representative embodiment of communication systems 100-200,
500-600 and 800. System 1000 can include a server 530 in
communication with media device or processor 506 (such as a set-top
box at location 849) and mobile wireless communication device 512
(such as mobile phones, smart phones, etc.), as well as other
mobile and/or fixed communication devices (not shown). The mobile
communication device 512 and the media processor 506 can be
communicatively coupled to the server 530 by way of a network 550
(e.g., a mobility network) coupled to the Internet or other
communication means. Personal content can be presented at location
849 using the display device 508.
[0095] A storage device 1010 can be utilized for storing various
personal content, which is generated in association with the mobile
communication device 512 and/or association with a user. The
storage device 1010 can be of various types and can be located in
various places. For example, the storage device 1010 can be a
database coupled with the server 530 for storing personal content
of various users of a service provider network. As another example,
the storage device 1010 can be integrated with the media processor
506 and/or can be a separate device (e.g., a DVR) connected with
the media processor at location 849. As yet another example, a
distributed storage environment can be utilized for storing the
personal content, which can include local and/or remote storage
devices, such as a cloud computing environment.
[0096] In this example, storage device 1010 can include personal
content 1015, 1020 and 1025, which are associated with a user of
mobile communication device 512. It should be understood that any
number of personal content can be stored by storage device 1010,
and the personal content can be associated with various entities,
including individuals, families, groups of friends, businesses, and
so forth.
[0097] Server 530 and/or media processor 506 (or another computing
device) can perform recognition analysis on personal content 1015
to identify person 1050. The recognition analysis can further be
performed on some or all of the other personal content in storage
device 1010, such as personal content 1020 and 1025 to identify
person 1050. In one embodiment, a user can provide a user input
1017 that highlights or otherwise indicates a person that is to be
flagged in other personal content. In one embodiment, a mouse arrow
can be placed over an individual and the individual in the image
can be highlighted. Data from the image (such as facial recognition
data) can be used as recognition information for performing image
recognition analysis on other personal content. In one embodiment,
a user can provide a user input 1019 that highlights or otherwise
indicates speech from a person that can be used to flag the person
in other personal content. In one embodiment, a selection of
sentences or phrases can be provided so that the user can identify
those sentences or phrases as being associated with an individual.
Data from the speech sample can be used as recognition information
for performing recognition analysis on audio content in other
personal content.
[0098] In one embodiment, a combination of image and audio
recognition can be performed. For example, a user can highlight a
person appearing in an image and the server 530 and/or the media
processor 506 can detect speech emanating from that person in the
personal content and use that speech as an audio sample for
subsequent recognition analysis. In one embodiment, the recognition
analysis can be performed without user intervention. For example,
image and/or audio samples can be obtained by the server 530 and/or
the media processor 506, such as in exchange for sensory
information, and the samples can be utilized to flag persons
without the user providing any user input.
[0099] As can be seen in system 1000, person 1050 can be identified
in each of the personal content 1015, 1020 and 1025. Once
identified or otherwise flagged, the personal content or portions
thereof can be made selectively accessible to the flagged
individuals. For example, a notice can be sent to communication
device(s) of flagged individuals so that they can access the
personal content and/or portions of the personal content. In one
embodiment, samples of flagged content, such as a still image taken
from video content that includes the flagged person, can be
provided to the flagged individual. The selective access can be
controlled by the user and/or according to various policies and/or
negotiated agreements, such as an agreement between two individuals
to share all personal content that includes both parties. In one
embodiment, only the portion of the personal content including the
flagged individual is made accessible while the remainder of the
personal content is rendered inaccessible. In one embodiment, all
of the personal content is made accessible to an individual flagged
in the particular personal content. In one embodiment, images of
the personal content can be made accessible while the audio content
is rendered inaccessible. Other variations of accessibility can
also be utilized. In one embodiment, the recognition analysis can
be utilized in combination with a social network. For example, a
flagged individual in the personal content of a user can be
provided with a greeting, introduction or other link to the user in
a social network environment as a result of the recognition
analysis.
[0100] FIG. 11 depicts an illustrative method 1100 that operates in
portions of the devices and methods of FIGS. 1-10. Method 1100 can
be performed by various devices or combinations of devices,
including the server 530 and/or the media processor 506. Method
1100 can begin with receiving sensory information at 1102 and
generating personal content at 1104. Additional detail as to these
steps is described with respect to methods 700 and 900.
[0101] At step 1106, it can be determined whether recognition
information was received in conjunction with the sensory
information. The recognition information can be of various types,
such as data associated with images of a person and/or audio
samples of the individual's speech. The recognition information can
be transmitted with the sensory information being gathered by a
particular communication device, such as another mobile
communication device that is in proximity to the environment. In
one embodiment, the recognition information can be stored in a
central database and distributed or otherwise accessed when sharing
of sensory information from a device associated with the
recognition information is detected. For instance, users of a
personal content service can each provide recognition information,
such as images and/or audio samples of their speech, which are
stored by the server 530 and which are associated with a particular
user and communication devices of the particular user. When
particular sensory information is gathered from a communication
device of a particular user and/or when personal content is
generated based on that particular sensory information, the server
530 can access the recognition information of the particular
user.
[0102] Recognition information can be shared between devices to
enable sharing of personal content or portions thereof with users
that have contributed to the generation of the personal content.
For instance, if two friends go to the park and capture sensory
information, such as images and audio of a football game being
played in the park, then the sharing of the recognition information
can enable sharing of those portions of the football game that
include one and/or the other of the two friends. Similarly, sharing
of recognition information can be selectively performed between
users that are not acquaintances. In this example, the sharing of
recognition information can be limited, such as preventing further
distribution of the recognition information, preventing display of
the recognition information (e.g., the recognition information can
be utilized by a device for performing image and/or audio
recognition analysis but cannot be presented by the same device),
providing the recognition information for a limited time, and so
forth.
[0103] At step 1108, if recognition information is available, then
one or more personal content can be analyzed to detect or otherwise
identify a person using the recognition information. The identified
person(s) can be tagged or flagged, including tagging frames of
video content and/or providing identification information of the
person in one or more images. In one embodiment, the identification
information can specify the person among other persons in an
image(s). In one embodiment, metadata can be inserted or otherwise
added to the personal content to facilitate the flagging of persons
based on the recognition information. The particular personal
content that is analyzed can vary. In one embodiment, all personal
content associated with a user be analyzed. In another embodiment,
personal content is analyzed subject to permission of the user,
such as limiting the analysis to events or environments that have
not been highlighted by the user as being family-only. In one
embodiment, personal content for analysis can be selected from a
group of personal content based on the environment from which the
sensory information was captured. For instance, personal content
can be provided with metadata that summarizes the particular
environment. The server 530 and/or the media processor 506 can
determine if there is a likelihood that a person flagged in
personal content of one environment might be present in personal
content of another environment. Various criteria, including
location, time, typical user activities, and so forth, can be used
in comparing the different environments to determine a likelihood
of presence of the flagged person.
[0104] At step 1110, one or more personal content can be analyzed
to detect or otherwise identify a person based on user input(s).
The user input(s) can be based on selecting persons being presented
in one or more images and/or selecting persons based on speech
recorded in the personal content. For example, the server 530
and/or the media processor 506 can present samples of images and/or
speech, and can request the user to identify the person (e.g.,
provide a name) or can request the user to select one or more of
the samples. The sample(s) can then be used in recognition analysis
applied to other personal content, as well as the same personal
content (e.g., selecting an image of the person and having other
portions of the same personal content highlighted where the person
again appears). In one embodiment, the recognition analysis can be
performed based on updated user input(s), newly received
recognition information that accompanied the sensory information,
and/or newly generated personal content.
[0105] At step 1112, notice can be provided based on the
recognition analysis and flagging of persons in the personal
content. The particular form of the notice can vary and can be
provided with or without samples of the personal content. For
example, where recognition information is received with sensory
information from a mobile communication device of a particular user
and that particular user is flagged in other personal content, then
the mobile communication device can receive notice, such as via
email, SMS and so forth, to notify the particular user of the
availability of the personal content. The notice can include a
summary of the environment of the personal content and/or can
include one or more images with or without the corresponding audio
in which the particular user is present.
[0106] The personal content in which a flagged person appears can
then be selectively shared with the flagged person, including
various communication devices of the flagged person. In one
embodiment, the sharing can be based on a negotiated agreement
between the user who generated the personal content and the flagged
person who provided the sensory information for generating the
personal content. Other criteria for sharing of the personal
content can also be utilized, including applying DRM rules to the
personal content and/or limiting or otherwise preventing further
distribution of the personal content (or portions thereof) to
others.
[0107] FIG. 12 depicts an illustrative embodiment of a system 1200
for managing advertisement content that is used with personal
content generated by a user. As described elsewhere herein, the
personal content can be generated based at least in part on
gathered sensory information associated with an environment, such
as from a group of mobile and/or fixed communication devices.
System 1200 can be overlaid or operably coupled to communication
systems 100-200, 500-600, 800 and 1000 as another representative
embodiment of communication systems 100-200, 500-600, 800 and
1000.
[0108] System 1200 can include a server 530 in communication with
media device or processor 506 (such as a set-top box at location
849) and mobile wireless communication device 512 (such as mobile
phones, smart phones, etc.), as well as other mobile and/or fixed
communication devices (not shown). The mobile communication device
512 and the media processor 506 can be communicatively coupled to
the server 530 by way of a network 550 (e.g., a mobility network)
coupled to the Internet or other communication means. Personal
content 1015 can be presented at location 849 using the display
device 508. The personal content 1015 be stored at a storage device
1010, which can be a local device (e.g., a memory of STB 506 and/or
a DVR coupled with the STB) and/or can be a remote device (e.g., a
database in communication with the server 530). While the exemplary
embodiment illustrates one personal content, it should be
understood that the exemplary embodiment can include any number of
personal content that is associated with or generated by various
entities, which may or may not include the user at location
849.
[0109] In one embodiment, server 530 can access personal content
1015 associated with the user of location 849 to determine targeted
advertisement content that can be provided with the personal
content. The server 530 can utilize a number of different factors
to select the advertisement content. In one embodiment, access to
the personal content 1015 by other subscribers or users can be
monitored. Information can be derived from the access to facilitate
selection of targeted advertisements, such as identification
information for the users (e.g., age, gender, occupation, salary,
and so forth). The identification information can be determined
from sources in combination with the monitoring of the access. For
instance, monitoring the user access to the personal content 1015
can enable the server 530 to determine the identity of the users.
User profiles associated with the users can then be consulted to
determine other identification information, such as the examples
described above. Sources other than user profiles can also be
utilized for gathering identification information, such as
subscriber plans of the users.
[0110] The monitoring of the access to the personal content can
also be utilized to determine the popularity of the personal
content (e.g., based on a number of times accessed and/or based on
a number of times distributed to others) and/or the popularity of
the user who generated the personal content (e.g., based on a
number of followers associated with the user such as in a social
network application). The popularity can be used for selecting the
advertisement content, such as providing different grades of
advertisement content to be used for different grades of
popularity.
[0111] In one embodiment, the server 530 can determine demographic
information for the personal content 1015, such as typical
characteristics of individuals that would desire to consume the
personal content. The demographic content can be used to predict
the types of users that will access the personal content 1015. In
one embodiment, the demographic information can be used as a
predictor of the users that will access the personal content 1015
and the identification information can be used to confirm the
predictor, such as correlating the demographic information with the
identification information, and adjusting the advertisement content
when there is a lack of correlation. The determined demographic
information and/or the identification information determined from
the access monitoring can be compared to demographic information
for the advertisement content for selecting the content, such as
selecting advertising images and/or audio that appeal to a certain
age group or other demographic group.
[0112] In one embodiment, server 530 can present offers to
advertise to various entities, such as an advertisement agency via
server 1230 and/or a merchant via server. In one embodiment, some
or all of the data gathered by the server 530 based on the
monitoring of the access and/or the demographic information can be
provided to the advertising agency server 1230 to select particular
advertisement content which can then be offered to the merchant via
server 1235.
[0113] An offer, whether made from the server 530 or from another
entity, can be based on the demographic information and/or the
identification information as described above. In one embodiment, a
value can be determined for the advertising and the offer can be
based on the value. For instance, personal content of a user that
has been accessed by several thousand people may be deemed as a
higher value than personal content that has been accessed by less
than one hundred people. Continuing with this example, a
relationship between the users who are accessing the personal
content can also be considered when determining a value of the
personal content for purposes of advertising. For instance, if the
users accessing the personal content are limited to a small
geographic area and are a part of a single family, then the
personal content for purposes of valuing the advertising may not be
as valuable as a situation where the users are spread over a vast
geographic area and are unrelated since there may be a higher
likelihood that the personal content will propagate to others. The
exemplary embodiments enable various devices to perform the
analysis described above, including the server 530, the media
processor 506, as well as the server 1230 and/or the server
1235.
[0114] In one embodiment, a spread rate for the personal content
can be determined and utilized in presenting an offer to an entity,
including placing a value on the advertisement content. For
instance, the spread rate can be determined based on various
factors such as the time it takes for the personal content to be
accessed by a threshold number of users and/or the time it takes
for the personal content to be accessed by particular categories of
users (e.g., based on demographics and a likelihood of purchasing
products) and/or the time it takes for the personal content to be
accessed by users in target locations. Server 530 can place a
higher value of personal content that has a desired spread rate
while placing a lower value on personal content having an undesired
spread rate. The exemplary embodiments enable various devices to
perform the analysis described above, including the server 530, the
media processor 506, as well as the server 1230 and/or the server
1235.
[0115] Advertisement content 1215 can be provided for presentation
with the personal content 1015. The particular type of
advertisement content can vary. For instance, the advertisement
content 1215 can be still or moving images that are overlayed on
the images of personal content 1015, such as in an area so as not
to block points of interest in the images (e.g., an upper corner of
the images). In one embodiment, the advertisement content can be
placed in the images of the personal content 1015 to simulate the
advertisement content being present in the environment. For
example, image recognition can be applied to the images to detect a
billboard that was captured in the images of the personal content
1015. The advertisement content can be overlayed on the billboard
to simulate the advertisement content having been on the billboard
at the time of capturing the images of the personal content 1015.
In one embodiment, other structures captured in the images and/or
open space captured in the images can be detected and the
advertisement content 1215 can be positioned on the detected
structures and/or open space, such as simulating the advertisement
content being on the side of a brick building and/or simulating an
advertisement sign being in a field captured by the images of
the
[0116] odiment, the advertisement content 1215 can include audio
content either alone or in combination with still or moving images.
In one embodiment, the advertisement content 1215 or a portion
thereof can be selectable in order to purchase product or service
associated with the advertisement content. In embodiment, a
selection of the advertisement content 1215 or a portion thereof
can provide a link to a website of a merchant (e.g., being run by
the server 1235) so that the user who selected the advertisement
content can purchase product or service associated with the
advertisement content. The exemplary embodiments can include other
configurations and methodologies to enable purchase of product via
a selection of the advertisement content 1215 or a portion thereof,
such as providing a link to a third party that manages the purchase
process. The exemplary embodiments can also include revenue sharing
for the user that generated the personal content and/or the service
provider that provided the access to the personal content. For
example, sales of products that are associated with a selection of
the advertisement content and/or access to a website of a merchant
via the selection of the advertisement content can used to
calculate revenue to be provided to one or both of the user that
generated the personal content and the service provider that
provided the access to the personal content.
[0117] In one embodiment, the advertisement content can be an
object 1225 that was captured in the images of the personal content
1015 and is adjusted to be advertisement content. For example,
object 1225 can be a soda bottle that is detected by the server 530
utilizing image recognition. The image and/or text recognition can
be used to identify a merchant associated with the particular brand
of soda. The personal content 1015 can be adjusted so that object
1225 can be selectable by a user who accesses the personal content.
A selection by the user can enable purchasing of the product (or
service depending on the object that is made selectable) and/or
reviewing additional information related to the selectable object
1225. In one embodiment, a symbol or highlighting of the perimeter
of the object 1225 can be used to indicate to the user who has
accessed the personal content 1225 that the object is selectable.
In one embodiment, a selection of the object 1225 can provide a
link to a website of a merchant that sells the product or service
(e.g., server 1235). The exemplary embodiments can include any
number and type of objects being made selectable. In one
embodiment, the number and/or types of objects made selectable can
be controlled by the user that generated the personal content. In
one embodiment, the type of objects made selectable can be based on
demographic information and/or the identification information
gathered from monitoring the access to the personal content 1015.
In one embodiment, all of the advertisement content included with
the personal content is selectable objects so that there are no
superimposed content overlaying the images. In another embodiment,
all of the advertisement content is superimposed content that
overlays the images so that there is no selectable objects that
were captured in the images. In another embodiment, a combination
of objects and superimposed advertisement content can be utilized
for the targeted advertisement.
[0118] FIG. 13 depicts an illustrative method 1300 that operates in
portions of the devices and methods of FIGS. 1-12. Method 1300 can
be performed by various devices or combinations of devices,
including one or more of the servers 530, 1230 and 1235 and/or the
media processor 506. Method 1300 can begin with step 1302 where
access to personal content associated with a user (e.g., generated
by that user) is monitored. The monitoring can include tracking the
identity of the users accessing the personal content and/or
counting the number of times the personal content has been
accessed. In one embodiment, the monitoring of the access can
include tracking the number of times particular portions of the
personal content have been accessed (e.g., to target advertising
space in a particular portion of the personal content). The
monitoring can be used in conjunction with other techniques for
gathering identification information beyond the name of the user,
such as accessing user profiles, subscriber plans, information
databases, social network profiles, and so forth, based on an
identity of the user. Various types of identification information
can be retrieved so as to facilitate selection of advertisement
content for the personal content and/or for evaluating a value of
advertising space in the personal content, such as demographics
associated with the users that are accessing and/or likely will
access the personal content.
[0119] At step 1304, demographic information can be determined for
the personal content. The demographic information can be determined
based on various factors and techniques. In one embodiment, the
user that generated the personal content can provide summary
information for the personal content that is utilized for
determining the demographics. For instance, the summary information
can be a description of the personal content environment, the
activities that occurred in the environment and/or individuals
captured in the personal content. The summary information can be
parsed or otherwise analyzed to determine demographic information
for the personal content. In one embodiment, image, text and/or
audio/speech recognition can be applied to the personal content to
determine the demographic information. The recognition analysis can
also be used to provide an automated summary of the personal
content that is generated without user intervention. In one
embodiment, other information can be utilized for determining
demographic information for the personal content. For instance, the
sensory information that is gathered to generate the personal
content can include metadata, such as identifying or categorizing
the sensory information. This metadata can be used to determine the
demographic information
[0120] At step 1306, the follower base (e.g., the users accessing
the personal content) can be evaluated, such as based on the
identification information and/or the demographic information
described above. The follower base can also be evaluated based on
other considerations. For instance, a particular user that
generated personal content may have a limited number of users that
have accessed the personal content but may have a large amount of
followers in a social network application. This can be taken into
consideration when determining the exposure that any advertisement
content included with the personal content may receive.
[0121] At step 1308, advertisement content can be provided for
presentation with the personal content, including based on various
factors that may include the access monitoring, the identification
information, the demographic information, the follower base
evaluation, the value of the advertising space, and so forth. The
advertisement content can take many forms, including adjusting
objects already present in the images to be selectable by a user,
overlaying advertisement content in areas of the images to minimize
blocking of the personal content and/or replacing objects in the
images with advertising content (e.g., replacing a sign in the
image with
[0122] , the advertisement content can be monitored for access. For
instance, selectable advertisement content can be monitored to
determine how many times users have selected and accessed the
advertisement content. In one embodiment, traffic being directed to
a merchant's website from the personal content via the selectable
advertisement content can be used to share revenue with the user
that generated the personal content and/or with the service
provider that provides the access to the personal content.
[0123] In one embodiment, sales resulting from access to the
advertisement content can also be monitored. For instance, a
selectable advertisement content can provide a link to a website
that sells a particular product. Whether or not the user purchases
the product using the website can be determined. Other monitoring
techniques are also contemplated by the exemplary embodiments,
including promotional codes being provided with the advertisement
content which can then be used when purchasing a product to obtain
a lower price and triggering an association between the purchase
and the advertisement content of the personal content. In one
embodiment, unique promotional codes can be used for the personal
content so that the monitoring of the purchase of a product is not
limited to use of a website link of the advertisement content, but
rather can be an independent purchase, such as by another user
directly accessing a merchant's website or going to the merchant's
store to purchase the product. In one embodiment, the unique
promotional codes can be dynamic, such as changing after a
pre-determined time period and/or after being accessed by a
different user.
[0124] At step 1312, revenue for the user that generated the
personal content and/or the service provider that provides access
to the personal content can be determined. The revenue can be
shared based on various criteria, including sales, traffic to a
website, negotiated agreement between the user and the merchant,
and so forth. If it is determined that the advertisement content
has not been accessed or the access fails to satisfy a minimum
threshold, then at step 1314 the advertisement content can be
adjusted. The adjustment to the advertisement content can be of
many types. For instance, the products being advertised can be
changed, the types of advertisements can be changed, the position
of the advertisement content in the personal content can be
changed, and so forth.
[0125] In one embodiment, the gathering of sensory information from
other communication devices can be based on a location of the
particular other devices. example, priority of retrieving sensory
information can be utilized in order to enable gathering sensory
information from a limited viewing angle. For instance, where a
first group of devices are located in close proximity and a second
group of devices are located remotely from each other, gathering of
the sensory information from the second group of devices can be
given priority to enable obtaining a wider perspective or viewing
angle of the environment. The use of prioritization for gathering
sensory information can be particularly helpful where a mobile
communication device has limited storage capacity and/or limited
time to negotiate an exchange of sensory information. By
prioritizing the gathering of sensory information to target devices
having better perspective over an environment, personal content can
be generated improved perspective.
[0126] Upon reviewing the aforementioned embodiments, it would be
evident to an artisan with ordinary skill in the art that said
embodiments can be modified, reduced, or enhanced without departing
from the scope and spirit of the claims described below. In one
embodiment, multiple personal content generated by a single user
can be processed together. For example, advertisement content can
be selected for the group and a value for the advertising space can
be determined for the entire group.
[0127] The exemplary embodiments can perform collection of sensory
information utilizing a number of different sources, including
sensors carried by a user, fixed sensors in building and street
infrastructure, sensors carried by friends and family, sensors
carried by bystanders, and so forth. The device discovery and/or
sharing of sensory data can be performed utilizing a number of
different techniques, such as a user broadcasting a mid-range
signal (e.g., 100 meters) indicating the user is recording his/her
life, devices with sensors in proximity can detect the broadcast,
devices can respond with opt-in or opt-out response to share data,
user of device may identify him/herself or remain private (e.g.,
friends and family can identify themselves while bystanders and
fixed infrastructure sensors remain anonymous.
[0128] In one embodiment, discovery of devices can be daisy chained
so that a user can build up a broader collection of sensory data,
such as a user in a sports complex or concert hall discovering
devices nearby and requesting a recording of the experience in the
user's proximity and throughout the sports complex or concert hall.
In this example, sensors can opt-in or opt-out, and information can
be shared by daisy chain or by transmission on a data channel
(e.g., 3G or 4G) to a server. Organizing of the collected sensory
information can be performed in a number of different ways, such as
based on an order of priority including vicinity to user, friends
and family, perspectives (e.g., front, back, side, rear). The
filtering of sensory information can be performed in a number of
different ways such as quality of sensory data, including clarity
visually, clarity audibly, identify portions that can be live, and
identify that can be still images.
[0129] FIG. 14 depicts an exemplary diagrammatic representation of
a machine in the form of a computer system 1400 within which a set
of instructions, when executed, may cause the machine to perform
any one or more of the methods discussed above. One or more
instances of the machine can operate, for example, as the server
530, 1230, 1235, the media processor 506, the display 508, the
fixed communication device 516, mobile devices 512, 514, or
combinations thereof as described above. In some embodiments, the
machine may be connected (e.g., using a network) to other machines.
In a networked deployment, the machine may operate in the capacity
of a server or a client user machine in server-client user network
environment, or as a peer machine in a peer-to-peer (or
distributed) network environment.
[0130] The machine may comprise a server computer, a client user
computer, a personal computer (PC), a tablet PC, a smart phone, a
laptop computer, a desktop computer, a control system, a network
router, switch or bridge, or any machine capable of executing a set
of instructions (sequential or otherwise) that specify actions to
be taken by that machine. It will be understood that a
communication device of the present disclosure includes broadly any
electronic device that provides voice, video or data communication.
Further, while a single machine is illustrated, the term "machine"
shall also be taken to include any collection of machines that
individually or jointly execute a set (or multiple sets) of
instructions to perform any one or more of the methods discussed
herein.
[0131] The computer system 1400 may include a processor 1402 (e.g.,
a central processing unit (CPU), a graphics processing unit (GPU,
or both), a main memory 1404 and a static memory 1406, which
communicate with each other via a bus 1408. The computer system
1400 may further include a video display unit 1410 (e.g., a liquid
crystal display (LCD), a flat panel, or a solid state display. The
computer system 1400 may include an input device 1412 (e.g., a
keyboard), a cursor control device 1414 (e.g., a mouse), a disk
drive unit 1416, a signal generation device 1418 (e.g., a speaker
or remote control) and a network interface device 1420.
[0132] The disk drive unit 1416 may include a tangible
computer-readable storage medium 1422 on which is stored one or
more sets of instructions (e.g., software 1424) embodying any one
or more of the methods or functions described herein, including
those methods illustrated above. The instructions 1424 may also
reside, completely or at least partially, within the main memory
1404, the static memory 1406, and/or within the processor 1402
during execution thereof by the computer system 1400. The main
memory 1404 and the processor 1402 also may constitute tangible
computer-readable storage media.
[0133] Dedicated hardware implementations including, but not
limited to, application specific integrated circuits, programmable
logic arrays and other hardware devices can likewise be constructed
to implement the methods described herein. Applications that may
include the apparatus and systems of various embodiments broadly
include a variety of electronic and computer systems. Some
embodiments implement functions in two or more specific
interconnected hardware modules or devices with related control and
data signals communicated between and through the modules, or as
portions of an application-specific integrated circuit. Thus, the
example system is applicable to software, firmware, and hardware
implementations.
[0134] In accordance with various embodiments of the present
disclosure, the methods described herein are intended for operation
as software programs running on a computer processor. Furthermore,
software implementations can include, but not limited to,
distributed processing or component/object distributed processing,
parallel processing, or virtual machine processing can also be
constructed to implement the methods described herein.
[0135] While the tangible computer-readable storage medium 1422 is
shown in an example embodiment to be a single medium, the term
"tangible computer-readable storage medium" should be taken to
include a single medium or multiple media (e.g., a centralized or
distributed database, and/or associated caches and servers) that
store the one or more sets of instructions. The term "tangible
computer-readable storage medium" shall also be taken to include
any non-transitory medium that is capable of storing or encoding a
set of instructions for execution by the machine and that cause the
machine to perform any one or more of the methods of the present
disclosure.
[0136] The term "tangible computer-readable storage medium" shall
accordingly be taken to include, but not be limited to: solid-state
memories such as a memory card or other package that houses one or
more read-only (non-volatile) memories, random access memories, or
other re-writable (volatile) memories, a magneto-optical or optical
medium such as a disk or tape, or other tangible media which can be
used to store information. Accordingly, the disclosure is
considered to include any one or more of a tangible
computer-readable storage medium, as listed herein and including
art-recognized equivalents and successor media, in which the
software implementations herein are stored.
[0137] Although the present specification describes components and
functions implemented in the embodiments with reference to
particular standards and protocols, the disclosure is not limited
to such standards and protocols. Each of the standards for Internet
and other packet switched network transmission (e.g., TCP/IP,
UDP/IP, HTML, HTTP) represent examples of the state of the art.
Such standards are from time-to-time superseded by faster or more
efficient equivalents having essentially the same functions.
Wireless standards for device detection (e.g., RFID), short-range
communications (e.g., Bluetooth, WiFi, Zigbee), and long-range
communications (e.g., WiMAX, GSM, CDMA) are contemplated for use by
computer system 1400.
[0138] The illustrations of embodiments described herein are
intended to provide a general understanding of the structure of
various embodiments, and they are not intended to serve as a
complete description of all the elements and features of apparatus
and systems that might make use of the structures described herein.
Many other embodiments will be apparent to those of skill in the
art upon reviewing the above description. Other embodiments may be
utilized and derived therefrom, such that structural and logical
substitutions and changes may be made without departing from the
scope of this disclosure. Figures are also merely representational
and may not be drawn to scale. Certain proportions thereof may be
exaggerated, while others may be minimized. Accordingly, the
specification and drawings are to be regarded in an illustrative
rather than a restrictive sense.
[0139] Although specific embodiments have been illustrated and
described herein, it should be appreciated that any arrangement
calculated to achieve the same purpose may be substituted for the
specific embodiments shown. This disclosure is intended to cover
any and all adaptations or variations of various embodiments.
Combinations of the above embodiments, and other embodiments not
specifically described herein, will be apparent to those of skill
in the art upon reviewing the above description.
[0140] The Abstract of the Disclosure is provided with the
understanding that it will not be used to interpret or limit the
scope or meaning of the claims. In addition, in the foregoing
Detailed Description, it can be seen that various features are
grouped together in a single embodiment for the purpose of
streamlining the disclosure. This method of disclosure is not to be
interpreted as reflecting an intention that the claimed embodiments
require more features than are expressly recited in each claim.
Rather, as the following claims reflect, inventive subject matter
lies in less than all features of a single disclosed embodiment.
Thus the following claims are hereby incorporated into the Detailed
Description, with each claim standing on its own as a separately
claimed subject matter.
* * * * *