U.S. patent application number 12/874022 was filed with the patent office on 2010-12-30 for methods and apparatuses for sequestering content.
This patent application is currently assigned to SONY CORPORATION. Invention is credited to Eric Edwards, Clay Fisher, Neal Manowitz, Robert Sato.
Application Number | 20100333210 12/874022 |
Document ID | / |
Family ID | 33493321 |
Filed Date | 2010-12-30 |
United States Patent
Application |
20100333210 |
Kind Code |
A1 |
Fisher; Clay ; et
al. |
December 30, 2010 |
METHODS AND APPARATUSES FOR SEQUESTERING CONTENT
Abstract
In one embodiment, the methods and apparatuses sequester content
receiving content for use in an application; review the content;
automatically sequester the content from the application based on
the reviewing; and form a reason associated with the sequestering
the content. In another embodiment, the methods and apparatuses
receive content for use with an application; determine whether the
content is one of acceptable content and unacceptable content;
remove the unacceptable content from the application; form an
explanation for the unacceptable content; and store the
unacceptable content and the explanation in an off-line storage
device.
Inventors: |
Fisher; Clay; (San Diego,
CA) ; Edwards; Eric; (Milolii, HI) ; Manowitz;
Neal; (Montebello, NY) ; Sato; Robert; (San
Diego, CA) |
Correspondence
Address: |
O'BANION & RITCHEY LLP/ SONY ELECTRONICS, INC.
400 CAPITOL MALL, SUITE 1550
SACRAMENTO
CA
95814
US
|
Assignee: |
SONY CORPORATION
TOKYO
NJ
SONY ELECTRONICS INC.
PARK RIDGE
|
Family ID: |
33493321 |
Appl. No.: |
12/874022 |
Filed: |
September 1, 2010 |
Related U.S. Patent Documents
|
|
|
|
|
|
Application
Number |
Filing Date |
Patent Number |
|
|
11207253 |
Aug 19, 2005 |
7809645 |
|
|
12874022 |
|
|
|
|
10771903 |
Feb 4, 2004 |
|
|
|
11207253 |
|
|
|
|
60472691 |
May 22, 2003 |
|
|
|
Current U.S.
Class: |
726/26 ;
715/751 |
Current CPC
Class: |
G06Q 10/10 20130101 |
Class at
Publication: |
726/26 ;
715/751 |
International
Class: |
G06F 3/01 20060101
G06F003/01; G06F 15/16 20060101 G06F015/16; G06F 21/00 20060101
G06F021/00 |
Claims
1-30. (canceled)
31. An electronic photo album, comprising: at least one processor,
and associated memory, configured for processing submitted content
containing photographic images and/or video; programming executable
on said at least one processor for receiving content from a
submitter, as submitted content, within an electronic photo album
application for, reviewing the submitted content for unacceptable
material, automatically removing unacceptable material, within said
submitted content, from said electronic photo album application and
sequestering it so that it is not accessible to third party users
of said electronic photo album application, and notifying the
submitter of why the submitted content is unacceptable and has been
sequestered.
32. The electronic photo album as recited in claim 31, wherein said
processor is configured with a user interface and has a video
display.
33. The electronic photo album as recited in claim 31, wherein said
processor is configured for being conventionally coupled to a
camera device.
34. The electronic photo album as recited in claim 31, wherein said
electronic photo album application is executed on a server computer
coupled to a network.
35. The electronic photo album as recited in claim 31, wherein said
programming is further configured for providing electronic sharing
of submitted content, from said electronic photo album, over an
electronic network.
36. The electronic photo album as recited in claim 31, wherein said
network comprises a local area network, a home network, and/or the
Internet.
37. The electronic photo album as recited in claim 31, wherein said
programming of said electronic photo album application is
configured for viewing, and/or editing, of digital image data.
38. The electronic photo album as recited in claim 37, wherein said
editing is selected from the group of photo editing operations
consisting of: formatting, adding captions, positioning, and adding
backgrounds.
39. The electronic photo album as recited in claim 31, wherein said
programming is further configured to allow the submitter to share
with others digital image data from said electronic photo album
over a network.
40. The electronic photo album as recited in claim 31, wherein said
reviewing comprises programming configured to analyze the submitted
content in relation to a threshold and identifying unacceptable
content which exceeds said threshold.
41. The electronic photo album as recited in claim 31, wherein said
undesirable material is selected by said programming from the group
of content consisting of content over a predetermined size,
copyrighted material, pornographic material, violent material,
illegal material, illegal material for a given jurisdiction,
content which violates a terms of service agreement, content which
contains a watermark, and any standard determined by the operator
of said photo album application.
42. The electronic photo album as recited in claim 31, wherein said
sequestered content is stored by said programming in an off-line
storage device.
43. The electronic photo album as recited in claim 31, wherein the
programming further comprises annotating the content with the
reason for the content being sequestered.
44. The electronic photo album as recited in claim 31, wherein said
notifying of the submitter is performed in response to programming
for sending the submitter a notification of sequestered content
within an electronic mail message.
45. The electronic photo album as recited in claim 31, wherein the
programming further comprises returning, to the submitter,
unacceptable material from within said submitted content.
46. The electronic photo album as recited in claim 31, wherein the
programming further comprises sequestering the unacceptable
material within said submitted content to prevent it from being
accessed by the submitter through the electronic photo album
application.
47. An electronic photo album, comprising: at least one processor,
and associated memory, configured for processing submitted content
containing digital image data; a user interface and video display
coupled to said at least one processor for controlling said photo
album application; programming executable on said at least one
processor for receiving content from a submitter, as submitted
content, within an electronic photo album application configured
for electronic sharing of submitted content over an electronic
network for, reviewing the submitted content for unacceptable
material in response to analyzing the submitted content based on a
threshold and identifying unacceptable content which exceeds said
threshold, automatically removing unacceptable material, within
said submitted content, from said electronic photo album
application and sequestering it so that it is not accessible to
third party users of said electronic photo album application, and
notifying the submitter of why the submitted content is
unacceptable.
48. The electronic photo album as recited in claim 47, wherein said
undesirable material is selected by said programming from the group
of content consisting of content over a predetermined size,
copyrighted material, pornographic material, violent material,
illegal material, illegal material for a given jurisdiction,
content which violates a terms of service agreement, content which
contains a watermark, and any standard determined by the operator
of said photo album application.
49. The electronic photo album as recited in claim 47, wherein said
notifying of the submitter is performed in response to programming
for sending the submitter an electronic mail message or annotating
the content with the reason for the content being sequestered.
50. A method of sequestering content from an electronic photo album
application executing on a computer processor and associated
memory, comprising the steps of: receiving submitted content
containing photographic images and/or video within an electronic
photo album application; reviewing the submitted content for
unacceptable material; automatically removing unacceptable
material, within said submitted content, from said electronic photo
album application and sequestering it so that it is not accessible
to third party users of said electronic photo album application;
and notifying the submitter of why the submitted content is
unacceptable.
Description
CROSS-REFERENCE TO RELATED APPLICATIONS
[0001] This application is a continuation of U.S. patent
application Ser. No. 11/207,253 filed on Aug. 19, 2005, now U.S.
Pat. No. ______, incorporated herein by reference in its entirety,
which is a continuation of U.S. patent application Ser. No.
10/771,903 filed on Feb. 4, 2004, incorporated herein by reference
in its entirety, which is a nonprovisional of U.S. provisional
patent application Ser. No. 60/472,691 filed on May 22, 2003,
incorporated herein by reference in its entirety. Priority is
claimed to each of the foregoing U.S. patent applications.
FIELD OF THE INVENTION
[0002] The present invention relates generally to sequestering
content and, more particularly, to automatically sequestering
content.
BACKGROUND
[0003] There has been a proliferation of on-line applications that
utilize content uploaded by users. For example, there are many
electronic photo album and file sharing applications available to
users over the Internet.
[0004] In some instances, the electronic photo album applications
allow the user to submit their own content to create their own
photo album. In addition to submitting content, the submitter is
also capable of formatting the content and creating their own
customized electronic photo album by adding captions, positioning
the content, adding backgrounds, and the like. In many instances,
the submitter displays the customized electronic photo album to
others. In some examples, the submitter can restrict access to the
customized electronic photo album by others.
[0005] The submitter typically agrees to a "terms of service"
agreement with an operator of the application. The terms of service
agreement usually informs the submitter that certain content is
considered unacceptable for use with the application. Examples of
unacceptable content includes pornography, obscene materials,
copyrighted materials, illegal materials, and the like.
[0006] Some operators of these on-line applications enforce their
terms of service agreement in an ad hoc manner relying on spot
checks of content or complaints from other users viewing
unacceptable content. Other operators automatically check all
content uploaded from submitters.
SUMMARY
[0007] In one embodiment, the methods and apparatuses sequester
content receiving content for use in an application; review the
content; automatically sequester the content from the application
based on the reviewing; and form a reason associated with the
sequestering the content. In another embodiment, the methods and
apparatuses receive content for use with an application; determine
whether the content is one of acceptable content and unacceptable
content; remove the unacceptable content from the application; form
an explanation for the unacceptable content; and store the
unacceptable content and the explanation in an off-line storage
device.
BRIEF DESCRIPTION OF THE DRAWINGS
[0008] The accompanying drawings, which are incorporated in and
constitute a part of this specification, illustrate and explain one
embodiment of the methods and apparatuses for sequestering content.
In the drawings,
[0009] FIG. 1 is a diagram illustrating an environment within which
the methods and apparatuses for sequestering content are
implemented;
[0010] FIG. 2 is a simplified block diagram illustrating one
embodiment in which the methods and apparatuses for sequestering
content are implemented;
[0011] FIG. 3 is a simplified block diagram illustrating a system,
consistent with one embodiment of the methods and apparatuses for
sequestering content;
[0012] FIG. 4 is a flow diagram consistent with one embodiment of
the methods and apparatuses for sequestering content;
[0013] FIG. 5 is a flow diagram consistent with one embodiment of
the methods and apparatuses for sequestering content;
[0014] FIG. 6 is a flow diagram consistent with one embodiment of
the methods and apparatuses for sequestering content;
[0015] FIG. 7 is an exemplary notification consistent with one
embodiment of the methods and apparatuses for sequestering content;
and
[0016] FIG. 8 is an exemplary notification consistent with one
embodiment of the methods and apparatuses for sequestering
content.
DETAILED DESCRIPTION
[0017] The following detailed description of the methods and
apparatuses for sequestering content refers to the accompanying
drawings. The detailed description is not intended to limit the
methods and apparatuses for sequestering content. Instead, the
scope of the methods and apparatuses for sequestering content is
defined by the appended claims and equivalents. Those skilled in
the art will recognize that many other implementations are
possible, consistent with the present invention.
[0018] References to "content" includes data such as audio, video,
text, graphics, and the like, that are embodied in digital or
analog electronic form. References to "applications" includes
programs accessible through a network such as the Internet for
tasks such as word processing, audio output or editing, video
output or editing, digital still photograph viewing or editing,
file sharing, and the like, that are embodied in hardware and/or
software.
[0019] References to an "operator" include individuals,
organizations, or entities that manage and maintain the
applications. References to a "submitter" includes individuals,
organizations, or entities that utilize the applications.
References to a "user" refers to either the submitter or the
operator.
[0020] FIG. 1 is a diagram illustrating an environment within which
the methods and apparatuses for sequestering content are
implemented. The environment includes an electronic device 110
(e.g., a computing platform configured to act as a client device,
such as a personal computer, a personal digital assistant, a
cellular telephone, a paging device), a user interface 115, a
network 120 (e.g., a local area network, a home network, the
Internet), and a server 130 (e.g., a computing platform configured
to act as a server).
[0021] In one embodiment, one or more user interface 115 components
are made integral with the electronic device 110 (e.g., keypad and
video display screen input and output interfaces in the same
housing as personal digital assistant electronics (e.g., as in a
Clie.RTM. manufactured by Sony Corporation). In other embodiments,
one or more user interface 115 components (e.g., a keyboard, a
pointing device (mouse, trackball, etc.), a microphone, a speaker,
a display, a camera) are physically separate from, and are
conventionally coupled to, electronic device 110. The user utilizes
interface 115 to access and control content and applications stored
in electronic device 110, server 130, or a remote storage device
(not shown) coupled via network 120.
[0022] In accordance with the invention, embodiments of
sequestering content as described below are executed by an
electronic processor in electronic device 110, in server 130, or by
processors in electronic device 110 and in server 130 acting
together. Server 130 is illustrated in FIG. 1 as being a single
computing platform, but in other instances are two or more
interconnected computing platforms that act as a server.
[0023] The methods and apparatuses for sequestering content are
shown in the context of exemplary embodiments of applications in
which a user posts content on the application. In some embodiments,
the submitter posts the content through the electronic device 110
and the network 120. In some embodiments, the content posted by the
submitter is utilized by the application which is located within
the server 130. Exemplary applications include a file sharing
application, an electronic photo album application, and the
like.
[0024] In one embodiment, the methods and apparatuses for
sequestering content automatically reviews the content provided by
the submitter and prevents the application from accessing selected
unacceptable content based on this review. In some instances, the
selected unacceptable content is moved to an off-line storage and
is unavailable to both the submitter that posted the selected
unacceptable content and other third party users attempting to view
the selected unacceptable content through the application. In other
instances, the selected unacceptable content is returned back to
the submitter or is made available to the application depending on
the subject matter of the content.
[0025] In one embodiment, the methods and apparatuses for
sequestering content automatically records a reason to explain why
the content is considered unacceptable. In some embodiments, this
reason is documented and linked to the unacceptable content for
future reference.
[0026] In one embodiment, the methods and apparatuses for
sequestering content automatically notifies the submitter that the
content is unacceptable. In some embodiments, the submitter is also
given the reason that the content is considered unacceptable.
[0027] In one embodiment, the methods and apparatuses for
sequestering content are located within the server 130. In some
embodiments, an operator monitors and provides instructions to the
methods and apparatuses for sequestering content through the
electronic device 110 and the network 120.
[0028] FIG. 2 is a simplified diagram illustrating an exemplary
architecture in which the methods and apparatuses for sequestering
content are implemented. The exemplary architecture includes a
plurality of electronic devices 110, a server device 130, and a
network 120 connecting electronic devices 110 to server 130 and
each electronic device 110 to each other. The plurality of
electronic devices 110 are each configured to include a
computer-readable medium 209, such as random access memory, coupled
to an electronic processor 208. Processor 208 executes program
instructions stored in the computer-readable medium 209. A unique
user operates each electronic device 110 via an interface 115 as
described with reference to FIG. 1.
[0029] Server device 130 includes a processor 211 coupled to a
computer-readable medium 212. In one embodiment, the server device
130 is coupled to one or more additional external or internal
devices, such as, without limitation, a secondary data storage
element, such as database 240.
[0030] In one instance, processors 208 and 211 are manufactured by
Intel Corporation, of Santa Clara, Calif. In other instances, other
microprocessors are used.
[0031] The plurality of client devices 110 and the server 130
include instructions for a customized application for sequestering
content. In one embodiment, the plurality of computer-readable
medium 209 and 212 contain, in part, the customized application.
Additionally, the plurality of client devices 110 and the server
130 are configured to receive and transmit electronic messages for
use with the customized application. Similarly, the network 120 is
configured to transmit electronic messages for use with the
customized application.
[0032] One or more user applications are stored in memories 209, in
memory 211, or a single user application is stored in part in one
memory 209 and in part in memory 211. In one instance a stored user
application, regardless of storage location, is made customizable
based on sequestering content as determined using embodiments
described below.
[0033] FIG. 3 illustrates one embodiment of a sequestering system
300. The sequestering system 300 includes a review module 310, a
capture module 320, a storage module 330, an interface module 340,
and a control module 350. In some embodiments, the control module
350 communicates with the review module 310, the capture module
320, the storage module 330, and the interface module 340.
[0034] In one embodiment, the control module 350 coordinates tasks,
requests and communications between the review module 310, the
capture module 320, the storage module 330, and the interface
module 340.
[0035] In one embodiment, the review module 310 analyzes content
via the capture module 320. in many embodiments, the review module
310 is configured to analyze the content and identify undesirable
content such as copyrighted material, pornographic material,
violent material and the like. In some cases, this undesirable
content is illegal or violates a terms of service agreement between
the submitter and the application. For example, in one instance,
the review module 310 analyzes and identifies content as being
copyrighted material, because the review module 310 detects that
the content contains a watermark. In other embodiments, the review
module 310 is customized to identify any type of content.
[0036] In another embodiment, the review module 310 is substituted
with an operator that reviews the content and determines if the
content is unacceptable.
[0037] In one embodiment, the review module 310 also annotates the
content that is found unacceptable. For example, in one instance,
the review module 310 finds that the content is undesirable because
the content is copyrighted. In this example, this content is
annotated with a label "copyrighted material".
[0038] In one embodiment, the capture module 320 identifies
specific content for use by the sequestering system 300. In some
embodiments, the capture module 320 identifies content that is
posted by a submitter. In addition, the capture module 320 supplies
content to the review module 310.
[0039] In one embodiment, the capture module 320 also identifies
descriptive information associated with the content. For example,
in one instance, the capture module identifies the location of the
content within a photo album, the background color behind the
content, a caption describing the content, and the like.
[0040] In one embodiment, the off-line storage module 330 stores
the content and associated information such that the content and
associated information is not available to the submitter or the
general public. In some embodiments, the associated information
includes formatting information relating to the content and
annotations which describe why the content is unacceptable.
[0041] In one embodiment, the information stored within the
off-line storage module 330 is encrypted. In addition, the content
stored within the off-line storage module 330 is stored in an
abbreviated form in some embodiments. For example, instead of the
content being stored at fully resolution, the content is stored as
a thumbnail.
[0042] In one embodiment, the annotations explaining reasons why
the content is unacceptable which are stored within the off-line
storage module are made available to multiple applications. For
example, when the annotations are shared with multiple
applications, other applications are forewarned about unacceptable
content originating from the submitter.
[0043] In one embodiment, the interface module 340 receives
instructions from an operator of the sequestering system 300. For
example, in one instance, the operator instructs the sequestering
system 300 to return the content stored within the off-line storage
module 330 to the submitter.
[0044] In another embodiment, the interface module 340 displays
content and information associated with the content to the
operator. In some instances, the operator manually reviews content
stored within the off-line storage module 330.
[0045] In yet another embodiment, the interface module 340
interacts with the submitter regarding the submitter's content that
is considered unacceptable. For example, in one instance, the
interface module 340 notifies the submitter when the submitter's
content is considered unacceptable. In another instance, the
submitter provides feedback to the sequestering system 300
regarding the submitter's content via the interface module 340.
[0046] In an additional embodiment, the interface module 340
interacts with other applications. For example, in some instances,
when content that is submitted to an application from a submitter
that is considered unacceptable, the interface module 340 instructs
the application to remove the unacceptable content or prevent the
unacceptable content from being utilized.
[0047] The sequestering system 300 in FIG. 3 is shown for exemplary
purposes and is merely one embodiment of the methods and
apparatuses for sequestering content. Additional modules may be
added to the system 300 without departing from the scope of the
methods and apparatuses for sequestering content. Similarly,
modules may be combined or deleted without departing from the scope
of the methods and apparatuses for sequestering content.
[0048] The flow diagrams as depicted in FIGS. 4, 5, and 6 are one
embodiment of the methods and apparatuses for sequestering content.
The blocks within the flow diagrams can be performed in a different
sequence without departing from the spirit of the methods and
apparatuses for sequestering content. Further, blocks can be
deleted, added, or combined without departing from the spirit of
the methods and apparatuses for sequestering content.
[0049] The flow diagram in FIG. 4 illustrates sequestering content
according to one embodiment of the invention. In Block 410, content
is received. In one embodiment, the content is sensed by the
capture module 320 (FIG. 3). In some instances, the submitter
offers the content to be published or viewed on an application such
as a file sharing program, a photo album, and the like.
[0050] In Block 415, the content is available for use by the
application. In one embodiment, the content is published by the
application such as a photograph within an electronic on-line photo
album application. In another embodiment, the content is made
available to others via a file sharing application.
[0051] In Block 420, the content is reviewed based on the subject
matter of the content. In one embodiment, the content is analyzed
by the review module 310 (FIG. 3). In many instances, the content
is determined to be either acceptable or unacceptable based on a
variety of criteria. For example, if the content is pornography or
copyrighted, the content is considered unacceptable. In another
example, the content is determined to be unacceptable based on a
violation of the terms of service agreement.
[0052] In Block 425, the content is deemed either acceptable or
unacceptable based on the review of the content in the Block
420.
[0053] If the content is considered acceptable, the content remains
available to the application in Block 430. For example, if the
content is a photograph that was posted in an on-line album in the
Block 415, the photograph continues to be available within the
Block 430.
[0054] If the content is considered unacceptable, the content is
annotated with a reason that the content is unacceptable in Block
435. In one instance, the reasons for the content being
unacceptable include pornography, copyrighted material, illegal
material, and the like.
[0055] In Block 440, the unacceptable content and the annotated
reason are removed from the application and moved to an off-line
storage facility. In one embodiment, the off-line storage facility
is the off-line storage module 330 (FIG. 3). When the unacceptable
content is removed from the application, the application cannot
gain access to the unacceptable content. For example, if the
unacceptable content is displayed within the submitter's electronic
photo album by an electronic on-line photo album application, once
the unacceptable content is removed, the application is notified;
and the unacceptable content is no longer displayed within the
submitter's electronic photo album.
[0056] In Block 445, the unacceptable content, the annotated
reason, and any formatting information associated with the
unacceptable content are stored in the off-line storage facility.
In one embodiment, access to any content within the off-line
storage facility is restricted to the operator. In other words,
access to content within the off-line storage facility is not
available to any applications, the submitter, or other third
parties. In another embodiment, access to content within the
off-line storage facility is available to others with authorization
from the operator.
[0057] The flow diagram in FIG. 5 illustrates notifying a submitter
of unacceptable content according to one embodiment of the
invention. In Block 520, the user is notified that the submitter's
content is unacceptable. The Block 520 is a continuation from the
Block 445 in FIG. 4. In some embodiments, the submitter is notified
via an electronic mail message. In some embodiments, the
notification identifies the particular content that is
unacceptable. In other embodiments, the notification also includes
the reason why the content is deemed unacceptable.
[0058] Additionally, in some embodiments, the notification requests
feedback from the submitter if the submitter wishes to dispute the
classification of the unacceptable content and to clarify the
nature of the content. For example, if the notification informs the
submitter that the submitter's content appears to be copyrighted
material. In this case, the submitter has several options such as
not responding if the content is copyrighted; verifying that the
submitter has permission to use this copyrighted material; and
contesting that the content is not copyrighted.
[0059] If feedback from the submitter is not received, the content,
associated information, and reason for the content being
unacceptable remain stored within the off-line storage and are not
accessible to the submitter, any applications, or third parties as
shown in Block 560.
[0060] In one embodiment, if feedback from the submitter is
received, the content is reviewed again in light of the feedback
from the submitter. In another embodiment, the content is reviewed
by the operator. In yet another embodiment, the content is reviewed
again by the review module 310. If the content is considered to
violate any laws in the Block 540, the content and associated
information remain stored within the off-line storage and are not
accessible to the submitter, any applications, or third parties as
shown in Block 560. For example, in one instance, public display of
pornography is a type of material which can violate the law.
Violations of law can vary from jurisdiction to jurisdiction and
continually change.
[0061] In another embodiment, the threshold in the Block 540 of
violating the law is substituted by any standard determined by the
operator that prevents the content from being returned to the
submitter.
[0062] If the content is found to not exceed a threshold as
determined in the Block 540, the content is reviewed for compliance
with the terms of service as agreed upon by the submitter in Block
550. For example, some terms of service agreements prohibit posting
copyrighted works, files over a predetermined size, and the like.
If the content is determined to violate the terms of service
agreement but otherwise considered returnable content, the content
is returned to the submitter with an additional notification to the
submitter as shown in Block 570.
[0063] In another embodiment, the threshold in the Block 550 of
violating the terms of service agreement is substituted by any
standard determined by the operator that prevents the content from
being posted by the application but returnable to the
submitter.
[0064] If the content is considered to conform to the terms of
service agreement in the Block 550 and the threshold in the Block
540, the content and associated information is recovered from the
off-line storage and made available for use by applications as
shown in Block 580.
[0065] For example, a photograph posted in an electronic photo
album application is originally found unacceptable in the Block 425
and moved to the off-line storage such that the posted photograph
was no longer available to the electronic photo album application
in the Block 440. The reason for the content being found
unacceptable is documented and stored in the Block 435.
[0066] In this example, the submitter is notified in the Block 520
of the content being found unacceptable. However, this photograph
was subsequently found acceptable because upon re-evaluation, the
photograph conforms to legal standards and the "terms of service"
in the Blocks 540 and 550. In this example, this photograph and
formatting information associated with the photograph is returned
to the electronic photo album application so that the photograph is
re-displayed in the electronic photo album application in the same
format as originally specified by the submitter in the Block
580.
[0067] In one embodiment, the photograph is automatically
re-displayed in the electronic photo album application without
input from either the submitter or the operator.
[0068] In another example, if the photograph was found to conform
with legal standards (Block 540) but fails to conform to the terms
of service agreement, the photograph is returned to the submitter
in the Block 570.
[0069] The flow diagram in FIG. 6 illustrates interacting with a
submitter regarding sequestering content according to one
embodiment of the invention. In Block 610 content submitted by a
submitter is found to be unacceptable.
[0070] In Block 620, an automated notification form is sent to the
submitter identifying the unacceptable content and the reason for
finding the content unacceptable. In addition, in some embodiments,
the automated notification form also allows the submitter to make
an inquiry regarding the unacceptable content. In FIG. 7, the
automated notification form 700 is shown for exemplary purposes.
The form 700 includes a title field 710 to identify the
unacceptable content, a reason field 720 to indicate the reason for
finding the content unacceptable, and an inquiry button 730 to
allow the submitter to inquire further regarding the unacceptable
content. In some embodiments, when the submitter selects the
inquiry button 730, an inquiry is automatically transmitted to
initiate a review process.
[0071] In Block 625, the inquiry from the user regarding the
automated notification form is received and the review process of
the unacceptable content is initiated.
[0072] In Block 630, the submitter is sent a follow-up form which
requests additional details from the submitter regarding the
unacceptable content. In FIG. 8, a sample follow-up form 800 is
shown. The form 800 includes a title field 810 to identify the
unacceptable content, a reason field 820 to indicate the reason for
finding the content unacceptable, and a justification field 830 to
indicate the justification that the unacceptable content should be
considered acceptable. In this example, the content is considered
unacceptable, because the content is believed to contain
copyrighted materials. Under the justification field 830, the
submitter is able to assert one of the following: the content does
not contain copyrighted materials; the content contains copyrighted
materials but the submitter has permission to utilize the
copyrighted materials; or "other reason" which the submitter types
in his/her own justification.
[0073] In Block 640, the follow-up form is received from the
submitter.
[0074] In Block 650, the follow-up form is attached to the
corresponding unacceptable content.
[0075] In Block 660, the unacceptable content is reassessed based
on the follow-up form and the unacceptable content.
[0076] The foregoing descriptions of specific embodiments of the
invention have been presented for purposes of illustration and
description. For example, the invention is described within the
context of creating profiles for modifying digital images as merely
one embodiment of the invention. The invention may be applied to a
variety of other applications.
[0077] They are not intended to be exhaustive or to limit the
invention to the precise embodiments disclosed, and naturally many
modifications and variations are possible in light of the above
teaching. The embodiments were chosen and described in order to
explain the principles of the invention and its practical
application, to thereby enable others skilled in the art to best
utilize the invention and various embodiments with various
modifications as are suited to the particular use contemplated. It
is intended that the scope of the invention be defined by the
Claims appended hereto and their equivalents.
* * * * *