U.S. patent application number 11/633624 was filed with the patent office on 2007-07-26 for screen image presentation apparatus and method for mobile phone.
Invention is credited to Jong Kerl Lee, Wei Jin Park, Jeong Wook Seo.
Application Number | 20070171192 11/633624 |
Document ID | / |
Family ID | 37629899 |
Filed Date | 2007-07-26 |
United States Patent
Application |
20070171192 |
Kind Code |
A1 |
Seo; Jeong Wook ; et
al. |
July 26, 2007 |
Screen image presentation apparatus and method for mobile phone
Abstract
An idle mode screen image presentation apparatus and method is
provided for a mobile phone. A screen image presentation apparatus
for a mobile phone according to the present invention includes a
memory storing a character house including a background and
function items for executing functions of the mobile phone; a
controller presenting the character house in an idle mode of the
mobile phone and providing information, when one of the function
items is selected, on a function associated with the selected
function item; and a display displaying the character house under
the control of the controller.
Inventors: |
Seo; Jeong Wook; (Daegu
Metropolitan City, KR) ; Lee; Jong Kerl; (Gumi-si,
KR) ; Park; Wei Jin; (Seoul, KR) |
Correspondence
Address: |
ROYLANCE, ABRAMS, BERDO & GOODMAN, L.L.P.
1300 19TH STREET, N.W.
SUITE 600
WASHINGTON,
DC
20036
US
|
Family ID: |
37629899 |
Appl. No.: |
11/633624 |
Filed: |
December 5, 2006 |
Current U.S.
Class: |
345/156 ;
345/531; 345/629 |
Current CPC
Class: |
H04W 52/027 20130101;
Y02D 30/70 20200801; H04M 2250/56 20130101; H04M 1/72427 20210101;
H04M 1/72451 20210101; H04M 1/7243 20210101 |
Class at
Publication: |
345/156 ;
345/629; 345/531 |
International
Class: |
G09G 5/00 20060101
G09G005/00 |
Foreign Application Data
Date |
Code |
Application Number |
Dec 6, 2005 |
KR |
2005-0118428 |
Claims
1. A screen image presentation apparatus for a mobile phone,
comprising: a memory for storing a character house comprising a
character for presenting a condition of the mobile phone; a
controller for controlling the presentation of the character house
at a presentation time; and a display for displaying the character
house under the control of the controller.
2. The screen image presentation apparatus of claim 1, wherein the
character house comprises: items arranged in the character house;
and a background arranged behind the character house.
3. The screen image presentation apparatus of claim 2, wherein the
items comprise function items that represent conditions of specific
functions of the mobile phone and illustrate respective menus when
selected.
4. The screen image presentation apparatus of claim 3, wherein the
function items comprise: a message item for presenting a number of
unread messages; an alarm clock item for presenting scheduled
alarms; and a memory monitor item for presenting memory
utilization.
5. The screen image presentation apparatus of claim 3, wherein the
items comprise: motion items for moving along movement paths in the
character house; and decoration items for decorating the character
house at fixed positions in the character house.
6. The screen image presentation apparatus of claim 3, wherein the
background comprises a plurality of background images changing in
accordance with the progress of time.
7. The screen image presentation apparatus of claim 2, wherein the
controller controls presentation of the character house when the
mobile phone operates in an idle mode.
8. The screen image presentation apparatus of claim 7, wherein the
controller controls presentation of the character house in a time
sequence of the background, items and character.
9. The screen image presentation apparatus of claim 8, wherein the
controller verifies the current time and presents a background
image corresponding to the current time.
10. The screen image presentation apparatus of claim 8, wherein the
controller verifies the types of the items and presents the items
differently according to the type of the item, the types of the
items comprising decoration items that are fixed at positions of
the character house, motion items that are moving along movement
paths in the character house, and the function items that
illustrate conditions of the mobile phone.
11. The screen image presentation apparatus of claim 10, wherein
the controller analyzes a condition of a function represented by
the function item and presents the function item together with an
analysis result in the form of a popup window.
12. The screen image presentation apparatus of claim 11, wherein
the function items comprise a message item presented together with
the number of unread messages in the form of a popup window.
13. The screen image presentation apparatus of claim 11, wherein
the function item comprises an alarm item presented together with
scheduled alarms in the form of the popup window.
14. The screen image presentation apparatus of claim 11, wherein
the motion item moves along the path in units of time.
15. The screen image presentation apparatus of claim 10, wherein
the controller determines a position of the character in the
character house and animates the character.
16. The screen image presentation apparatus of claim 15, wherein
the controller starts presenting the character in a transparency,
outputs sound effects and texts according to progression of the
transparency, and animates the character.
17. The screen image presentation apparatus of claim 16, wherein
the controller verifies the current time and presents a greeting
sound and text if the current time is in at least one of greeting
times.
18. A screen image presentation method for a mobile phone,
comprising: verifying configuration information on a character
house comprising a character representing a condition that is
determined by usage pattern of the mobile phone; and presenting the
character house in a specific operation state of the mobile phone
on the basis of the configuration information.
19. The screen image presentation method of claim 18, wherein the
character house comprises: items arranged in the character house;
and a background arranged behind the character house.
20. The screen image presentation method of claim 19, wherein the
specific operation state comprises an idle state of the mobile
phone.
21. The screen image presentation method of claim 20, wherein the
character house is presented in a time sequence of the background,
items, and character.
22. The screen image presentation method of claim 21, wherein the
items comprise function items that represent conditions of specific
functions of the mobile phone and illustrate respective menus when
selected.
23. The screen image presentation method of claim 22, wherein the
items further comprise: motion items that move along movement paths
in the character house; and decoration items for decorating the
character house at fixed positions in the character house.
24. The screen image presentation method of claim 23, wherein the
background comprises a plurality of background images changing in
accordance with the progress of time.
25. The screen image presentation method of claim 23, wherein
presenting the character house comprises: determining whether an
item comprises a function item: determining whether the function
item is set to operating if an item comprises a function item; and
presenting the function item together with information collected by
a function represented by the function item in the form of a popup
window if the function item is set to operating.
26. The screen image presentation method of claim 25, wherein the
function items comprise a message item which is presented together
with the number of unread messages in the form of the popup
window.
27. The screen image presentation method of claim 25, wherein the
function items comprise an alarm item which is presented together
with scheduled alarms in the form of a popup window.
28. The screen image presentation method of claim 23, wherein each
motion item moves along the path in units of movement time.
29. The screen image presentation method of claim 23, wherein
presenting the character house comprises: determining a position of
the character in the character house; and animating the character
at the position.
30. The screen image presentation method of claim 23, wherein
presenting the character house comprises: starting presentation of
the character at an appearance position in a transparency;
outputting sound effects and texts according to progressing the
transparency; and animating the character at the appearance
position.
31. The screen image presentation method of claim 29, wherein
presenting the character house further comprises: determining
whether a current time is within at lest one of a preset greeting
time zone of a day; and presenting a greeting sound and text if the
current time is within one of the preset greeting time zones.
32. The screen image presentation method of claim 21, wherein
presenting the character house comprises: verifying the current
time; and presenting a background image corresponding to at least
one of a preset time zone to which the current time belongs.
33. The screen image presentation method of claim 21, wherein
presenting the character house comprises: determining whether an
item comprises a function item; presenting a function item image
corresponding to the function item at a position in the character
house together with information collected by a function represented
by the function item in the form of a popup window, if an item
comprises a function item; determining whether the function item
comprises a motion item if the item is not a function item;
presenting a motion item image corresponding to the motion item and
moving the motion item along a path in units of movement time if
the item comprises a motion item; and determining that the item
comprises a decoration item and presenting the decoration item at a
position in the character house if the item does not comprise a
motion item.
34. The screen image presentation method of claim 33, wherein
presenting the character house further comprises: determining
whether the function item comprises a mailbox item if the item
comprises a function item; verifying a number of unread messages
and presenting the number of the unread messages together with the
mailbox item if the function item comprises a mailbox item;
determining whether the function item comprises an alarm item if
the function item does not comprise a mailbox item; verifying
scheduled alarms and presenting the alarm scheduled at the closest
scheduled alarm time together with the alarm item if the function
item comprises an alarm item; determining that the function item
comprises a memory item if the function item does not comprise an
alarm item; and verifying a memory utilization and presenting the
memory utilization together with the memory item.
35. The screen image presentation method of claim 33, wherein
presenting the character house further comprises: determining
whether a character information presentation function is set to
operating; verifying character quotients if a character information
presentation function is set to operating; determining an emotion
representing the character quotients; presenting an emoticon and
text corresponding to the emotion; and displaying the character
quotients in a character information presentation region.
36. The screen image presentation method of claim 21, wherein
presenting the character house comprises: presenting the character
to gradually appear by reducing a transparency of the character in
preset steps after the background and the items are presented;
presenting a sound and text corresponding to a condition of the
mobile phone; and moving the character in a reference
direction.
37. The screen image presentation method of claim 36, wherein
presenting the character comprises: verifying the current time;
determining whether the character has been presented in a preset
time zone to which the current time belongs; and presenting a
greeting behavior of the character if the character is presented in
a time zone to which the current time belongs.
38. The screen image presentation method of claim 37, wherein
presenting a greeting behavior comprises: animating the character;
presenting a text, in the form of a text balloon; notifying that
the greeting behavior corresponding to the preset time zone is
performed; determining whether the current time comprises a
greeting event time; and illustrating the greeting event if the
current time comprises a greeting event time.
39. The screen image presentation method of claim 38, wherein
showing the greeting event comprises: determining at least one of
an item and character money to present to the user as a gift;
presenting the at least determined item and character money; and
presenting a greeting event text in the form of a text balloon
together with an animation of the character.
40. A screen image presentation method for a mobile phone,
comprising: registering configuration information of a character
house in a program mode; editing the character house comprising a
background and items in an editing mode; verifying the
configuration information of the character house; and presenting
the character house with a character, items, and background on the
basis of the configuration information of the character house.
41. The screen image presentation method of claim 40, wherein
registering configuration information comprises setting a character
house presentation on/off option and a character information
presentation on/off function.
42. The screen image presentation method of claim 41, wherein
editing the character house comprises: illustrating a background
image together with a title while editing the background of the
character house; registering the background image as the background
of the character house; adding an item in an item editing mode;
presenting a plurality of sections defined by dividing the
character house; assigning section numbers to the sections;
positioning the item in any of the sections by selecting the
corresponding section numbers; and adjusting a position of the item
finely in the section.
43. The screen image presentation method of claim 42, wherein
adding an item comprises: displaying an item image and title of the
item image; showing the sections with section numbers on the
character house; presenting, if a section number is selected, the
item in the section to which the section number is assigned;
adjusting the position of the item finely in the section; and
registering a type and position of the item in the character
house.
44. The screen image presentation method of claim 42, wherein
adjusting a position of the item comprises: defining sections by
partitioning the character house if an item to be edited is
selected and assigning section numbers to the sections;
positioning, if an item number is selected, the item in the section
corresponding to the section number; adjusting the position of the
item finely in the section; and registering a type and position of
the item in the character house.
45. The screen image presentation method of claim 40, wherein
presenting the character house comprises: determining whether the
character house presentation on/off option is set for operating;
extracting identification and position information of the
background and items of the character house from the configuration
information of the character house if the character house
presentation on/off option is set for operating; and presenting the
background and items on the basis of the identification and
position information and then presenting the character.
46. A screen image presentation apparatus for a mobile phone,
comprising: a memory for storing an artificial intelligence-based
character house in which an artificial-intelligence-based character
resides for presenting a condition of the mobile phone; wherein the
character house comprises items arranged in the character house and
a background arranged behind the character house; a controller for
controlling the presentation of the character house at a
presentation time; and a display for displaying the character house
under the control of the controller.
47. The screen image presentation apparatus of claim 46, wherein
the items comprise function items that represent conditions of
specific functions of the mobile phone and illustrate respective
menus when selected.
48. The screen image presentation apparatus of claim 47, wherein
the function items comprise: a message item for presenting a number
of unread messages; an alarm clock item for presenting scheduled
alarms; and a memory monitor item for presenting memory
utilization
49. The screen image presentation apparatus of claim 47, wherein
the items comprise: motion items for moving along movement paths in
the character house; and decoration items for decorating the
character house at fixed positions in the character house.
50. The screen image presentation apparatus of claim 47, wherein
the background comprises a plurality of background images changing
in accordance with the progress of time.
51. The screen image presentation apparatus of claim 8, wherein the
controller verifies the types of the items and controls to present
the items differently according to at leat one of the type of the
item, the types of the items comprising decoration items that are
fixed at positions of the character house, motion items that are
moving along movement paths in the character house, and the
function items that illustrate conditions of the mobile phone.
Description
CROSS-REFERENCE TO RELATED APPLICATION
[0001] This application claims the benefit under 35 U.S.C.
.sctn.119(a) of Korean Patent Application filed in the Korean
Intellectual Property Office on Dec. 6, 2005 and assigned Serial
No. 2006-0118428, the entire disclosure of which is hereby
incorporated by reference.
BACKGROUND OF THE INVENTION
[0002] 1. Field of the Invention
[0003] The present invention relates to portable phone. More
particularly, the present invention relates to an idle mode screen
image presentation apparatus and method for a mobile phone.
[0004] 2. Description of the Related Art
[0005] A portable terminal is typically a cordless electronic
device such as a mobile phone. Recent mobile phones are implemented
with various functions in addition to basic voice communication.
Such phones include a camera phone equipped with a camera module, a
digital TV phone equipped with a Digital Multimedia Broadcast (DMB)
or Digital Video Broadcast (DVB) module, and an MP3 phone equipped
with an MPEG-1 Audio Layer 3 (MP3) module. That is, the mobile
phone has developed into a multipurpose electronic device.
[0006] Typically, an advanced mobile phone with an advanced
multimedia processing ability is provided with a liquid crystal
display (LCD) to display multimedia data. Various screen display
techniques have been researched and developed to effectively
display information. That is, a screen display scheme is designed
to support viewing multimedia data while simultaneously displaying
information related to a basic task. For example, a mobile phone
can be implemented to display a preset image while the mobile phone
is in an idle state. The image of the idle screen view is
downloaded from a website or taken by a built in camera. However,
conventional mobile phones limit designs to screen images to be
presented on the display by themselves. Although there has been a
screen image design method in which the screen image is designed
with a background image, text, and clock selected by a user, most
of the conventional screen image design techniques do not
facilitate the design of a dynamic screen image with various
character images.
[0007] Accordingly, there is a need for an improved screen image
presentation apparatus and method for mobile phones.
SUMMARY OF THE INVENTION
[0008] An aspect of exemplary embodiments of the present invention
is to address at least the above problems and/or disadvantages and
to provide at least the advantages described below. Accordingly, an
aspect of exemplary embodiments of the present invention is to
provide a screen image presentation apparatus and method for a
mobile phone.
[0009] It is another object of an exemplary embodiment of the
present invention to provide a screen image presentation apparatus
and method for a mobile phone that are capable of improving
usability of mobile phone by using a character as an artificial
intelligent agent.
[0010] According to another object of an exemplary embodiment of
the present invention, a screen image presentation apparatus and
method for a mobile phone are provided. This facilitates the
presentation of a character house with an artificial intelligent
character, items, and background as a screen image.
[0011] It is another object of an exemplary embodiment of the
present invention to provide a screen image presentation apparatus
and method for a mobile phone that are capable of making a
presentation on a screen of the mobile phone dynamic by displaying
an artificial intelligence-based character house, in which an
intelligent character resides, as a screen image in a specific
operation mode of the mobile phone.
[0012] It is yet another object of an exemplary embodiment of the
present invention to provide a screen image presentation apparatus
and method for a mobile phone that are capable of providing an
intelligent character with an artificial intelligence-based
character house. The intelligent character reacts and adapts to a
change in the conditions of the mobile phone in the form of an
animation, sound, and text.
[0013] It is another object of an exemplary embodiment of the
present invention to provide a screen image presentation apparatus
and method for a mobile phone that are capable of improving
aesthetics of a screen image by means of a character house
consisted of various visible items.
[0014] It is another object of an exemplary embodiment of the
present invention to provide a screen image presentation apparatus
and method for a mobile phone that are capable of informing
progress of time by means of a character house having a background
which is changed according to various time periods within a
day.
[0015] According to another object of an exemplary embodiment of
the present invention, a screen image presentation apparatus and
method for a mobile phone is provided. This screen image
presentation apparatus and method is capable of providing a
character house as an animated screen image for a specific
operation mode of a mobile phone. The character house includes a
character, background, function items, decoration items and motion
items. The character illustrates conditions and emotions reflecting
states of the mobile phone in the form of an animation, sound, and
text. The function items illustrate system information on the
mobile phone. Decoration items improve aesthetics of the screen
image and motion items move along respective movement lines in the
screen house.
[0016] It is another object of an exemplary embodiment of the
present invention to provide a screen image presentation apparatus
and method for a mobile phone that are capable of producing current
time zone information by means of a character house including a
background and items and especially an intelligent character which
illustrates greeting behaviors with an animation, sound, and text
in the time zone.
[0017] It is yet another object of an exemplary embodiment of the
present invention to provide a screen image presentation apparatus
and method for a mobile phone that are capable of producing
information on conditions and emotions representing states of a,
mobile phone by means of character quotients, emoticons, and text
expressions provided in a character house with a character,
background, and items.
[0018] It is still another object of an exemplary embodiment of the
present invention to provide a screen image presentation apparatus
and method for a mobile phone that are capable of presenting a
character house with a character, background, and items, as an idle
mode screen image of the mobile phone.
[0019] It is another object of an exemplary embodiment of the
present invention to provide a screen image presentation apparatus
and method for a mobile phone that are capable of generating a
character house with a character, background, and items, as a
screen image for an idle mode of the mobile phone.
[0020] According to another object of an exemplary embodiment of
the present invention, a screen image presentation apparatus and
method for a mobile phone are provided. This apparatus and method
facilitates a user's ability to design a character house, by
selecting a character, background, and items, to be used as a
screen image in an idle mode of the mobile phone.
[0021] It is another object of an exemplary embodiment of the
present invention to provide a screen image presentation apparatus
and method for a mobile phone that facilitates a user's ability to
edit a character house consisting of a character, background, and
items, by adding and deleting the items. The character house is
used as a screen image for a specific operation mode of the mobile
phone.
[0022] It is yet another object of an exemplary embodiment of the
present invention to provide a screen image presentation apparatus
and method for a mobile phone that are capable of adding new items
or editing the items constituting a character house while the
character house is presented as a screen image for a specific
operation mode of the mobile phone.
[0023] It is still another object of an exemplary embodiment of the
present invention to provide a screen image presentation apparatus
and method for a mobile phone that are capable of adding new items
to or removing or repositioning items presented with a character
house. The character house consists of a background, character, and
items and is used as a screen image for a specific operation mode
of the mobile terminal.
[0024] According to another object of an exemplary embodiment of
the present invention, a screen image presentation apparatus and
method for a mobile phone is provided. The screen image
presentation apparatus and method facilitates a user's ability to
select a partitioned sectors of a character house to see the
selected sector in an enlarged form such that the user can perform
a fine adjustment of a position of an item located in the selected
sector. The character house consists of a background, character,
and items, and is used for a screen image for a specific operation
mode of the mobile terminal.
[0025] In accordance with an aspect of an exemplary embodiment of
the present invention, the above and other objects are accomplished
by a screen image presentation apparatus of a mobile phone. The
screen image presentation apparatus includes a memory, a controller
and a display. The memory stores a character house including a
character for presenting a condition of the mobile phone. The
controller controls to present the character house at a
predetermined presentation time and the display displays the
character house under the control of the controller.
[0026] In accordance with another aspect of an exemplary embodiment
of the present invention, the above and other objects are
accomplished by a screen image presentation method for a mobile
phone. According to the screen image presentation method,
configuration information on a character house is verified with a
character that represents a condition which is determined by usage
pattern of the mobile phone. The character house is presented in a
specific operation state of the mobile phone on the basis of the
configuration information.
[0027] In accordance with another aspect of an exemplary embodiment
of the present invention, the above and other objects are
accomplished by a screen image presentation method for a mobile
phone. According to the screen image presentation method,
configuration information of a character house is registered in a
program mode. The character house with a background and items is
edited in an editing mode. The configuration information of the
character house is checked and the character house is presented
with a character, items, and background on the basis of the
configuration information of the character house.
[0028] Other objects, advantages and salient features of the
invention will become apparent to those skilled in the art from the
following detailed description, which, taken in conjunction with
the annexed drawings, discloses exemplary embodiments of the
invention.
BRIEF DESCRIPTION OF THE DRAWINGS
[0029] The above and other exemplary objects, features and
advantages of certain exemplary embodiments of the present
invention will be more apparent from the following description
taken in conjunction with the accompanying drawings, in which:
[0030] FIG. 1 is a block diagram illustrating a configuration of a
mobile phone employing a screen image presentation apparatus and
method according to an exemplary embodiment of the present
invention;
[0031] FIGS. 2a, 2b and 2c are views illustrating a concept of a
character house of a screen image presentation apparatus according
to an exemplary embodiment of the present invention;
[0032] FIG. 3 is a screen image illustrating an exemplary character
house according to an exemplary embodiment of the present
invention;
[0033] FIGS. 4a to 4n are screen images illustrating steps for
implementing exemplary character animations in a screen image
presentation technique according to an exemplary embodiment of the
present invention;
[0034] FIGS. 5a to 5c are screen images illustrating steps for
implementing an animated background in a screen image presentation
technique according to an exemplary embodiment of the present
invention,
[0035] FIG. 6 illustrates screen images that illustrate background
images for implementing an animated background in a screen image
presentation technique according to an exemplary embodiment of the
present invention;
[0036] FIG. 7a is screen image illustrating a character quotient
module on the idle mode screen in a screen image presentation
technique according to an exemplary embodiment of the present
invention;
[0037] FIG. 7b is an enlarged view illustrating the character
quotient module of FIG. 7a;
[0038] FIG. 8 is a drawing illustrating a table of emoticons to be
displayed with the quotient module of FIG. 7b.
[0039] FIGS. 9a to 9l are views illustrating the presentation of
function items in a screen image presentation technique according
to an exemplary embodiment of the present invention;
[0040] FIGS. 10a to 10c are views illustrating steps of presenting
motion items in a screen image presentation technique according to
an exemplary embodiment of the present invention;
[0041] FIGS. 11a to 11c are views illustrating steps of presenting
decoration items in a screen image presentation technique according
to an exemplary embodiment of the present invention;
[0042] FIG. 12 illustrates screen images illustrating character
behaviors in the screen image presentation technique according to
an exemplary embodiment of the present invention;
[0043] FIG. 13 is a screen image illustrating a character animation
with an event in a screen image presentation technique according to
an exemplary embodiment of the present invention;
[0044] FIG. 14 is a flowchart illustrating a character house
presentation procedure of a screen image presentation method
according to an exemplary embodiment of the present invention;
[0045] FIG. 15 is a flowchart illustrating the background
presentation procedure of FIG. 14;
[0046] FIG. 16 is a flowchart illustrating the item presentation
procedure of FIG. 14;
[0047] FIG. 17 is a flowchart illustrating a procedure for
presenting a motion item in the screen image presentation technique
according to an exemplary embodiment of the present invention;
[0048] FIGS. 18a to 18d are flowcharts illustrating a procedure for
presenting function items in the screen image presentation
technique according to an exemplary embodiment of the present
invention;
[0049] FIG. 19 is a flowchart illustrating a character presentation
procedure of the screen image presentation method according to an
exemplary embodiment of the present invention;
[0050] FIG. 20 is a flowchart illustrating a greeting behavior
presentation procedure of FIG. 19;
[0051] FIG. 21 is a flowchart illustrating a greeting event
presentation procedure that occurred during a greeting behavior of
the character of the screen image presentation technique according
to an exemplary embodiment of the present invention;
[0052] FIGS. 22a and 22b are flowcharts illustrating an animation
behavior presentation procedure of the screen image presentation
technique according to an exemplary embodiment of the present
invention;
[0053] FIGS. 23a and 23b illustrate the character shift information
determination procedure in detail;
[0054] FIG. 24 is a flowchart illustrating a screen house
presentation procedure of a screen image presentation method for a
mobile phone according to another exemplary embodiment of the
present invention;
[0055] FIG. 25 is a state diagram illustrating a character house
design procedure of a screen image presentation method according to
an exemplary embodiment of the present invention;
[0056] FIG. 26 is a state diagram illustrating transitions between
the character house configuration mode and the character house
editing mode of FIG. 25;
[0057] FIG. 27 is a screen image illustrating a character house
configuration screen in the character house configuration mode of
FIG. 25;
[0058] FIG. 28 is a state diagram illustrating transitions between
the character house configuration mode and the character house
editing mode of FIG. 25;
[0059] FIG. 29 illustrates screen images illustrating a procedure
for setting a "character house presentation on/off" option on the
character house configuration screen of FIG. 25;
[0060] FIG. 30 is a state diagram illustrating transitions among
the character house editing mode and specific item editing modes of
FIG. 25;
[0061] FIG. 31 illustrates a screen image illustrating an example
character house editing screen in the character house editing mode
of FIG. 25;
[0062] FIG. 32 is a state diagram illustrating transitions between
the character house editing mode and the background selection mode
of FIG. 25;
[0063] FIG. 33 illustrates screen images illustrating a procedure
for selecting a background on the background selection mode of FIG.
32;
[0064] FIG. 34 is a state diagram illustrating transitions between
the character house editing mode and item editing modes of FIG.
25;
[0065] FIGS. 35a to 35d illustrate screen images illustrating a
procedure for adding an item to the character house in the
character house editing mode of FIG. 34;
[0066] FIG. 36 is a state diagram illustrating transitions between
the character house editing mode and the selected item editing mode
via the edited item shortcut selection mode of FIG. 25;
[0067] FIGS. 37a to 37c illustrate screen images illustrating a
procedure for assigning shortcut key in the edited item shortcut
selection mode of FIG. 36.
[0068] Throughout the drawings, the same drawing reference numerals
will be understood to refer to the same elements, features and
structures.
DETAILED DESCRIPTION OF EXEMPLARY EMBODIMENTS
[0069] The matters defined in the description such as a detailed
construction and elements are provided to assist in a comprehensive
understanding of the embodiments of the invention. Accordingly,
those of ordinary skill in the art will recognize that various
changes and modifications of the embodiments described herein can
be made without departing from the scope and spirit of the
invention. Also, descriptions of well-known functions and
constructions are omitted for clarity and conciseness.
[0070] A "screen image" is an image presented on a display screen
of the mobile phone.
[0071] "House," "avatar house," and "character house" are used
interchangeably to refer to a visible object presented on screen in
a specific mode of the mobile phone. The house is displayed with
characters, items, and background.
[0072] "Item" is a visible object for enriching the house and
includes decoration items, function items for notifying states of
specific functions, and motion items that move along preset
paths.
[0073] "Background" is an image forming a background graphics of
the house.
[0074] "Character" is a visible object that resides in the
character house and can express a state of the mobile phone in an
emotional manner. The character includes fictional characters
represented by a human, avatar, or animal, among others that can
express various emotions. Accordingly, the character can express
emotions.
[0075] In the exemplary embodiment of the present invention, the
character house can be correlated with a background, a character,
and at least one item.
[0076] FIG. 1 is a block diagram illustrating a configuration of a
mobile phone employing a screen image presentation apparatus and
method according to an exemplary embodiment of the present
invention.
[0077] Referring to FIG. 1, the mobile phone includes a controller
110, data processing unit 120, radio frequency (RF) unit 123, audio
processing unit 125, voice synthesizer 127, memory 130, keypad unit
141, and display unit 143.
[0078] The RF unit 123 facilitates communication between the mobile
phone and other terminals over a wireless channel. The RF unit 123
includes an RF transmitter for up-converting and amplifying the
frequency of a transmitted signal, and an RF receiver for low-noise
amplifying a received signal and down-converting the frequency of
the received signal.
[0079] The data processing unit 120 includes a transmitter for
coding and modulating the transmitted signal, and a receiver for
demodulating and decoding the received signal. The data processing
unit 120 may include a modem and a codec. The voice synthesizer 129
performs a function of converting a text message output from the
controller 110 or received from outside into a voice signal.
[0080] The keypad unit 141 is provided with a plurality of
alphanumeric keys for input data and functions key for setting
various functions.
[0081] The memory 130 includes a program memory and a data memory.
The program memory stores programs for processing incoming and
outgoing calls, and programs for presenting the character house for
a specific state of the mobile phone and controlling reactions of
the characters in response to events caused by operations of the
mobile phone and for structuring the screen image comprising the
character house with the characters, according to an exemplary
embodiment of the present invention. The data memory includes a
character memory for storing character images and information on
the characters for performing the exemplary embodiment of the
present invention and a text memory for storing text data for
emotional expressions of the characters. The data memory also
temporarily stores data generated while the programs are executed.
The term "character house image" is interchangeably used to refer
to the screen image presented in a specific mode of the mobile
phone. The specific mode or specific state may refer to the idle
mode or idle state.
[0082] The character memory may include a character storage, an
item storage and a background storage. The character storage stores
information on the structures of the characters and images thereof.
The item storage stores information on the structures of the items
and image thereof. The background storage stores background images.
The character house structure may be stored in the form of
information on a combination of the character, item, and background
image, and the images stored in the respective storages may be used
individually.
[0083] A character is provided with information on character
animations expressing emotional states for the mobile phone,
character images for expressing emotions in response to text
messages (for example, SMS, EMS, and MMS, among others), and
character images for expressing emotions in response to events
occurring in the mobile phone. According to an exemplary embodiment
of the present invention, a plurality of character images
expressing the emotional states for the mobile phone, a plurality
of character images expressing the emotions in accordance with the
contents of the text messages, and a plurality of character images
expressing the emotions in accordance with the events occurring in
the mobile phone is assumed. The memory for storing the images
operates in a file structure. A character file structure includes a
header and character images, and the header contains information on
identifiers, start delimiters, a number of images, display
positions and names of animations. In a character transmission
mode, the headers and character images are transmitted after being
processed in the form of a transmission file.
[0084] The character house is displayed on the screen together with
a background image, a character, and at least one item in a
specific mode (for example, an idle mode) of the mobile phone. That
is, the character house is designed by arranging the items on the
background image for a specific mode screen (for example, an idle
mode screen), the items and the background images being selected by
the user. At this time, the file structure of the character house
may include a header and a plurality of images, and the header may
include a header of the character house and the item headers of the
items associated with the character house. The character house
header contains information on the character house identifier, the
positions of the characters on the screen, the sizes of the
characters, and the names of the characters; and each item header
contains information on the item index, the position of the item in
relation with the character house, and the name of the item. In a
character house transmission mode, the headers of the character
house and the items associated with the character house and the
background and item images are transmitted in the form of a
compressed file.
[0085] The items means the images presented in association with the
character house on the screen. Each item may be displayed in a
specific association with the character house or regardless of the
character house. If the item is not associated with the character
house, the item should be transmitted individually. The item
includes a header and an image, and the item header contains
information on the item index, the position of the item in relation
to the character house, and the name of the item. According to an
exemplary implementation, if the item is not associated with the
character house, the information on the position of the item has a
default value.
[0086] In this exemplary embodiment of the present invention, the
character is used for presenting the information on the states and
event occurrences of the mobile phone. The character house is used
to indicate the specific operation mode (for example, the idle
mode) of the mobile phone. The item is used to show the functions
of the mobile phone, decorate the character house, and show the
screen dynamic. The background image is used to provide a
background of the characters and items associated with the
character house.
[0087] The controller 110 controls general operations of the mobile
phone. The controller 110 may be integrated with the data
processing unit 120. According to an exemplary implementation, the
controller 110, data processing module 120, audio processing unit
125, and voice synthesizer 127 can be integrated into a mobile
station modem (MSM) chip. In this exemplary embodiment of the
present invention, the controller 110 controls to select the
characters, items, background image, character house, texts, and
sounds. The controller arranges the selected images and texts at
the present positions on the screen of the display unit 143 and
outputs the texts as voice sounds under the support of the voice
synthesizer 127.
[0088] The display unit 143 displays the state of the mobile phone
under the control of the controller 110. The display unit 143 may
be implemented with a liquid crystal display (LCD) together with a
LCD controller, extra memory for storing the display data, and LCD
display device. If the display unit 143 is implemented with a
touchscreen function, the display unit 143 can operate an input
device.
[0089] The characters represent a graphical user interface (GUI)
including the character displayed on the screen of the display unit
143 in the specific mode of the mobile phone. The character house
is a space in which the character with an artificial intelligence
resides such that the user can experience a virtual space with a
virtual talking character. The user can quickly learn the functions
of the mobile phone with the assistances of the character house and
enjoy with the vivid characters.
[0090] The character house has a function that expresses a dynamic
character. First, the character house can express the condition and
emotion of the current character in the animation, sound, and text
according to a character scenario associated with the character.
Second, the character house has a function to present a background
of the screen for a specific mode (especially, the idle mode) of
the mobile phone that is changed in accordance with a progress of
time. Third, the character house has a function for displaying
items. The items displayed in association with the character house
may be classified by function. In this exemplary embodiment of the
present invention, the items can be classified into three
categories. The first is the function item for presenting the
functions and system information on the mobile phone in order to
enhance the user's convenience. The second is the decoration item
for increasing aesthetic satisfaction by adding various images to
the character house. The third is the motion items that make the
character house screen dynamic by adding the motion images (for
example, the cloud or bird, among others).
[0091] The character house may further provide a greeting function
with which the character associated with the character house gives
greetings with appropriate sounds at preset times. An additional
function such as allowing a user to give a present to the character
may also be implemented.
[0092] FIGS. 2a to 2c are views illustrating a concept of a
character house of a screen image presentation apparatus according
to an exemplary embodiment of the present invention.
[0093] FIG. 2a is an illustration of a house setting screen for
selecting a character house to be displayed on the screen. The
house setting screen is displayed on the screen under the control
of a character setting program. If the character setting program is
executed, 4 character house images are displayed in corresponding
tiled windows on the screen of the display unit 143. When a number
of the character house images are less than 4, the windows that are
not allocated for the character house image will be empty.
[0094] If a character house is selected from the house setting
screen, the controller 110 operates a character engine which is
structured as illustrated in FIG. 2b. The character engine stores
information of the selected character house into a character house
information storage module 210 and executes a character house
expression module 220 upon activation. That is, the character
engine includes the character house information storage module 210
and the character house expression module 220. The character house
information storage module 210 is a module for storing information
on the character houses and the character to be displayed in
association with each other. According to an exemplary embodiment
of the present invention, the character house information storage
module 210 includes character house information storage 213, system
information storage 215, and character information storage 217.
[0095] The character information storage 217 stores emotion
information, status information, quotient information, and text and
sound information associated with the character's emotions and
statuses. The character information storage 217 also stores
animation information for animating the character in a specific
status. The mobile phone processes various events which have
occurred while operating. The mobile phone also updates the
quotients of the characters that correspond to the events.
[0096] The quotient information includes a physical quotient (HQ),
emotion quotient (EQ), intelligent quotient (IQ), and social
quotient (SQ), that are used for determining the status and emotion
of the character. The controller 110 accumulates the quotients
associated with the events whenever an event is processed and
determines the status of the mobile phone on the basis of the
quotient analysis results. The emotion of the character is
determined according to the quotients and user's feedback analysis
results. The emotion of the character can be determined in
accordance with a content related to the event (for example, a text
message) that is obtained during the process. The text and sound
information associated with the emotion can be stored together with
the character expression. The character information storage 217
includes information on the animation of the character that is
expressed in a specific mode (for example, an idle mode) of the
mobile phone.
[0097] The system information storage 215 stores system information
on the function items for displaying a memory capacity of the
mobile phone, scheduled alarms and unread mail. The items
associated with the character house includes the decoration items,
motion items, and function items. The function items present the
operations and functions of the mobile phone. According to an
exemplary embodiment of the present invention, the items for
presenting a memory status, preset alarm times, and unread mail are
described as examples of the function items.
[0098] The character house information storage 213 stores the
information on the character house that is defined by a character
house setting program such that the character house information
includes the types and positions of the items and the type of the
background associated with the character house. The character house
information storage 213 stores the information on the character
house images except for the character and the function items. Thus,
the character house information 213 stores the information on the
decoration items, motion items and the background images.
[0099] The character house expression module 220 includes a
character display module 223, an item display module 225 and a
background display module 227.
[0100] The character display module 223 renders the character
animation of the character house. The character display includes a
mobility and entrance of the character and a text balloon or a
sound effect associated with the behavior of the character. The
character display module 223 can make the character express special
motions such as a greeting. The item display module 225 displays
the items at preset positions with reference to the character house
information on the current character house and the system
information. The items are displayed in a different manner by item
category (decoration, motion, and function). The background display
module 227 displays the background image in accordance with the
information stored in the character house information storage 213.
The background image can be configured such that the brightness of
the background image changes over time.
[0101] FIG. 2c illustrates an exemplary character house screen
displayed by the character engine of FIG. 2b. The character house
is displayed in a specific mode (for example, an idle mode) of the
mobile phone and includes a character, items and a background.
[0102] FIG. 3 is a screen image illustrating an exemplary character
house according to an exemplary embodiment of the present
invention. The character house screen provides information on the
detailed status, which is not provided with a normal idle mode
screen, for the mobile phone together with decoration effects, as
illustrated in FIG. 3.
[0103] Referring to FIG. 3, a character 230 is displayed in the
form of an animated character in association with the character
house. The character 230 behaves in a motion preset for the idle
mode of the mobile phone to alarm at preset times in corresponding
animation together with a text and sound effect. The character 230
can be set to animate for giving items as presents for the
character house in a random manner. In order to express the state
of the mobile phone and the emotion of the character representing
the state of the mobile phone, the character 230 expresses the
emotion in the animation according to the state of the mobile phone
and an emotion model together with the sound and text. The
background 240 is represented by a background image of the
character house and the brightness of the background can be changed
over time.
[0104] The items 250 give effects for decorating the character
house and inform the operations of the mobile phone. The function
items 251, 253, 255, and 259 decorate the character house and
present the statuses of the preset functions and operations. For
example, the function item 251 is a clock item that illustrates a
preset alarm time, and the function item 253 is a mailbox item for
illustrating the amount of unread mail. If multiple alarms are set,
the closest alarm time is displayed. The function item 255 is a
memory monitor item for displaying the state of the memory. The
decoration item 259 is an item for decorating the character house.
The motion item (not shown) can be activated for decorating the
character house.
[0105] The character quotient module 260 includes a character
quotient, a physical quotient, an intelligent quotient (IQ) and an
emotional quotient (EQ). In an exemplary embodiment of the present
invention, the social quotient (SQ) is represented by values
assigned for the events related to the communication. The physical
quotient (HQ) is represented by values assigned for events related
to the performance for the mobile phone. The intelligent quotient
(IQ) is represented by values assigned for an amount of information
stored in the mobile phone and the emotional quotient (EQ) is
represented by values assigned for events related to the
multimedia. The character quotient module 260 can be configured to
be fixed at a position or floats across the screen.
[0106] The character 230 of the character house has the following
three basic expression factors: animation, text and speech.
Animation facilitates the character's 230 ability to express a
state and an emotion while moving. A text facilitates the
character's 230 expression of a state and emotion in text. The text
can be presented in the form of a text balloon. Speech facilitates
the character's 230 expression of a state and emotion through voice
or sound effects. The character 230 can express various states and
emotions, and the state and emotion of the character 230 are
determined in accordance with a state decision model algorithm and
an emotion mode decision algorithm as a character scenario
dependent on the variation of the quotient values of the character.
The decision on the state and emotion mode of the character on the
basis of the character scenario is disclosed in the Korean Patent
Application No. 2005-107174. The character 230 of the character
house expresses the state information and the emotion mode using
the character image together with a unique voice effect and text
balloon.
[0107] The character 230 can be implemented with an animation in
which the character 230 moves around the character house and
disappears outside the screen and then appears from outside. That
is, the character 230 expresses the state and emotion with a
motion, sound effect, and text, if the time to express the state
and emotion has arrived. The state and emotion of the character 230
is expressed when the quotients reach respective threshold values,
rather than always. However, the character house is always
displayed in the specific mode screen. For example, if the mobile
phone enters an idle mode, the character house is displayed on the
screen. The character house screen disappears when a preset display
time expires or when the mobile phone enters another operation
mode. If the character house is displayed at the idle mode, the
character 230 expresses the state and emotion in motion and with
other animations.
[0108] FIGS. 4a to 4h are screen images illustrating steps for
implementing an exemplary character animation in a screen image
presentation technique according to an exemplary embodiment of the
present invention. FIGS. 4i to 4n are screen images illustrating
steps of another exemplary character animation in a screen image
presentation technique according to an exemplary embodiment of the
present invention.
[0109] Referring to FIGS. 4a to 4h, the character 230 moves in a
preset behavior pattern in the character house. The character 230
appears at a random position on the screen, makes a speech
expression and then moves around. If the character house disappears
on the screen while the character moves around the character house,
the character 230 appears at a random position and repeats the
behavior after a predetermined time has elapsed. The motion, sound
effect, and text can be selectively disabled. The character house
can be expressed through six steps as followings.
[0110] As illustrated in FIG. 4a, if the mobile phone enters the
idle mode, the items and background image of the character house
are displayed. Next, as illustrated in FIG. 4b, a character 230
appears at a position in the character house. The character 230 can
appear with a fade-in effect. That is, the character 230 can be
implemented with an alpha blending technique so that the
transparency of the character 230 changes from 100% to 0%
gradually. For example, the transparency reduces by 25% per 0.5
second such that the character 230 is gradually faded in to
completely appear in 2 seconds. The character 230 starts appearing
at a random position. The character 230 is preferably implemented
so that the character is at a position in which the character does
not completely appear. That is, the character 230 is configured to
appear at a random position but where the character 230 can be
completely shown across the border of the screen.
[0111] After the character 230 is displayed, a sound effect is
output in accordance with the state of the character and with a
notification icon, as illustrated in FIG. 4c. The sound effect is
output at once and repeated one more time when the character
reappears. The sound effect may also be cancelled.
[0112] After the sound effect and the text are output, the
character 230 is animated as illustrated in FIG. 4d. The character
animation process is repeated until the character 230 disappears.
The character animation is performed at a fixed position.
[0113] As illustrated in FIG. 4e, the state and emotion of the
character 230 is expressed with a text after the character
animation starts. The text is expressed in the text balloon in
consideration of the time taken for the user to read the text. The
text balloon is displayed in the form of a semitransparent image
with about 30% of transparency and is open in about 3 seconds in
consideration of the reading time.
[0114] After the text balloon disappears, the character 230 moves
as illustrated in FIGS. 4f and 4g. The movement direction of the
character 230 can be changed according to the position at which the
character 230 appears. That is, the position at which the character
230 appears is verified such that the character 230 moves in a
direction to which the character can secure the longest distance
for movement. The character 230 moves at a predetermined speed. The
character 230 is preferably set to move at the speed of about 20
pixels per second. The character 230 disappears across the border
of the screen. If the character 230 disappears completely, an
initial idle mode screen is displayed as illustrated in FIG. 4h
identical with FIG. 4a. After a predetermined time (about 2
seconds) elapses, steps 4a to 4g are repeated.
[0115] The character animation illustrated in FIGS. 4i to 4h is an
example in which the character does not move. Referring to FIGS. 4i
to 4h, the character is displayed in the idle mode screen. In the
initial state of the idle mode illustrated in FIG. 4i, the preset
items and background images are displayed without the character.
Next, as shown in FIG. 4j, the character appears on the screen. The
character appears at a predetermined position on the screen. The
appearance position is changed to the bottom right position in
relation to the center of the screen. The appearance position can
be changed according to the size of the character. The character
can be set to appear with a fade-in effect in such a manner that
the transparency of the character changes from 100% to 0%
gradually. For example, the transparency reduces by 25% per 0.5
second such that the character is gradually faded in to completely
appear in 2 seconds.
[0116] After the character 230 is completely displayed, a sound
effect is output in accordance with the state of the character and
with a notification icon as illustrated in FIG. 4k. The sound
effect is output at once and repeated one more time when the
character reappears.
[0117] According to an exemplary implementation, a text is
displayed to express the current state and emotion of the
character. The text is presented in the text balloon with a
transparency of 30%. The text can be displayed in synchronization
with a sound. The text balloon is expressed in consideration of the
time taken for the user to read the text. For example, the text
balloon can be set to disappear after 1.5 seconds.
[0118] After the sound effect ends and the text balloon disappears,
the character is animated as illustrated in FIG. 4m. The character
animation process is repeated until the character disappears in the
character house as illustrated in FIG. 4n.
[0119] The character animation process is performed in a certain
order so that the items 250 and background image 240 are displayed,
the character 230 appears in a fade-in effect, the sound effect is
output for notifying the state of the mobile phone, the text is
presented for notifying the state of the mobile phone, and the
character animation is performed in the character house.
[0120] The character animation is a motion behavior of the
character without moving around the screen, and the character
movement is a behavior of the character moving around the screen.
The character animation and the movement may be performed
simultaneously. The character 230 may also appear simultaneously
with the items 250 and the background. The character 230 may appear
in the form of an opaque object, such as with a 0% transparency.
The output of sound effect and text may be skipped.
[0121] According to an exemplary implementation, the presentation
of the background can be performed so that the background image
changes according to the progress of the time. A fixed background
image can be used.
[0122] FIGS. 5a to 5c are screen images illustrating steps for
implementing an animated background in a screen image presentation
technique according to an exemplary embodiment of the present
invention. FIG. 6 illustrates screen images illustrating background
images for implementing an animated background in a screen image
presentation technique according to an exemplary embodiment of the
present invention. The background images of FIG. 6 can be used for
the animated background or the fixed background.
[0123] Referring to FIGS. 5a and 5c, the animated background is
implemented with several background images that are presented in
series over a period of time. According to an exemplary embodiment
of the present invention, a day is divided into time periods, such
as morning, afternoon, evening, and night. The background is placed
with different background images for the respective time periods.
In this exemplary embodiment of the present invention, the
background images of the character house change according the
specific time period if an animation background 240 is set by the
user. Since the change of the background is performed in accordance
with the progress of the time, the background of the character
house that changed with different brightness background images is
assigned for the morning, day, and night as illustrated in FIGS. 5a
to 5c. According to an exemplary embodiment of the present
invention, the time periods are defined as illustrated in table 1.
According to an exemplary implementation, the character can be set
to present corresponding greetings at the beginning of each time
period. TABLE-US-00001 TABLE 1 Class Time Period Description
Morning 06:00.about.12:00 A little Bright Image having sunshine
Afternoon 12:00.about.18:00 Warm Image having Sun in the middle sky
Evening, 18:00.about.06:00 Dark image with no sun Night
[0124] The background may be set on the character house setting
screen. FIG. 6 illustrates the background images that may be
selected for the background. The background images may be used for
the animated background in which background images are changed in
accordance with the time period or may be used for a fixed
background. The background images can be used for the animated
background of the character house or can be used for the fixed
background of the character house. Each background image can be
used for the animated background by only changing the brightness of
the background image for expressing the time periods.
[0125] The character house can present information on items that
illustrate the system information and on the quotients of the
character, as well as the character 230 and the background 240.
[0126] FIG. 7a is screen image illustrating a character quotient
module on the idle mode screen in a screen image presentation
technique according to an exemplary embodiment of the present
invention. FIG. 7b is an enlarged view illustrating the character
quotient module of FIG. 7a.
[0127] Referring to FIG. 7a, the character quotient module 260
notifies the current status of the character 230. In this exemplary
embodiment of the present invention, the character quotient module
260 illustrates the state and emotion of the character 230 with the
HQ, EQ, IQ, and SQ status gauges. Since the state and emotion of
the character are presented with the quotient status gauges, the
character can attract the user's interest.
[0128] The character quotient module 260 can be presented at a
preset position in the character house. In this exemplary
embodiment of the present invention, the character quotient module
is presented at the top left area of the screen, as illustrated in
FIG. 7a. The character quotient module 260 can be set to be visible
or invisible according to a user selection.
[0129] Referring to FIG. 7b, the character quotient module 260
includes a quotient information display region 267, an emotion icon
display region 263, and a status display region 265. The emotion
icon display region 263 can be implemented to display emoticons
representing the emotions of the character. The status display
region 265 presents the status of the character in text and the
quotient information display region 267 displays the levels of the
quotients with color bars (gauge). The emotions that can be
presented in the emotion icon display region 263 includes "happy,"
"normal," "angry," "sad," and "tired," and the emotions are linked
to corresponding images in the form of a table as illustrated in
FIG. 8. Accordingly, when implementing the character house, the
emotion of the character can be presented by the emoticon of the
character quotient module corresponding to the current emotion of
the character in addition with the character animations. The status
information of the character can be presented at the status display
region 265 in the form of a text. Each character can have various
statuses that can be expressed in the form of a text. The character
quotient information is displayed in the form of gauges in the
quotient information display region 267. In this exemplary
embodiment of the present invention, four quotients, such as, HQ,
EQ, IQ, and SQ are measured and presented in the quotient
information display region 267. In order to distinguish the
quotients from each other, the quotients are presented by means of
colored gauges, such as a red gauge for SQ, a green gauge for HQ, a
purple gauge for IQ, a blue gauge for EQ, as shown in FIG. 7b. Each
gauge presents the level of the quotient in a percentage (%)
without a text based notification. The emotion icon display region
263 and a status display region 265 can be deleted.
[0130] The items and expressions such as the greetings can be
bought in an item shop on the web or received as a gift. The items
that can be presented in the character house include the function
items, decoration items, and motion items. The positions of the
items on the screen can be set in the character house setting
screen. Presentation of the items is explained based on an order of
the function item, motion item, and decoration item.
[0131] The function items are items that provide specific system
information of the mobile phone. In this exemplary embodiment of
the present invention, the function items include a memory monitor
item, an alarm item, and a mailbox item. Additional function items
may also be added when required. The mailbox item can be a message
item, which is used to explain the function item.
[0132] The function items can be implemented for a signal function
or can be an item grouping the function items comprising similar
functions. As a group-wise function item, the mailbox item can
present a variety of the messages. According to an exemplary
implementation, the function item integrates a short message
service (SMS) item, an email message service (EMS) item, and a
multimedia message service (MMS) item. When there is an alarm item,
a schedule item, a memo item, and an alarm generation item, among
others may be integrated. If such a multifunction item is selected,
the controller 110 can display the function items integrated in the
multifunction item such that the user can select one of the
items.
[0133] FIGS. 9a to 9l are views illustrating how to present
function items in a screen image presentation technique according
to an exemplary embodiment of the present invention.
[0134] The functional item can be displayed in a normal mode (see
FIG. 9a), a selection mode (see FIG. 9b), and an enable mode (see
FIG. 9c). Presentation of the function item with a single function
is described with reference to FIGS. 9a to 9c.
[0135] Referring to FIG. 9a, when the function items are presented
with the character house, the function items are arranged at the
preset positions on the screen. In FIG. 9a, the mailbox item, the
clock item, and the memory monitor item are examples displayed on
the screen. In the normal mode, the function items are displayed as
basic item images. If an event occurs while the function items are
displayed in the normal mode, the basic item image of the function
item related to the event is animated. In the normal mode, the
mobile phone enters a selection mode for displaying the function
item information if an item display mode key (a special key such as
"clear key" provided to the mobile phone) is input.
[0136] Referring to FIG. 9b, the function items are displayed with
a semitransparent layer in the selection mode, and the selected
item is zoomed in to be displayed in an enlarged form. In the
selection mode, other function items can be selected by keypad
manipulation. If an OK key is input while an item is selected, the
item is activated in the enable mode.
[0137] Referring to FIG. 9c, upon entering the enable mode, the
item image of the selected function item is displayed together with
the information on the selected function item. By manipulating the
up/down direction keys provided in the keypad unit, it is possible
to navigate to other applets or return to the normal mode shown in
FIG. 9a. Left/right direction keys allow navigating to and
activating other function items.
[0138] A structure and presentation of a function item are
described with the memory monitor item as an example hereinafter
with reference to FIGS. 9d to 9f.
[0139] The memory monitor item is a function item for display
memory utilization information. FIG. 9d illustrates a memory
monitor item image in the normal mode. If the memory usage is over
50% in the normal mode, the memory monitor item is animated as
illustrated in FIG. 9e. If the memory monitor item is selected by
the keypad manipulation, the memory monitor item is displayed on
the screen in the form of an enlarged memory monitor item image
presenting the memory utilization with a text and gauge. The memory
utilization information is displayed in the text form of a used
memory/total memory. Also, the memory utilization information is
presented with a colored bar of the gauge. The gauge bar can be
placed in left or right side of the memory monitor item according
the position of the memory monitor item on the screen and changed
in color (for example, red) if the memory utilization percentage is
over a predetermined percentage (for example, 95%). If a "go to
memory status" command is input through a menu screen, a detailed
memory utilization screen is displayed to provide information on
the usage of the memory.
[0140] A structure and presentation of a function item are
described with the clock item as an example hereinafter with
reference to FIGS. 9g to 9i.
[0141] The clock item is a function item for alarming at preset
times. FIG. 9g illustrates the clock item image in the normal mode.
If a preset time has arrived, the clock item is animated as
illustrated in FIG. 9h. If the clock item is selected by the keypad
manipulation, the clock item is presented on the screen in the form
of an enlarged clock item image with the present time. The preset
time can be presented in the form of a text and image. The preset
time can be of a variety of functions such as "morning call,"
"schedule," "to do,". The text can be presented in various formats.
For example, a text format 1 can be used for a title of the
schedule and a text format 2 for date and time. If the schedule
title is not set, a default text "no title" is presented. In the
text format, the time is presented in the form of TT/MM, and the
date is presented in the form of today/tomorrow/a day of the week.
The text can be positioned at a left or right side of the clock
item according to the position of the clock item on the screen. If
multiple schedules are preset, the most closely arranged schedule
is shown. If a "go to alarm status" command is input through a menu
screen, a detailed schedule screen is displayed for informing
preset alarm schedule details.
[0142] A structure and presentation of a function item are
described with the mailbox item as an example hereinafter with
reference to FIGS. 9j to 9l.
[0143] The mailbox item is a function item for informing the number
of the unread messages among the received messages. FIG. 9j
illustrates the mailbox item image in the normal mode. If a message
is received in the normal mode, the mailbox item is animated as
illustrated in FIG. 9k. If the mailbox item is selected by the
keypad manipulation, the mailbox item is presented on the screen in
the form of an enlarged mailbox image with the information on the
received and unread messages. The information on the received and
unread messages can be presented with a text and image. The text
can be presented
[0144] The text can be presented in various formats. For example, a
text format 1 can be used for a category of the information (in
this case, New Message), and a text format 2 for the numbers of the
messages in the form of "new message/total message." The text can
be positioned at a left or right side of the mailbox item according
to the position of the mailbox item on the screen. If a "go to
inbox" command is input through a menu screen, a message list
screen is displayed for illustrating the received messages.
[0145] The motion items are items for making the screen
presentation of the character house dynamic. The motion items move
around on the screen according to movement lines by item.
[0146] FIGS. 10a to 10c are views illustrating steps of presenting
motion items in a screen image presentation technique according to
an exemplary embodiment of the present invention.
[0147] Referring to FIGS. 10a to 10c, the motion items of the
character house include animals and movable objects as shown in
FIG. 10a. Each motion item moves in accordance with a unique
movement line and moves at a preset animation time.
[0148] The motion items are animated on the basis of the position
information contained in the character house information such that
each motion item starts moving at the preset time along the preset
movement line as illustrated in FIG. 10b. The motion item can be
configured to move at a random speed and turn the progress
direction in random probability.
[0149] As illustrated in FIG. 10c, when moving across each other,
the motion items are overlapped. According to an exemplary
implementation, the motion items are illustrated in an order of the
presentation priorities (depth) that are preset for the respective
items. All the motion items are assigned presentation priorities so
that the motion item with the highest priority is presented when at
least two motion items are overlapped while on the move. Some
motion items have higher priorities than that of the character, and
others have lower priorities. In FIG. 10c, the motion item "collie"
(a dog) has a priority higher than the character such that the legs
of the character are covered by the collie. Also, a special motion
item (in FIG. 10a, a butterfly) is assigned the highest priority
such that it always appears on the character and other items.
[0150] As described above, the motion items are animated to move at
a certain speed (for example, 5 pixels/sec) so they may disappear
out of the character house (cross the border of the screen) and
then re-appear. If a motion item encounters the character or other
item while moving along the movement line, the motion item or the
character with a higher priority is displayed on top. If a motion
item with a priority lower than that of the character passes across
the character, the motion item is shaded by the character.
[0151] The decoration items are objects for decorating the
character house such that the decoration items have no special
functions. FIGS. 11a to 11c are views illustrating steps in the
presentation of decoration items in a screen image presentation
technique according to an exemplary embodiment of the present
invention. FIG. 11a illustrates exemplary decoration items that can
be presented for decorating the character house, FIG. 11b
illustrates exemplary character houses decorated with the
decoration items, and FIG. 11c illustrates specific decoration
items shown in an enlarged form on the screen.
[0152] Referring to FIGS. 11a to 11c, the decoration items to be
applied to the character house include flowers, furniture, and
other ornamental things. Such decoration items can be presented
with the character in the character house as illustrated in FIG.
11b. Some decoration items have special meanings. The special
decoration items such as a trophy and a clover shown in FIG. 11c
are obtained as a reward for winning a game or for a certain
achievement of a specific state of the mobile phone, but bought in
cash. The decoration item "trophy" is used for informing a number
of wins of the game, such that all wins of games may change the
trophy in color and type and increases the number of the
trophies.
[0153] Until now, how to present the character 230, the background
240, the items, and the character quotient module 260 is explained.
Here, the character quotient module 260 can be selectively set to
be visible or invisible. All the elements constituting the
character house are presented in association with the character
house. The character house also can be set to differently operate
at a specific time or situation. For example, the character house
can be configured such that the character gives greetings at
specific times or when a specific event occurs.
[0154] When the greeting behavior is set, the character gives the
user greetings at the preset times and an item or character money
can be given as a reward for the behavior.
[0155] FIG. 12 shows screen images illustrating character behaviors
in the screen image presentation technique according to an
exemplary embodiment of the present invention. As shown in FIG. 12,
when the character greets at a specific time, the greeting
expression is displayed in the form of a text balloon and the
greeting expression changes according to the time periods of the
day.
[0156] The greeting behavior of the character can be configured
such that the character greets at a predetermined number of times
per day (such as, 4 times a day). If a determination is made that
the character greets 4 times a day, 4 greeting time zones are
configured: morning (08:30.about.09:30), afternoon
(12:00.about.13:30), evening (17:30.about.18:30), and night
(23:30.about.00:30). Accordingly, the character greets once at each
greeting time zone. If the mobile phone is off or in character
house disabled modes at each greeting time zone, the greeting
behavior is disabled. Therefore, the mobile phone is not in the
idle mode where the character house screen is presented (for
example, when a clam shell type phone is unfolded or a slide type
phone is opened), the greeting behavior function is disabled. When
the greeting behavior starts, the greeting behavior remains during
a predetermined time period (for example, 10 seconds). If a
specific event (for example, an incoming call or a key input for
activating a specific function) occurs while the character greets,
the greeting is performed if the greeting behavior is maintained
over a predetermined amount of time (for example, 3 seconds). If
the greeting behavior stops before 3 seconds elapse and the
character house is activated at the greeting times zone, the
greeting behavior starts again. The character is preferably
associated with different greeting behaviors for the respective
greeting time zones as in table 2. TABLE-US-00002 TABLE 2 Class
Time Zone Text Morning 08:30.about.09:30 Hello! Afternoon
12:30.about.13:30 Good.about. Evening 17:30.about.18:30 Bye.about.
Night 23:30.about.00:30 See U.about.
[0157] When the greeting behavior starts, the mobile phone can
generate an event. The event may occur in a probability set
according to the times of the greeting behaviors. In this exemplary
embodiment of the present invention, the probability is about 10%.
If the event does not occur during the greeting time zone (that is,
the mobile phone is off during the greeting time zone or the
character house is not activated), the event will become
invalid.
[0158] FIG. 13 is a screen image illustrating a character animation
with an event in a screen image presentation technique according to
an exemplary embodiment of the present invention.
[0159] Referring to FIG. 13, if the character animation starts, a
text is presented in the form of a text balloon. The text
expression as an even is presented in the text balloon. The text
balloon is displayed with a 30% transparency. The text balloon for
expressing the event can be presented together with an image.
According to an exemplary embodiment of the present invention, a
gift event text is presented with a gift image as shown in FIG. 13.
The gift image is preferably presented at a bottom left or bottom
light portion of the screen. The gift event can be replaced by
money or other item event.
[0160] The items and background are selected and arranged on the
screen for decorating the character house.
[0161] FIG. 14 is a flowchart illustrating a character house
presentation procedure of a screen image presentation method
according to an exemplary embodiment of the present invention.
[0162] Referring to FIG. 14, the controller 110 controls the
operations of the mobile phone and determines whether the mobile
phone enters a character house presentation mode (S311). If the
mobile phone enters the character house presentation mode, the
controller verifies the character house configuration information
(S313). The character house presentation mode can be identical with
the idle mode of the mobile phone. In the idle mode, the controller
110 displays the character house on the screen of the display unit
143 during a predetermined display period. If the display period
ends, the controller 110 turns off the display unit 143 to save
power. If a clam shell or a slide of the mobile phone is open in
the idle mode, the controller 110 turns on the display unit 143.
The character house configuration information includes parameters
configured using a setting application. The parameters may include
information for activating or deactivating the character house
function. The character house configuration information may include
character house on/off information, character information display
on/off information. If the character house configuration
information includes the character house function related
information, the controller 110 determines whether the character
house function is turned on (S315). If it is determined that the
character house function is turned off, the controller 110 displays
a normal idle mode screen.
[0163] If it is determined that the character house function is
turned on, the controller 110 verifies the character house
information (S319). The character house information is on the
currently set character house. The character house information
includes the information on the background and items associated
with the character house. The background information includes a
background image ID, and the item information includes item IDs and
positions of the items on the screen (X, Y coordinate). The
character house can be displayed with one or more items, and the
items are arranged in an order of the presentation priorities such
that if two items are overlapped, the item with a higher priority
is displayed on the other item. By verifying the character house
information, the controller 110 recognizes the background and items
with their positions on the screen of the display unit 143.
[0164] After verifying the character house information, the
controller 110 controls to present the background, items, and
character (S321, S323, and S325). In FIG. 14, the character house
presentation is performed in the order of the background, items,
and character. However, the presentation order of the elements can
be changed. For example, the elements of the character house can be
presented in an order of the background, character, and items; an
order of the character, background, and items; and an order of the
character, items, and background. That is, the character house
presentation can be performed in various orders. Through the steps
S321 to S325, the character house presentation is completed.
[0165] During the presentation of the character house, the mobile
phone determines whether an event (for example, a function is
selected by a key input) occurs or a predetermined display time
ends (S327). If an event occurs or the display time ends, the
controller 110 ends the presentation of the character house and
then performs the selected function or turns off the display unit.
Also, it is possible to perform the selected function while
maintaining the character house screen. For example, if a message
is received while the character house is displayed or a message
transmission mode is activated, the controller 110 can maintain the
presentation of the character while reading the received message
and writing a message to transmit. The controller 110 displays a
message received or a message to be transmitted in a transparent
message window on the character house screen. Also, the controller
110 can increase the transparency of the character house
presentation and display the message on the character house
presentation. Other events can be processed while the character
house is presented in the similar manner.
[0166] FIG. 15 is a flowchart illustrating the background
presentation procedure of FIG. 14. Referring to FIG. 15, the
controller 110 verifies a background ID of the character house
(S361), and determines whether the background ID is 0 (S363). The
background ID "0" means the character house does not use any
background images. If the background ID is 0, the controller 110
loads a normal background image set as a default image (S365). If
the default background image is not set, no background image is
presented.
[0167] If the background ID is not 0, the controller 110 retrieves
a background corresponding to the background ID and presents the
retrieved background on the screen. According to an exemplary
implementation, the controller 110 selects one of the backgrounds
as shown in FIG. 6 on the basis of the background ID.
[0168] The background of the character house can be represented by
multiple images that are presented at different time periods. The
background of the character house can be represented by a fixed
single image.
[0169] Accordingly, after retrieving the background, the controller
110 determines whether the retrieved background is a multiple image
background (S364). If the retrieved background is a multiple image
background, the controller verifies the current time (S367) and
determines that the current time is greater than or equal to 06:00
and less that 12:00 (S369). If the current time is greater than or
equal to 06:00 and less than 12:00, the controller 110 selects a
morning image assigned for the morning or afternoon from the
multiple image background and then displays the morning image on
the screen (S371). If the current time is not greater than or equal
to 06:00 or if the current time us not less than 12:00, the
controller 110 determines whether the current time is greater than
or equal to 12:00 and less than 18:00 (S373). If the current time
is greater than or equal to 12:00 and less than 18:00, the
controller 110 selects an afternoon image assigned for the
afternoon from the multiple image background and then displays the
afternoon image on the screen (S375). If the current time is not
greater than or equal to 12:00 or not less than 18:00, the
controller 110 determines whether the current time is greater than
or equal to 18:00 and less than 23:59, or greater than or equal to
00:00 or less than 06:00 (S377). If the current time is greater
than or equal to 18:00 and less than 23:59, or greater than or
equal to 00:00 or less than 06:00, the controller 110 selects an
evening or night image assigned for the evening or night from the
multiple image background and displays the evening or night image
on the screen (S379). At step S364, the retrieved background is not
a multiple image background, the controller regards the retrieved
image as a fixed single image background and displays the fixed
signal image as the background (S381).
[0170] When the multiple image background is used, a signal fixed
image can be used by changing the brightness of the image for the
respective time periods of a day.
[0171] FIG. 16 is a flowchart illustrating the item presentation
procedure of FIG. 14. The item presentation procedure of FIG. 16 is
described with the character information presentation of FIG. 7b.
However, the character information presentation can be performed
with the background presentation procedure, or the character
presentation procedure of FIG. 19.
[0172] In this exemplary embodiment of the present invention, it is
assumed that the character information presentation is included in
the item presentation procedure as shown in FIG. 16. The emoticons
for expressing the emotion of the character, text for expressing
the state of the character, and the character quotients, shown in
FIG. 7b, included in the character information are explained.
[0173] Referring to FIG. 16, the controller 110 verifies whether
character house configuration information includes information on a
character quotient presentation function (S401) and determines
whether the character information presentation function is on
(S403). If the character information presentation function is on,
the controller 110 verifies the character quotient information
(S405) and presents a character emoticon on the basis of the
character quotient information in the emotion icon display region
263 of FIG. 7b (S407). The emotion of the character can be
represented by 5 emotion modes: happy, normal, angry, sad, and
tired. The emoticon to be displayed in the emotion icon display
region 263 can be one of the emoticons corresponding to the emotion
modes, as shown in FIG. 8. Once the emoticon is presented, the
controller 110 presents a text corresponding to the current state
of the character in the status display region 265 of FIG. 7b (S409)
and then presents the character quotients (SQ, HQ, IQ, and EQ) in
the quotient information display region 267 of FIG. 7b in the form
of bar graphs (S411). The values of character quotients are
presented as different colored bars. As mentioned above, if the
character information presentation function is enabled, the
controller 110 verifies the character quotients and determines the
emotion mode on the basis of the character quotients. If the
emotion mode is determined, the controller 110 presents an emoticon
corresponding to the emotion mode, a text appropriate for the state
of the character, and the quotients in the form of the bar
graphs.
[0174] The character information based on the character quotients
can be presented at a fixed position in the character house (for
example, a top left portion) as shown in FIG. 7a in the form of
FIG. 7b. The character quotient information cannot be presented
according to the user's configuration. The user decides whether to
present the character information. If the user does not activate
the character information presentation function, the controller
recognizes the configuration at step S403 such that the steps S405
to S411 are skipped. According to an exemplary implementation, the
character information based on the character quotients information
shown in FIG. 7b is not presented.
[0175] After presenting the character quotients, the controller
verifies the information on an item with a lowest priority (depth)
(S413). The item information includes an item ID and display
position. The item can be one of the motion items, function items,
and decoration items. In some cases, two or more items can be
overlapped. The items are presented in an order of the presentation
priorities assigned to the items such that the item with a lowest
priority is presented first. That is, when two or more items are
overlapped, only the item with the highest priority is completely
presented. If the item on top is disabled, the item with the second
highest priority is presented.
[0176] Accordingly, after verifying the item information, the
controller 110 determines whether the item information exists
(S415). If the item information exists, the controller 110
determines whether the item is a motion item on the basis of the
item information (S417). If the item is a motion item, the
controller 110 presents the motion item (S419) and verifies another
item with the next lowest priority (S421). If it is determined that
the item is not a motion item at step S417, the controller 110
determines whether the item is a function item (S423). If a
determination is made that the item is a function item, the
controller 110 presents the function item (S425). If the item is
not a function item, the controller 110 presents the item as a
decoration item (S427).
[0177] The item presentation is performed in a different manner
based on the type of item. That is, the motion item, function item,
and decoration item are presented in different procedures.
[0178] FIG. 17 is a flowchart illustrating a procedure for
presenting a motion item in the screen image presentation technique
according to an exemplary embodiment of the present invention.
[0179] Referring to FIG. 17, the controller 110 verifies the motion
item information on the basis of the item ID (SS431). The motion
item information includes an item ID, a sojourn time for staying
the item at the initial position, a coordinate represented by a
horizontal distance (X axis distance) for moving the item to the
next position and a vertical distance (Y axis distance) for moving
the item to the next position, and initial position information.
The initial position information contains one of 8 coordinates
listed in table 3. The motion item is assumed to start moving from
an edge of the screen as the initial position (such as 4 sides and
4 corners of the screen of the display unit 143) in table 3. The
motion item may also be configured to start moving at a random
position in the screen of the display unit 143. TABLE-US-00003
TABLE 3 1. (X=0) 2. (X=LCD_X_MAX) 3. (Y=0) 4. (Y=LCD_Y_MAX) 5.
(X=0, Y=0) 6. (X=0, Y=LCD_Y_MAX) 7. (X=LCD_X_MAX, Y=0) 8.
(X=LCD_X_MAX, Y=LCD_Y_MAX)
[0180] After verifying the motion item information, the controller
110 obtains the initial position of the motion item to animate the
motion item at the initial position (S433). The animation is
maintained before the sojourn time is ended and then the motion
item is positioned at the next position through the steps S435 to
S445. After animating the motion item, the controller 110
determines whether the sojourn time expires (S435). If the sojourn
time is expired, the controller 110 calculates a coordinate of the
next position to which the motion item moves (S437). After
calculating the next position coordinate, the controller 110
determines whether the calculated next position is in a available
display range on the screen of the display unit 143 (S439). If the
calculated next position is in the available display range, the
controller 110 moves the motion item to the next position (S441)
and then returns to step S433 to animate the motion item at the
next position. At step S439 the controller 110 determines that the
movement of the motion item is completed and moves positions the
motion item at the initial position (S443) and repeats the step
S433 if the calculated next position is not in the available
display range.
[0181] The motion item is configured with one of the initial
positions described in table 3 and moves to a predetermined
coordinate on the screen after a predetermined period of time. The
motion item is animated at the initial position, moved to the next
position after the sojourn time, and then animated again, under the
control of the controller 110. The movement time is determined on
the basis of the sojourn time and a distance between the two
positions. If the sojourn time is short and the distance between
the two positions is close, the motion item moves fast.
Alternatively, if the sojourn time is long and the distance between
the two positions is great, the motion item moves slowly. The
movement direction of the motion item depends on the relationship
of the two positions (coordinate in the screen). The motion item
can move in horizontal, vertical and diagonal directions and can be
determined in consideration of the distance between the two
positions. The motion item can be one of the items depicted in FIG.
10 and move in the manner illustrated in FIG. 10b. The motion items
move on the screen as illustrated in FIG. 10b in accordance with
the procedure of FIG. 17.
[0182] The movement of the motion item stops once the character
house presentation ends. The movement of the motion item may also
be set to stop at a time different from the character house
presentation end time. That is, the motion item presentation time
can be shorter than the character house presentation time.
According to an exemplary implementation, the motion item
disappears before the character house presentation stops.
[0183] Accordingly, if the sojourn time is not expired at step
S435, the controller 110 maintains the animation of the motion item
(S445) and determines whether the motion item presentation time is
ended, (S447). If the motion presentation time is ended, the
controller stops presentation of the motion, and otherwise, repeats
the step S435.
[0184] FIGS. 18a to 18d are flowcharts illustrating a procedure for
presenting function items in the screen image presentation
technique according to an exemplary embodiment of the present
invention.
[0185] Referring to FIGS. 18a and 18b, the controller 110 verifies
a type of function item (S452). In this exemplary embodiment of the
present invention, three function items are assumed to exist. This
items are the mailbox item, alarm clock item, and memory monitor
item. The number and type of the function items can be changed.
[0186] After the function item is verified, the controller 110
determines whether the function item is a mailbox item (S453). If
the function item is the mailbox item, the controller verifies a
presentation position of the mailbox item and presents mailbox item
at the presentation position on the screen (S455). Next, the
controller 110 verifies a number of received but unread messages
(S457) and determines whether the number of unread messages are
greater than 0 (S459). If the number of the unread messages is not
greater than 0, the controller presents an empty mailbox image.
[0187] If the number of the unread messages is greater than 0, the
controller 110 displays the number of the unread messages with the
mailbox image (S461 to S467). The number of the unread messages can
be presented in various manners. For example, the number of unread
messages can be displayed on the mailbox image. Also, the number of
unread messages can be displayed by means of a text balloon.
According to an exemplary implementation, the text balloon has a
shape of a postcard with the number of the unread messages. In this
exemplary embodiment of the present invention, the mailbox item
presentation is explained with the text balloon image for
displaying the number of the unread messages.
[0188] According to the position of the mailbox item, the
controller 110 determines an unread message number presentation
position. The controller 110 verifies the left and right side
spaces of the mailbox items and arranges the text balloon at the
left or right side. If a determination is made that the text
balloon is arranged at the right side, the controller 110
determines whether the text balloon presentation position is in the
available display range of the screen (S463). If the text balloon
presentation position is in the available display range, the
controller 110 presents the text balloon at the right side of the
mail box (S467). If the text balloon presentation position is not
in the available display range, the controller 110 determines to
the text balloon at the left side of the mailbox item (S465) and
repeats the step S463.
[0189] At step S453, if it is determined that the function item is
not the mail box item, the controller determines whether the
function item is an alarm clock item (S469) in FIG. 18b. If the
function item is the alarm clock item, the controller 110 verifies
scheduled alarms on the basis of alarm information (S471). The
alarm information includes the presentation position of the alarm
clock item on the screen of the display unit 143. After verifying
the scheduled alarms, the controller determines whether the there
is scheduled alarms (S473). If no alarm is scheduled, the
controller 110 presents an empty alarm image. If there is at least
one scheduled alarm, the controller 110 retrieves a next scheduled
alarm and displays an alarm clock image with the next scheduled
alarm (S475). The alarm clock item can be implemented with a clock
image. According to an exemplary implementation, the alarm clock
image may have the needles arranged to show the scheduled alarm
time.
[0190] Next, the controller 110 verifies an alarm time presentation
position and determines whether there is enough space to present
the alarm time at the right side of the alarm clock item in the
screen. If there is enough space to present the alarm time, the
controller 110 presents the next scheduled alarm time at the right
side of the alarm clock item (S477). If it is determined that there
is not enough space to present the alarm time at the right side of
the alarm clock item (S479), the controller 110 presents the alarm
time at the left side of the alarm clock item (483). The alarm can
be represented by a morning call, a meeting schedule, and a
reservation, among others. The scheduled alarms are activated at
the scheduled times with the specific events. The alarm time can be
shown in the form of YYYY/MM and HH/MM as a text. The alarm time
text can be placed at the left or right side of the alarm clock
item according to the position of the alarm clock item.
[0191] If two or more alarms are scheduled, the alarm time of the
next closest alarm is presented. The alarm time can be presented by
means of the alarm clock image. When the alarm time is presented by
means of the alarm clock image. The alarm time text may not be
presented.
[0192] If it is determined that the function item is not the alarm
clock item at step S469, the controller 110 determines that the
function is a memory item. Next, the controller 110 verifies a
presentation position of the memory item from the character house
information and presents the memory item at a verified presentation
position (S484). After presenting the memory item on the screen,
the controller 110 verifies a total memory capacity and used memory
capacity (S485) and then displays the information on the total
memory capacity and the user memory capacity (S487 to S493). That
is, after verifying the memory information, the controller 110
verifies the side of the memory item that has enough space for
presenting the memory information. If it is determined that the
right side of the memory item has enough space for presenting the
memory information, the controller 110 presents the memory
information at the right side of the memory item. Alternatively, if
it is determined that the right side of the memory item does not
have enough space for presenting the memory information, the
controller 110 presents the memory information at the left side of
the memory item. The memory item may inform of the memory
information of the memory unit 130 of FIG. 1 and the memory
information can be shown in the form of a gauge bar. The gauge bar
can be located at the left or right side according to the position
of the memory item on the screen. The gauge bar shows the memory
utilization and changes in red if the memory utilization reaches
over 90%.
[0193] As describe above, a function item is presented in such a
manner that the controller verifies the presentation position of
the function and presents the function item at the presentation
position on the screen of the display unit 143 together with the
function information obtained by the function item. The function
information is presented in such a manner that the controller
verifies which side of the function item has enough space for
presenting the function information and then presents the function
information at a left or right side of the function item, where
there is enough space. The function information can be shown as a
text or a graphical gauge. When the function information is
provided in the text, the graphical gauge may or may not be
used.
[0194] FIGS. 18c and 18d are flowcharts illustrating a procedure
for presenting function items in the screen image presentation
technique according to another exemplary embodiment of the present
invention.
[0195] Referring to FIGS. 18c and 18d, the controller 110 verifies
a type of function item (S751). In this exemplary embodiment of the
present invention, three function items are assumed to exist. The
function items include a mailbox item, alarm clock item, and memory
monitor item. The number and types of the function items can be
changed. If function information is associated with a function
item, the controller presents the function item in the form of an
animated function item for attracting the user's interest and
presents the function information in a popup window during a popup
display time. In this exemplary embodiment of the present
invention, the function items are presented in the following order:
mailbox item, alarm clock item, and memory item.
[0196] After the function item is verified, the controller 110
determines whether the function item is the mailbox item (S753). If
the function item is the mailbox item, the controller 110 verifies
an amount of the unread mail (S755) and then determines whether the
amount of the unread mail is greater than 0 (S757). If the amount
of the unread mail is 0, the controller 110 presents an empty
mailbox image on the screen of the display unit 143 as shown in
FIG. 9j (S767) to notify that there is no unread message.
[0197] If the amount of the unread mail is greater than 0, the
controller 110 presents an animated mailbox image as shown in FIG.
9k. After presenting the animated mailbox image, the controller
determines whether a popup time is arrived (S761). If the popup
time has arrived, the controller 110 presents a popup window as
shown in FIG. 9l (S763). The popup window is displayed during a
predetermined popup time and then disappears. During the
presentation of the popup window, the controller 110 simultaneously
starts a popup counter. After the popup window is presented, the
controller determines whether the popup counter is expired (S765).
If the popup counter is expired, the popup window disappears.
[0198] If the function item is not the mailbox item as in step
S753, the controller 110 determines whether the function item is an
alarm clock item (S771). If the function item is the alarm clock
item, the controller 110 verifies scheduled alarms (S773) and
determines whether there are scheduled alarms (S775). If there is
no scheduled alarm, the controller 110 presents an empty alarm
image on the screen of the display unit 143 for notifying that no
scheduled alarm exists as shown in FIG. 9g (S785). If there is at
least one scheduled alarm, the controller 110 presents an animated
alarm clock image as shown in FIG. 9h (S777) and determines a popup
time has arrived (S779). If the popup time is arrived, the
controller 110 presents a popup window as shown in FIG. 9i (S781).
The popup window is displayed during a predetermined popup time and
then disappears. During the presentation of the popup window, the
controller 110 simultaneously starts a popup counter. After the
popup window is presented, the controller 110 determines whether
the popup counter is expired (S783). If the popup counter is
expired, the popup window disappears. The popup window presentation
time may be configured in consideration of the alarm time.
According to an exemplary embodiment of the present invention, the
scheduled alarm is assumed to be an alarm set with a specific date
and time for "meeting" and "to do" rather than an everyday
occurrence alarm (for example, morning call), and displays alarms
before 6 hours.
[0199] If the function item is not the alarm clock item, the
controller 110 determines whether the function item is a memory
monitor item (S786). If the function item is the memory monitor
item, the controller 110 verifies memory utilization (S787) and
determines whether the memory utilization is in a predetermined
normal range (S789). If the memory utilization is in the normal
range, the controller 110 presents only the memory monitor item
image on the screen of the display unit 143 for notifying that the
memory utilization is normal, as shown in FIG. 9d (S799). If the
memory utilization exceeds the normal range, the controller 110
presents an animated memory monitor as shown in FIG. 9e (S791) and
determines a popup time has arrived (S793). If the popup time is
arrived, the controller 110 presents a popup window as shown in
FIG. 9f (S795). The popup window is displayed during a
predetermined popup time and then disappears. During the
presentation of the popup window, the controller 110 simultaneously
starts a popup counter. After the popup window is presented, the
controller 110 determines whether the popup counter is expired
(S797). If the popup counter is expired, the popup window
disappears. The memory monitor item can be displayed regardless of
the memory utilization. According to an exemplary implementation,
the controller 110 verifies the memory utilization and then
presents the memory utilization in the form of a popup window.
[0200] As described above, the function items can be presented
together with the character house. According to an exemplary
implementation, the function items are displayed in a preset order.
While presenting the function items in the preset order, the
controller determines whether to present the function item in an
active state. If it is required to present the function item in the
activate state, the controller 110 presents the function item as an
activated function item image and then provides the information
related to the function in the form of a popup window after a
while. The popup window is presented during a predetermined popup
time. For example, when all the function items are presented in the
activate state, the controller 110 presents in the order of the
mailbox item as an animated mailbox image followed by the mailbox
popup window, the alarm clock item as an animated alarm clock image
followed by the alarm popup window, and the memory monitor item as
an animated memory monitor image followed by the memory utilization
popup window. The function items are presents in the character
house as shown in FIGS. 9j, 9g, and 9d after the popup windows
disappear.
[0201] The decoration items are presented on the basis of the
decoration item information contained in the character house
information. The item information includes an item ID that provides
information on the type of the decoration item and a presentation
position on the screen. The controller 110 retrieves a decoration
item image corresponding to the item ID and presents the retrieved
decoration item image at the presentation position on the screen as
shown in FIG. 11a. The decoration item is used for improving
aesthetic factor of the character house, regardless of any
function. The decoration item can be added or removed according to
the user's configuration.
[0202] In the character house, the character is implemented with
three behavior properties.
[0203] First, the character greets the user of the mobile phone at
preset times. Second, the character irregularly gives the user a
gift while greeting the user. Third, the character expresses a
state of the mobile phone or an emotion reflecting the state of the
mobile phone in an animation or a motion. In this exemplary
embodiment of the present invention, it is assumed that the mobile
phone has various emotions that represent a user's contribution to
the mobile phone. The character house can be presented only with
the third behavior property. That is, the greeting and gift-giving
behavior properties can be disabled when the character house is
presented. In this exemplary embodiment of the present invention,
it is assumed that the greeting, gift-giving, and animation
behavior properties are enabled.
[0204] The character house can consist of the items and background
as well as the character, and the character house is presented in a
specific operation mode of the mobile phone. When the character
house is presented in the idle mode, the character house can be
shown as an idle mode screen (for example, when a clam shell of the
mobile terminal is unfolded or a slide mobile phone is slid open).
If an event occurs while the character house is presented, the
character house disappears. In other words, the idle screen is
replaced by other function screen. Also, the character house can be
maintained when a specific event occurs (for example, a message
processing event).
[0205] FIG. 19 is a flowchart illustrating a character presentation
procedure of the screen image presentation method according to an
exemplary embodiment of the present invention.
[0206] Referring to FIG. 19, the controller 110 verifies a current
time (S501) and then determines whether the current time is in any
of preset greeting behavior time zones (S503, S513, S517, and
S521). In this exemplary embodiment of the present invention, the
greeting behavior time zones are set to four time zones as in the
table 2. The greeting behavior is shown once in a greeting behavior
time zone.
[0207] Accordingly, the controller 110 determines whether the
current time zone is in the greeting behavior time zones (S503,
S513, S517, and S521). If the current time zone is in the greeting
behavior time zones, the controller 110 determines the greeting
behavior is shown (S505, S515, S519, and S523).
[0208] If it is determined that the current time zone is not in the
greeting behavior time zones at the steps of S503, S513, S517, and
S521 or greeting behavior is performed already in the greeting
behavior time zone at the steps S505, S515, S519, and S523, the
controller 110 skips the greeting behavior and controls the
character to be in a normal behavior (S511).
[0209] If it is determined that the current time zone is in the
greeting behavior time zones at the steps of S503, S513, S517, and
S521 and the greeting behavior is not performed yet in the greeting
behavior time zones at the steps S505, S515, S519, and S523, the
controller 110 records the current time (S507) and controls the
character to show the greeting behavior (S509). Recording the
current time zone helps to avoid situations in which the greeting
behavior is not performed in the same greeting behavior time zone.
After the greeting behavior is performed, the controller 110
controls the character to be in the normal behavior (S511).
[0210] The greeting behavior presentation procedure is performed as
shown in FIG. 20. The greeting behavior is presented as the
character greets to the user and can be performed with a salutation
and an addition event for giving an item or character money in a
probability.
[0211] FIG. 20 is a flowchart illustrating a greeting behavior
presentation procedure of FIG. 19.
[0212] Referring to FIG. 20, the controller 110 verifies a current
greeting time zone (S531) and verifies current character
information (S533). Next, the controller 110 controls the character
to show the greeting behavior in the form of an animation (S535)
and retrieves a greeting text corresponding to the current greeting
time zone (S537). After retrieving the greeting text, the
controller 110 presents the retrieved greeting text in the form of
a text balloon (S539). The greeting behavior can be presented as
shown in FIG. 12. The greeting text can be presented in the form of
the text balloon of FIG. 12. The greeting text can be changed,
depending on the greeting time zones.
[0213] When the character house presentation is terminated due to
an event that occurs during the character's greeting behavior, the
controller 110 determines that the greeting behavior has been
performed if the greeting behavior is performed for a tolerable
time. According to an exemplary implementation, the tolerable time
is assumed to be 3 seconds.
[0214] Accordingly, the controller 110 determines whether the
greeting behavior is continued over the tolerable time (in this
embodiment, 3 seconds) (S541). If the greeting behavior is not
continued over the tolerable time, the controller 110 maintains the
greeting behavior and text balloon presentation (S557). While
maintaining the greeting behavior and the text balloon
presentation, the controller determines an event occurs such that
the character house presentation is terminated in 3
second-tolerable time (S559). If an event occurs in tolerable time,
the controller 110 stops the greeting behavior and then returns.
According to an exemplary implementation, the greeting behavior is
performed again, if the character house is activate in the same
greeting behavior time zone.
[0215] If the greeting behavior is performed over the tolerable
time (3 seconds), the controller determines that the greeting
behavior is completely performed and records the completion of the
greeting behavior in the greeting behavior time zone (S543). Next,
the controller 110 determines whether the event presentation time
has arrived (S545). The event may occur in a random probability set
according to the times of the greeting behaviors.
[0216] If the event presentation time has arrived, the controller
110 presents the greeting event (S547).
[0217] FIG. 21 is a flowchart illustrating a greeting event
presentation procedure that occurred during a greeting behavior of
the character of the screen image presentation technique according
to an exemplary embodiment of the present invention.
[0218] The greeting event can randomly occur while the character
performs the greeting behavior. The greeting event may also be
presented in the same manner with the character's greeting behavior
except for the screen layout and the text.
[0219] Referring to FIG. 21, the controller 110 selects an item or
character money as a gift at the greeting event presentation time
(S561) and presents an image corresponding to the item or the
character money to give as the gift (S563). Next, the controller
110 verifies the information of the character (S565) and animates
the character (S567). Subsequently, the controller 110 presents the
greeting text in the form of the text balloon (S569).
[0220] When presenting the character according to the greeting
event, the controller 110 animates the character with the text
and/or sound as shown in FIG. 13. At this time, the emotion mode of
the character is fixed to "happy" and the character is presented
with the item image and/or the character money image as the gift.
The item image and the character money image provided in the
greeting event presentation procedure can be displayed at the
bottom left or bottom right portion of the character house as shown
in FIG. 13, and the text expression related to the gift is
presented in the form of a text balloon having a 30%
transparency.
[0221] After performing the greeting event presentation procedure
as in FIG. 21 at step S547 of FIG. 20, the controller 110
determines whether 10 seconds are elapsed after the presentation of
the greeting expression text balloon (S549). If 10 seconds have
elapsed, the controller 110 makes the text balloon and the
character animation image disappear (S551).
[0222] If 10 seconds have not elapsed, the controller maintains the
presentation of the current character animation and the greeting
expression text balloon (S553). The greeting behavior of the
character is assumed to be maintained for 10 seconds. Accordingly,
before 10 seconds have elapsed during the presentation of the
greeting behavior of the character, the controller 110 maintains
the greeting behavior of the character (S549, S553, and S555). If
10 seconds have elapsed for the start of the greeting behavior of
the character, the controller 10 makes the greeting expression text
balloon and the character animation image disappear (S551).
[0223] As described above, if the current time is in a
predetermined greeting time zone, the controller 110 controls the
character of the character house to show the greeting behavior
corresponding to the greeting time zone. The greeting behavior is
for greeting to the user of the mobile phone and is performed on in
every greeting time zone. If the mobile phone is turned off or the
character house is not activated, the greeting behavior of the
character is skipped. In this exemplary embodiment of the present
invention, the greeting behavior of the character is performed in a
random probability to give the user an unexpected gift. In this
exemplary embodiment of the present invention, the gift-giving
behavior is performed during the greeting behavior time. However,
the gift-giving behavior may be performed at any time during the
presentation of the character house. If the gift-giving behavior is
performed at any time during the character house presentation, the
controller 110 can control the gift-giving behavior to be performed
when the character is presented.
[0224] In FIG. 19, the controller 110 controls the behavior of the
character so that it is normal (S511) after performing the greeting
behavior of the character out of the greeting time zones or after
performing the greeting behavior. The character behaviors are
presented in the procedure of FIG. 22a or FIG. 22b.
[0225] FIGS. 22a and 22b are flowcharts illustrating an animation
behavior presentation procedure of the screen image presentation
technique according to an exemplary embodiment of the present
invention.
[0226] In FIG. 22a, the character is moving and animated together
with a text and sound for expressing the state and emotion of the
character while the character house is presented. In FIG. 22b, the
character is animated after the text and sound expression is
presented in the character house.
[0227] The character behavior can be vividly expressed in the
character house in various manners. For example, the character can
be represented by a plurality of animation images such that the
animation images are continuously displayed to form an animation of
the character. Also, the character can be presented to move along a
preset movement line. The character animation is implemented by
presenting the animation images in series at regular intervals,
similar to presenting the function items as shown in FIG. 9e, 9h,
and 9k. The character movement is implemented by moving the
character on the basis of a preset movement direction, movement
distance, and movement time. In order for the character to
seamlessly move, the movement can be implemented with a plurality
of images. FIG. 23 shows a procedure for presenting a character
movement.
[0228] FIG. 22a shows a character movement behavior presentation
procedure, and FIG. 22b shows a character gesture behavior
presentation procedure.
[0229] Referring to FIG. 22a, the controller 110 verifies
information on the current character (S571) and determines an
initial appearance position of the character (an X, Y coordinate in
the screen of the display unit 143). The initial appearance
position of the character can be randomly determined in a display
range of the screen of the display unit 143. After the initial
appearance position is determined, the controller 110 presents the
character with 100% transparency (S575). The 100% transparency
produces character invisibility so that the background and items
images are displayed on the screen of the display unit 143 as shown
in FIG. 4a.
[0230] According to an exemplary implementation, the controller 110
decreases the transparency of the character image at a
predetermined rate so that the character gradually appears in the
character house. The controller 110 controls such that the
character appears gradually by repeating the steps S577 and S579.
If the transparency of the character image is assumed to decrease
by 10%, the controller 110 controls to gradually reduce the
transparency of the character image to be opaque (alpha blending
effect). Accordingly, the transparency of the character image
decreases from 100% to 0% by 10% such that the character appears in
the character as shown in FIG. 4b.
[0231] According to an exemplary implementation, the controller 110
outputs a sound effect representing the state of the character as
shown in FIG. 4c (S581), animates the character as shown in FIG. 4d
(S583), presents a text expressing the state and emotion of the
character in the text balloon during a predetermined presentation
time as shown if FIG. 4e, and then terminates the animation of the
character. According to an exemplary implementation, the
presentation time is set to 3 seconds. During the presentation of
the character house as discussed above, the text balloon is
presented after appearance of the character in the order of FIGS.
4a to 4e and presentation of the sound effect and the character
animation.
[0232] The controller 110 determines a next position of the
character and then performs movement of the character (S589, S597,
and S559). In order to present the character movement, the
controller determines position shift information (S587). The
position shift information is determined according to a procedure
as shown if FIGS. 23a and 23b.
[0233] FIGS. 23a and 23b show the character shift information
determination procedure in detail.
[0234] Referring to FIGS. 23a and 23b, the controller 110 verifies
an initial appearance position of the character (S601) and
determines a movement type of the character (S603). The movement
type includes a vertical movement, a horizontal movement, and a
diagonal movement. The movement type is randomly selected among the
three movement types at step S603. It is assumed that the character
moves by 5 pixels.
[0235] If the movement type is selected, the controller determines
whether the movement type is the vertical movement (S605). If the
movement type is the vertical movement, the controller 110
determines whether the character's appearance position locates at
an upper half part of the screen of the display unit 143 (S607). If
the character's appearance position locates at an upper half part
of the screen of the display unit 143, the controller 110 generates
a character position down shift information ((A, B)=(0, -5))
(S609). If the character's appearance position locates at a lower
half part of the screen of the display unit 143, the controller 110
generates a character position up shift information ((A, B)=(0, 5))
(S611).
[0236] If it is determined, at step S605, that the movement type is
not the vertical movement, the controller 110 determines whether
the movement type is the horizontal movement (S613). If the
movement type is the horizontal movement, the controller determines
whether the character's appearance position is located at a right
half part of the screen of the display unit 143 (S615). If the
character's appearance position locates the right half part of the
screen of the display unit 143, the controller generates a
character position left shift information ((A, B)=(-5, 0)) (S617).
If the character's appearance position is located at a left half
part of the screen of the display unit 143, the controller 110
generates character position right shift information ((A, B)=(5,
0)).
[0237] If it is determined, at step S615, that the movement type is
not the horizontal movement, the controller 110 regards the
movement type as the diagonal movement.
[0238] The diagonal movement includes a top-right to bottom-left
movement, a top-left to bottom-right movement, a bottom-right to
top-left movement, and a bottom-left to top-right movement.
[0239] If the movement type is the diagonal movement, the
controller determines whether the character's appearance position
is located at the right half part (X>LCD_X_MAX/2) of the screen
(S621). If the character's appearance position is located at the
right half part of the screen, the controller determines whether
the character's appearance position is located at the upper half
part (Y>LCD_Y_MAX/2)) of the screen (S623). If the character's
appearance position is located at the upper half part, the
controller 110 generates character position left-down shift
information ((A,B)=(-5,-5)). If the character's appearance position
is not located at the upper half part, the controller 110 generates
character position left-up shift information ((A,B)=(-5,5)).
[0240] If it is determined, at step S621, that the character's
appearance position is not located at the right half part of the
screen, the controller determines whether the character's
appearance position is located at the upper half part
(Y>LCD_Y_MAX/2)) of the screen (S629). If the character's
appearance position is located at the upper half part, the
controller 110 generates character position right-down shift
information ((A,B)=(5,-5)) (S631). If the character's appearance
position is not located at the upper half part, the controller 110
generates character position right-up shift information
((A,B)=(5,5)) (S633).
[0241] At step S587 of FIG. 22a, the controller 110 verifies the
appearance position and movement type of the character and
calculates the position shift information on the basis of the
appearance position and the movement type of the character. The
movement direction of the character is changed according to the
character's initial appearance position and the movement type of
the character is randomly selected among the three movement types
(the vertical, horizontal, and diagonal movement). The movement
direction is determined such that the character secures the longest
length to move at an axis of the center of the screen.
[0242] If the position shift information is obtained, the
controller 10 controls such that the character moves in the
movement direction and according to movement type (S589 to S599).
The appearance position shift is performed in a predetermined
interval. It is assumed that the character moves 20 pixels per
second. The character's appearance position shift is performed by 5
pixels every 0.25 seconds. The character is moving and animated at
the same time. That is, the character moves in a direction
determined at every predetermined times by a unit of predetermined
number of pixels. The character can be animated while moving or not
moving. The controller 110 determines whether the position shift
time arrived (S589). If the position shift time has not arrived,
the controller 110 animates the character (S597).
[0243] If the position shift time has arrived, the controller 110
calculates a next position of the character (S591). The next
position is obtained according to the procedure shown in FIGS. 23a
and 23b. After obtaining the next position, the controller 110
determines whether the next position of the character is in a
display range of the screen of the display unit 143 (S593). If the
next position of the character is in the display range of the
screen, the controller 110 controls to move the character to the
next position and animate the character. The controller 110 moves
the character on the screen of the display unit 143 in the
predetermined movement direction at the predetermined movement
times by unit of predetermined number of pixels as shown in FIGS.
4f and 4g.
[0244] If the character deviates from the display range of the
screen of the display unit 143 while repeating the above operation,
the controller 110 recognizes the deviation of the character and
returns to the step S571 to show the idle mode screen as shown in
FIG. 4h. Consequently, the presentations of FIGS. 4b to 4g are
repeated.
[0245] The above operation is repeatedly performed while presenting
the character house. If the presentation of the character ends
(S599), the controller 110 terminates the behavior of the
character.
[0246] The movement of the character is performed after the text
expression of the character disappears as shown in FIGS. 4f and 4g.
The movement direction of the character is randomly selected among
the vertical, horizontal, and diagonal direction. That is the
movement direction is determined such that the character secures
the longest length to move at an axis of the center of the screen
and the movement speed is determined at 20 pixels per second. If
the character moves across the border of the screen of the display
unit 143 to disappear, the character house is presented with the
character. After about 2 seconds have passed, the operation
depicted in FIG. 4b is performed.
[0247] As described above, the presentation of the character
behavior in FIGS. 22a and 22b is performed by repeating the
operations of FIGS. 4a to 4g. The character can be presented in 6
steps. The first is an invisible character step as shown in FIGS.
4a and 4h, in which the character is not presented yet such that
the character house is displayed with the items and background. The
second is a character presentation step, in which the character
appears at a random position in the character house. When the
initial appearance position of the character is close to the border
of the screen of the display unit 143 such that the whole image of
the character cannot be shown, the character is not presented. The
character can appear in a fade-in effect by decreasing the
transparency of the character image. The transparency decreased
from 100% to 0% (alpha blending image). The third is a sound effect
output step as shown in FIG. 4c, in which a unique sound effect is
output according to the state of the character after the character
is completely presented. The sound effect is output once. The
fourth is an animation step as shown in FIG. 4d, in which the
character is animated after the sound effect is output. The
character animation is repeated until the character moves to
disappear out of the screen of the display unit 143. The fifth is a
text presentation step as shown in FIG. 4e, in which the state and
emotion of the character are presented in text. The text can be
presented in the form of a text balloon as shown in FIG. 4e. The
text balloon is presented with a semitransparent balloon with a 30%
transparency in consideration of the time taken to read the text.
The text balloon disappears after about 3 seconds. The sixth step
is a character movement step as shown in FIGS. 4f to 4g, in which a
movement direction of the character is randomly determined. After
the movement direction and distance is determined, the character
moves according to the movement direction and distance. The
character moves to disappear out of the screen of the display unit
143.
[0248] Since the character appears at a random position, the
character may appear at different position at every appearance
time. The movement direction is randomly determined, the character
may move in different directions at every movement time. The sound
effects and the texts can be differently prepared for expressing
various states and emotions of the character. Since the appearance
position, movement direction, sound effect, and text are presented
differently at every presentation timings, various character
behaviors can be provided to the user.
[0249] At the initial character presentation time, the operations
illustrated in FIGS. 4a to 4g are performed and then the operations
of FIGS. 4c and 4e can be skipped after the character moves to
disappear out of the screen of the display unit 143. According to
an exemplary implementation, the sound effect and the text balloon
appear once and the character appears at a random position and
moves in a random direction.
[0250] Referring to FIG. 22b, the controller 110 verifies
information of the current character (S651) and determines a
character appearance position (S653). The character appearance
position can be a preset position on the screen of the display unit
143. In this embodiment, the character appearance position is
preset at a bottom right portion of the screen relative to the
center of the screen of the display unit 143. After determining the
character appearance position, the controller 110 presents the
character with a 100% transparency (S655). Since the transparency
is 100%, the character is invisible. Accordingly, the character
house is displayed only with the background image and item images
as shown in FIG. 4i.
[0251] According to an exemplary implementation, the controller 110
decreases the transparency of the character at a predetermined rate
such that the character appears gradually. That is, the controller
110 repeats the steps S657 and S659 for gradually presenting the
character. If it is assumed that the transparency of the character
image decreases 25% every 0.5 seconds, the character is completely
presented in 2 seconds (alpha blending effect). The transparency of
the character image decreases in an order of 100%, 75%, 50%, 25%,
and 0% such that the character house is displayed as shown in FIG.
4j. After the character has completely appeared, the controller 110
outputs a sound effect corresponding to the state of the character
as shown in FIG. 4c (S661) and then presents a text associated with
the state and emotion of the character in the form of a text
balloon during a predetermined text presentation time as shown in
FIG. 4l. The text presentation time can be set to about 1.5 seconds
in consideration of the time taken for reading the text.
[0252] After the sound effect is output and after the text has been
presented, the controller 110 presents a character animation at
steps S665 to S669. The character animation is repeated until the
character disappears from the character house. As described above,
the character presentation is performed in a specific order
according to the character appearance, output of sound effects,
presenting text balloon, and character animation.
[0253] As described above, the screen image presentation method of
an exemplary embodiment of the present invention presents the
character house (screen) in a specific operation (idle mode) of the
mobile phone. In the idle mode, the controller of the mobile phone
presents a character house by arranging a background and items in
the character house and presents a character. The character
expresses a state and emotion in the forms of a sound effect and
text. The character moves in random direction and shows an
animation on the screen of the display unit 143. The background
image is used as a background of the house and can be replaced with
another image to graphically illustrate the progress of the time
using character house. The items presented with the character house
includes movement items, decoration items, and function items. The
function items are used for improving user convenience by showing
system information of the mobile phone. The decoration items are
used for improving the aesthetic effect of the screen of the mobile
phone. The motion items are used to make dynamic the screen by
moving around the screen house. Also, the character house can be
implemented with a special expression function such that the
character gives special information at predetermined time zones or
greets to the user. Also, the character house can be implemented
such that the character gives an item or character money as a gift
at unexpected timing while the character greets, resulting in
user's enjoyment.
[0254] In this exemplary embodiment of the present invention, the
character house is structured by presenting the elements in an
order of the background, items, and character. However, the
presentation order can be changed. In this exemplary embodiment of
the present invention, the character house is configured such that
the greeting and event presentation are performed when the item is
presented, however, the greeting and event presentation can be
performed when the background or character is presented.
[0255] FIG. 24 is a flowchart illustrating a screen house
presentation procedure of a screen image presentation method for a
mobile phone according to another exemplary embodiment of the
present invention.
[0256] In association with the character house presentation, the
mobile phone can be in a character house configuration mode,
character house editing mode, or character house presentation
mode.
[0257] The character house configuration mode and character house
editing mode can be accessed by a corresponding menu selection or a
specific key (including a soft key) manipulation, and the character
house presentation mode activates for a specific operation mode of
the mobile terminal when the character house presentation function
is enabled.
[0258] Referring to FIG. 24, if a character house configuration
mode is executed by a mode execution key or menu selection, the
controller 110 detects the execution of the character house
configuration mode (S719) and then displays a character house
configuration menu screen for allowing a user to configure the
character house presentation options (S721). The character options
include a character house presentation on/off option and character
quotients presentation on/off option. If the character house
presentation on/off option is set to off, the controller 110
disables the character house presentation function. If the
character quotients presentation on/off option is set to off, the
controller 110 disables the presentation of the character quotients
shown in FIGS. 7b and 8. The character house option may include a
greeting function on/off option. If the greeting function on/off
option is set to off, the controller 110 disables the greeting
behavior presentation shown in FIG. 12 and greeting event
presentation shown in FIG. 13.
[0259] If the character house editing mode is executed by a mode
executing key or menu selection, the controller 110 detects the
executing of the character house editing mode (S715) and then
displays a character house editing menu screen for allowing the
user to edit the character house (S717). In the character house
editing mode, it is possible to change, add, and delete the
background, items, and character.
[0260] FIG. 25 is a state diagram illustrating a character house
design procedure of a screen image presentation method according to
an exemplary embodiment of the present invention.
[0261] Referring to FIG. 25, the controller 110 displays a
character house configuration screen in the character house
configuration mode 810, in which it is allowed to delete a
character house or to set a default character house. If a "new
character house" option is selected or a "character house editing"
option is selected on the character house configuration screen, the
controller 110 displays a character house editing screen in the
character house editing mode 820. From the character house editing
mode 820, it is possible to enter modes 830 to 870 for editing a
background and items.
[0262] FIG. 26 is a state diagram illustrating transitions between
the character house configuration mode and the character house
editing mode of FIG. 25. FIG. 27 is a screen image illustrating a
character house configuration screen in the character house
configuration mode of FIG. 25.
[0263] Referring to FIGS. 26 and 27, the character house
configuration screen displays character houses that can be used for
a specific mode screen image. The mobile phone includes a plurality
of character houses. These character houses are displayed on the
character house configuration screen in the form of tiled windows
as shown in FIG. 27. In this exemplary embodiment of the present
invention, the character house configuration screen is assumed to
display four character houses at once. If there are three character
houses stored in the mobile phone, one of the four tiled window is
shown empty. In relation with the empty character house window, it
is impossible to set a default character house, delete a character
house, and edit a character house. Also, a new character house can
be put into the empty window. If an OK key is input while the empty
window is selected, a character creation function is enabled. If
the OK key is input while one of the tiled windows having the
character houses is selected, the character house is set for a
screen image for a specific operation mode of the mobile phone.
[0264] In order to select one of the character houses on the
character house configuration screen, the cursor navigates on the
tiled windows in accordance with a direction key (navigation key)
manipulation. When the cursor is positioned on one of the tiled
windows, the controller 110 displays a menu having options shown in
table 4. TABLE-US-00004 TABLE 4 Empty window 2. new character house
Non-empty window 1. set default 2. edit character house 3. delete
character house
[0265] FIG. 28 is a state diagram illustrating transitions between
the character house configuration mode and the character house
editing mode of FIG. 25, and FIG. 29 shows screen images
illustrating a procedure for setting a "character house
presentation on/off" option on the character house configuration
screen of FIG. 25.
[0266] Referring to FIGS. 28 and 29. the "character house
presentation on/off" option can be set to "on" or "off." The
controller 110 displays the character houses on the character house
configuration screen in the form of tiled windows as shown in a
first screen image of FIG. 29. If one of the windows is selected
while the "character house presentation on/off" option is set to
off, the controller 110 highlights the selected window and displays
an announcement message asking whether to use the character house
presentation function as shown in a second screen image of FIG. 29.
If a user command is input for using the character house
presentation function, the controller 110 enables the character
house presentation function and displays the character house of the
selected window as shown in a third screen image of FIG. 29.
[0267] As described above, if the character house configuration
screen is loaded while the character house presentation function is
disabled, the controller 110 displays an announcement message
asking whether to use the character house presentation function and
then enables the character house presentation function according to
the user command for using the character house presentation
function.
[0268] FIG. 30 is a state diagram illustrating transitions among
the character house editing mode and specific item editing modes of
FIG. 25, and FIG. 31 shows a screen image illustrating an example
character house editing screen in the character house editing mode
of FIG. 25.
[0269] Referring to FIGS. 30 and 31, the character house editing
screen is displayed when a "new character house" option or "edit
character house" option is selected on the character house
configuration screen. The character house editing screen provides a
menu with options for editing the background and item (1.
Background, 2. New item). When the "new character house" option is
selected, a selected background is basically presented without
items. The character house editing screen displays a key input
guide as shown in FIG. 31. If a menu option is selected, the menu
with options for editing the background and items is displayed.
[0270] FIG. 32 is a state diagram illustrating transitions between
the character house editing mode and the background selection mode
of FIG. 25. FIG. 33 shows screen images illustrating a procedure
for selecting a background on the background selection mode of FIG.
32.
[0271] Referring to FIGS. 32 and 33, if a "background" option is
selected on the character house editing screen, the controller 110
controls to present the background selection screen in a background
selection mode 830 as shown in a first screen image of FIG. 33. The
background selection screen enables navigating background images
using direction keys (navigation keys). If a selection key (OK key)
is input while one of the background images is presented, the
controller 110 returns to display the character house editing mode
820 and displays the selected background image as the background on
the screen of the display unit 143 as shown in a second screen
image of FIG. 33.
[0272] As described above, if the "background" option is selected
on a character house editing screen, the controller 110 presents
available background images stored in the mobile phone with titles
of the background images. The background images are individually
shown according to the navigation key manipulation. If a selection
key is input while a background image is displayed, the background
image is stored as the background and the character house editing
screens are redisplayed.
[0273] FIG. 34 is a state diagram illustrating transitions between
the character house editing mode and item editing modes of FIG. 25.
FIGS. 35a to 35d illustrate screen images illustrating a procedure
for adding an item to the character house in the character house
editing mode of FIG. 34.
[0274] Referring to FIGS. 34 and 35a to 35b, if a "new item" option
is selected from option menu on the character house editing screen,
as shown in FIG. 35a, in the character house editing mode 820, the
controller 110 displays a new item selection screen in a new item
selection mode 860 as shown in FIG. 35b. The new item selection
screen enables navigating item images using direction keys
(navigation keys). Each item image is presented with a title of the
item image.
[0275] If a selection key (OK key) is input during the presentation
of one of the item images, the controller 110 displays an item
positioning screen, as shown in FIG. 35c, in an item positioning
mode 870. The item positioning screen is partitioned to define a
plurality of sections on the screen of the display unit 143 and
each section is assigned a unique section number. The item
positioning screen facilitates the movement of an item from one
place to another with reference to the section number. If a section
number is input while the item positioning screen is displayed, the
controller 110 displays a selected item editing screen, in which
the selected item is positioned at the section having the input
section number as shown in FIG. 35d, in a selected item editing
mode 850. The selected item editing screen facilitates fine
adjustment of the selection item position by manipulating the
navigation keys.
[0276] The selected item editing screen also provides an
item-fixing key (for example, # key) for fixing the selected item
at the adjusted position and a fast move key (for example * key)
for quickly moving the item to another section in the item
positioning mode 870. If an additional item is selected from the
option menu of the selected item editing mode 850, the controller
110 displays the new item selection screen of the new item
selection mode 860 so as to allow selecting another item. If the
fast move key is input while the selected item editing screen is
presented, the controller 110 performs displays the item position
screen of the item position mode 870 so as to move the item to
another section.
[0277] As described above, the selected item editing mode 850
facilitates fine adjustment of the position of the selected item on
the selected item editing screen. The items can be overlapped each
other. The overlapped presentation of the items can be performed
with a "send to front," "send to back," and "message adjustment
options displayed when an option key is input. If the "send to
front" (a "volume up" key) or the "send to back" (a "volume down"
key), the controller 110 places the item on or beneath the item in
the selected section.
[0278] If a fast move key (* key) is selected while the mobile
phone is in the selected item editing mode 850, the mobile phone
enters the item positioning mode 870. If the item-fixing key is
selected, the mobile phone enters an edited item shortcut selection
mode 840. If an "add item" option is selected, the mobile phone
enters the new item selection mode 860. If a "delete item" option
is selected, the mobile phone enters the character house editing
mode 820. If a "background" option is selected, the mobile phone
enters the background selection mode 830. The transition routes
from the selected item editing mode are summarized as shown in
table 5. TABLE-US-00005 TABLE 5 Functions in screen Fine position
adjustment function (navigation key) having a selected Item
overlapping display function (option >> item to be edited
send to front, send to back, background)- background selection
screen
[0279] FIG. 36 is a state diagram illustrating transitions between
the character house editing mode and the selected item editing mode
via the edited item shortcut selection mode of FIG. 25, and FIGS.
37a to 37c show screen images illustrating a procedure for
assigning shortcut key in the edited item shortcut selection mode
of FIG. 36.
[0280] Referring to FIGS. 36 and 37, an item to be edited is
selected on the edited item shortcut selection screen in the edited
item shortcut selection mode 840. The edited item shortcut
selection mode 840 is transited by from the character house editing
mode 820 represented by the character house editing screen as shown
in FIG. 37a and the selected item editing mode 850 represented by
the selected item editing screen as shown in FIG. 37b. If the
character house editing screen is open, the controller 110 displays
the items included in the character house as shown in FIG. 37b. A
character house can be provided with maximal 10 items that are
numbered from 0 to 9. Accordingly, if an item fixing key (#) is
input while the character house editing screen is displayed, the
controller 110 displays the edited item shortcut selection screen
to assign the section numbers to the items. If a section number of
an item is in the edited item shortcut selection screen, the
controller 110 displays the selected item editing screen to display
the selected item in a bracket as shown in FIG. 37c. The selected
item in the bracket can be finely adjusted in position. If the fast
move key (*) is selected to quickly move the selected item at other
section, the controller 110 displays the item positioning screen
defining the sections.
[0281] As described above, the screen image presentation method for
a mobile phone according to an exemplary embodiment of the present
invention facilitates the design of a character house that is used
as a screen image in a specific operation mode of the mobile phone.
The character house can be edited by adding, deleting, and changing
positions of the background and items.
[0282] The screen image presentation method provides the mobile
phone with the character house editing mode 820 in which the
character houses stored in the mobile phone can be selected,
deleted, and edited. The mobile phone can switch between the
character house-related editing modes for creating a new character
house.
[0283] The background selection mode 830 is represented by the
background selection screen which is accessed from the selected
item editing screen of the selected item editing mode 850 or by
selecting the "background" option in the menu provided in the
selected item editing mode 850. The controller 110 presents the
background image together with the title of the background image as
shown in the first screen image of FIG. 33. The background image
slides according to the direction key (left and right direction
keys) manipulation to make the selection of one of the background
images as the background of the character house possible.
[0284] The new item selection mode 860 facilitates the addition of
a new item to the character house and transitions to the item
positioning mode 870 and the selected item editing mode 850. If the
"new item" option is selected in the menu provided in the character
house editing mode 820, the controller 110 displays the new item
selection screen, as shown in FIG. 35b, representing the new item
selection mode 860. The controller 110 presents the item images
together with the title of the items. The new item selection screen
facilitates the navigation of item images using direction keys
(left and right direction keys). If one of the items is selected
from the new item selection screen, the controller 110 displays the
item positioning screen, as shown in FIG. 35c, representing the
item positioning mode 870. The item positioning screen 870 is
partitioned so as to define several sections as shown in FIG. 35c.
In this exemplary embodiment of the present invention, the screen
is divided into 9 sections and assigned unique section numbers,
respectively. If a section number is selected while the item
positioning screen, the controller 110 displays the selected item
editing screen as shown in FIG. 35d. The selected item editing
screen facilitates fine adjustment of the position of the selected
item. If the addition of another item is required, the controller
110 displays the new item selection screen again, i.e. enters the
new item selection mode 860.
[0285] As described above, the controller 110 presents the items
stored in the mobile phone if there is a command input for adding
an item to the character house. The screen is divided into several
sections for quickly positioning an item on the screen and the item
is finely adjusted in a section.
[0286] The controller 110 operates the edited item shortcut mode
840 and the selected item editing mode 850 to facilitate a change
in the position of the item in the character house. That is, the
controller 110 assigns the item numbers to the items presented with
the character house on the edited item shortcut screen and
highlights the item selected by the item number as a target item.
The target item is edited in the selected item editing mode 850 and
the item positioning mode 870.
[0287] As described above, the controller 110 facilitates the
editing of the character house at the steps 719 and 721 of FIG. 24.
In these steps, the controller 110 also determines whether to
present the character quotients and the state and emotion of the
character. Also, the controller 110 facilitates the setting of
whether to activate the greeting behavior and random greeting event
of the character.
[0288] The controller 110 also allows editing the background and
items and adjusting the position of the items at steps 715 and 717
of FIG. 24. As described above, the character house includes the
character, background, and items. The character house can be
created in the mobile phone and store at least one character house.
The controller 110 of the mobile phone facilitates the selection of
one of the character houses in order for a selected character house
to be used as a screen image in a specific operation mode of the
mobile phone. The background and items constituting the character
house can be edited. The background of the character house can be
deleted, added, or modified in the character house editing mode
under the control of the controller 110. Also, the items of the
character house can be deleted or added, and repositioned on the
screen in the character house editing mode.
[0289] If the character house is created with the character,
background, and items, and the character house presentation
function is activated, the controller 110 of the mobile phone
performs the operations for presenting the character house at steps
S711 and S713 of FIG. 24. The character house presentation is
performed in the idle mode of the mobile phone at the same or a
different time with a start of the idle mode. That is, the
character house presentation can be set to start at the time when
the display unit 143 is turned on by opening the clam shell phone
or sliding up open the slide phone, or after completing an event of
the mobile phone. The character house presentation also can be set
to start when a special event occurs. For example, the character
house can be presented when verifying an incoming message, writing
an outgoing message, or verifying a phone book.
[0290] If the message or information is displayed with the
character house, the controller 110 controls such that the
character house is presented in a transparency higher than that of
the message or information, using the alpha blending method. For,
example, the character house is presented in 50% transparency while
the message or information in 0% transparency. Also, the character
house can be implemented with a display window such the message or
information is presented in the display window of the character
house. The character house is preferably presented during a preset
time to reduce power consumption of the mobile phone.
[0291] If the mobile phone is in the character house presentation
mode (S711), the controller 110 controls to present the character
house (S713).
[0292] The character house is presented while the controller 110
performs the steps S313 to 315 of FIG. 14. That is, the controller
110 verifies the character house configuration information (S313)
and determines whether the character house function is turned on
(S315). If the character house function is turned on, the
controller 110 verifies the character house information (S319) and
then presents the character house in accordance with the character
house information. As described above, the character house includes
a background, items, and character. The controller 110 presents the
background of the character house according to the procedure of
FIG. 15. If the background is a time-varying background, the
controller 110 verifies the current time and presents a background
image corresponding to the current time. The controller 110
presents the items of the character house according to the
procedure of FIG. 16. The controller 110 determines whether the
character quotients presentation function is turned on. If the
character quotients presentation function is turned on, the
controller presents the character quotients. The controller 110
analyzes the types of items and presents the items according to the
procedures corresponding to the types of the items. If the item is
a motion item, the motion item is presented according to the
procedure of FIG. 17. If the item is a function item, the function
item is presented according to the procedure of FIG. 18. The
presentation of the character is performed according to the
procedure of FIG. 19. If the greeting behavior function is set to
"on," the controller 110 verifies the current time and controls the
character to show the greeting behavior if the current time is in a
greeting time zone of a day. If a greeting event occurs while the
character shows the greeting behavior, the controller 110 controls
to preset the greeting event according to the procedure of FIG. 21.
The behavior of the character house is performed according to the
procedure of FIG. 22 and a character movement is performed
according to the procedure of FIG. 23.
[0293] The controller 110 facilitates the setting of the character
house functions, editing the character house, and presenting the
character house according to the preset functions, according to the
procedure of FIG. 24.
[0294] As described above, the character house of the present
invention is a virtual space in which an artificial intelligent
character resides, talks, and moves such that a user recognize a
state of a mobile by means of the character house. The character
house facilitates learning functions of the mobile phone and
attracts the user's interest to the mobile phone. The character
house of an exemplary embodiment of the present invention is
presented with a character, background image, items, and animation.
The character can present a sound effect and text expression
reflecting its condition and emotion according to a character
scenario, moves around the character house, and shows an animation.
The character can express the state and emotion of the mobile phone
in various manners so as to make the character house dynamic. The
background is used to lively express the character house by
changing the background image according to a progress of the time,
whereby the user can guess the current time from the character
house. The items includes decoration items, motion items, and
function items. The function items show system information on the
mobile phone to improve user convenience. The decoration items
increase the aesthetic sense of the character house. The motion
items move around in the character so as to make the character
house dynamic.
[0295] The character house can be implemented with a special
expression method in which the character greets the user of the
mobile phone with an appropriate sound, text, and behavior at
predetermined time zones of a day, to make the mobile phone
user-friendly. Also, the character house can be implemented such
that the character randomly gives a gift to the user, resulting in
enjoyment of the user.
[0296] While the present invention has bee shown and described with
reference to certain exemplary embodiments thereof, it will be
understood by those skilled in the art that various changes in form
and details may be made therein without departing from the spirit
and scope of the invention as defined by the appended claims and
their equivalents.
* * * * *