U.S. patent application number 15/405265 was filed with the patent office on 2017-07-06 for mobile terminal and method of controlling a mobile terminal to display image upon receiving proximity touch input.
This patent application is currently assigned to LG ELECTRONICS INC.. The applicant listed for this patent is LG ELECTRONICS INC.. Invention is credited to Yeerang YUN.
Application Number | 20170195473 15/405265 |
Document ID | / |
Family ID | 45218136 |
Filed Date | 2017-07-06 |
United States Patent
Application |
20170195473 |
Kind Code |
A1 |
YUN; Yeerang |
July 6, 2017 |
MOBILE TERMINAL AND METHOD OF CONTROLLING A MOBILE TERMINAL TO
DISPLAY IMAGE UPON RECEIVING PROXIMITY TOUCH INPUT
Abstract
A mobile terminal and a method of controlling a mobile terminal
are provided. According to one embodiment, a mobile terminal
includes: a touch screen for displaying and receiving information;
and a controller for: controlling the touch screen to display a
lock screen upon the mobile terminal entering a lock screen state;
and controlling the touch screen to display a preview image of each
of at least one application executed before entering the lock
screen state upon receiving a proximity touch input on the
displayed lock screen.
Inventors: |
YUN; Yeerang; (Seoul,
KR) |
|
Applicant: |
Name |
City |
State |
Country |
Type |
LG ELECTRONICS INC. |
Seoul |
|
KR |
|
|
Assignee: |
LG ELECTRONICS INC.
Seoul
KR
|
Family ID: |
45218136 |
Appl. No.: |
15/405265 |
Filed: |
January 12, 2017 |
Related U.S. Patent Documents
|
|
|
|
|
|
Application
Number |
Filing Date |
Patent Number |
|
|
13029070 |
Feb 16, 2011 |
|
|
|
15405265 |
|
|
|
|
Current U.S.
Class: |
1/1 |
Current CPC
Class: |
G06F 3/044 20130101;
H04M 1/67 20130101; G06F 3/0414 20130101; G06F 3/0481 20130101;
H04M 1/673 20130101; H04M 1/72519 20130101; H04M 2250/22 20130101;
G06F 3/0416 20130101; H04M 2250/60 20130101; G06F 3/04842 20130101;
G06F 3/0488 20130101; G06F 3/04886 20130101; G06F 2203/04101
20130101; G06F 3/04883 20130101; G06F 3/0483 20130101; H04M 1/72583
20130101 |
International
Class: |
H04M 1/725 20060101
H04M001/725; G06F 3/044 20060101 G06F003/044; G06F 3/0488 20060101
G06F003/0488; G06F 3/0481 20060101 G06F003/0481; G06F 3/041
20060101 G06F003/041; G06F 3/0484 20060101 G06F003/0484 |
Foreign Application Data
Date |
Code |
Application Number |
Sep 30, 2010 |
KR |
10-2010-0095071 |
Claims
1. A mobile terminal comprising: a touchscreen; and a controller
configured to: cause the touchscreen to display a screen including
a plurality of items; cause the touchscreen to display a detailed
screen when a touch received at a first item of the plurality of
items is a first touch type, wherein the detailed screen includes
detailed amount of data associated with the first item; cause the
touchscreen to display preview information when a touch received at
a first item of the plurality of items is a second touch type,
wherein the preview information is only a partial amount of data
associated with the first item, wherein the partial amount of data
is less than the detailed amount of data of the detailed screen,
wherein an area of the preview information is less than an area of
the screen including the plurality of items, and wherein the first
touch type is different from the second touch type; cause the
touchscreen to maintain the displaying of the preview information
while the touch of the second touch type is continued to be
received; and cause the touchscreen to stop the displaying of the
preview information when the touch of the second touch type is
released.
2. The mobile terminal of claim 1, wherein the controller is
further configured to: change a display characteristic of the
screen while the touch of the second touch type is continued to be
received.
3. The mobile terminal of claim 2, wherein the controller is
further configured to: change the display characteristic of the
screen back to an initial display characteristic when the touch of
the second touch type is released.
4. The mobile terminal of claim 1, further comprising: a touch
sensor configured to provide data according to a touch received at
the touchscreen; wherein the controller is further configured to:
recognize touch pressure of the touch of the first touch type and
touch pressure of the touch of the second touch type according to
the data provided by the touch sensor, wherein the touch pressure
of the touch of the first touch type and the touch pressure of the
touch of the second touch type are different.
5. The mobile terminal of claim 1, wherein location of the touch of
the first touch type and location of the touch of the second touch
type are substantially same.
6. The mobile terminal of claim 1, wherein the touch of the second
touch type comprises a touch and drag input.
7. The mobile terminal of claim 1, wherein the preview information
is displayed in a pop-up window that obscures a portion of the
screen including the plurality of items.
8. The mobile terminal of claim 7, wherein the preview information
in the pop-up window comprises more detailed information than
information displayed in screen including the first item of the
plurality of items.
9. The mobile terminal of claim 7, wherein a size of the pop-up
window is larger than a displayed size of first item of the
plurality of items.
10. The mobile terminal of claim 7, wherein the pop-up window is
displayed using a portion of an execution screen of the first item
of the plurality of items.
11. The mobile terminal of claim 7, wherein a size of the pop-up
window is variable according to a type, or quantity, of the preview
information.
12. The mobile terminal of claim 7, wherein the type of the preview
information comprises at least one of text or image.
13. The mobile terminal of claim 1, further comprising: a wireless
communication unit, wherein each of the plurality of items comprise
a task that was performed previously using the wireless
communication unit, wherein the task include at least one call task
or a message task.
14. The mobile terminal of claim 1, wherein the first touch type is
a proximity touch received at the first item, and the second touch
type is a contact touch received at the first item.
15. A mobile terminal comprising: a touchscreen; and a controller
configured to: cause the touchscreen to display a plurality of
items; cause the touchscreen to display a detailed screen when a
touch received at a first item of the plurality of items is a first
touch type, wherein the detailed screen includes a first amount of
data associated with the first item; cause the touchscreen to
display a preview when a touch received at a first item of the
plurality of items is a second touch type, wherein the preview is a
partial amount of data associated with the first item, wherein the
partial amount of data is less than the first amount of data of the
detailed screen, wherein a displayed area of the preview is less
than an area within which the plurality of items are displayed, and
wherein the first touch type is different from the second touch
type; cause the touchscreen to maintain the displaying of the
preview while the touch of the second touch type is continued to be
received; and cause the touchscreen to stop the displaying of the
preview when the touch of the second touch type is no longer
received.
Description
[0001] This application is a continuation of U.S. patent
application Ser. No. 13/029,070, filed on Feb. 16, 2011, currently
pending, which claims the benefit of earlier filing date and right
of priority to Korean Patent Application No. 10-2010-0095071 filed
on Sep. 30, 2010, the contents of which are all hereby incorporated
by reference herein in their entirety.
FIELD OF THE INVENTION
[0002] Embodiments of the present invention relate to a mobile
terminal and a method of controlling a mobile terminal.
RELATED ART
[0003] Terminals can be divided into mobile terminals and
stationary terminals. The mobile terminals can be classified as
handheld terminals or vehicle mounted terminals according to
whether users can personally carry the terminals.
[0004] In view of various types of services that are provided by
mobile terminals, an improvement in the structure and/or software
of mobile terminals is considered.
SUMMARY
[0005] Embodiments of the present invention relate to a mobile
terminal and a method of controlling a mobile terminal, and more
particularly to a mobile terminal (and a method of controlling a
mobile terminal) for allowing a user to check required preview
information and history information without influence over a
current state of the mobile terminal such that user convenience may
be enhanced.
[0006] According to one embodiment, a mobile terminal includes: a
touch screen for displaying and receiving information; and a
controller for: controlling the touch screen to display a lock
screen upon the mobile terminal entering a lock screen state; and
controlling the touch screen to display a preview image of each of
at least one application executed before entering the lock screen
state upon receiving a proximity touch input on the displayed lock
screen.
[0007] According to one embodiment, a mobile terminal includes: a
touch screen for displaying and receiving information; and a
controller for controlling the touch screen to display history
information or preview information corresponding to a specific
region of the touch screen upon receiving a proximity touch input
of the specific region.
[0008] According to one embodiment, a method of controlling a
mobile terminal includes: controlling a touch screen to display a
lock screen upon the mobile terminal entering a lock screen state;
receiving a proximity touch input via the displayed lock screen;
and controlling the touch screen to display a preview image of at
least one application executed before entering the lock screen
state on the displayed lock screen.
BRIEF DESCRIPTION OF THE DRAWINGS
[0009] The accompanying drawings, which are included to provide a
further understanding of described embodiments of the present
invention and are incorporated in and constitute a part of this
specification, illustrate embodiments of the present invention and
together with the description serve to explain aspects and features
of the present invention.
[0010] FIG. 1 is a block diagram illustrating a configuration of a
mobile terminal according to embodiments of the present
invention;
[0011] FIGS. 2 and 3 are perspective views of a mobile terminal
according to embodiments of the present invention.
[0012] FIG. 4 is a conceptual diagram for illustrating a proximity
depth of a proximity sensor;
[0013] FIG. 5 is a flowchart illustrating a method of controlling a
mobile terminal according to an embodiment of the present
invention;
[0014] FIGS. 6A, 6B and 7A-7C are views of screen displays for
illustrating examples of displaying preview images of at least one
application that was executed before entering a lock screen state
based on a proximity touch input;
[0015] FIGS. 8 to 12 are views of screen displays illustrating
examples of preview images of an application in the lock screen
state;
[0016] FIGS. 13A-13C, 14A, 14B, 15A-15C, 16A-16C, 17A-17D, 18A and
18B are views of screen displays illustrating examples of ending
the lock screen when preview images of an application are displayed
on a locked screen;
[0017] FIG. 19 is a flowchart illustrating a method of controlling
a mobile terminal according to another embodiment of the present
invention;
[0018] FIGS. 20A, 20B, 21A-21D, 22A-22C, 23A-23C and 24A-24D are
views of screen displays illustrating examples of displaying
preview information in a mobile terminal;
[0019] FIG. 25 is a flow chart illustrating a method of controlling
a mobile terminal according to still another embodiment of the
present invention; and
[0020] FIGS. 26A, 26B, 27A, 27B, 28A, 28B, 29A-29C, 30A, 30B and
31A-31D are views illustrating of screen displays illustrating
examples of history information.
DETAILED DESCRIPTION
[0021] Embodiments of the present invention will now be described
more fully with reference to the accompanying drawings, in which
certain embodiments of the invention are illustrated. The invention
may, however, be embodied in many different forms and should not be
construed as being limited to the embodiments set forth herein;
rather, these embodiments are described and/or illustrated so that
this disclosure will be more thorough and complete, and will more
fully convey the aspects of the invention to those skilled in the
art.
[0022] Hereinafter, a mobile terminal according to embodiments of
the present invention will be described below in more detail with
reference to the accompanying drawings. In the following
description, the suffixes "module" and "unit" are used in reference
to components of the mobile terminal for convenience of description
and do not have meanings or functions different from each
other.
[0023] The mobile terminals described herein may include a cellular
phone, a smart phone, a laptop computer, a digital broadcasting
terminal, a personal digital assistant (PDA), a portable multimedia
player (PMP), and a navigation system.
[0024] FIG. 1 is a block diagram of a mobile terminal 100 according
to an embodiment of the present invention. It is understood that
other embodiments, configurations and arrangements may also be
provided. With reference to FIG. 1, the mobile terminal 100 may
include a wireless communication unit 110, an audio/video (A/V)
input unit 120, a user input unit 130, a sensing unit 140, an
output unit 150, a memory 160, an interface unit 170, a controller
180, and a power supply 190. Not all of the components shown in
FIG. 1 are essential, and the number of components included in the
mobile terminal 100 may be varied. The components of the mobile
terminal 100, as illustrated with reference to FIG. 1 will now be
described.
[0025] The wireless communication unit 110 may include at least one
module that enables wireless communication between the mobile
terminal 100 and a wireless communication system or between the
mobile terminal 100 and a network in which the mobile terminal 100
is located. For example, the wireless communication unit 110 may
include a broadcast receiving module 111, a mobile communication
module 112, a wireless Internet module 113, a local area (or
short-range) communication module 114, and a location information
(or position-location) module 115.
[0026] The broadcast receiving module 111 may receive broadcasting
signals and/or broadcasting related information from an external
broadcasting management server through a broadcasting channel. The
broadcasting channel may include a satellite channel and a
terrestrial channel, and the broadcasting management server may be
a server that generates and transmits broadcasting signals and/or
broadcasting related information or a server that receives
previously created broadcasting signals and/or broadcasting related
information and transmits the broadcasting signals and/or
broadcasting related information to a terminal.
[0027] The broadcasting signals may include not only TV
broadcasting signals, wireless broadcasting signals, and data
broadcasting signals, but also signals in the form of a combination
of a TV broadcasting signal and a radio broadcasting signal. The
broadcasting related information may be information on a
broadcasting channel, a broadcasting program or a broadcasting
service provider, and may be provided even through a mobile
communication network. In the latter case, the broadcasting related
information may be received by the mobile communication module
112.
[0028] The broadcasting related information may exist in any of
various forms. For example, the broadcasting related information
may exist in the form of an electronic program guide (EPG) of a
digital multimedia broadcasting (DMB) system or in the form of an
electronic service guide (ESG) of a digital video
broadcast-handheld (DVB-H) system.
[0029] The broadcast receiving module 111 may receive broadcasting
signals using various broadcasting systems. More particularly, the
broadcast receiving module 111 may receive digital broadcasting
signals using digital broadcasting systems such as a digital
multimedia broadcasting-terrestrial (DMB-T) system, a digital
multimedia broadcasting-satellite (DMB-S) system, a media forward
link only (MediaFLO.TM.) system, a DVB-H system, and an integrated
services digital broadcast-terrestrial (ISDB-T) system. The
broadcast receiving module 111 may receive signals from
broadcasting systems providing broadcasting signals other than the
above-described digital broadcasting systems.
[0030] The broadcasting signals and/or broadcasting related
information received through the broadcast receiving module 111 may
be stored in the memory 160. The mobile communication module 112
may transmit/receive a wireless signal to/from at least one of a
base station, an external terminal and a server on a mobile
communication network. The wireless signal may include a voice call
signal, a video call signal or data in various forms according to
the transmission and reception of text/multimedia messages.
[0031] The wireless Internet module 113 may correspond to a module
for wireless Internet access and may be included in the mobile
terminal 100 or may be externally attached to the mobile terminal
100. Wireless LAN (WLAN or Wi-Fi), wireless broadband (Wibro.TM.)
world interoperability for microwave access (Wimax.TM.), high speed
downlink packet access (HSDPA) and other technologies may be used
as a wireless Internet technique.
[0032] The local area communication module 114 may correspond to a
module for local area communication. Further, Bluetooth.TM., radio
frequency identification (RFID), infrared data association (IrDA),
ultra wideband (UWB) and/or ZigBee.TM. may be used as a local area
communication technique.
[0033] The position-location module 115 may confirm or obtain the
position of the mobile terminal 100. The position-location module
115 may obtain position information by using a global navigation
satellite system (GNSS). The GNSS refers to a radio navigation
satellite system that revolves around the earth and transmits
reference signals to predetermined types of radio navigation
receivers such that the radio navigation receivers may determine
their positions on the earth's surface or near the earth's surface.
The GNSS may include a global positioning system (GPS) of the
United States, Galileo of Europe, a global orbiting navigational
satellite system (GLONASS) of Russia, COMPASS of China, and a
quasi-zenith satellite system (QZSS) of Japan among others.
[0034] A global positioning system (GPS) module is one example of
the position-location module 115. The GPS module 115 may calculate
information regarding distances between one point or object and at
least three satellites and information regarding a time when the
distance information is measured and apply trigonometry to the
obtained distance information to obtain three-dimensional position
information on the point or object according to latitude, longitude
and altitude at a predetermined time. A method of calculating
position and time information using three satellites and correcting
the calculated position and time information using another
satellite may also be used. In addition, the GPS module 115 may
continuously calculate the current position in real time and
calculate velocity information using the location or position
information.
[0035] As shown in FIG. 1, the A/V input unit 120 may input an
audio signal or a video signal and include a camera 121 and a
microphone 122. The camera 121 may process image frames of still
images or moving pictures obtained by an image sensor in a video
call mode or a photographing mode. The processed image frames may
be displayed on a display module 151 which may be a touch
screen.
[0036] The image frames processed by the camera 121 may be stored
in the memory 160 or may be transmitted to an external device
through the wireless communication unit 110. The mobile terminal
100 may also include at least two cameras 121.
[0037] The microphone 122 may receive an external audio signal in a
call mode, a recording mode or a speech recognition mode and
process the received audio signal into electronic audio data. The
audio data may then be converted into a form that may be
transmitted to a mobile communication base station through the
mobile communication module 112 and output in the call mode. The
microphone 122 may employ various noise removal algorithms (or
noise canceling algorithms) for removing or reducing noise
generated when the external audio signal is received.
[0038] The user input unit 130 may receive input data required for
controlling the mobile terminal 100 from a user. The user input
unit 130 may include a keypad, a dome switch, a touch pad (e.g.,
constant voltage/capacitance), a jog wheel, and a jog switch.
[0039] The sensing unit 140 may sense a current state of the mobile
terminal 100, such as an open/closed state of the mobile terminal
100, a position of the mobile terminal 100, whether a user touches
the mobile terminal 100, a direction of the mobile terminal 100,
and acceleration/deceleration of the mobile terminal 100, and
generate a sensing signal required for controlling the mobile
terminal 100. For example, if the mobile terminal 100 is a slide
phone, the sensing unit 140 may sense whether the slide phone is
opened or closed. Further, the sensing unit 140 may sense whether
the power supply 190 supplies power and/or whether the interface
unit 170 is connected to an external device. The sensing unit 140
may also include a proximity sensor 141.
[0040] The output unit 150 may generate visual, auditory and/or
tactile output and may include the display module 151, an audio
output module 152, an alarm unit 153 and a haptic module 154. The
display module 151 may display information processed by the mobile
terminal 100. The display module 151 may display a user interface
(UI) or a graphic user interface (GUI) related to a voice call when
the mobile terminal 100 is in the call mode. The display module 151
may also display a captured and/or received image and a UI or a GUI
when the mobile terminal 100 is in the video call mode or the
photographing mode.
[0041] In addition, the display module 151 may include at least a
liquid crystal display, a thin film transistor liquid crystal
display, an organic light-emitting diode display, a flexible
display or a three-dimensional display. Some of these displays may
be of a transparent type or a light transmissive type. That is, the
display module 151 may include a transparent display.
[0042] The transparent display may include a transparent liquid
crystal display. The rear of the display module 151 may include a
light transmissive type display. Accordingly, a user may be able to
see an object located behind the body of the mobile terminal 100
through the transparent portion of the display unit 151 on the body
of the mobile terminal 100.
[0043] The mobile terminal 100 may also include at least two
display modules 151. For example, the mobile terminal 100 may
include a plurality of display modules 151 that are arranged on a
single face of the mobile terminal 100 and spaced apart from each
other at a predetermined distance or that are integrated together.
The plurality of display modules 151 may also be arranged on
different sides of the mobile terminal 100.
[0044] Further, when the display module 151 and a touch-sensing
sensor (hereafter referred to as a touch sensor) form a layered
structure that is referred to as a touch screen, the display module
151 may be used as an input device in addition to an output device.
The touch sensor may be in the form of a touch film, a touch sheet,
or a touch pad, for example.
[0045] The touch sensor may convert a variation in pressure,
applied to a specific portion of the display module 151, or a
variation in capacitance, generated at a specific portion of the
display module 151, into an electric input signal. The touch sensor
may sense pressure, position, and an area (or size) of the
touch.
[0046] When the user applies a touch input to the touch sensor, a
signal corresponding to the touch input may be transmitted to a
touch controller. The touch controller may then process the signal
and transmit data corresponding to the processed signal to the
controller 180. Accordingly, the controller 180 may detect a
touched portion of the display module 151.
[0047] The proximity sensor 141 of the sensing unit 140 may be
located in an internal region of the mobile terminal 100,
surrounded by the touch screen, or near the touch screen. The
proximity sensor 141 may sense the presence of an object
approaching a predetermined sensing face or an object located near
the proximity sensor using an electromagnetic force or infrared
rays without mechanical contact. The proximity sensor 141 may have
a lifetime longer than a contact sensor and may thus be more
appropriate for use in the mobile terminal 100.
[0048] The proximity sensor 141 may include a transmission type
photoelectric sensor, a direct reflection type photoelectric
sensor, a mirror reflection type photoelectric sensor, a
high-frequency oscillating proximity sensor, a capacitive proximity
sensor, a magnetic proximity sensor, and/or an infrared proximity
sensor. Meanwhile, a capacitive touch screen may be constructed
such that proximity of a pointer is detected through a variation in
an electric field according to the proximity of the pointer.
Therefore, the touch screen (touch sensor) may be considered as a
proximity sensor 141 if the touch screen is implemented as the
capacitive touch screen.
[0049] For the convenience of description, an action in which a
pointer approaches the touch screen without actually touching the
touch screen may be referred to as a proximity touch, and an action
in which the pointer is brought into contact with the touch screen
may be referred to as a contact touch. The proximity touch point of
the pointer on the touch screen may correspond to a point of the
touch screen at which the pointer is perpendicular to the touch
screen.
[0050] The proximity sensor 141 may sense the proximity touch and a
proximity touch pattern (e.g., a proximity touch distance, a
proximity touch direction, a proximity touch velocity, a proximity
touch time, a proximity touch position, a proximity touch moving
state). Information corresponding to the sensed proximity touch
action and proximity touch pattern may then be displayed on the
touch screen.
[0051] The audio output module 152 may output audio data received
from the wireless communication unit 110 or stored in the memory
160 in a call signal receiving mode, a call mode or a recording
mode, a speech recognition mode and a broadcast receiving mode. The
audio output module 152 may output audio signals related to
functions performed in the mobile terminal 100, such as a call
signal incoming tone and a message incoming tone. The audio output
module 152 may include a receiver, a speaker, and/or a buzzer. The
audio output module 152 may output sounds through an earphone jack.
The user may listen to the sounds by connecting an earphone to the
earphone jack.
[0052] The alarm unit 153 may output a signal indicating generation
(or occurrence) of an event of the mobile terminal 100. For
example, alarms may be generated when a call signal or a message is
received and when a key signal or a touch is input. The alarm unit
153 may also output signals different from video signals or audio
signals, for example, a signal indicating generation of an event
through vibration. The video signals or the audio signals may also
be output through the display module 151 or the audio output module
152.
[0053] The haptic module 154 may generate various haptic effects
that the user may feel. One of the haptic effects is vibration. The
intensity and/or pattern of a vibration generated by the haptic
module 154 may also be controlled. For example, different
vibrations may be combined with each other and output or may be
sequentially output.
[0054] The haptic module 154 may generate a variety of haptic
effects including an effect attributed to an arrangement of pins
vertically moving against a contact skin surface, an effect
attributed to a jet force or a suctioning force of air through a
jet hole or a suction hole, an effect attributed to a rubbing of
the skin, an effect attributed to contact with an electrode, an
effect of stimulus attributed to an electrostatic force, and an
effect attributed to a reproduction of cold and warmth using an
element for absorbing or radiating heat in addition to
vibrations.
[0055] The haptic module 154 may not only transmit haptic effects
through direct contact but may also allow the user to feel haptic
effects through the user's fingers or arms. The mobile terminal 100
may also include a plurality of haptic modules 154.
[0056] The memory 160 may store a program for operating the
controller 180 and temporarily store input/output data such as a
phone book, messages, still images, and/or moving pictures. The
memory 160 may also store data regarding various patterns of
vibrations and sounds that are output from when a touch input is
applied to the touch screen.
[0057] The memory 160 may include at least a flash memory, a hard
disk type memory, a multimedia card micro type memory, a card type
memory such as SD or XD memory, a random access memory (RAM), a
static RAM (SRAM), a read-only memory (ROM), an electrically
erasable programmable ROM (EEPROM), a programmable ROM (PROM)
magnetic memory, a magnetic disk, or an optical disk. The mobile
terminal 100 may also operate in association with a web storage
performing the storage function of the memory 160 on the
Internet.
[0058] The interface unit 170 may serve as a path to external
devices connected to the mobile terminal 100. The interface unit
170 may receive data or power from the external devices, transmit
the data or power to internal components of the mobile terminal
100, or transmit data of the mobile terminal 100 to the external
devices. For example, the interface unit 170 may include a
wired/wireless headset port, an external charger port, a
wired/wireless data port, a memory card port, a port for connecting
a device having a user identification module, an audio I/O port, a
video I/O port, and/or an earphone port.
[0059] The interface unit 170 may also interface with a user
identification module that is a chip that stores information for
authenticating authority to use the mobile terminal 100. For
example, the user identification module may be a user identity
module (UIM), a subscriber identity module (SIM) and a universal
subscriber identify module (USIM). An identification device
including the user identification module may also be manufactured
in the form of a smart card. Accordingly, the identification device
may be connected to the mobile terminal 100 through a port of the
interface unit 170.
[0060] The interface unit 170 may also be a path through which
power from an external cradle is provided to the mobile terminal
100 when the mobile terminal 100 is connected to the external
cradle or a path through which various command signals input by the
user through the cradle are provided to the mobile terminal 100.
The various command signals or power input from the cradle may be
used as signals for checking whether the mobile terminal 100 is
correctly settled (or loaded) in the cradle.
[0061] The controller 180 may control overall operations of the
mobile terminal 100. For example, the controller 180 may control
and process voice communication, data communication and/or a video
call. The controller 180 may also include a multimedia module 181
for playing a multimedia file. The multimedia module 181 may be
included in the controller 180 as shown in FIG. 1 or may be
separated from the controller 180.
[0062] The controller 180 may perform a pattern recognition process
of recognizing handwriting input or picture-drawing input applied
to the touch screen as characters or images. The power supply 190
may receive external power and internal power and provide power
required for operating the components of the mobile terminal 100
under the control of the controller 180.
[0063] According to a hardware implementation, embodiments of the
present invention may be implemented using at least application
specific integrated circuits (ASICs), digital signal processors
(DSPs), digital signal processing devices (DSPDs), programmable
logic devices (PLDs), field programmable gate arrays (FPGAs),
processors, controllers, micro-controllers, microprocessors, and/or
electrical units for executing functions. The embodiments may be
implemented using the controller 180.
[0064] According to a software implementation, embodiments
including procedures or functions may be implemented using a
separate software module executing at least one function or
operation. Software code may be implemented according to a software
application written in an appropriate software language. The
software codes may be stored in the memory 160 and executed by the
controller 180.
[0065] FIG. 2 is a front perspective view of a mobile terminal or a
handheld terminal 100 according to an embodiment of the present
invention. The handheld terminal 100 has a bar type terminal body.
However, embodiments of the present invention are not limited to a
bar type terminal and may include various types of terminals such
as slide type, folder type, swing type and swivel type terminals
having at least two bodies coupled such that they can move relative
to each other.
[0066] The terminal body includes a case (a casing, a housing, or a
cover) forming the exterior of the terminal 100. In the embodiment
of FIG. 2, the case may be divided into a front case 101 and a rear
case 102. Various electronic components are arranged in the space
(volume) defined between the front case 101 and the rear case 102.
At least one middle case may be additionally arranged between the
front case 101 and the rear case 102. The cases may be made of
plastics through injection molding or made of a metal material such
as stainless steel (STS) or titanium (Ti).
[0067] The display module 151, the audio output module 152, the
camera 121, user input unit 130 (e.g., operating units 131, 132),
the microphone 122 and the interface unit 170 may be arranged in
the terminal body, specifically, in the front case 101.
[0068] The display module 151 occupies a majority of the area of
the main face of the front case 101. The audio output module 152
and the camera 121 are arranged in a region in proximity to an end
of the display module 151, and the operating unit 131 and the
microphone 122 are located in a region in proximity to the other
end of the display module 151. The operating unit 132 and the
interface unit 170 are arranged on the lateral sides of the front
case 101 and the rear case 102.
[0069] The user input unit 130 is operated to receive commands
controlling the handheld terminal 100 and may include a plurality
of operating units 131 and 132. The first and second operating
units 131 and 132 may be referred to as manipulating portions and
may employ a tactile mechanism by which a user operates the
operating units 131 and 132 by touch.
[0070] The first and second operating units 131 and 132 may receive
various inputs. For example, the first operating unit 131 receives
commands such as `START,` END,' and `SCROLL,` and the second
operating unit 132 receives commands such as `CONTROL` (the volume
of sound output from the audio output module 152) or `CONVERT` (a
mode of the display module 151 into a touch recognition mode).
[0071] FIG. 3 is a rear perspective view of the handheld terminal
shown in FIG. 2 according to an embodiment of the present
invention.
[0072] Referring to FIG. 3, a camera 121' may be additionally
mounted at the rear side of the terminal body, that is, the rear
case 102. The camera 121' captures images along a direction
opposite to that of the camera 121 shown in FIG. 2 and may have a
pixel resolution different from that of the camera 121.
[0073] For example, according to one embodiment, the camera 121 has
a relatively low resolution suitable for capturing an image of the
face of a user and transmitting the captured image in a video call,
while the camera 121' has a relatively high resolution suitable for
capturing a photograph of a general subject. The cameras 121 and
121' may be mounted to the terminal body such that they may be
rotated or popped-up.
[0074] A flash 123 and a mirror 124 may be additionally arranged in
proximity to the camera 121'. The flash 123 lights a subject when
the camera 121' takes a picture of the subject. The mirror 124 may
be used by the user to view his/her face when the user wants to
self-photograph himself/herself using the camera 121'.
[0075] An audio output module 152' may be additionally provided on
the rear side of the terminal body. The audio output module 152'
may facilitate a stereo function in conjunction with the audio
output module 152 shown in FIG. 2 and may be used in a speaker
phone mode when the terminal is used for a voice call.
[0076] A broadcasting signal receiving antenna 116 may be
additionally attached to the side of the terminal body in addition
to an antenna for voice calls. The antenna, which may constitute a
part of the broadcast receiving module 111 shown in FIG. 1, may be
mounted in the terminal body such that the antenna may be pulled
out from (and retracted into) the terminal body.
[0077] The power supply 190 for providing power to the handheld
terminal 100 is set in the terminal body. The power supply 190 may
be provided in the terminal body or detachably installed on the
terminal body.
[0078] A touch pad 135 for sensing a touch may be additionally
attached to the rear case 102. The touch pad 135 may be of a light
transmission type similar to the display module 151. In this
configuration, if the display module 151 outputs visual information
through both of its sides (or faces), the visual information may be
viewable via the touch pad 135. The information output through both
sides of the display unit 151 may be controlled by the touch pad
135. Alternatively (or in addition), a display is additionally
attached to the touch pad 135 such that a touch screen may be
arranged in the rear case 102.
[0079] The touch pad 135 operates in connection with the display
module 151 of the front case 101. The touch pad 135 may extend
parallel to the display module 151 behind the display module 151.
The touch panel 135 may have a size equal to or smaller than the
size of the display module 151.
[0080] FIG. 4 is a conceptual diagram for explaining a proximity
depth of the proximity sensor. Referring to FIG. 4, when a pointer
such as a user's finger approaches the touch screen 151, the
proximity sensor located inside or near the touch screen senses the
approach and outputs a proximity signal. The proximity sensor may
be configured to output a proximity signal according to the
distance between the pointer approaching the touch screen and the
touch screen (referred to as a "proximity depth").
[0081] The distance, according to which the proximity signal is
output when the pointer approaches the touch screen, is referred to
as a detection distance. The proximity depth may be detected by
using a plurality of proximity sensors having different detection
distances and by comparing respective proximity signals output from
the proximity sensors. Referring to FIG. 4, proximity sensors for
sensing three proximity depths are arranged in a section of the
touch screen. Alternatively, proximity sensors suitable for sensing
less than three proximity depths or more than four proximity depths
may be arranged in the touch screen.
[0082] Referring to FIG. 4, when the pointer comes into contact
(e.g., complete contact) with the touch screen (distance of DO), it
is recognized as a contact touch. When the pointer is located
within a distance D1 from the touch screen, it is recognized as a
proximity touch of a first proximity depth. When the pointer is
located in a range between the distance D1 and a distance D2 from
the touch screen, it is recognized as a proximity touch of a second
proximity depth. When the pointer is located in a range between the
distance D2 and a distance D3 from the touch screen, it is
recognized as a proximity touch of a third proximity depth. When
the pointer is located at a position further than the distance D3
from the touch screen, no proximity touch or a cancellation (e.g.,
termination) of a proximity touch is recognized.
[0083] Accordingly, the controller 180 may recognize the proximity
touch as various input signals according to the proximity distance
and the proximity position of the pointer with respect to the touch
screen. The controller 180 may perform various control operations
according to the input signals.
[0084] Embodiments of the present invention may be implemented in
the mobile terminal(s) 100 described with reference to FIGS. 1 to
4. Hereinafter, a method of controlling the mobile terminal 100
according to embodiments of the present invention and operation of
the mobile terminal 100 performing the method will be described in
more detail with reference to FIGS. 5, 6A, 6B, 7A-7C, 8-12,
13A-13C, 14A, 14B, 15A-15C, 16A-16C, 17A-17D, 18A and 18B.
[0085] FIG. 5 is a flowchart of a method of controlling the mobile
terminal 100 according to an embodiment of the present invention.
FIGS. 6A, 6B, 7A-7C, 8-12, 13A-13C, 14A, 14B, 15A-15C, 16A-16C,
17A-17D, 18A and 18B are views of display screens illustrating the
method of controlling the mobile terminal 100 according to
embodiments of the present invention.
[0086] Referring to FIG. 5, the controller 180 enters a lock screen
state and displays a lock screen on the touch screen 151 indicating
that the mobile terminal 100 has entered the lock screen state
(S101). The lock screen function is for preventing the mobile
terminal 100 from being operated when an undesired touch input
occurs on the touch screen 151. According to an embodiment of the
present invention, the controller 180 performs a function
corresponding to a specific touch input in controlling the mobile
terminal 100 to enter the lock screen state. For example, the
controller 180 enters the lock screen state when the entry into the
lock screen state is requested through the user input unit 130 or
when a control input is not received over a predetermined time.
[0087] When the touch screen 151 is proximity-touched and the
proximity touch input is received (S102), the controller 180
displays a preview image (or preview information) of at least one
application on the touch screen 151 based on the received proximity
touch input (S103). For example, the at least one application for
which the preview image is displayed may be an application that was
being executed before the mobile terminal 100 entered the lock
screen state.
[0088] In the operation S102, the proximity touch input that
triggers the display of the preview image of the at least one
application in the lock screen state may be input (or generated in
various manners as described below. FIGS. 6A, 6B, 7A, 7B and 7C are
views of screen displays illustrating examples of the display of
preview images of at least one application based on the proximity
touch input, where the at least one application was being executed
before entry into the lock screen state.
[0089] The controller 180 may display the preview image of the
application when at least one point of the touch screen 151 is
proximity-touched when the mobile terminal 100 is in the lock
screen state. Referring to FIG. 6A, the controller 180 receives the
proximity touch input through the touch screen on which a lock
screen 5 is displayed.
[0090] Referring to FIG. 6B, the controller 180 then displays a
preview image 6a of an application, for example, of a gallery
application that was being executed before the mobile terminal 100
entered the lock screen state. The controller 180 displays (or
causes the touch screen 151 to display) the preview image 6a such
that the preview image 6a overlaps with the lock screen 5 on a
portion of the touch screen 151. That is, the preview image 6a is
superimposed over the lock screen 5 on a portion of the touch
screen. As illustrated in FIG. 6B, the preview image 6a includes an
execution screen of the gallery application.
[0091] In addition, the controller 180 may display a preview image
of an application on the touch screen 151 when a specific region of
the touch screen 151 is selected and dragged by a proximity touch.
Referring to FIG. 7A, the controller 180 displays a slide type lock
release region 7a, which may be dragged on the lock screen 5. The
lock release region 7a is for receiving a touch input for ending
(or exiting) the lock screen state. Referring to FIG. 7B, the lock
release region 7a is touched by a proximity touch.
[0092] Referring to FIG. 7C, the controller 180 displays the
preview image 6a of the application that was being executed before
the lock screen state was entered when the lock release region 7a
is dragged toward an upper end of the screen. In this situation,
the controller 180 may steadily (e.g., incrementally) move the
display of the lock release region 7a toward the upper end of the
screen based on the drag path of the proximity touch. The
controller 180 may display the preview image 6a at a location from
which the lock release region 7a is dragged and moved as the
display of the lock release region 7a is moved upwards. Further,
the display region of the preview image 6a may be expanded as the
lock release region 7a is further dragged toward the upper end of
the screen.
[0093] For example, the controller 180 may increase a size of the
execution screen of the application included in the preview image
6a as the display region of the preview image 6a is expanded. To
this end, the controller 180 may match the display region of the
preview image 6a to a region of the execution screen of the
application and may display the region of the execution screen of
the application that is matched to the display region of the
preview image 6a in the preview image 6a. In addition, the
controller 180 may expand (or enlarge) the displayed preview image
6a (relative to the initially displayed preview image 6a) as the
display region of the preview image 6a is expanded.
[0094] In the operation of S103 (see FIG. 5), the controller 180
may generate a preview image of an application in various manners
as will be described in more detail below. FIGS. 8 to 12 are views
of screen displays illustrating examples of a preview image of an
application that is displayed when the mobile terminal 100 is in
the lock screen state.
[0095] The controller 180 may generate a preview image to include
some region of the execution screen of the application. In more
detail, when an entry into the lock screen state is requested, the
controller 180 stores a final execution screen (e.g., a screen
snapshot) of the application that is being executed in the memory
160. At least a portion of the final execution screen stored in the
memory 160 may be used as a preview image of a corresponding
application.
[0096] Referring to FIG. 8, when the mobile terminal 100 enters the
lock screen state while a gallery application displays an image
list (or series of images), the controller 180 displays a portion
of the image list (which was displayed directly before the entry
into the lock screen state) as the preview image 6a on the lock
screen 5. Referring to FIG. 9, when the mobile terminal 100 enters
the lock screen state while a predetermined application displays a
calendar image, the controller 180 displays a reduced (or shrunken)
image of the calendar image (which was displayed directly before
the entry into the lock screen state) as a preview image 6a on the
lock screen 5.
[0097] As illustrated in FIG. 8, when the preview image 6a is
displayed using a portion of the execution screen of the
application according to a proximity touch of at least one point of
the touch screen 151, the controller 180 may display different
preview images according to the point of the touch screen 151 that
is proximity-touched. In more detail, when a specific point of the
touch screen 151 is proximity-touched, the controller 180 matches
the proximity-touched point to (a correspondingly positioned
portion of) the final execution screen of the application that was
being executed before the entry into the lock screen state. In
addition, the controller 180 displays the corresponding portion of
the final execution screen of the application as a preview image
based on the matched location of the proximity-touched point.
[0098] When the mobile terminal 100 is operating in a multi-tasking
mode before the entry into the lock screen state, the controller
180 may display all of a plurality of executing screens of
applications that were being executed as preview images. Referring
to FIG. 10, the controller 180 displays preview images 6a, 6a', and
6a'' by reducing (or shrinking) respective execution screens of a
scheduling application, a contacts application, and a music
reproducing application that were being executed in the
multi-tasking mode before the entry into the lock screen state. In
addition, referring to FIG. 11, when the lock release region 7a is
selected and dragged upwards, the controller 180 displays at least
portions of the execution screens of the scheduling, contacts, and
music reproducing applications (which were being executed in the
multi-tasking mode before the entry into the lock screen state) as
preview images 6a, 6a', and 6a''. The controller 180 may increase
portions of the execution screens displayed as the preview images
6a, 6a', and 6a'' as the sizes of the displayed preview images 6a,
6a', and 6a'' are increased as the lock release region 7a is
progressively dragged upwards by the proximity touch.
[0099] When the mobile terminal 100 has usage restrictions or when
access of an application that was being executed before the entry
into the lock screen state is restricted, the controller 180 may
display at least a portion of a password input screen as a preview
image instead of the execution screen of the application. Referring
to FIG. 12, the controller 180 displays a portion of a password
input screen (instead of the execution screen of the application)
as the preview image 6a. The preview image 6a overlaps the lock
screen 5.
[0100] Referring back to FIG. 5, if ending of (or exit out of) the
lock screen state is requested when a preview image of a previously
executed application is displayed (S104), the controller 180 ends
(or controls the mobile terminal 100 to exit) the lock screen state
(S105). In the operation of S104, an input requesting ending of the
lock screen state may be received in various manners as described
below. FIGS. 13A-13C, 14A, 14B, 15A-15C, 16A-16C, 17A-17D, 18A and
18B are views of screen displays illustrating examples of ending
the lock screen state when preview images of an application are
displayed on a locked screen.
[0101] The controller 180 may end the lock screen state when the
preview image that is displayed in the lock screen state is
contacted and touched. Referring to FIG. 13A, the controller 180
displays a plurality of preview images 6a, 6a', and 6a''
corresponding to applications that were being executed before the
entry into the lock screen state. With continued reference to FIG.
13A, the lock screen 5 is proximity-touched. Consequently,
referring to FIG. 13B, the controller 180 ends the lock screen
state when any one of the preview images 6a, 6a', and 6a'' that are
displayed on the screen is contacted and touched. Referring to FIG.
13C, the controller 180 executes the application(s) that was being
executed before the entry into the lock screen state and enters an
application corresponding to the contacted and touched preview
image. That is, the controller 180 displays an execution screen of
an application corresponding to the preview image that was
contacted and touched on the touch screen 151.
[0102] In more detail, the controller 180 stores a final entry step
(or operation) of at least one application that is being executed
before entering the lock screen state in the memory 160. When the
application is executed again at the ending of the lock screen
state, the controller 180 executes the application from the state
that it was in before the entry into the lock screen state based on
the final entry step that is stored in the memory 160. Referring to
FIG. 14A, the controller 180 ends the lock screen state when a
specific item 14a included in the preview image 6a is contacted and
touched. In addition, the controller 180 re-executes (or again
executes) a corresponding application and executes an application
function related to the contacted and touched specific item 14a.
For example, with reference to FIGS. 14A and 14B, when a calendar
screen of a predetermined application is displayed in a preview
image 6a in the lock screen state and when a specific item 14a
corresponding to the schedule of a particular day is contacted and
touched, the controller 180 displays detailed information 14 of the
corresponding schedule on the screen.
[0103] When a specific portion of the lock screen is dragged longer
(or farther) than a predetermined distance by the proximity touch,
the controller 180 may end the lock screen state. Referring to FIG.
15A, the controller 180 displays the preview image 6a of the
previously executed application on the screen when a specific
region, for example, the lock release region 7a, of the lock screen
is dragged by proximity touch. Referring to FIG. 15B, the
controller 180 ends the lock screen state when the lock release
region 7a is dragged farther than a predetermined distance.
Consequently, referring to FIG. 15C, the controller 180 re-executes
an application corresponding to the preview image 6a. That is, the
controller 180 displays the execution screen 15 of the application
corresponding to the preview image 6a on the touch screen 151.
[0104] The controller 180 may end the lock screen state when the
preview image of the application displayed on the lock screen is
dragged into a specific region by the proximity touch. Referring to
FIG. 16A, the controller 180 displays the preview image 6a of the
application that was being executed before the entry into the lock
screen state when the lock screen 5 is proximity-touched. The
preview image 6a is selected by a proximity touch. Referring to
FIG. 16B, the controller 180 ends the lock screen state when the
preview image 6a is dragged into a specific region of the lock
screen 5, for example, a lock release button region 16a, by the
proximity touch. Consequently, referring to FIG. 16C, the
controller 180 displays an execution screen of the application
corresponding to the dragged preview image 6a. That is, the
controller 180 displays an execution screen 16 of the application
corresponding to the preview image 6a on the touch screen 151.
[0105] The controller 180 may end the lock screen state when a size
of the preview image displayed on the lock screen is increased to
become larger than a predetermined size based on a multi-touch
input by a proximity touch. Referring to FIG. 17A, the controller
180 displays the preview image 6a of the application that was being
executed before the entry into the lock screen state when the lock
screen 5 is proximity-touched. The preview image 6a is selected by
a proximity touch. Referring to FIG. 17B, the controller 180
changes the size of the displayed preview image 6a when the preview
image 6a is touched again by the proximity touch and when a drag
input that starts from at least one point is received via the
proximity touch.
[0106] Referring to FIGS. 17C and 17D, the controller 180 ends the
lock screen state when the size of the displayed preview image 6a
is increased to become larger than a predetermined size. In
addition, referring to FIG. 17D, the controller 180 re-executes an
application that was being executed before the entry into the lock
screen state. That is, the controller 180 displays an execution
screen 17 of an application corresponding to the preview image 6a
on the touch screen 151.
[0107] In the operation of S105 (see FIG. 5), when the lock screen
state is ended as illustrated in FIGS. 13C, 14B, 15C, 16C and 17D,
the controller 180 re-executes at least one of the applications
that was being executed before the entry into the lock screen state
and displays an execution screen of the at least one of the
re-executed applications on the touch screen 151. However, in a
situation where the mobile terminal 100 has a usage restriction or
where access to the application that was being executed before the
entry into the lock screen state is restricted, the controller 180
may display a password input screen instead of the execution screen
of the application when the lock screen state is ended.
[0108] Referring to FIG. 18A, the controller 180 displays a portion
of the password input screen as a preview image 6a overlapping with
the lock screen 5, instead of a portion of the execution screen of
the application. Subsequently, referring to FIG. 18B, when the lock
screen state is ended, the controller 180 displays the password
input screen 18 on the screen.
[0109] According to previously-described embodiments of the present
invention, it is possible for a user to check a previously executed
application in advance using the proximity touch without ending the
lock screen state. Moreover, it is also possible for a user to
check an application executed before the lock screen state was
entered and to proceed toward requesting an ending of the lock
screen state without a break. In addition, when the mobile terminal
100 has a usage restricted or when access to the application that
was being executed before the entry into the lock screen state is
restricted, the controller 180 may display the password input
screen as a preview image instead of the execution screen of the
application so that user convenience is enhanced and privacy
protection is reinforced.
[0110] Hereinafter, a method of controlling a mobile terminal 100
according to other embodiments of the present invention and
operation of the mobile terminal for performing the method will be
described in more detail with reference to FIGS. 19, 20A, 20B,
21A-21D, 22A-22C, 23A-23C and 24A-24D.
[0111] FIG. 19 is a flowchart illustrating a method of controlling
a mobile terminal according to another embodiment of the present
invention. FIGS. 20A, 20B, 21A-21D, 22A-22C, 23A-23C and 24A-24D
are views of screen displays illustrating the method according to
embodiments of the present invention.
[0112] Referring to FIG. 19, the controller 180 receives a
proximity touch input on the touch screen 151 (S1001). When there
is preview information corresponding to a region where the
proximity touch occurs (S1002), the controller 180 displays the
corresponding preview information on the touch screen 151 (S1003).
Regarding the operation of S1003, different preview information may
be selected based on an operation mode of the mobile terminal 100,
a function of an application that is currently executed, and an
item displayed on the region where the proximity touch occurs.
[0113] FIGS. 20A, 20B, 21A-21C, 22A-22C, 23A-23C, 24A-24D and 25
show examples of screen displays displaying preview information in
the mobile terminal 100. The controller 180 may display preview
information regarding alarm information when the alarm information
is proximity-touched while the mobile terminal 100 is operating in
the lock screen state. The alarm information includes information
of which a user is to be notified such as a missed call, an
unchecked received message, and a schedule.
[0114] Referring to FIG. 20A, while the mobile terminal 100 is
operating in the lock screen state, the controller 180 displays
alarm information 20a and 20b for notifying the user of a missed
voice call and a missed video call, respectively, and alarm
information 20c for notifying the user of a particular schedule
(e.g., today's schedule) on the lock screen 5. As illustrated with
reference to FIG. 20A, among the alarm information 20a, 20b and 20c
that is displayed on the lock screen 5, the alarm information 20a
for notifying the user of a missed voice call number is
proximity-touched (S1101). Subsequently, referring to FIG. 20B, the
controller 180 displays preview information 20d corresponding to
the missed calls on the lock screen 5. When the displayed preview
information 20d is contacted and touched, the controller 180 may
end the lock screen state and display more detailed information
corresponding to the contacted-and-touched preview information 20d
on the screen.
[0115] When a specific region of a message writing screen is
proximity-touched during the writing (or creation) of a message,
the controller 180 may display preview information (e.g., a
preview) of the in-progress message on the screen.
[0116] Referring to FIG. 21A, the controller 180 displays a message
writing screen 21 on the touch screen 151. Referring to FIG. 21B,
when a specific region of the message writing screen 21, for
example, a region where a send button 21a is displayed, is
proximity-touched, the controller 180 displays preview information
21b of the message that is being written. That is, the controller
180 displays a preview image of the current message on the screen.
The controller 180 may activate or terminate display of the preview
information according to a proximity touch depth of the send button
21a.
[0117] For example, referring to FIG. 21C, when a distance between
a pointer causing the proximity touch (e.g., the user's finger) and
the touch screen 151 becomes shorter than a predetermined distance,
the controller 180 deletes the preview information 21b displayed on
the screen. Further, referring to FIG. 21D, when the send button
21a is contacted and touched, the controller 180 executes a send
function, which is an inherent function mapped (or corresponding)
to the send button 21a. That is, the controller 180 transmits the
current written message to a designated recipient.
[0118] For example, with reference back to FIG. 4, the controller
180 activates the displaying of the preview information 21b when a
distance between the pointer and the touch screen 151 is within a
range from D1 to D3 (e.g., a range from greater than or equal to D1
to less than or equal to D3). When the distance between the pointer
and the touch screen 151 becomes shorter than D1, the controller
180 inactivates (or terminates) the displaying of the preview
information 21b. In addition, when the distance between the pointer
and the touch screen 151 is reduced to D0, that is, when the send
button 21a displayed on the touch screen 151 is directly contacted
and touched, the controller 180 executes an inherent function
mapped to the send button 21a.
[0119] The controller 180 may display preview information (e.g., a
preview) of a photographed image when a specific region of a
photographing screen is proximity-touched while the mobile terminal
100 is operating in a photographing mode. That is, the controller
180 displays a photographed image of an external view that is input
through the camera 121 as a preview image.
[0120] Referring to FIG. 22A, the controller 180 displays a
photographed image 22 on the screen when the mobile terminal 100
enters the photographing mode. When a specific region of the
photographing screen 22, for example, a photograph button 22a is
proximity-touched, the controller 180 displays an image generated
by photographing an external view that is input through the camera
121 as preview information 22b. That is, the controller 180
displays a preview image, generated by photographing an external
view that is currently input through the camera 121, on the
screen.
[0121] The controller 180 may activate or inactivate the displaying
of the preview information 22b according to the proximity touch
depth with respect to the photograph button 22a. For example,
referring to FIG. 22B, the controller 180 inactivates the
displaying of the preview information 22b when the proximity touch
depth of the photograph button 22a becomes shorter than a
predetermined distance. In addition, referring to FIG. 22C, the
controller 180 executes a photographing function (an inherent
function mapped (or corresponding) to the photograph button 22a)
when the photograph button 22a is contacted and touched. That is,
the controller 180 photographs an external view input through the
camera 121 and stores the photographed image in the memory 160.
[0122] For example, with reference back to FIG. 4, when a distance
between a pointer causing the proximity touch (e.g., the user's
finger) and the touch screen 151 is within a range from D1 to D3,
the controller 180 activates the displaying of the preview
information 22b. When the distance between the pointer and the
touch screen 151 becomes shorter than D1, the controller 180
inactivates the displaying of the preview information 22b. In
addition, when the distance between the pointer and the touch
screen 151 is reduced to D0, that is, when the photograph button
22a displayed on the touch screen 151 is directly contacted and
touched, the controller 180 executes an inherent function mapped to
the photograph button 22a.
[0123] The controller 180 may display preview information of a
previous channel or a next channel relative to a channel that is
currently being received when a specific region is
proximity-touched while the mobile terminal 100 is operating in a
broadcast receiving mode. The preview information of the previous
channel or the next channel may be schedule information acquired
from a preview image of broadcasting content that is currently
being broadcasted through a corresponding channel or an electronic
program guide (EPG). In the former situation, the controller 180
receives broadcasting content of a previous channel or a next
channel in the background during the reception of broadcasting
content through a current channel. When a user requests a display
of preview information, the controller 180 displays a preview image
of the received broadcasting content. In the latter situation, the
controller 180 receives EPG information of a previous channel or a
next channel in the background during the reception of broadcasting
content through a current channel. When a user requests a display
of preview information, the controller 180 displays a broadcasting
schedule using the received EPG information on the screen.
[0124] Referring to FIG. 23A, the controller 180 displays a
broadcast receiving screen 23 on the screen when the mobile
terminal 100 enters a broadcast receiving mode. When a specific
region of the broadcast receiving screen 23, for example, a region
where a channel changing button 23a is displayed is
proximity-touched, the controller 180 displays preview information
23b of a channel previous to or next to the current channel. For
example, when a channel changing button 23a for requesting a change
to a previous channel (e.g., the lower pointer of changing button
23a) is proximity-touched, the controller 180 may display preview
information of the previous channel on the screen. In addition,
when the channel changing button 23a for requesting a change to a
next channel (e.g., the upper pointer of changing button 23a) is
proximity-touched, the controller 180 may display preview
information of the next channel on the screen.
[0125] The controller 180 may activate or inactivate the displaying
of preview information according to the proximity touch depth with
respect to the channel changing button 23a. For example, referring
to FIG. 23B, when a proximity depth at which the channel changing
button 23a is proximity-touched is less than a predetermined
distance, the controller 180 inactivates the displaying of the
preview information 23b that was displayed earlier. Referring to
FIG. 23C, when the channel changing button 23a is contacted and
touched, the controller 180 executes a channel changing function,
which is an inherent function mapped (or corresponding) to the
channel changing button 23a. That is, the controller 180 displays
broadcasting content received for the previous channel or the next
channel on the broadcast receiving screen 23.
[0126] For example, referring back to FIG. 4, the controller 180
activates the displaying of the preview information 23b when a
distance between a pointer causing the proximity touch (e.g., the
user's finger) and the touch screen 151 is within a range from D1
to D3. When a distance between the pointer and the touch screen 151
becomes shorter than D1, the controller 180 inactivates the
displaying of the preview information 23b. In addition, when a
distance between the pointer and the touch screen 151 is reduced to
DO, that is, when the channel changing button 23a displayed on the
touch screen 151 is directly contacted and touched, the controller
180 executes an inherent function mapped to the channel changing
button 23a.
[0127] The controller 180 may display preview information of a web
page linked to a specific region when the specific region is
touched during the displaying of a web page through a web
browser.
[0128] Referring to FIG. 24A, the controller 180 displays a web
page P1 through a web browser on the screen. An item linking to
another web page P2 is located in a region 24a of the web page P1.
When the region 24a is proximity-touched, the controller 180
displays preview information 24b of the web page P2 (see, e.g.,
FIG. 24D) linked to the corresponding region. For example, the
controller 180 displays a portion (e.g., an image) of the web page
P2 linked to the region that is proximity-touched as preview
information 24b.
[0129] The controller 180 may activate or inactivate the displaying
of the preview information 24b according to the proximity-touch
depth with respect to the item 24a to which the web page P2 is
linked. For example, referring to FIG. 24B, when the proximity
depth where the item 24a is proximity-touched is shorter than a
predetermined distance, the controller 180 inactivates the
displaying of the preview information 24b that was displayed
earlier. Referring to FIG. 24C, when the item 24a is contacted and
touched, the controller 180 accesses the web page P2 linked to the
corresponding item 24a. Referring to FIG. 24D, the controller 180
displays the corresponding web page P2 on the screen.
[0130] For example, referring back to FIG. 4, when a distance
between a pointer causing the proximity touch (e.g., the user's
finger) and the touch screen 151 is within a range from D1 to D3,
the controller 180 activates the displaying of the preview
information 24b. When a distance between the pointer and the touch
screen 151 becomes shorter than D1, the controller 180 inactivates
the displaying of the preview information 24b. In addition, when a
distance between the pointer and the touch screen 151 is reduced to
D0, that is, when the item 24a displayed on the touch screen 151 is
directly contacted and touched, the controller 180 displays the web
page P2 linked to the corresponding item 24a on the entire
screen.
[0131] According to another embodiment of the present invention, it
is possible for a user to check (in advance) a missed call, a
missed message, and today's schedule by proximity-touching the
corresponding alarm information without ending the lock screen
state. In addition, a user may check (in advance) a written message
in a message writing screen before transmission of the written
message and may then send the written message. Also, a user may
check (in advance) a photographed image in a photographing screen
without switching screens. Also, a user may check (in advance)
broadcasting content of another channel and may then select a
desired channel. Moreover, a user may check content of a linked web
page without (fully) accessing the linked web page.
[0132] Hereinafter, a method of controlling a mobile terminal 100
according to another embodiment of the present invention and
operation of the mobile terminal 100 performing the control method
will be described in more detail with reference to FIGS. 25, 26A,
26B, 27A, 27B, 28A, 28B, 29A-29C, 30A, 30B and 31A-31D.
[0133] FIG. 25 is a flowchart illustrating a control method of the
mobile terminal 100 according to still another embodiment of the
present invention. FIGS. 26A, 26B, 27A, 27B, 28A, 28B, 29A-29C,
30A, 30B and 31A-31D are views illustrating a control method of the
mobile terminal 100 according to still another embodiment of the
present invention.
[0134] Referring to FIG. 25, the controller 180 receives a
proximity touch input (e.g., via detection of a particular
pressure) through the touch screen 151 (S1601). The controller 180
determines whether there is history information corresponding to a
proximity-touched region (S1602). If such information exists, the
controller 180 displays the corresponding history information on
the touch screen 151 (S1603).
[0135] In the operation of S1603, different parts of the history
information may be selected based on the operation mode of the
mobile terminal 100, a function of an application that is being
executed currently, and an item displayed in the proximity-touched
region.
[0136] FIGS. 26A, 26B, 27A, 27B, 28A, 28B, 29A-29C, 30A, 30B and
31A-31D are views of screen displays illustrating examples of
displaying the history information in the mobile terminal 100
according to various embodiments.
[0137] When contact information displayed on the screen is
proximity-touched, the controller 180 may display history
information of a task that was performed previously (e.g.,
recently) using the proximity-touched contact information.
Referring to FIG. 26A, the controller 180 displays a contact screen
26 on the touch screen 151. Referring to FIG. 26B, when specific
contact information 26a is proximity-touched on the contact screen
26, the controller 180 displays history information 26b related to
a task performed recently using the contact information 26a that is
proximity-touched.
[0138] For example, when a call to/from the contact of contact
information 26a has occurred recently, the controller 180 may
display call history or detail information of the corresponding
call as the history information 26b using the proximity-touched
contact information 26a on the screen. In addition, when a message
to/from the contact has been transmitted recently, the controller
180 may display history of the message transmission or detail
information of a corresponding message (see, e.g., FIG. 26B) as the
history information 26b using the proximity-touched contact
information 26a on the screen. The history information may be
displayed such that it overlaps with the contact screen 26 in the
form of a pop-up window, or it may be displayed in a
semi-transparent manner. Also, the history information may be
displayed in response to only receipt (e.g., detection) of the
proximity-touch input in a particular region.
[0139] When receiver information (information regarding an intended
recipient) is proximity-touched on a message writing screen, the
controller 180 may display history information of a message that
was received recently in association with the proximity-touched
receiver information on the screen. Referring to FIG. 27A, the
controller 180 displays a message writing screen 27 on the touch
screen 151. Receiver information 27a is also displayed. Referring
to FIG. 27B, when the receiver information 27a is proximity-touched
on the message writing screen 27, the controller 180 displays a
message that was received recently in association with the receiver
information 27a as the history information 27b on the screen. The
history information may be displayed such that it overlaps with the
message writing screen 27 in the form of a pop-up window, or it may
be displayed in a semi-transparent manner. In addition, the history
information may be displayed in response to only receipt (e.g.,
detection) of the proximity-touch input in a particular region.
[0140] When a specific region is proximity-touched on a web browser
screen, the controller 180 may display history information
regarding web pages to be accessed using a web browser.
[0141] Referring to FIG. 28A, the controller 180 displays web page
WP1 on a web browser screen. The web page WP1 includes an empty (or
blank) region 28a. Referring to FIG. 28B, when a specific region of
the web browser screen, for example, the empty region 28a of the
web page WP1 that is being displayed, is proximity-touched, the
controller 180 displays a portion of another web page WP2 accessed
through the web browser as history information 28b. For example,
the controller 180 may display the other web page WP2 on the web
page WP1 in a layered manner.
[0142] FIGS. 29A-29C, 30A and 30B are views illustrating examples
of switching web pages using the history information described
above with reference to FIGS. 28A and 28B. Referring to FIG. 29A,
when an empty region 28a (see, e.g., FIG. 28A) of a web page WP1
that is currently displayed is proximity-touched, the controller
180 displays history information 28b using a portion of another web
page WP2 that is accessed through the web browser. Referring to
FIG. 29B, when the web page WP2 that is displayed as the history
information 28b is dragged by the proximity-touch, the controller
180 changes a size of the displayed web page WP2 based on a drag
path. Referring to FIG. 29C, when the web page WP2 is dragged into
a specific region farther than a predetermined distance, the
controller 180 ends the displaying of the history information and
displays the web page WP2 on the entire screen.
[0143] Referring to FIG. 30A, a web page WP2 that is displayed as
the history information 28b is directly contact and touched. In
response, referring to FIG. 30B, the controller 180 ends the
displaying of the history information 28b and displays the web page
WP2 on the whole screen.
[0144] Although the figures of FIGS. 28A, 28B, 29A-29C, 30A and 30B
illustrate the displaying of history information by displaying web
pages in a layered manner, the controller 180 may display the
history information using reduced (e.g., shrunken) images of the
respective web pages.
[0145] Referring to FIG. 31A, the controller 180 displays web page
WP1 on a web browser screen. The web page WP1 includes an empty (or
blank) region 28a. Referring to FIG. 31B, when a specific region of
the web browser screen, for example, the empty region 28a of the
web page WP1 that is currently displayed, is proximity-touched, the
controller 180 displays the history information 28b using reduced
images of other web pages WP2, WP3, WP4 that are accessed through
the web browser. With reference to FIG. 31C, one of the other web
pages (e.g., WP2) that are displayed on the screen is directly
contacted and touched. In response, referring to FIG. 31D, the
controller 180 ends the displaying of the history information 28b
and displays the web page WP2 on the entire screen.
[0146] According to still another embodiment of the present
invention, it is possible for a user to check history information
related to contact information without selecting a menu related to
the contact information. Since only the history information is
displayed using the proximity touch, the likelihood of initiating a
highly undesirable action may be reduced. Since only the history
information is displayed in response to the proximity touch,
inconvenience caused by displaying the history information before
actually performing another function on the contact screen is
minimal.
[0147] A user may check a message that was recently sent by a
recipient in the message writing screen. Therefore, while composing
a message with reference to the recently received message, a user
may check the recently received message in the message writing
screen without closing the message writing screen in order to check
the previously received message. As a result, user convenience is
enhanced.
[0148] In addition, when a specific region is proximity-touched in
a web browser screen, the controller 180 may display history
information on a web page that was previously accessed on the
screen. A user may check web pages that are currently being
accessed only by proximity-touching a specific region of a web
browser screen. Moreover, since only the history information is
displayed using the proximity touch, the likelihood of
unintentionally opening an undesired web page or triggering the
execution of undesired content may be reduced.
[0149] The disclosed methods of controlling the mobile terminal may
be written as computer programs and may be implemented in digital
microprocessors that execute the programs using a computer readable
recording medium. The methods of controlling the mobile terminal
may be executed through software. The software may include code
segments that perform required tasks. Programs or code segments may
also be stored in a processor readable medium or may be transmitted
according to a computer data signal combined with a carrier through
a transmission medium or communication network.
[0150] The computer readable recording medium may be any data
storage device that may store data and may be read by a computer
system. Examples of the computer readable recording medium may
include read-only memory (ROM), random-access memory (RAM),
CD-ROMs, DVD.+-.ROM, DVD-RAM, magnetic tapes, floppy disks, and
optical data storage devices. The computer readable recording
medium may also be distributed over network coupled computer
systems such that the computer readable code is stored and executed
in a distributed manner.
[0151] The foregoing embodiments and features are merely exemplary
in nature and are not to be construed as limiting the present
invention. The disclosed embodiments and features may be readily
applied to other types of apparatuses. The description of the
foregoing embodiments is intended to be illustrative, and not to
limit the scope of the claims. Many alternatives, modifications,
and variations will be apparent to those skilled in the art.
* * * * *