U.S. patent application number 14/193398 was filed with the patent office on 2014-06-26 for multi display apparatus and multi display method.
This patent application is currently assigned to Samsung Electronics Co., Ltd.. The applicant listed for this patent is Samsung Electronics Co., Ltd.. Invention is credited to Dae-myung KIM, Ji-yeon KWAK.
Application Number | 20140181700 14/193398 |
Document ID | / |
Family ID | 49513716 |
Filed Date | 2014-06-26 |
United States Patent
Application |
20140181700 |
Kind Code |
A1 |
KIM; Dae-myung ; et
al. |
June 26, 2014 |
MULTI DISPLAY APPARATUS AND MULTI DISPLAY METHOD
Abstract
A multi display apparatus includes a first body including a
first display, a second body including a second display, a hinge
configured to rotatably connect the first body and the second body
to support the first body and the second body, a sensor configured
to sense a first user manipulation made in a nondecisive location
of a nonstructured first screen displayed on the first display and
to sense a second user manipulation made in a nondecisive location
of a nonstructured second screen displayed on the second display,
and a controller configured to combine the sensed first user
manipulation and second user manipulation to perform an operation
corresponding to a result of the combination.
Inventors: |
KIM; Dae-myung;
(Hwaseong-si, KR) ; KWAK; Ji-yeon; (Seoul,
KR) |
|
Applicant: |
Name |
City |
State |
Country |
Type |
Samsung Electronics Co., Ltd. |
Suwon-si |
|
KR |
|
|
Assignee: |
Samsung Electronics Co.,
Ltd.
Suwon-si
KR
|
Family ID: |
49513716 |
Appl. No.: |
14/193398 |
Filed: |
February 28, 2014 |
Related U.S. Patent Documents
|
|
|
|
|
|
Application
Number |
Filing Date |
Patent Number |
|
|
14050936 |
Oct 10, 2013 |
|
|
|
14193398 |
|
|
|
|
Current U.S.
Class: |
715/761 |
Current CPC
Class: |
G06F 3/04883 20130101;
G06F 3/0486 20130101; G06F 3/0484 20130101; G06F 3/04845 20130101;
G06F 2203/04808 20130101; G06F 1/1616 20130101; G06F 3/0488
20130101 |
Class at
Publication: |
715/761 |
International
Class: |
G06F 3/0488 20060101
G06F003/0488; G06F 3/0486 20060101 G06F003/0486 |
Foreign Application Data
Date |
Code |
Application Number |
Oct 10, 2012 |
KR |
10-2012-0112638 |
Claims
1. A multi display apparatus comprising: a first body comprising a
first display; a second body comprising a second display; a hinge
configured to rotatably connect the first body and the second body
to support the first body and the second body; a sensor configured
to sense a first user manipulation made in a nondecisive location
of a nonstructured first screen displayed on the first display and
to sense a second user manipulation made in a nondecisive location
of a nonstructured second screen displayed on the second display;
and a controller configured to combine the sensed first user
manipulation and the sensed second user manipulation to perform an
operation corresponding to a result of the combination, to extract
contents displayed on the first display when the first user
manipulation is sensed, and to display a screen corresponding to
the extracted contents on the second display when the second user
manipulation is sensed.
2. The multi display apparatus of claim 1, wherein: the controller
captures the first screen displayed on the first display according
to the first user manipulation; and the controller displays the
captured first screen in a location where the second user
manipulation is made in the second display.
3. The multi display apparatus of claim 1, wherein: the controller
extracts the contents displayed on the first display and second
display when a third user manipulation made in a nondecisive
location of the second screen is sensed while the first user
manipulation is being sensed; and the controller displays the
screen corresponding to the extracted contents on the second
display when the second user manipulation is sensed.
4. A multi display apparatus comprising: a first body comprising a
first display; a second body comprising a second display; a hinge
configured to rotatably connect the first body and the second body
to support the first body and the second body; a sensor configured
to sense a first user manipulation made in a nondecisive location
of a nonstructured first screen displayed on the first display and
to sense a second user manipulation made in a nondecisive location
of a nonstructured second screen displayed on the second display;
and a controller configured to combine the sensed first user
manipulation and the sensed second user manipulation to perform an
operation corresponding to a result of the combination, wherein the
sensed first user manipulation comprises a touch input or a
proximity input for a predetermined time.
5. The multi display apparatus of claim 4, wherein: the sensed
second user manipulation is a touch input for a predetermined time;
the controller captures the first screen and the second screen; and
the controller displays the captured first screen and the second
screen on a location on the second screen where a third user
manipulation is input, when the third user manipulation made in a
nondecisive location of the second screen.
6. The multi display apparatus of claim 4, wherein: when the second
user manipulation is a drag input intersecting the first screen and
the second screen, the controller extracts information from the
first screen, and displays the extracted information on the second
screen.
7. The multi display apparatus of claim 6, wherein the controller
displays the information searched based on the extracted
information on the second screen.
8. The multi display apparatus of claim 4, wherein the second user
manipulation is an input proximate to or touching the second
display using an input pen.
9. A multi display method of a multi display apparatus comprising a
first body comprising a first display, a second body comprising a
second display, and a hinge configured to rotatably connect the
first body and the second body to support the first body and the
second body, the multi display method comprising: displaying a
nonstructured first screen on the first display; displaying a
nonstructured second screen on the second display; sensing a first
user manipulation made in a nondecisive location on the first
screen and a second user manipulation made in a nondecisive
location on the second screen; extracting contents displayed on the
first display when the first user manipulation is sensed; and
combining the sensed first user manipulation and the sensed second
user manipulation and performing an operation corresponding to a
result of the combination, wherein performing the operation
comprises displaying a screen corresponding to the extracted
contents on the second display when the second user manipulation is
sensed.
10. The multi display method of claim 9, wherein: extracting the
contents comprises capturing the first screen displayed on the
first display according to the first user manipulation; and
displaying the screen corresponding to the extracted contents
comprises displaying the captured first screen in a location on the
second display where the second user manipulation is made.
11. A multi display method of a multi display apparatus comprising
a first body comprising a first display, a second body comprising a
second display, and a hinge configured to rotatably connect the
first body and the second body to support the first body and the
second body, the multi display method comprising: displaying a
nonstructured first screen on the first display; displaying a
nonstructured second screen on the second display; sensing a first
user manipulation made in a nondecisive location on the first
screen and a second user manipulation made in a nondecisive
location on the second screen; combining the sensed first user
manipulation and the sensed second user manipulation and performing
an operation corresponding to a result of the combination, wherein
the sensed first user manipulation comprises a touch input or a
proximity input for a predetermined time.
12. The multi display method of claim 11, wherein performing the
operation comprises: extracting the contents displayed on the first
display and second display when a third user manipulation made in a
nondecisive location of the second screen is sensed while the first
user manipulation is being sensed; and displaying the screen
corresponding to the extracted contents on the second display when
the second user manipulation is sensed.
13. The multi display method of claim 11, wherein: the sensed
second user manipulation is a touch input for a predetermined time;
and performing the operation comprises: capturing the first screen
and second screen; and displaying the captured first screen and
second screen in a location of the second screen where a third user
manipulation is input, when the third user manipulation is made in
a nondecisive location of the second screen.
14. The multi display method of claim 11, wherein performing the
operation comprises: extracting information from the first screen;
and displaying the extracted information on the second screen when
the first user manipulation is a touch input on a nondecisive
location of the first screen for a predetermined time and the
second user manipulation is a drag input intersecting the first
screen and second screen.
15. The multi display method of claim 14, wherein performing the
operation comprises displaying information searched based on the
extracted information on the second screen.
16. The multi display method of claim 11, wherein the second user
manipulation is an input adjacent to or touching the second display
using an input pen.
Description
CROSS-REFERENCE TO RELATED APPLICATIONS
[0001] This application is a Continuation Application of prior
application Ser. No. 14/050,936, filed on Oct. 10, 2013 in the
United States Patent and Trademark Office, which claims priority
under 35 U.S.C. .sctn.119(a) from Korean Patent Application No.
10-2012-0112638, filed in the Korean Intellectual Property Office
on Oct. 10, 2012, the disclosure of which is incorporated herein by
reference.
BACKGROUND OF THE INVENTION
[0002] 1. Field of the Invention
[0003] Methods and apparatuses consistent with the exemplary
embodiments of the present general inventive concept relate to a
multi display apparatus, and more particularly to a multi display
apparatus configured to receive inputs through each display thereof
and to perform outputting accordingly, and a multi display method
thereof.
[0004] 2. Description of the Related Art
[0005] A portable electronic apparatus in recent use such as a
smart phone generally has one display.
[0006] However, as the performance of a Central Processing Unit
(CPU) mounted in a portable electronic apparatus increases and
multi-tasking environments capable of executing various
applications at the same time are provided, it becomes difficult to
effectively utilize the performance a display apparatus having just
one display.
[0007] In addition, due to the development of battery technologies
and thermal treatment systems, the design paradigm that a portable
electronic apparatus may only have one display is not an acceptable
truth any more.
[0008] Furthermore, the need has been raised to expand touch
screens and introduce more various user experiences to portable
electronic apparatuses in user experience oriented interface
environment.
[0009] Against this background, portable electronic apparatuses
having two or more displays, that is, multi display apparatuses,
are emerging.
[0010] A multi display apparatus is receiving attention as a next
generation display apparatus since it may execute and display
various applications on multiple displays based on powerful
hardware performance, and thus is not only appropriate to a
multi-tasking environment, but also provides various and abundant
information to a user, thereby providing new user experience.
[0011] In a multi display apparatus, not only outputting but also
inputting may be performed through a plurality of displays.
Therefore, the need has been raised for the technology which may
perform more various inputs and outputs according to inputs of each
display in a multi display apparatus.
SUMMARY OF THE INVENTION
[0012] Exemplary embodiments of the present general inventive
concept provide a multi display apparatus configured to perform an
unstructured input through each display of the multi display
apparatus, and to combine each input to perform an output
corresponding thereto, and a multi display method thereof.
[0013] Additional features and utilities of the present general
inventive concept will be set forth in part in the description
which follows and, in part, will be obvious from the description,
or may be learned by practice of the general inventive concept.
[0014] Exemplary embodiments of the present general inventive
concept provide a multi display apparatus including a first body
including a first display, a second body including a second
display, a hinge configured to rotatably connect the first body and
the second body to support the first body and the second body, a
sensor configured to sense a first user manipulation made in a
nondecisive location of a nonstructured first screen displayed on
the first display and to sense a second user manipulation made in a
nondecisive location of a nonstructured second screen displayed on
the second display, and a controller configured to combine the
sensed first user manipulation and the sensed second user
manipulation to perform an operation corresponding to a result of
the combination.
[0015] The controller may change a function matching the second
user manipulation according to the first user manipulation, and the
controller may perform an operation corresponding to the second
user manipulation according to the changed function, when the
second user manipulation is sensed.
[0016] The second user manipulation may be a line input by a drag
operation, the controller may increase a thickness or transparency
of the line input of the second user manipulation when the first
user manipulation is a pinch out input, the controller may decrease
a thickness or transparency of the line input of the second user
manipulation when the first user manipulation is a pinch in input,
and the controller may change a texture or thickness of the line
input of the second user manipulation when the first user
manipulation is a touch and rotate input.
[0017] The first user manipulation may be a touch and hold input
and the second user manipulation may be a touch input, and the
controller may display a menu corresponding to the second user
manipulation when the first user manipulation and the second user
manipulation are sensed at the same time.
[0018] The controller may change the second screen displayed on the
second display according to the first user manipulation when the
first user manipulation is sensed, and the controller may display
the second screen changed by the first user manipulation according
to the second user manipulation, when the second user manipulation
is sensed.
[0019] The controller may expand the second screen displayed on the
second display when the first user manipulation is a pinch out
input, the controller may reduce the second screen displayed on the
second display when the first user manipulation is a pinch in
input, and the controller may rotate the second screen displayed on
the second display when the first user manipulation is a touch and
rotate input.
[0020] Exemplary embodiments of the present general inventive
concept also provide a multi display apparatus including a first
body including a first display, a second body including a second
display, a hinge configured to rotatably connect the first body and
the second body to support the first body and the second body, a
sensor configured to sense a first user manipulation made in a
nondecisive location of the first display and a second user
manipulation made in a nondecisive location of the second display,
when the first display and the second display respectively display
a nonstructured first screen and a nonstructured second screen, and
a controller configured to combine the sensed first user
manipulation and second user manipulation to recognize the
combination as one user gesture, and to display an execution screen
corresponding to the one user gesture on at least one of the first
display and the second display.
[0021] The first user manipulation may be an input proximate to or
touching the first display using a user's body, and the second user
manipulation may be an input proximate to or touching the second
display using an input pen.
[0022] Exemplary embodiments of the present general inventive
concept also provide a multi display method of a multi display
apparatus including a first body including a first display, a
second body including a second display, and a hinge configured to
rotatably connect the first body and the second body to support the
first body and the second body, the multi display method including
displaying a nonstructured first screen on the first display,
displaying a nonstructured second screen on the second display,
sensing a first user manipulation made in a nondecisive location on
the first screen and a second user manipulation made in a
nondecisive location on the second screen, and combining the sensed
first user manipulation and the sensed second user manipulation and
performing an operation corresponding to a result of the
combination.
[0023] Performing the operation may include changing a function
matching the second user manipulation according to the first user
manipulation, and performing an operation corresponding to the
second user manipulation according to the changed function when the
second user manipulation is sensed.
[0024] The second user manipulation may be a line input by a drag
operation. Changing the function may include increasing a thickness
of the line input of the second user manipulation when the first
user manipulation is a pinch out input, reducing a thickness of the
line input of the second user manipulation when the first user
manipulation is a pinch in input, and changing a texture of the
line input of the second user manipulation when the first user
manipulation is a touch and rotate input.
[0025] The first user manipulation may be a touch and hold input
and the second user manipulation may be a touch input. Performing
the operation may include displaying a menu corresponding to the
second user manipulation when the first user manipulation and the
second user manipulation are sensed at the same time.
[0026] The multi display method may further include changing the
second screen displayed on the second display according to the
first user manipulation, when the first user manipulation is
sensed. Performing the operation may include displaying the second
screen changed by the first user manipulation according to the
second user manipulation, when the second user manipulation is
sensed.
[0027] The changing the second screen may include expanding the
second screen displayed on the second display when the first user
manipulation is a pinch out input, reducing the second screen
displayed on the second display when the first user manipulation is
a pinch in input, and rotating the second screen displayed on the
second display when the first user manipulation is a touch and
rotate input.
[0028] Exemplary embodiments of the present general inventive
concept also provide a multi display method of a multi display
apparatus including a first body including a first display, a
second body including a second display, and a hinge configured to
rotatably connect the first body and the second body to support the
first body and the second body, the multi display method including
sensing a first user manipulation made in a nondecisive location of
a nonstructured first screen displayed on the first display,
sensing a second user manipulation made in a nondecisive location
of a nonstructured second screen displayed on the second display,
and combining the sensed first user manipulation and second user
manipulation to recognize the combination as one user gesture, and
displaying an execution screen corresponding to the one user
gesture on at least one of the first display and the second
display.
[0029] The first user manipulation may be an input proximate to or
touching the first display using a user's body, and the second user
manipulation may be an input proximate to or touching the second
display using an input pen.
[0030] The controller may extract contents displayed on the first
display when the first user manipulation is sensed, and display a
screen corresponding to the extracted contents on the second
display when the second user manipulation is sensed.
[0031] The controller may capture a screen displayed on the first
display according to the first user manipulation, and display the
captured screen on a location on the second display where the
second user manipulation is made.
[0032] The controller may extract contents displayed on the first
display and the second display and display a screen corresponding
to the extracted contents on the second display a third user
manipulation is sensed on a nondecisive location of the second
screen while the first user manipulation is sensed.
[0033] Performing the operation may include extracting contents
displayed on the first display when the first user manipulation is
sensed, and displaying a screen corresponding to the extracted
contents on the second display when the second user manipulation is
sensed.
[0034] Extracting contents may include capturing a screen displayed
on the first display according to the first user manipulation, and
displaying the screen corresponding to the extracted contents may
include displaying the captured screen on the second display on a
location where the second user manipulation is made.
[0035] A non-transitory computer-readable recording medium may
contain computer-readable codes as a program to execute the multi
display method.
[0036] Exemplary embodiments of the present general inventive
concept also provide a multi display apparatus including a
plurality of displays, a sensor configured to sense a first user
manipulation on a first display of the plurality of displays and to
sense a second user manipulation made in a nondecisive location on
at least one second display of the plurality of displays, and a
controller to perform an operation on at least one of the plurality
of displays according to a combination of the sensed first user
manipulation and the sensed second user manipulation.
[0037] The sensor may also be configured to sense a location of
each of the plurality of displays relative to each other.
[0038] The second user manipulation may be made in a nondecisive
location on a plurality of second displays, and the controller may
perform the operation on each of the plurality of second displays
on which the second user manipulation is made.
BRIEF DESCRIPTION OF THE DRAWINGS
[0039] These and/or other features and utilities of the present
general inventive concept will become apparent and more readily
appreciated from the following description of the embodiments,
taken in conjunction with the accompanying drawings of which:
[0040] FIG. 1 is a block diagram illustrating a configuration of a
multi display apparatus according to an exemplary embodiment of the
present general inventive concept;
[0041] FIG. 2 is a flowchart illustrating operations of a multi
display apparatus according to an exemplary embodiment of the
present general inventive concept;
[0042] FIG. 3 is a view illustrating a structured screen according
to an exemplary embodiment of the present general inventive
concept;
[0043] FIG. 4 is a flowchart illustrating an operation of a multi
display apparatus according to an exemplary embodiment of the
present general inventive concept configured to change a function
matching a second user manipulation, when there is a first user
manipulation;
[0044] FIG. 5 is a reference view illustrating a display screen of
an exemplary embodiment of the multi display apparatus of FIG.
4;
[0045] FIG. 6 is a flowchart illustrating another exemplary
embodiment of the multi display apparatus of FIG. 4;
[0046] FIG. 7 is a reference view illustrating a display screen of
an exemplary embodiment of the multi display apparatus of FIG.
6;
[0047] FIG. 8 is a flowchart illustrating another exemplary
embodiment of the multi display apparatus of FIG. 4;
[0048] FIG. 9 is a reference view illustrating a display screen of
an exemplary embodiment of the multi display apparatus of FIG.
8;
[0049] FIG. 10 is a flowchart illustrating another exemplary
embodiment of the multi display apparatus of FIG. 2;
[0050] FIG. 11 is a reference view illustrating a display screen of
an exemplary embodiment of the multi display apparatus of FIG.
10;
[0051] FIG. 12 is a flowchart illustrating operations of a multi
display apparatus according to an exemplary embodiment of the
present general inventive concept;
[0052] FIG. 13 is a flowchart illustrating an exemplary embodiment
of the multi display apparatus of FIG. 12;
[0053] FIG. 14 is a reference view illustrating a display screen of
an exemplary embodiment of the multi display apparatus of FIG.
13;
[0054] FIG. 15 is a flowchart illustrating another exemplary
embodiment of the multi display apparatus of FIG. 12;
[0055] FIG. 16 is a reference view illustrating a display screen of
an exemplary embodiment of the multi display apparatus of FIG.
15;
[0056] FIG. 17 is a flowchart illustrating another exemplary
embodiment of the multi display apparatus of FIG. 12;
[0057] FIG. 18 is a reference view illustrating a display screen of
an exemplary embodiment of the multi display apparatus of FIG.
17;
[0058] FIG. 19 is a flowchart illustrating an operation of a multi
display apparatus according to an exemplary embodiment of the
present general inventive concept;
[0059] FIG. 20 is a flowchart illustrating an exemplary embodiment
of the multi display apparatus of FIG. 19;
[0060] FIGS. 21 to 23 are views illustrating a display screen of an
exemplary embodiment of the multi display apparatus of FIG. 20;
[0061] FIG. 24 is a flowchart illustrating a specific another
exemplary embodiment of the multi display apparatus of FIG. 19;
[0062] FIG. 25 is a reference view illustrating a display screen of
an exemplary embodiment of the multi display apparatus of FIG.
24;
[0063] FIG. 26 is a block diagram illustrating a configuration of a
multi display apparatus according to an exemplary embodiment of the
present general inventive concept;
[0064] FIG. 27 is a block diagram illustrating a hardware
configuration of a controller according to an exemplary embodiment
of the present general inventive concept;
[0065] FIG. 28 is a view illustrating a system hierarchy structure
of a multi display apparatus according to an exemplary embodiment
of the present general inventive concept;
[0066] FIG. 29 is a view illustrating a circuit configuration of an
image outputter of the first display or the second display
according to an exemplary embodiment of the present general
inventive concept;
[0067] FIG. 30 is a view illustrating a circuit structure
configuring R, G, and B pixel areas within the display panel of
FIG. 29;
[0068] FIG. 31 is a cross-sectional view of a first display or
second display including a touch screen according to an exemplary
embodiment of the present general inventive concept;
[0069] FIG. 32 is a block diagram illustrating a display driver of
a multi display apparatus according to an exemplary embodiment of
the present general inventive concept;
[0070] FIG. 33 is a block diagram illustrating a configuration of a
display driver according to another exemplary embodiment of the
present general inventive concept;
[0071] FIG. 34 is a view illustrating a proximate touch according
to an exemplary embodiment of the present general inventive
concept;
[0072] FIG. 35 is a detailed perspective view of a multi display
apparatus according to an exemplary embodiment of the present
general inventive concept;
[0073] FIGS. 36 to 38 are perspective views illustrating states of
a multi display apparatus changed in various ways with reference to
a hinge according to an exemplary embodiment of the present general
inventive concept;
[0074] FIGS. 39 and 40 are perspective views illustrating standing
states of a multi display apparatus according to an exemplary
embodiment of the present general inventive concept;
[0075] FIGS. 41 and 42 are views illustrating an arrangement of two
cameras of a multi display apparatus according to an exemplary
embodiment of the present general inventive concept;
[0076] FIGS. 43 and 47 are flowcharts illustrating a multi display
method according to various exemplary embodiments of the present
general inventive concept; and
[0077] FIG. 48 is a view illustrating an arrangement of displays of
the multi display apparatus wirelessly connected according to
another exemplary embodiment of the present general inventive
concept.
DETAILED DESCRIPTION OF THE EMBODIMENTS
[0078] Reference will now be made in detail to the embodiments of
the present general inventive concept, examples of which are
illustrated in the accompanying drawings, wherein like reference
numerals refer to the like elements throughout. The embodiments are
described below in order to explain the present general inventive
concept while referring to the figures.
[0079] In the present specification, a multi display apparatus is
an apparatus having a plurality of displays including one or more
touch screens, and configured to execute an application or display
contents. Examples include a tablet personal computer (PC),
portable multimedia player (PMP), personal digital assistant (PDA),
smart phone, mobile phone, and digital photo frame, etc. The
exemplary embodiments of the present general inventive concept
described below and illustrated in the drawings are directed to a
multi display apparatus such as a cellular phone or smart phone.
However, it will be understood that the present general inventive
concept is not limited to these exemplary embodiments, and may be
applied to any multi display apparatus.
[0080] FIG. 1 is a block diagram illustrating a configuration of a
multi display apparatus 100 according to an exemplary embodiment of
the present general inventive concept, and FIG. 2 is a flowchart
illustrating an operation of a multi display apparatus 100
according to an exemplary embodiment of the present general
inventive concept.
[0081] With reference to FIG. 1, a multi display apparatus 100
according to an exemplary embodiment of the present general
inventive concept includes a first body 2, a second body 4, a hinge
185, a sensor 150, and a controller 130.
[0082] The first body 2 may include a first display 190a of the
multi display apparatus 100. In addition, as illustrated in FIG. 1,
the first body 2 may include the sensor 150, and the controller 130
to be explained hereinbelow. However, the sensor 150 and controller
130 may alternatively be included in the second body 3. Besides the
above, the first body 2 may further include circuit components (not
illustrated) to drive the first display 190a and a bezel (not
illustrated) to accept the circuit components, but further
configuration and explanation is omitted for convenience of
explanation.
[0083] The second body 4 may include a second display 190b of the
multi display apparatus 100. Besides the above, the second body 4
includes circuit components (not illustrated) to drive the second
display 190b and a bezel (not illustrated) to accept the circuit
components. The circuit components may be included in either one of
the first body 2 or the second body 4, or in both the first body 2
and the second body 4. The first body 2 and the second body 4 are
connected to each other by the hinge 185, to be explained
hereinbelow.
[0084] The first display 190a and the second display 190b may be
referred to as a multi display (dual display) 190.
[0085] The hinge 185 is configured to connect the first body 2 and
the second body 4 to support the first body 2 and the second body 4
so that they are rotatable. More specifically, the hinge 185
combines the first body 2 and the second body 4 physically so that
they could be connected to each other, and plays the role of a
shaft so that the first body 2 and the second body 4 may rotate
while they are connected to each other. An exterior configuration
of the first and second body 4 and the hinge 185 will be explained
hereinbelow with reference to the figures. However, the hinge 185
may be embodied as a part of a flexible connector or flexible touch
screen (not illustrated) besides the structure to be explained
hereinbelow.
[0086] The sensor 150 is configured to sense a user manipulation of
the multi display apparatus 100. For example, when there is a touch
input of a user object 50 (illustrated for example in FIG. 34)
through a display 190a, 190b, the sensor 150 senses a location
(coordinate) of a touch input and a type of input. Especially, when
the first display 190a and the second display 190b display a
nonstructured first screen and second screen, respectively, as
described below with reference to FIG. 2, the sensor 150 senses a
first user manipulation regarding a nondecisive location of the
first screen and a second user manipulation regarding a nondecisive
location of a second screen.
[0087] A structured state refers to a state where various menus and
other objects that a user may select are arranged according to a
predetermined layout. That is, a structured screen is a screen
configured to provide an interface structured by such a layout.
Examples of a structured screen include templates 390 having
various shapes, keypads 391 where numbers or text data could be
input, and screens where tool bars 392, menu buttons 393, scroll
bars 394, etc. are displayed, as illustrated in FIG. 3. A user
becomes able to perform an input by simply touching a button of a
menu at a predetermined location on a screen or by an operation
(for example, dragging their finger) while touching the button. On
the other hand, a nonstructured screen refers to a screen that does
not provide a structured interface. When a nonstructured screen is
displayed, the user may simply drag a nondecisive location of the
screen or may touch for a predetermined time (touch & hold
input) and perform an input. Various types of user input will be
explained hereinbelow in tables 1 to 3.
[0088] Herein, a nondecisive location denotes any location on the
screen. In other words, a nondecisive location means that a same
result can be obtained regardless of the location of an input on
the screen.
[0089] In various exemplary embodiments of the present general
inventive concept, the sensor 150 may sense whether or not a user
object 50 touched or is proximate to the multi display apparatus
100, and the location where the touch or the proximity is made. The
definitions of "touch" and "proximate" for the purposes of the
present general inventive concept are given below. In addition, the
sensor 150 may distinguish a case where the user object 50 is the
user's body and a case where the user object 50 is an input pen
200. Technological means which may support the aforementioned
exemplary embodiments will be explained in further detail
hereinbelow.
[0090] The controller 130 combines a first user manipulation and a
second user manipulation, and performs a control operation
corresponding to a result of that combination. For example, when
there is a touch input in the first display 190a and there is also
a proximate input in the second display 190b, the controller 130
may perform operations corresponding to each of the touch input and
the proximate input, but the controller 130 may also perform a
calculation corresponding to the result of that combination. In
addition, the controller 130 may display information corresponding
to at least one calculation result of the first display 190a and
the second display 190b. Operations of the controller 130 will be
explained in further detail hereinbelow based on each exemplary
embodiment of the present general inventive concept, and the
technological means configuring the controller 130 may be explained
in further detail hereinbelow.
[0091] First of all, various input types to be used in the
exemplary embodiments of the present general inventive concept are
as follows. Herein, inputs using the user object 50 include a
touch, proximity, and motion. The user object 50 may be part of the
user's body such as fingers or palm, or other objects such as an
input pen 200. The input pen 200 may be made from conductive
material such as metal, but it may also be made from other
materials that may be detected by the sensor 150. For example, an
input pen 200 may be configured to include magnetic coil, and the
sensor 150 of the multi display apparatus 100 may include a
magnetic sensor.
[0092] As such, the input pen 200 including the magnetic coil may
also be the user object 50. As such, the user object 50 includes
all types of objects and bodies that may be recognized within
contact or a predetermined range. These various objects may be used
independently or used in combinations.
[0093] A touch denotes an input of contacting the aforementioned
user object 50. For example, it is a user's operation of contacting
one location or a plurality of consecutive locations on a touch
screen 192 (illustrated in FIG. 31) using a user object 50 may
include the user's fingers (especially, index finger) of a left
hand and right hand, thumb, or a stylus pen 200 contacting the
touch screen 192. Technological means to recognize a touch will be
explained hereinbelow.
[0094] Proximity denotes an input of locating a user object 50
within the predetermined range 5 of the sensor 150 of the multi
display apparatus 100 without directly touching or pressing the
multi display apparatus 100. Various exemplary embodiments of the
present general inventive concept presuppose differentiating
between the aforementioned proximity and touch. Technological means
to recognize proximity will be explained in further detail
hereinbelow.
[0095] A motion denotes an input of a predetermined identifiable
movement while proximate to the multi display apparatus 100.
"Motion" may be included in the definition of "proximity," and thus
a detailed explanation is omitted herein.
[0096] Various input types used in the present disclosure are
explained with reference to the table below. The explanation below
is made based on a touch input, but proximate inputs may be based
on the same reference. In addition, as aforementioned, the user
object 50 may be the user's body part or input pen 200.
TABLE-US-00001 TABLE 1 Type Mark Single Tap Finger Touch & Hold
Gesture Double Tap Drag Drag & Drop Flick
[0097] Table 1 illustrates exemplary gesture types which use one
finger. Although hereinbelow is explanation in the case where the
user object 50 is a finger, cases where other user objects 50 are
used are also defined as the same operation.
[0098] With reference to table 1, examples of a gesture type using
one finger include Tap, Touch & Hold, Double Tap, Drag, Drag
& Drop, and Flick etc.
[0099] A tap denotes an operation where the user contacts the user
object 50 to the screen for a predetermined time and then separates
it therefrom. A touch & hold is an operation where the user
object 50 touches the touch screen 192 for more than a
predetermined time. A double tap denotes an operation of performing
a tap twice quickly and consecutively within a predetermined time.
A drag is an operation of moving to a predetermined direction while
a touch is made, and a drag & drop is an operation of touching
any object on the touch screen 192 and performing a drag to a
predetermined location and then separating the user object 50
therefrom. A flick is an operation of quickly dragging.
TABLE-US-00002 TABLE 2 Type Mark Two Two finger Tap Finger Touch
& Spread Gesture Pinch Out Pinch In Two finger Drag Cross Two
Finger Touch & Rotate
[0100] Table 2 illustrates exemplary gesture types using two
fingers. Although hereinbelow is explanation in the case where the
user object 50 is a finger, cases where other user objects 50 are
used are also defined as the same operation.
[0101] With reference to table 2, gesture types using two fingers
include a Two Finger Tap, Touch & Spread, Pinch Out, Pinch In,
Two Finger Drag, Cross Two Finger, and Touch & Rotate etc.
[0102] A Two Finger Tap is an operation of two fingers tapping at
the same time. A Touch & Spread is an operation of two fingers
pressing the touch screen 192 at the same time, one finger moving,
and the other finger moving in a straight line. A Pinch Out is an
operation of two fingers touching the screen at the same time and
then dragging away from each other, while a Pinch In is an
operation of two fingers touching the touch screen 192 and then
dragging in a direction towards each other. A Two Finger Drag is an
operation of two fingers dragging in a same direction, a Cross Two
Finger is an operation of dragging in an approaching direction at
the same time and then dragging away from each other again.
Finally, a Touch & Rotate is an operation with one finger
touching the touch screen 192 (touch & hold) and the other
rotating around the still finger.
TABLE-US-00003 TABLE 3 Type Mark Multi Finger Three Finger Touch
Gesture Four Finger Touch Five Finger Touch Palm Palm
[0103] Table 3 illustrates exemplary gesture types using two or
more fingers and a gesture type using a palm.
[0104] With reference to table 3, examples of a gesture type using
two or more fingers include a Three Finger Touch, Four Finger
Touch, and Five Finger Touch etc. In addition, it is possible to
perform gesture operations such as a Tap, Drag, and Rotate, as
illustrated in the aforementioned tables 1 and 2.
[0105] The sensor 150 identifies each input type of the
aforementioned tables 1 to 3, and transmits the identification
information to the controller 130. Each aforementioned input type
may be sensed independently and be transmitted to the controller
130. The controller 130 may process calculations corresponding to
combination results of each input type or a plurality of input
types sensed by the sensor and display information corresponding to
the calculation results on the first display 190a or second display
190b.
[0106] Operations of the aforementioned multi display apparatus 100
may be performed according to the flowchart illustrated in FIG.
2.
[0107] First of all, the first display 190a and second display 190b
each displays the nonstructured first screen and second screen
(operation S210). In this state, the multi display apparatus 100
may sense a first user manipulation made in a nondecisive location
of the first screen and a second user manipulation made in a
nondecisive location of a second screen (operations S220, S230). In
FIG. 2, it is explained that the first user manipulation is made
first and then the second user manipulation is made next, but the
first and second user manipulations may be made independently and
sensed. When the first and second user manipulations are each
sensed, the multi display apparatus 100 combines the sensed first
user manipulation and second user manipulation (operation S240),
and performs one or more operations corresponding to that
combination result (operation S250). Operations performed by the
combination of the user manipulations may be embodied in various
ways according to exemplary embodiments of the present general
inventive concept.
[0108] Hereinbelow is explanation on various examples of operations
performed by the first and second user manipulations made in
nondecisive locations.
[0109] First of all, hereinbelow is explanation of an exemplary
embodiment of changing a function matching a second user
manipulation, when there is a first user manipulation through the
first display 190a of the multi display apparatus 100.
[0110] In a case of performing a user manipulation on a display to
perform a particular operation, it is necessary to set a function
matching the user manipulation. For example, in a case of
performing an operation of drawing a line according to a user
manipulation on a screen of a display, there may be a need to
adjust the thickness of the line according to circumstances.
Herein, it is possible to use a menu screen to adjust the thickness
of the line through a structured interface and adjust the thickness
of the line. However, such an operation is cumbersome, and may
cause inconvenience if there is frequent changing of functions.
Therefore, there is a need to reduce the setting process for user
manipulation.
[0111] FIG. 4 is a flowchart illustrating operations of the multi
display apparatus 100 changing functions matching the second using
manipulation when there is a first user manipulation, and FIG. 5 is
a reference view illustrating the display screen of the multi
display apparatus 100 of FIG. 4.
[0112] With reference to FIG. 4, first of all, the sensor 150
senses a point touched in the first display 190a or the second
display 190b and notifies the controller 130. The controller 130
determines whether a pinch in or pinch out is made according to the
number and location of the touch point (operation S410). The
controller 130 determines a pinch out when a touch is sensed at two
or more points and the straight line distance between the sensed
coordinates increases, and determines a pinch out when the straight
line distance between the sensed coordinates decreases. The pinch
in and pinch out may be the first user manipulation
aforementioned.
[0113] The controller 130 may perform operations corresponding to
the determined pinch input. That is, as illustrated in FIG. 4, it
is possible to adjust the thickness of the line or change the
transparency or other characteristics according to the pinch
input.
[0114] In this state, when a drag is made in a nondecisive location
of another display, through the sensor 150 the controller 130 may
sense a touch input by a drag defining a drag route (operation
S420). Herein, the drag may be the aforementioned second user
manipulation.
[0115] The controller 130 may recognize, according to the detected
pinch in or pinch out of the first user input, whether a command to
adjust characteristics of the line, such as thickness or
transparency, has been input (operation S430). The controller 130
performs an operation of displaying a line along the drag route
according to the converted characteristics (operation S440).
[0116] View (1) of FIG. 5 is a view illustrating a state of
adjusting the thickness of a line.
[0117] According to view (1) of FIG. 5, the first display 190a,
located on the left, displays the nonstructured first screen, and
the sensor 150 senses the first user manipulation at the
nondecisive location of the first screen. Such an operation is
performed by a touch sensor, but this will be explained in further
detail below. In addition, the second display 190b, located on the
right, also displays the nonstructured second screen. It will be
understood that the first and second displays 190a and 190b are not
necessarily located on the left and right as illustrated in FIG. 5,
and may have different orientations to each other depending on the
configuration of the multi display apparatus 100.
[0118] As illustrated in FIG. 5, the user may perform a touch input
such as a pinch in or pinch out using fingers in the first display
190a. At the same time, the user may perform a touch input using an
input pen 200 or one's fingers, or other user objects 50 on the
second display 190b. For convenience of explanation, FIG. 5
illustrates a case of an input pen 200, and also in various
exemplary embodiments of the present general inventive concept to
be explained below, explanation is based on a limitation that
manipulations are made using an input pen 200 regarding one
display, and using the user's body regarding the other display.
However, as aforementioned, manipulations may obviously be made
using only the input pen 200 or only the user's body regarding both
displays. Explanation hereinbelow will be made in consideration of
specific situations and convenience of explanation, but the
exemplary embodiments of the present general inventive concept may
be performed in all types of manipulation methods.
[0119] In the exemplary embodiment of view (1) of FIG. 5, when
there is a touch input by a pinch in using fingers on the first
display 190a, the controller 130 displays a line according to the
changed thickness along the input route of the input pen 200 (or
user's body) regarding the second display 190b. Therefore, a
gradually thinner line is displayed along the drag route of the
input pen 200 on the second display 190b. On the other hand, when
there is a touch input by a pinch out using fingers on the first
display 190a, the controller 130 displays a line according to the
changed thickness along the input route of the input pen 200
regarding the second display 190b. Accordingly, a gradually thicker
line is displayed along the drag route of the input pen 200 on the
second display 190b.
[0120] View (2) of FIG. 5 illustrates an exemplary embodiment of
the present general inventive concept similar to (view 1),
illustrating operations of the multi display apparatus 100 of
changing the transparency of the line output by the second user
manipulation when the first user manipulation is a touch input by a
pinch in or pinch out. For example, in a case where the first user
manipulation is a touch input by a pinch in, when performing an
operation of drawing a line on the second display 190b with the
input pen 200, the transparency of the line being drawn by the
input pen 200 decreases. On the other hand, when the first user
manipulation is a touch input by a pinch out, the transparency of
the line being drawn by the input pen 200 increases. Of course, the
outputs corresponding to the pinch in and pinch out may be set to
be opposite to each other from this particular exemplary embodiment
of the present general inventive concept.
[0121] In this exemplary embodiment of the present general
inventive concept, the user becomes able to change the thickness or
transparency of an input expression regarding the second display
190b with simply a touch input by a pinch in or pinch out into the
first display 190a of the multi display apparatus 100 only. That
is, there is no need to perform numerous manipulations using an
additional structured interface in order to change the functions
matching the second user manipulation. By performing an input at a
certain point on a nonstructured screen, the user becomes able to
easily change the function corresponding to the user
manipulation.
[0122] FIG. 6 is a flowchart illustrating another exemplary
embodiment of the multi display apparatus 100 of FIG. 4, and FIG. 7
is a reference view illustrating a display screen of an exemplary
embodiment of the multi display apparatus 100 of FIG. 6.
[0123] With reference to FIG. 6, the sensor 150 senses a first user
manipulation which is maintained for a predetermined time with an
initial touch made in the first display 190a and the touch is not
removed, that is, a touch input by a touch & hold (operation
S610). That is, the display 190a displays the nonstructured first
screen, and the sensor 150 senses that there is a touch input by a
touch & hold at a nondecisive location of the first screen.
Such an operation may be performed by a touch sensor, but this will
be explained in more detail below.
[0124] After the first user manipulation is sensed, a second user
manipulation may be performed on the second display 190b while the
first user manipulation is maintained. This second user
manipulation may be a touch input by a drag on the second display
190b (operation S620). This drag operation defines a drag route on
the second display 190b.
[0125] In this case, the controller 130 recognizes that the first
user manipulation is to convert the tools corresponding to the
second user manipulation, and that the second user manipulation is
an operation to perform an input along the drag route (operation
S630). In addition, the controller 130 performs an input along the
drag route according to the changed tool (operation S640). In the
exemplary embodiment of the present general inventive concept
illustrated in FIG. 7, when there is a touch input by a touch &
hold using fingers in the first display 190a, the controller 130
converts the tool corresponding to the input of the input pen 200
regarding the second display 190b into an eraser. In addition, the
controller 130 recognizes that there is an eraser input along the
drag route of the input pen 200 on the second display 190b, and
deletes the display on the route.
[0126] Unlike the above, when the user stops a touch & hold
input regarding the first display 190a, that is, when the user
detaches one's hand from the first display 190a, the sensor 150
does not sense the first user manipulation any more, but senses
only the second user manipulation, and thus the controller 130
performs only operations corresponding to the second user
manipulation. As an example, in the exemplary embodiment of the
present general inventive concept illustrated in FIG. 4, the tool
corresponding to the input of the input pen 200 returns from an
eraser to the original tool again, and an input is performed
according to the original tool.
[0127] In such an exemplary embodiment of the present general
inventive concept, the user becomes able to change the input tool
regarding the second display 190b with only a touch input (such as
for example a touch & hold) in the first display 190a. That is,
it is not necessary to perform numerous manipulations using an
additional structured interface in order to change the function
matching the second user manipulation. The user becomes able to
easily change the function corresponding to the user manipulation
by performing an input at any point on the nonstructured
screen.
[0128] FIG. 8 is a flowchart illustrating another exemplary
embodiment of the multi display apparatus 100 of FIG. 4. FIG. 9 is
a reference view illustrating a display screen of an exemplary
embodiment of the multi display apparatus 100 of FIG. 8.
[0129] With reference to FIG. 8, first of all, the sensor 150
senses the first user manipulation where a touch input by rotation,
that is where an initial touch is made in the first display 190a
and a drag is made by a rotations in one direction, corresponding
to a touch & rotate (operation S810). That is, the first
display 190a displays the nonstructured first screen, and the
sensor 150 senses that there is a touch input by a touch &
rotate at a nondecisive location of the first screen.
[0130] After the first user manipulation is sensed, a second user
manipulation of a touch input by a drag regarding the second
display 190b is sensed (operation S820). Herein, the touch input by
the drag may be made by the input pen 200. In the exemplary
embodiment of the present general inventive concept illustrated in
FIG. 9, a drag input is performed in the second display 190b using
the input pen 200. The drag input defines a drag route on the
second display 190b.
[0131] In this case, the controller 130 recognizes that the first
user manipulation is to convert the texture of the line displayed
by the second user manipulation, and recognizes that the second
user manipulation is an operation to perform an input along the
drag route. In addition, the controller 130 combines both user
manipulations and recognizes the input of a line, the texture of
which is changed along the drag route (operation S830). The line is
then displayed according to the changed texture along the drag
route (operation S840).
[0132] Herein, texture denotes characteristics of the material
being displayed. For example, in the case of an embossing, the
texture denotes a case where protruded parts and recessed parts
exist in turns so that each part is shown as an emboss with
protruded parts and an engrave with recessed parts. Another example
may be expressions of a touch of a thin brush or a touch of a thick
brush.
[0133] As in FIG. 9, when the first user manipulation is a touch
input by a rotation in one direction (for example, in a clockwise
direction), the texture of the line or dot displayed on the second
display 190b by the input pen 200 may be changed (operation S840).
On the other hand, when the first user manipulation is a touch
input by a rotation in another direction (for example, a
counterclockwise direction), the texture of the line or dot
displayed by the input pen 200 may return to the original state or
be changed to another shape. That is, when there is a first user
manipulation of a rotation in one direction, the line or dot
displayed on the second display 190b by the input pen 200 may be
expressed as a watercolor painting brush, and when there is a first
user manipulation of a rotation in another direction, the line or
dot may return to the original state, or may be expressed to show a
more precise brush. However, this is merely an exemplary
embodiment, and thus more various textures may be expressed.
[0134] FIG. 10 is a flowchart specifically illustrating a tool menu
display of a multi display apparatus 100 according to an exemplary
embodiment of the present general inventive concept, and FIG. 11 is
a reference view illustrating a display screen of the multi display
apparatus 100 of FIG. 10.
[0135] In the case of performing an operation of drawing a picture
on the multi display apparatus 100 through a touch input, or
selecting a certain tool and performing an operation, there may be
a need to change the setting of the tool. FIGS. 10 and 11
illustrate methods of easily displaying a menu of the tool in such
cases.
[0136] With reference to FIG. 10, first the sensor 150 senses a
first user manipulation where a touch input by a touch & hold,
that is, an initial touch is made in the first display 190a and the
touch is maintained for a predetermined time and the touch is not
removed (operation S1010). The first display 190a displays the
nonstructured first screen, and the controller 130 senses that
there is a touch input by a touch & hold at a nondecisive
location of the first screen using the sensor 150 (operation
S1010). While the touch & hold is maintained, a second user
manipulation, such as a touch input by a tap on the second display
190b, is sensed (operation S1020). That is, the second display 190b
also displays the nonstructured second screen, and the sensor 150
senses that there is a touch input at a nondecisive location of the
second screen. Herein, the touch input by the tap may be made by
the input pen 200, as illustrated in FIG. 11.
[0137] In this case, the controller 130 may display a menu 1110
(operation S1030), which may change the setting of the tool, in
response to a touch & hold input performed on the first screen
190a and a second user manipulation on the second screen 190b.
Furthermore, when a first user manipulation by a touch & hold
is performed on the first screen 190a while the menu 1110 to set
the tool is displayed, and there is a second user manipulation such
as a touch on the second display 190b, the already displayed menu
1110, which may change the setting of the tool, disappears from the
screen.
[0138] The above process is illustrated in FIG. 11. To be more
specific, when the user performs a touch & hold input on the
first display 190a with their hand while drawing a picture on the
multi display apparatus 100, and touches the second display 190b
with the input pen 200 (view 1), the menu 1110, which may set the
tool corresponding to the input pen 200, is displayed on the second
display 190b (view 2). The user may perform a setting of the tool
corresponding to the input pen 200 through the menu 1110. For
example, if the tool corresponding to the input pen 200 is a color
pencil, the user may adjust the color and thickness of the color
pencil. In addition, when the user performs a touch & hold
input on the first display 190a again with their hand, and then
touches the second display 190b with the input pen 200 (view 3),
the menu 1110 disappears from the second display 190b (view 4).
[0139] As such, it is possible to grant functions to the input pen
200 and use the functions without complicated option setting
operations.
[0140] Hereinbelow is an explanation on various operations of the
multi display apparatus 100 which changes a display of the second
display 190b according to a first user manipulation when an input
is sensed through the first display 190a of the multi display
apparatus 100.
[0141] FIG. 12 is a flowchart illustrating operations of the multi
display apparatus 100 according to an exemplary embodiment of the
present general inventive concept.
[0142] With reference to FIG. 12, the sensor 150 of the multi
display apparatus 100 senses the first user manipulation on the
first display 190a (operation S1210). That is, with the first
display 190a displaying the nonstructured first screen, the user
may touch a nondecisive location of the first screen and input a
first user manipulation. When a touch point is sensed by the sensor
150, the controller 130 changes the screen displayed on the second
display 190b according to the sensed first user manipulation and
displays the changed screen (operation S1220). A screen changing
operation may be embodied in various ways according to the type of
the first user manipulation. For example, depending on how the
first user manipulation is made, an operation of returning to the
previous screen, an operation of converting into the next screen,
and an operation of returning to the main screen may be selectively
performed.
[0143] In this state, when the user performs a second user
manipulation of touching the second display 190b, the sensor 150
senses the second user manipulation (operation S1230), and the
controller 130 changes the screen which has been changed by the
first user manipulation and then displayed according to the second
user manipulation, and displays the changed screen on the second
display 190b (operation S1240).
[0144] Hereinbelow is more detailed explanation on the exemplary
embodiment of the present general inventive concept illustrated in
FIG. 12. FIG. 13 is a flowchart illustrating a specific exemplary
embodiment of the multi display apparatus 100 of FIG. 12, and FIG.
14 is a reference view illustrating a display screen of an
exemplary embodiment of the multi display apparatus 100 of FIG.
13.
[0145] With reference to FIG. 13, when the sensor 150 of the multi
display apparatus 100 senses the first user manipulation such as a
drag on the first display 190a (operation S1310), the controller
130 changes the screen displayed on the second display 190b to the
previous operated screen according to the sensed first user
manipulation and displays the changed screen (operation S1320).
This may correspond for example to a scrolling command, moving
sequentially back through pictures which have been previously
displayed. In this state, when a second user manipulation such as a
tap is sensed in the second display 190b by the sensor 150
(operation S1330), the controller 130 changes the previously
operated screen being displayed on the second display 190b, the
change being according to the second user manipulation, and
displays the changed screen (operation S1340).
[0146] With reference to FIG. 14, a picture is displayed on the
second display 190b (view 1). Inputs may be performed on this
picture with a user object 50 such as an input pen 200. Such inputs
may perform operations such as modifying the picture. When a touch
input by a drag is performed in a downward direction on the first
display 190a (view 2), a previous picture from among the pictures
stored in the second display 190b is displayed (view 3). In
addition, a new input may be performed with an input pen 200 on the
second display 190b regarding the displayed picture, and perform an
operation such as modifying the picture.
[0147] Furthermore, when a touch input by a drag in an upwards
direction is performed on the first display 190a (view 4), a
picture stored after the present picture is displayed on the second
display 190b (view 5). In the exemplary embodiment of the present
general inventive concept illustrated in FIG. 5, such an upward
drag on the first display 190a returns the second display 190b to
the original picture illustrated in view 1. There may be various
touch input directions by the drag performed by the first display
190a. For example, as an alternative configuration, when there is a
touch input by a drag in the left direction (not illustrated) on
the first display 190a, the previous picture from among the
pictures stored in the second display 190b is displayed, and when
there is a touch input by a drag in the right direction (not
illustrated) on the first display 190a, the picture after the
picture displayed on the second display 190b from among the stored
pictures may be displayed.
[0148] FIG. 15 is a flowchart illustrating another exemplary
embodiment of the multi display apparatus 100 of FIG. 12, and FIG.
16 is a reference view illustrating a display screen of an
exemplary embodiment of the multi display apparatus 100 of FIG.
15.
[0149] With reference to FIG. 15, when the sensor 150 of the multi
display apparatus 100 first senses the first user manipulation
performing a touch input by a pinch in or pinch out at a certain
point on the first display 190a (operation S1510), the controller
130 reduces or expands the screen displayed on the second display
190b according to the sensed first user manipulation and displays
the reduced or expanded screen (operation S1520). More
specifically, when the first user manipulation is a pinch out
input, the controller 130 may expand the screen displayed on the
second display 190b according to the direction and length of the
drag made by the pinch out input and displays the expanded screen,
and when the first user manipulation is a pinch in input, the
controller 130 may reduce the screen displayed on the second
display 190b according to the direction and length of the drag made
by the pinch out input and displays the reduced screen. In
addition, the controller 130 senses the second user manipulation
performing a touch input by a tap on the second display 190b
(operation S1530), and in this case, the controller 130 changes the
reduced or expanded screen displayed on the second display 190b
according to the second user manipulation result and displays the
changed screen (operation S1540).
[0150] In the example illustrated in FIG. 16, when a touch input by
a pinch out is performed on the first display 190a (view 1), for
example in a horizontal direction, the picture displayed on the
second display 190b is expanded to correspond to the direction and
length of the drag made by the pinch out and is displayed (view 2).
In addition, the user becomes able to perform a new input on the
second display 190b regarding the expanded and displayed picture
with the input pen 200, such as performing an operation of
modifying the picture. On the other hand, when a touch input by a
pinch in is performed on the first display 190a (view 3), the
picture displayed on the second display 190b is reduced and then
displayed (view 4).
[0151] FIG. 17 is a flowchart specifically illustrating another
exemplary embodiment of the multi display apparatus 100 of FIG. 12,
and FIG. 18 is a reference view illustrating a display screen of an
exemplary embodiment of the multi display apparatus 100 of FIG.
17.
[0152] With reference to FIG. 17, when the sensor 150 of the multi
display apparatus 100 first senses the first user manipulation
which performs a touch input by a touch & rotate on the first
display 190a (operation S1710), the controller 130 rotates the
screen displayed on the second display 190b according to the sensed
first user manipulation and displays the rotated screen (operation
S1720). More specifically, when the first user manipulation is a
touch & rotate input, the controller 130 controls so that the
screen displayed on the second display 190b is rotated and
displayed according to the rotate direction and length based on the
point of the touch. In addition, the controller 130 senses the
second user manipulation such as performing a touch input by a tap
by the second display 190b (operation S1730), and in this case, the
controller 130 changes the rotated screen on the second display
190b according to the second user manipulation result and displays
the changed screen (operation S1740).
[0153] In the example of FIG. 18, when a touch & rotate input
is performed on the first display 190a, or when there is a rotate
input based on the predetermined location at the touch point (views
1 and 3, respectively illustrating clockwise and counterclockwise
rotation), the picture displayed on the second display 190b rotates
corresponding to the rotation direction and length of the touch
& rotate input based on the point corresponding to the touch
point, and is displayed on the second display 190b (views 2 and 4,
respectively corresponding to the rotation illustrated in views 1
and 3). Regarding the rotated and displayed picture, the user may
perform a new input with an input pen 200 on the second display
190b, becoming able to perform operations such as zooming in on or
modifying the picture. In this case, since the picture prior to
rotation is displayed on the first display 190a, the user becomes
able to perform the desired operation on the picture rotated and
displayed on the second display 190b while checking the shape and
color of the entire picture. Of course, the picture displayed on
the first display 190a may also be rotated as well, depending on
the configuration of the particular embodiment of the present
general inventive concept.
[0154] Hereinbelow is explanation of an exemplary embodiment of
extracting contents from the first display 190a by the first user
manipulation and displaying the extracted contents on the second
display 190b by the second user manipulation, according to the
present general inventive concept.
[0155] FIG. 19 is a flowchart illustrating operations of a multi
display apparatus 100 according to an exemplary embodiment of the
present general inventive concept.
[0156] With reference to FIG. 19, the sensor 150 of the multi
display apparatus 100 first senses the first user manipulation on
the first display 190a (operation S1910). That is, the first
display 190a displays the nonstructured first screen, and the
sensor senses that there is a first user manipulation at a
nondecisive location of the first screen.
[0157] The controller 130 extracts contents displayed on the first
display 190a according to the sensed first user manipulation
(operation S1920). In addition, the sensor 150 senses a second user
manipulation on the second display 190a (operation S1930). That is,
the second display 190b displays the nonstructured second screen,
and the sensor 150 senses that there is a second user manipulation
at a nondecisive location of the second screen. The controller 130
displays the screen on the second display 190b corresponding to the
extracted contents from the first display 190a according to the
second user manipulation (operation S1940).
[0158] Hereinbelow is specific explanation on the exemplary
embodiment of the present general inventive concept illustrated in
FIG. 19.
[0159] FIG. 20 is a flowchart illustrating an exemplary embodiment
of the multi display apparatus 100 of FIG. 19, and FIGS. 21 to 23
are reference views of a display screen of an exemplary embodiment
of the multi display apparatus 100 of FIG. 20.
[0160] With reference to FIG. 20, first the sensor 150 senses a
first user manipulation where a touch input by a touch & hold,
that is, an initial touch is made on the first display 190a, and
the touch is maintained for a predetermined time (operation
S2010).
[0161] When the first user manipulation is sensed, the controller
130 captures the screen displayed on the first display 190a
(operation S2020).
[0162] After the first user manipulation is sensed, the sensor 150
senses a second user manipulation where a touch input by a double
tap or drag on the second display 190b is made (operation S2030).
Herein, a touch input by the double tap or drag may be made by an
input pen 200.
[0163] In this case, the controller 130 displays the screen
captured from the first display 190a or a screen related thereto on
the second display 190b where a touch is made by the second user
manipulation (operation S2040).
[0164] In the exemplary embodiment illustrated in FIG. 21, a screen
is displayed on the first display 190a and a second screen is
displayed on the second display 190b (view 1). When a point is
touched on the first display 190a and a predetermined time has
passed, such that a touch & hold input is detected, the screen
of the first display 190a is captured (view 2). When a touch input
by a double tap is made on the second display 190b, for example
with the input pen 200 (view 3), a memo 2110 including an image of
the screen captured at the touch point of the first display 190a is
inserted and displayed on the second display 190b (view 4). The
user may perform operations such as performing an input of writing,
storing, or uploading necessary contents together with the captured
screen in the memo 2110, for example with the input pen 200.
[0165] In the exemplary embodiment illustrated in FIG. 22, a screen
is displayed on the first display 190a and a second screen is
displayed on the second display 190b (view 1). When a point is
touched on the first display 190a and a predetermined time has
passed, such that a touch & hold input is detected (view 2),
the screen of the first display 190a is captured. When a touch
input by a drag is made in the second display 190b with the input
pen 200 (view 3), the image regarding the screen captured at the
touch point is uploaded to the message input window L 2210 (view
4).
[0166] In the exemplary embodiment illustrated in FIG. 23, a screen
is displayed on the first display 190a and a second screen is
displayed on the second display 190b (view 1). When a point is
touched in the first display 190a and a predetermined time has
passed, such that a touch & hold input is detected (view 2),
the screen of the first display 190a is captured. When a touch
input by a drag is made on the second display 190b with an input
pen 200 (view 3), an image similar to the captured image is
searched and a search result screen is displayed on the second
display 190b (view 4).
[0167] FIG. 24 is a flowchart specifically illustrating another
exemplary embodiment of the multi display apparatus 100 of FIG. 19,
and FIG. 25 is a reference view illustrating a display screen of an
exemplary embodiment of the multi display apparatus 100 of FIG. 24.
More specifically, FIGS. 24 and 25 illustrate an exemplary
embodiment of the present general inventive concept in which the
first user manipulation on the first display 190a and the second
user manipulation on the second display 190b are followed by a
third user manipulation.
[0168] With reference to FIG. 24, first the sensor 150 senses each
of a first user manipulation and a second user manipulation. Each
of the first and second user manipulations may correspond to a
touch input on the first display 190a and the second display 190b
by a touch & hold, that is, where an initial touch is made on
each of the first display 190a and the second display 190b, and
each touch is maintained for a predetermined time (operation
S2410).
[0169] When the first and second user manipulations are sensed, the
controller 130 captures each screen displayed on the first display
190a and the second display 190b (operation S2420). When the first
and second user manipulations are sensed, the controller 130 may
perform a capture operation corresponding to each of the first and
second user manipulations independently. In this case, the second
display 190b may display the nonstructured second screen, and the
second user manipulation may be made in a nondecisive location of
the second screen. Herein, the screens displayed in the first
display 190a and the second display 190b are both captured as part
of operation S2420.
[0170] After the first and second user manipulations are sensed,
the sensor 150 senses a third user manipulation where a touch input
by a double tap regarding the second display 190b is made
(operation S2430). Herein, a touch input by the double tap or drag
may be made by the input pen 200.
[0171] When there exists a third user manipulation unlike in the
aforementioned exemplary embodiment of the present general
inventive concept, the controller 130 displays the two captured
screens on a point on the second display 190b where the third user
manipulation is made (operation S2440).
[0172] In the exemplary embodiment illustrated in FIG. 25, a screen
is displayed on the first display 190a and a second screen is
displayed on the second display 190b (view 1). When a point is
touched on the first display 190a and a predetermined time passed,
such that a touch & hold input is detected, the screen of the
first display 190a is captured. Independently from this, when a
certain point is touched in the second display 190b and a
predetermined time passed, the screen of the second display 190b is
captured (view 2). When a touch input by a double tap is made in
the second display 190b with the input pen 200 (view 3), a memo
2510 is inserted at the touch point, including an image
corresponding to each screen captured (view 4). The user may
perform operations of performing inputting necessary contents
together with the capture screen, or storing or uploading the
contents to a sns (Social Network Service) together with the
capture screen with the input pen 200.
[0173] When there is a third user manipulation after the
aforementioned first user manipulation, the third user manipulation
may be differentiated from the second user manipulation. That is,
since there is a possibility that a touch corresponding to the
second user manipulation may be determined as the third user
manipulation, when a touch input regarding the second display 190b
is maintained for a predetermined time, the touch input may be
determined as the touch for the second user manipulation, and when
the touch input ends before the predetermined time passes, the
touch input may be determined as an input for the third user
manipulation.
[0174] When the first user manipulation and the second user
manipulation are input at the same time, unlike in the exemplary
embodiment of the present general inventive concept illustrated in
FIG. 25, the user manipulations may be recognized as a single user
manipulation. That is, when a touch input by a touch & hold is
made on the second display 190b at the same time as the touch &
hold input is made on the first display 190a, all the display
screens having a touch & hold by the user manipulation may be
captured and the aforementioned operations may be performed.
[0175] Meanwhile, in the aforementioned exemplary embodiments of
the present general inventive concept illustrated in FIGS. 2 to 25,
the controller 130 may connect the sensed first user manipulation
and the sensed second user manipulation and recognize them as one
user gesture, and may display the execution screen corresponding to
the user gesture on at least one of the first display 190a and the
second display 190b.
[0176] That is, in this case, the controller 130 senses the first
user manipulation and the second user manipulation, but the
controller 130 recognizes these as a single manipulation and
performs operations corresponding to this single manipulation,
instead of performing operations corresponding to each of the first
and second manipulations. In addition, the controller 130 may
perform each of the operations corresponding to the separate first
and second user manipulations when recognizing the first and second
user manipulations as a single manipulation.
[0177] For example, in an exemplary embodiment of the present
general inventive concept where the first user manipulation changes
the functions corresponding to the second user manipulation, the
controller 130 may not change the functions corresponding to the
second user manipulation solely in response to the first user
manipulation. Instead, when there is a second user manipulation
after the first user manipulation, the controller 130 may then
change the functions corresponding to the second manipulation, and
operations according to the changed functions.
[0178] As such, the controller 130 may recognize one user
manipulation according to the result of combining the first user
manipulation, second user manipulation, and third user
manipulation, and may control to perform operations corresponding
to the user manipulation according to the combination results.
[0179] Hereinbelow is explanation on the detailed configuration of
the multi display apparatus 100 which may be embodied differently
according to various exemplary embodiments of the present general
inventive concept.
[0180] FIG. 26 is a block diagram illustrating a configuration of a
multi display apparatus 100 according an exemplary embodiment of
the present general inventive concept.
[0181] With reference to FIG. 26, the multi display apparatus 100
of the exemplary embodiment of the present general inventive
concept includes a communication unit 110, multimedia unit 120,
controller 130, imaging unit 140, sensor 150, input/output unit
160, storage unit 170, power unit 180, the first display 190a, and
the second display 190b.
[0182] The communication unit 110 is configured to transmit and
receive information to and from an external apparatus (not
illustrated) using various wired/wireless communication methods.
Herein, the external apparatus may include at least one of another
multi display apparatus, mobile phone, smart phone, tablet PC
(Personal Computer), computer server, and digital TV. These
apparatuses may be connected to the multi display apparatus 100
through a communication network (not illustrated).
[0183] The communication unit 110 may include a connector 114 which
includes at least one of wireless communication modules such as for
example a cellular communication module 111, wireless LAN module
112, short distance communication module 113, GPS communication
module 115, and broadcast communication module 116. The
communication unit may also include wired communication modules
(not illustrated), such as for example an HDMI (High-Definition
Multimedia Interface), USB (Universal Serial Bus), IEEE (Institute
of Electrical and Electronics Engineers) 1394, and so on.
[0184] The cellular communication module 111 uses wireless access
technology according to a cellular communication protocol by a
control of the controller 130 so that the multi display apparatus
100 may be connected to an external apparatus (not illustrated),
such as a base station of the cellular system, through at least one
or a plurality of antennas (not illustrated).
[0185] In addition, the cellular communication module 111
transmits/receives wireless signals containing voice calls, video
calls, Short Messaging Services (SMS), messages or multimedia
(Multimedia Messaging Service: MMS) messages with compatible
apparatuses such as a mobile phone having telephone numbers input
into the multi display apparatus 100, a smart phone, a tablet PC,
or other apparatuses.
[0186] The wireless LAN module 112 is configured to access a
wireless AP (access point, not illustrated) existing within a
predetermined range and of be connected to the internet according
to the control of the controller 130. The wireless LAN module 112
may support for example the wireless LAN standard (IEEE802.11x) of
the IEEE.
[0187] The short distance communication module 113 is configured to
perform a short distance communication wirelessly between the multi
display apparatus 100 and the external apparatus according to the
control of the controller 130. The short distance communication
module 113 may include for example at least one of a Bluetooth
module, an IrDA (infrared data association) module, a NFC (Near
Field Communication) module, a Wi-Fi module, and a Zigbee
module.
[0188] As such, the communicator 110 may be embodied by the
aforementioned various short distance communication methods. Other
communication technologies not mentioned herein may be adopted when
necessary, according to the configuration of the particular
embodiment of the present general inventive concept.
[0189] The connector 114 is configured to provide an interface with
various apparatuses such as for example USB 2.0, USB 3.0, HDMI,
IEEE 1394, and so on.
[0190] The connector 114 may be used as an interface to connect the
multi display apparatus 100 and the external apparatus or a power
source (not illustrated). By the control of the controller 130,
data stored in the storage unit 170 of the multi display apparatus
100 may be transmitted to an external apparatus (not illustrated)
or data may be received from the external apparatus. Through for
example a wired cable connected to the connector 114, power may be
input from a power source (not illustrated) or may be charged in a
battery (not illustrated).
[0191] The GPS module 115 may receive electric waves from a
plurality of GPS satellites (not illustrated) in the Earth orbit,
and may calculate the location of the multi display apparatus 100
using Time of Arrival and GPS parameters from the GPS satellites
(not illustrated) to the multi display apparatus 100.
[0192] The broadcast communication module 116 may receive broadcast
signals (for example, TV broadcast signals, radio broadcast signals
or data broadcast signals) and broadcast additional information
(for example, EPS (Electric Program Guide) or ESG (Electric Service
Guide)) transmitted from a broadcasting station through broadcast
communication antennas (not illustrated) according to a control by
the controller 130.
[0193] The multimedia unit 120 is a configurative element to play
video contents, audio contents, and other various multimedia
contents. For the purposes of this application, "playing" contents
denotes executing and displaying the contents for a user of the
multi display apparatus 100. It may also denote playing or
replaying the contents. The multimedia unit 120 processes
multimedia contents using a parser or codec etc. to play contents.
The multimedia unit 120 may include for example an audio module 121
and a video module 122.
[0194] The audio module 121 may play digital audio files (for
example, files having mp3, wma, ogg or way as file extensions)
stored or received according to the control of the controller 130.
The video module 123 may play digital video files (for example,
files having mpeg, mpg, mp4, avi, mov, or mkv as file extensions)
stored or received according to the control of the controller
130.
[0195] The video module 122 supports codecs of various forms to
play the digital video files. That is, the video module 122 plays a
video file by the codec prestored in a format suitable to the
format of the video file to be played. In addition, the audio
module 122 or video module 123 of the multimedia unit 120 may be
included in the controller 130.
[0196] The controller 130 is configured to control the
communication unit 110, multimedia unit 120, imaging unit 140,
sensor 150, input/output unit 160, storage unit 170, power unit
180, first display 190a, and second display 190b. As illustrated in
FIG. 27, the controller 130 may include a CPU (Central Processing
Unit) 131 configured to provide a clock and to transmit a control
signal, a memory such as a RAM (Random Access Memory) 135 and ROM
(Read Only Memory) 137 configured to temporarily or
semi-permanently store a process, a GPU (Graphics Processing Unit)
133 for graphic processing, and a system bus 139 configured to
transmit data between the memory, CPU 131, and GPU 133. In
addition, the controller 130 also includes an operating system (OS)
420 (illustrated in FIG. 28) to drive the hardware configuration,
and an application configured to provide the user interface on the
OS 420 and delivering it to the framework. These components of the
controller 130 may be explained in further detail hereinbelow.
[0197] The imaging unit 140 may include at least one of the first
camera 141 and second camera 142, as illustrated in FIG. 26.
Although FIG. 26 only illustrates the first camera 141 and the
second camera 142, more cameras may be added according to exemplary
embodiments of the present general inventive concept.
[0198] Each camera includes a shutter (not illustrated), lens (not
illustrated), aperture (not illustrated) and CCD (Charge Coupled
Device) image sensor (not illustrated) and ADC (Analog/Digital
Converter). The lens receives light from an external source (not
illustrated) and processes the images. The shutter is an apparatus
configured to adjust the volume of light entering the camera,
together with the aperture. The aperture adjusts the quantity of
light (light quantity) entering the camera according to a degree by
which the aperture is opened or closed. The CCD image sensor
accumulates the quantity of light entering through the lens,
obtains an image, and outputs the image in accordance with a
vertical sync signal according to the accumulated quantity of
light. The CCD image sensor obtains the image by converting the
received light into electrical signals. In order to obtain a color
image using the CCD image sensor, a color filter is needed, and a
filter called CFA (Color Filter Array) may be adopted. The CFA
passes only light that represents just one color per each pixel,
has a regular arrangement structure, and has various forms
according to the arrangement structure. The ADC converts the analog
image signals output from the CCD image sensor into digital
signals. This is a mere example, and configurations of each camera
may be changed in various ways. For example, each camera may
photograph an image using a CMOS (Complementary Metal Oxide
Semiconductor) image sensor instead of a CCD image sensor.
[0199] The first camera 141 and second camera 142 may be provided
in a housing of the multi display apparatus 100 or may be connected
to the multi display apparatus 100 using additional connection
means. At least one of the first camera 141 and second camera 142
may include a subsidiary light source (for example, a flash, not
illustrated) which provides a quantity of light needed to properly
obtain an image.
[0200] In an exemplary embodiment of the present general inventive
concept, the first camera 141 may be arranged on a front surface of
the multi display apparatus 100, and the second camera 142 may be
arranged in the rear surface of the multi display apparatus 100. In
another exemplary embodiment of the present general inventive
concept, the first camera 141 and second camera 142 may be arranged
adjacent to each other (for example, so that the distance between
the first camera 141 and the second camera 142 is bigger than 1 cm
but smaller than 8 cm), and photograph 3-dimensional still images
or 3-dimensional videos. In another exemplary embodiment of the
present general inventive concept, the first camera 141 may be
arranged in the first display 190a, and the second camera 142 may
be arranged in the second display 190b.
[0201] The imaging unit 140 may detect a movement or shape of a
user object 50 through at least one of the first camera 141 and
second camera 142, and transmit the detected movement or shape to
the controller 130 as an input to execute or control an
application. In an exemplary embodiment of the present general
inventive concept, movement of the user object 50 denotes the
movement of the user's hand sensed through the first camera 141 or
second camera 142. The shape of the user object 50 may denote for
example the user's facial shape or orientation of fingers of their
hand, sensed through the first camera 141 or second camera 142.
[0202] In another exemplary embodiment, the multi display apparatus
100 may detect the user's movement using another means such as an
infrared sensor (not illustrated) and may control or execute the
application in response to the detected movement.
[0203] The sensor 150 is a configurative element to sense various
changes of state, such as for example the user's touch on the multi
display apparatus 100, the user's movement, or movement of the
multi display apparatus 100 itself. In the aforementioned exemplary
embodiment of the present general inventive concept, a touch sensor
151 or proximity sensor 155 was explained as an example of the
sensor 150. Besides the aforementioned exemplary embodiment of the
present general inventive concept, the sensor 150 may include at
least one of a touch sensor 151, a geomagnetic sensor 152, an
acceleration sensor 153, a hinge sensor 154 and a proximity sensor
155.
[0204] The touch sensor 151 is a sensor which may sense contact of
the user object 50 regarding the multi display (dual display) 190
of the multi display apparatus 100. That is, the touch sensor 151
denotes a sensor which may sense an input of touching the first
display 190a or the second display 190b and selecting an object
displayed on the touched display. The touch sensor 151 may be
classified as an electrostatic type and a piezoelectric type
depending on the method of sensing the touch of the user. The touch
sensor 151 according to an exemplary embodiment of the present
general inventive concept may be embodied as any one of the two
types. The touch sensor 151 may be included in the multi display
190 together with a display panel 235 (illustrated in FIG. 29).
[0205] The geomagnetic sensor 152 is a sensor which may detect the
flow of a magnetic field and detect the azimuth of the magnetic
field. The geomagnetic sensor 152 may detect the azimuth
coordinates of the multi display apparatus 100, and may detect the
direction of the multi display apparatus 100 when it is placed
based on the azimuth coordinates. The detected direction may be
recognized by the control input corresponding thereto and the
controller 130 may perform the corresponding output.
[0206] The acceleration sensor 153 is a sensor used to detect the
acceleration of a moving object (dynamic acceleration), but is also
utilized to detect the acceleration due to gravity as well. The
acceleration sensor 153 may set virtual x, y, z axes on the multi
display apparatus 100, detect the acceleration of the multi display
apparatus 100, and detect the value of acceleration due to gravity,
which changes according to an inclined degree of each axis.
[0207] The hinge sensor 156 may detect angles or movement of the
hinge 185 (illustrated in FIG. 1). The proximity sensor 155 may
detect whether or not a user object 50 approaches the multi display
apparatus 100. The proximity sensor 155 will be explained in
further detail below.
[0208] Although not illustrated in FIG. 26, the sensor 150 of the
multi display apparatus 100 may further include at least one of a
gravity sensor which may detect in which direction gravity acts, a
gyro sensor which may recognize a total of 6 axes due to rotations
applied to the acceleration sensor 153, an orientation sensor which
may automatically sense a width and length of contents, such as an
image, and align the sensed width and length, an illuminance sensor
which may detect the quantity of light surrounding the multi
display apparatus 100, a height measuring sensor which may measure
the air pressure, an RGB sensor which may detect the color of
objects, a distance measurement sensor which may measure the
distance using an ultrasound wave or infrared ray, and a hall
sensor which uses changes of voltage according to the intensity of
the magnetic field.
[0209] The sensor 150 may detect the state of each of the sensors
described above, generate a signal corresponding to the detection,
and transmit the generated signal to the controller 130. The
sensors of the sensor 150 described above may be added or deleted
according to the performance and specific configuration of the
multi display apparatus 100.
[0210] The input/output unit 160 is a configurative element to
perform an input/output using a screen or other external connection
ports. More specifically, the input/output unit 160 may receive an
input signal transmitted from an input means (not illustrated) such
as a mouse, keyboard, joystick connected to the multi display
apparatus 100, or from a wireless input means (not illustrated)
such as a remote control, etc., and transmit the received input
signal to the controller 130. Otherwise, the input/output unit 160
may output various signals or data generated by the controller 130
to the external apparatus. The input/output unit 160 may include a
button 161, a microphone 162, a speaker 163 and a vibration motor
164.
[0211] The at least one button 161 may be formed in a push type or
touch type in a front surface, lateral service, or rear surface of
the housing of the display apparatus 100, and may include at least
one of the power/lock button, volume adjusting button, menu button,
home button, back button and search button.
[0212] When the button 161 is pressed, for example by the user, a
corresponding control command is generated and transmitted to the
controller 130, and the controller 130 controls operations of the
multi display apparatus 100 according to the corresponding control
command.
[0213] The microphone 162 receives voice or sound according to the
control of the controller 130 and generates a corresponding
electrical signal.
[0214] The speaker 163 may output sound corresponding to various
signals (for example, a wireless signal, a broadcast signal, a
digital audio file, a digital video file, photography, etc.) of the
cellular communication module 111, wireless LAN module 112, short
distance communication module 113, multimedia unit 120 or imaging
unit 140 to an outside of the multi display apparatus 100.
[0215] The speaker 163 may output sound (for example, button
manipulating sound or ring back tone corresponding to telephone
calls) corresponding to the functions performed by the multi
display apparatus 100. The speaker 163 may be formed as a single
speaker or a plurality of speakers, in an appropriate location or
locations of the housing of the multi display apparatus 100. For
example, the speaker 164 may be configured to include an inner
speaker module (not illustrated) arranged in a location suitable to
approach the user's ear, and an external speaker module (not
illustrated) having a higher output suitable to be used in playing
audio/video files or when viewing broadcast programs, and may be
arranged in a suitable location of the housing of the multi display
apparatus 100.
[0216] The vibration motor 164 may convert electrical signals into
mechanical vibration according to a control of the controller 130.
For example, in the case of the multi display apparatus 100 in a
vibration mode, when a voice call is received from another
apparatus (not illustrated), the vibration motor 164 operates. The
vibration motor 164 may be formed as a single motor or a plurality
of motors within the housing of the multi display apparatus 100.
The vibration motor 164 may operate in response to the user's touch
gesture sensed on the first and second display 190a, 190b, and
consecutive movements of the touch sensed on the first and second
display 190a, 190b.
[0217] The storage unit 170 is configured to store data.
[0218] First of all, the storage unit 170 stores an operating
system program to control an operation of the multi display
apparatus 100. When the multi display apparatus 100 is turned on,
the stored operating system is read in the storage unit 170 and
compiled to drive each configuration of the multi display apparatus
100.
[0219] Secondly, the storage unit 170 is managed by the operating
system and uses the resources of the operating system to perform
operations of the multi display apparatus 100, and stores an
application program providing the user interface. The application
program is read in the storage unit 170 by the operating system by
the user's execution command, and is transferred to the controller
130 in an executable state to perform various operations.
[0220] Thirdly, the storage unit 170 stores various multimedia data
processed by the controller 130, contents data, and data received
from external sources. That is, the storage unit 170 may store
signals, information, or data being input/output correspondingly to
the operations of the cellular communication module 111, wireless
LAN module 112, short distance communication module 113, connector
114, GPS module 115, multimedia unit 120, imaging unit 140, sensor
150, input/output unit 160, first display 190a, and second display
190b according to the control of the controller 130.
[0221] The storage unit 170 may be embodied as at least one of a
memory card (for example, SD card, memory stick), nonvolatile
memory, volatile memory, hard disk drive (HDD) or solid state drive
(SSD) which are detachable/mountable to a ROM, RAM or multimedia
apparatus 100.
[0222] The power unit 180 supplies power used in the multimedia
apparatus 100. The power unit 180 may be embodied as a chargeable
battery, and may further include a voltage converter which converts
external supply power and supplies the converted external supply
power to a chargeable battery (not illustrated).
[0223] The power unit 180 may supply power to the multi display
apparatus 100 in various modes such as a maximum performance mode,
general mode, saving mode, waiting mode etc. according to the power
management control of the controller 130.
[0224] The first display 190a and the second display 190b are
included in the sensor 150 as aforementioned, and output images
stored in a frame buffer (described below). The first display 190a
and the second display 190b may display multimedia contents,
images, videos, and texts etc. by the control of the controller
130.
[0225] The first display 190a and the second display 190b may be
physically separated from each other. The screens displayed on the
first display 190a and the second display 190b may be controlled
independently from each other. For example, the resolution of the
first display 190a and the resolution of the second display 190b
may be set independently from each other. In addition, expansion,
rotation, screen movement, screen division etc. of the screens
displayed on the first display 190a and the second display 190b may
be executed independently from each other.
[0226] In addition, the first display 190a and the second display
190b may display a single screen using a virtual integrated frame
buffer.
[0227] The first and second display 190a, 190b are display
apparatuses which display various applications (for example,
telephone calls, data transmission, broadcasting, camera operation,
etc.) which may be executed by the controller 130 and which provide
user interfaces configured to be adaptable thereto. The first
display 190a and the second display 190b may include the
aforementioned touch sensor 151, in which case at least one touch
gesture may be input through a user object 50 such as for example
the user's body (for example fingers including the thumb) or
sensible input means (for example, stylus pen)
[0228] Such a user interface may include a predetermined touch
area, soft key and soft menu. The first and second display 190a,
190b may transmit electronic signals corresponding to at least one
touch gesture input through the user interface to the first display
190a and the second display 190b through an LCD controller (not
illustrated). In addition, the first display 190a and the second
display 190b may sense the consecutive movement of the touch, and
transmit electronic signals corresponding to the consecutive or
nonconsecutive movements of the touch to the LCD controller. As
aforementioned, the touch sensor 151 may be embodied with a
resistive method, capacitive method, infrared method, or acoustic
wave method.
[0229] The first display 190a and the second display 190b may
convert the sensed signals regarding the user operations sensed
through the touch sensor 151 into digital signals (for example, X
and Y coordinates), and transmit the converted digital signals to
the controller 130. The controller 130 uses the received digital
signals to perform control operations corresponding to the user
operation input through the first display 190a and the second
display 190b. For example, the controller 130 may have a soft key
displayed on the first display 190a and/or the second display 190b,
and may execute the application corresponding to the soft key in
response to the user operation.
[0230] The aforementioned user gesture is not limited to direct
contact between the first display 190a and the second display 190b,
and the user object 50 (including the user's body and other
touchable input means), but includes methods by noncontact as well.
The sensitivity of the user operations detectable in the first
display 190a and the second display 190b may be changed according
to the performance and structure of the multi display apparatus
100.
[0231] The multi display apparatus 100 is an apparatus configured
to execute applications, widgets and functions which may be stored
in the storage unit 170 and may be executed by the controller 130
through a touch screen 192. A general touch screen 192 may display
applications, widgets, functions and graphic objects (that is soft
keys or shortcut keys) that correspond to those groups through a
home screen or an application menu, and the multi display apparatus
100 executes the corresponding applications, widgets or functions
in response to the detected user's touch gesture on each displayed
graphic object.
[0232] Herein, a widget denotes a mini application which may be
downloaded by the user and then used, or may be generated by the
user, for example, a weather widget, stock widget, calculator
widget, alarm clock widget, dictionary widget etc. A shortcut icon
to execute a widget may provide simple dictionary information
through the widget application corresponding to the shortcut icon
to execute the widget. For example, the icon of the weather widget
simply provides the current temperature and weather symbols, and
the widget application executed through a touch of the icon
provides more information such as weather per period/region. In the
present application, an application includes a widget based
application and a non-widget based application.
[0233] The first body 2 including the first display 190a and the
second body 4 including the second display 190b may be connected to
each other by a connecting portion such as the hinge 185, and the
first body 2 and the second body 4 may be folded in or folded out
by a certain angle around the hinge 185.
[0234] The connecting portion may be embodied as a part of a
flexible connecting portion (not illustrated), or part of a
flexible touch screen (not illustrated), besides the hinge 185.
[0235] Hereinbelow is detailed explanation on the configuration of
the hardware of the aforementioned controller 130, with reference
to FIG. 27.
[0236] FIG. 27 is a block diagram illustrating a hardware
configuration of the controller 130.
[0237] With reference to FIG. 27, the controller 130 of the multi
display apparatus 100 may include a CPU 131, a GPU 133, a RAM 135,
a ROM 137 and a system bus 139 in a hardware perspective, and
provides an operating system (OS) to drive the hardware
configuration in a software (program) perspective, as well as an
application to provide the user interface on the operating system
and delivering it to the framework. However, operating systems,
frameworks, and applications will be explained in further detail
below.
[0238] The CPU 131 controls functions of various blocks while
performing data communication with each block through the system
bus 139, collects the result of the control, and transmits the
control signal to various peripheral apparatuses connected to the
controller 130 based on the collected result, to perform the
function of controlling the peripheral apparatuses. The "peripheral
apparatuses" include for example the communication unit 110, the
multimedia unit 120, the imaging unit 140, the sensor 150, the
input/output unit 160, the storage unit 170, and the power unit
180, as illustrated for example in FIG. 26. In addition, the CPU
131 controls the calculator to read the instruction and argument
per each process in the RAM 135, and to perform calculation.
[0239] First, the CPU 131 performs a booting using the booting
information prestored in the ROM 137. That is, when the power of
the system is turned on, the CPU 131 reads the command to operate
each peripheral apparatus from the ROM 137, and transmits the
control signal to each peripheral apparatus according to the
command. In addition, the CPU 131 reads the data stored in the
storage unit 170 to the RAM 135, and transmits data which need
graphic processing from among the data stored in the RAM 135 to the
GPU 133. The CPU 131 receives the data which has been graphically
processed by the GPU 133, and transmits the data to the LCD
controller (not illustrated) connected to the system bus 139 to
display the image on the multi display 190.
[0240] Herein, the CPU 131 temporarily stores the image data
processed by the GPU 133 in the virtual frame buffer allocated to
the predetermined area of the RAM 135. The CPU 131 allocates the
area of the virtual frame buffer so as to support the maximum
resolution (for example 1024*600) of the multi display 190. When
there are two displays in the multi display 190, the virtual frame
buffer area is allocated in a size of 1024*1200.
[0241] The CPU 131 inputs the data temporarily stored in the
virtual frame butter to the GPU 133 and performs digital signal
processing.
[0242] The GPU 133 performs graphic processing on the input data
under the control of the CPU 131. More specifically, the GPU 133
may generate a screen including various objects such as an icon, an
image, text, etc. using a calculator (not illustrated) and renderer
(not illustrated). The calculator calculates the characteristics
values such as coordinate values, shapes, sizes, and colors etc.
where each object is to be displayed according to the layout of the
screen. The rendering generates screens of various layouts
including the object based on the characteristics values calculated
in the calculator. The screen generated in the rendering is
transmitted to the first display 190a and/or the second display
190b through the bus 139, and is displayed within a display area of
the first display 190a and/or the second display 190b, or is stored
in the storage unit 170.
[0243] The CPU 131 may control to display the data graphic
processed by the GPU 133 through at least one of the first display
190a and the second display 190b, or may input the processed data
into a display controller (not illustrated).
[0244] The GPU 133 may include a decoder, renderer, and scaler,
etc. (not illustrated). Accordingly, the stored contents are
decoded, the decoded contents data is rendered to configure a
frame, and the size of the configured frame may be scaled to be
suitable to the display size by the control of the display
controller. If a screen is displayed on one of the first display
190a and the second display 190b, the scaling is made to suit the
corresponding display size, and if the screen is displayed in both
of the first display 190a and the second display 190b, the scaling
is made to suit the combination of the two display sizes. The GPU
133 provides the processed frame to the display and displays the
provided processed frame.
[0245] The controller 130 may further include an audio processor,
an interface, etc. (not illustrated), besides the above. The
interface is a configurative element to interface with the
surrounding configurative elements.
[0246] The audio processor (not illustrated) denotes a
configurative element which interfaces with the multimedia unit 120
through the audio interface (not illustrated) and processes the
audio data to provide the processed audio data to sound output
means such as a speaker. The audio processor may decode the audio
data stored in the storage unit 170 and the audio data received
through the communicator 110, noise filter the decoded audio data,
and then perform audio signal processing such as amplifying the
audio data to an appropriate decibel. In the aforementioned
example, when the contents that are played are video contents, the
audio processor may process the audio data demultiplexed from the
video contents, synchronize the processed audio data with the GPU
133, and provide the result to the speaker 163 to be output.
[0247] In ROM 137, a set of commands for system booting etc. is
stored. When a turn on command is input and power is supplied, the
CPU 131 copies the O/S stored in the storage unit 170 to the RAM
133 according to the commands stored in the ROM 137, and executes
the O/S to boot the system. When the booting is completed, the CPU
131 copies various application programs stored in the storage unit
170 in the RAM 133, and executes application programs copied in the
RAM 133 to perform various operations. As such, the CPU 131 may
perform various operations according to execution of the
application program stored in the storage unit 170.
[0248] As aforementioned, when a touch and other user manipulation
is sensed in the multi display apparatus 100, the controller 130
may determine whether or not the user manipulation is intended.
When it is determined that it is an intended user manipulation, the
controller 130 reads information on the operations corresponding to
that user manipulation from the storage unit 170, and then performs
operations corresponding to that information. Such operations of
the controller 130 may be embodied by execution of various programs
stored in the storage unit 170.
[0249] The imaging unit 140 is configured to perform photographing
operations according to the user manipulations. The imaging unit
140 may be provided in plural within the multi display apparatus
100. For example, a first imaging unit (not illustrated) may be
provided in the first body 2 where the first display 190a is
provided, and a second imaging unit (not illustrated) may be
provided in the second body 4 where the second display 190b is
provided. The controller 130 may store the image photographed by
the imaging unit 140 or attach the photographed image to an email,
document, message, etc. and transmit the attached image outside of
the multi display apparatus 100 via the communication unit 110. In
addition, the controller 130 may analyze the image photographed by
the imaging unit 140 and recognize the user's motion or gesture,
and perform a control operation corresponding to that motion or
gesture.
[0250] The CPU 131 may be embodied as at least one of a single core
processor, multi core processor, triple core processor, and quad
core processor. In the case of a multi core processor, each
processor included in the multi core processor may independently
control the first display 190a and second display 190b.
[0251] Hereinbelow is an explanation of a software (program)
hierarchy configuration of the controller 130.
[0252] FIG. 28 is a view illustrating a system hierarchy structure
of a multi display apparatus 100 according to an exemplary
embodiment of the present general inventive concept.
[0253] With reference to FIG. 28, the multi display apparatus 100
according to an exemplary embodiment of the present general
inventive concept includes an operating system (OS, 420) configured
to drive the hardware, application hierarchies 441-445 managed by
the operating system and configured to provide user services using
resources of the multi display apparatus 100, and a framework
hierarchy 430 relaying the communications between the operating
system and the applications.
[0254] The operating system 42 controls the overall operations of
the hardware 410 and performs functions such as managing the
process corresponding to each application. That is, OS 420 is a
hierarchy configured to play the basic functions such as hardware
management, memory, and security etc. The OS 420 includes modules
(not illustrated) such as a display driver to drive the multi
display, a communication driver to transceive data, a camera driver
to drive a camera, an audio driver to drive an audio, and a power
manager etc. In addition, an application programming interface
(API) library and runtime records (not illustrated) that may be
read by a developer (not illustrated) may be included. The OS 420
processes the calls of the applications, and operates the hardware
according to the processed results.
[0255] There exists a framework hierarchy 430 as a superior
hierarchy of OS 420. The framework hierarchy 430 performs a role of
connecting the application hierarchy 440 and the OS hierarchy 420.
That is, the framework hierarchy 430 includes a location manager,
notification manager, and a frame buffer to display an image on the
touch screen 192.
[0256] There exists an application hierarchy 440 where various
functions of the multi display apparatus 100 are embodied in the
superior hierarchy of the framework hierarchy 430. For example,
various application programs such as call applications 441,
multimedia applications 442, camera applications 443, browser
applications 444, gesture applications 445, etc., may be
included.
[0257] The applications 441-445 provide the user interface, and
receive the command from the user and transmit the command received
to OS 420 through the framework, or requests for resources of the
OS 420. The OS 420 processes the system call, and manages operation
of various applications. In addition, the applications 441-445
operate various hardware, such as the first display 190a and second
display 190b, according to the system call processing result.
[0258] Hereinbelow is explanation of the hardware configuration of
the multi display 190 according to an exemplary embodiment of the
present general inventive concept.
[0259] FIG. 29 is a view illustrating a circuit configuration of an
image outputter 191 of the first display 190a or the second display
190b. Hereinbelow, the explanation will be made based on the first
display 190a, but the second display 190b would include the same
configurations and would operate in the same manner. Furthermore,
FIG. 29 illustrates the circuit configuration of the image
outputter 191, whereas FIG. 31 illustrates the physical
construction of the image outputter 191.
[0260] According to FIG. 29, the image outputter 191 of the first
display 190a may include a timing controller 231, a gate driver
232, a data driver 233, a voltage driver 234, and a display panel
235.
[0261] The timing controller 231 receives a clock signal (DCLK) and
horizontal sync signal (Hsync), vertical sync signal (Vsync), etc.
and generates a gate control signal (injection control signal), and
data control signal (data signal), and realigns the received R, G,
B values of the image to be displayed, and provides them to the
data driver 233.
[0262] The timing controller 231 may generate a gate shift clock
(GSC), gate output enable (GOE), and gate start pulse (GSP) etc.
Herein, a GSC is a signal configured to determine the time when a
thin film transistor (TFT) connected to light emitting diodes such
as R, G, B (for example Organic Light Emitting Diodes, OLEDs) is
turned on/off, a GOE is a signal configured to control the output
of the gate driver 232, and a GSP is a signal configured to notify
the first driving line of the screen in one vertical sync
signal.
[0263] In addition, the timing controller 231 may generate Source
Sampling Clock (SSC), Source Output Enable (SOE), and Source Start
Pulse (SSP) etc. regarding the data control signal. Herein, SSC is
used as a sampling clock to latch data in the data driver 233, and
determines the driving frequency of the data driver 233. The SOE
delivers the data latched by the SSC to the display panel 235. The
SSP is a signal notifying a latch or sampling start of data during
one horizontal sync period.
[0264] The gate driver 232 is a configuration generating an
injection signal, and is connected to the display panel 235 through
the injection lines S1, S2, S3, . . . Sn. The gate driver 232
applies gate on/off voltage (Vgh/Vgl, not illustrated) provided
from the voltage driver 234 to the display panel 235 according to
the gate control signal generated by the timing controller 231. The
gate on voltage (Vgh) is provided consecutively from gateline 1
(GL1) to gateline N (GLn) for embodiment of unit frame images on
the display panel 235. The data driver 233 is a configuration of
generating a data signal, and is connected to the display panel 235
through data lines D1, D2, D3, . . . Dn. The data driver 233
finishes scaling according to the data control signal generated by
the timing controller 233, and inputs the RGB data of the image
into the display panel 235. The data driver 233 converts the RGB
image data provided in serial into parallel in the timing
controller 231, and converts the digital data into analog voltage
to provide the image data corresponding to one horizontal line
portion to the display panel 235. This process is made
consecutively per horizontal line.
[0265] The voltage driver 234 generates and delivers each driving
voltage of the gate driver 232, data driver 233, and display panel
235 etc. That is, the voltage driver 234 may receive prevailing
voltage from outside the multi display apparatus 100, for example,
alternating voltage of 110V or 220V, generate and provide source
voltage (VDD) necessary in the display panel 235, or provide ground
voltage (VSS). In addition, the voltage driver 234 may generate a
gate on voltage (Vgh) and provide it to the gate driver 232. To
this end, the driver 234 may include a plurality of voltage driving
modules (not illustrated) operating independently. Herein, the
plurality of voltage driving modules may operate to provide
different voltages from one another according to the control of the
controller 130, and the controller 130 may control the voltage
driver 234 so that a plurality of voltage driving modules could
provide different driving voltages from one another according to
the predetermined information. For example, each of the plurality
of voltage driving modules may provide a different first voltage
according to the information predetermined according to the control
of the controller 130 and a second voltage set in default.
[0266] According to an exemplary embodiment of the present general
inventive concept, the voltage driver 234 may include a plurality
of voltage driving modules corresponding to each area of the
display panel 235 divided in a plurality of areas. In this case,
the controller 130 may control a plurality of voltage driving
modules to provide first voltages which are different from one
another, that is, ELVDD voltages (not illustrated) according to
each screen information (or input image information) of the
plurality of areas. That is, the controller 130 may use the image
signal input into the data driver 233 to control the size of the
ELVDD voltage. Herein, the screen information may be at least one
of the brightness information and gradation information of the
input image.
[0267] The display panel 235 may be provided with a plurality of
gatelines GL1-GLn and datalines D1-Dn configured to intersect one
another and define pixel areas 236, and in the intersecting pixel
areas 236, R, G, B light emitting diodes such as OLEDs may be
formed in the pixel area 236 where the gatelines and datalines
intersect one another. In addition, in one area of the pixel area
236, more specifically in the corner, a switching element, that is
a TFT, is formed. During the turn on operation of the TFT, the
gradation voltage from the data driver 233 is provided to each
light emitting diode of R, G, B. Herein, R, G, B light emitting
diodes in accordance with the quantity of current provided based on
the gradation voltage. That is, when a large quantity of current is
provided, R, G, B light emitting diodes provide that much of the
current provided.
[0268] FIG. 30 is a view illustrating a circuit structure
configuring R, G, and B pixel areas 236 which configure the display
panel 235 of FIG. 29.
[0269] With reference to FIG. 30, the display panel 235 includes
three pixel areas 236 corresponding to R, G, and B colors. The R,
G, B pixel areas 236 may include a scan signal S1, switches M11,
M21, M31 that operate by the gate on voltage (Vgh, not
illustrated), switchers M12, M22, M32 that output current based on
the pixel values including the changed high gradation value
provided to the datalines D1-Dn, and switchers M13, M23, M33 that
adjust the quantity of current provided from the switchers M12,
M22, M32 to the R, G, B light emitting diodes according to the
control signal provided in the timing controller 231. In addition,
such switchers M13, M23, M33 are connected to the organic light
emitting diodes (OLEDs) to supply current to the OLEDs. Herein,
"OLED" denotes a display that emits light itself using the
principle of electric field light emission when current is applied
to fluorescence or phosphorescence organic matter thin film. The
anode electrode of each OLED is connected to the pixel circuit, and
the cathode electrode is connected to the ELVSS. Such an OLED
generates light of a predetermined brightness in response to the
current supplied from the pixel circuit. Herein, the gate electrode
of the switchers M11, M21, M31 are connected to the injection line
S1, while the other one of the source electrode and the drain
electrode is connected to the dataline D1. As such, the display
panel 235 may be embodied as an AM-OLED (Active Matrix Organic
Light-Emitting Diode) panel. However, the aforementioned exemplary
embodiment is merely one exemplary embodiment of the present
general inventive concept, and a PM OLED (Passive Matrix Organic
Light-Emitting Diode), which is a method where one line emits light
and drives at the same time, is obviously not to be excluded.
[0270] As such, in the case of embodying the multi display 190 of
the multi display apparatus 100 with OLEDs, there is no need to
provide an additional light emitting means (backlight), and thus
there is an advantage of making the thickness of the display
thinner and simplifying the configuration.
[0271] However, although OLEDs are described in the above exemplary
embodiment of the present general inventive concept, the display
may be embodied as various display technologies such as Liquid
Crystal Display Panel (LCD Panel), Plasma Display Panel (PDP),
Vacuum Fluorescent Display (VFD), Field Emission Display (FED), and
Electro Luminescence Display (ELD).
[0272] FIG. 31 is a cross-sectional view of a first display 190a
including a touch screen 192. Although the configuration
illustrated in FIG. 31 is described below with reference to the
first display 190a, the second display 190b may have the same
construction.
[0273] With reference to FIG. 31, the first display 190a includes
an image outputter 191 and a touch screen 192.
[0274] The image outputter 191 includes the display panel 235,
where an upper substrate 362 and a lower substrate 361 are
disposed, an upper polarizing plate 300 provided on the upper
substrate 362 of the display panel 235, and a lower electrode 310
provided in the edge area of the upper polarizing plate 300. In
addition, the image outputter 191 includes a lower polarizing plate
301 provided on the lower substrate 361.
[0275] There may be provided a liquid crystal layer 363 between the
upper substrate 362 and lower substrate 361. The upper substrate
362 and lower substrate 361 include a display area to display
images and a peripheral area provided at the edge of the display
area. Although not illustrated, in the display area of the lower
substrate 361, a thin film transistor is provided in an
intersecting area of a plurality of gatelines, and a plurality of
gatelines and datalines, as well as a pixel electrode connected to
the thin film transistor, are provided. There is provided a
sustained electrode where the pixel electrode and a portion thereof
overlap each other. In addition, in a peripheral area of the lower
substrate 361, there may be provided a plurality of pads (not
illustrated) connected to the gatelines and datalines.
[0276] The upper substrate 362 and the lower substrate 361 may be
sealed using a sealing member 364, such as silant. The sealing
member 364 is effective when provided in a peripheral area of the
upper substrate 362 and lower substrate 361.
[0277] The touch screen 192 includes a lower transparent electrode
330 connected to the lower electrode 310 and provided on the upper
polarizing plate 300, a touch pad 340 provided on the lower
transparent electrode 330, and a protrusion 350 provided between
the upper transparent electrode 342 and the lower transparent
electrode 330.
[0278] The touch pad 340 includes a base film 341, the upper
transparent electrode 342 provided on the base film 341, and an
upper electrode 343 provided on the upper transparent electrode
342.
[0279] In the case of pressing the touch pad 340 with a
predetermined force, the protrusion 350 of the area where the force
is applied contacts the lower transparent electrode 330, and
accordingly, the lower transparent electrode 330 and upper
transparent electrode 342 contact each other electrically. Through
the electrical contact, the current flowing through the upper
transparent electrode 342 and/or the lower transparent electrode
330 changes, and this change is sensed through an additional
sensing means (not illustrated). A signal output from the sensing
means is delivered to the controller 130. The controller 130 uses
the signal to generate coordinates of the area where the flow of
current is changed, and delivers the generated coordinates to a
predetermined driver (not illustrated). The driver may perform the
same operations as those used in manipulating the image displayed
on the display panel 235 using the input tool such as a mouse in
response to the input coordinates.
[0280] Meanwhile, due to the characteristics of the multi display
apparatus 100, it is possible to provide an image frame buffer
134a-1, 134b-1, or 134-1 in various methods. Hereinbelow is
explanation on image frame butters provided in various methods.
[0281] FIG. 32 is a block diagram illustrating a display driver 134
of the multi display apparatus 100 according to an exemplary
embodiment of the present general inventive concept, and FIG. 33 is
a block diagram illustrating a configuration of the display driver
134' according to another exemplary embodiment of the present
general inventive concept.
[0282] According to FIG. 32, the display driver 134 may include a
first frame buffer 134a-1, first display driver 1341-2, second
frame buffer 134b-1 and second display driver 134b-2. That is, each
of the displays 190a, 190b may have an additional frame buffer
134a-1 134b-1, and display driver 134a-2, 134-2.
[0283] The first frame buffer 134a-1 is a configuration to buffer
an image frame to be displayed on the display 190a, and the second
frame buffer 134b-1 is a configuration to buffer an image frame to
be displayed on the second display 190b.
[0284] For example, the image frame which has been digital signal
processed by the GPU 133 is stored in a bit map format to the first
and second frame buffers 134a-1, 134b-1. In this case, the
buffering area of each frame 134a-1, 134b-1 is allocated suitable
to the maximum pixel size that may be supported in each of first
display 190a and second display 190b. The first display driver
134a-2 analyzes the image frame stored in the first frame buffer
134a-1 and converts the image frame stored in the first frame
buffer 134a-1 into a first image source signal. The first display
driver 134a-2 provides the first image source signal to the first
display 190a, to drive the first display 190a to display the image
frame.
[0285] Likewise, the second display driver 134b-2 analyzes the
image frame stored in the second frame buffer 134b-1 and converts
the analyzed image frame into a second image source signal, and
provides the converted second image source signal to the second
display 190b to be displayed.
[0286] These two frame buffers 134a-1 and 134b-1 may be suitable to
process the image frames regarding each of the first display 190a
and the second display 190b in parallel. When a parallel processor
(not illustrated) outputs the image signal regarding each of first
display 190a and second display 190b, it is possible to increase
the image output efficiency.
[0287] However, unlike the above, the first frame buffer 134a-1 and
second frame buffer 134b-1 respectively corresponding to each of
first display 190a and second display 190b may be integrated to one
integrated frame buffer 134-1 and used instead of being provided
separately.
[0288] FIG. 33 is a block diagram illustrating a configuration of a
multi display apparatus 100 controlling operations of each display
190a, 190b using the integrated frame buffer 134-1. When the frame
buffer is embodied as an integrated frame buffer 134-1, the size of
the integrated frame buffer 134-1 may be allocated and embodied to
be bigger than the maximum resolution of the first display 190a and
the second display 190b.
[0289] For example, when each of the first display 190a and the
second display 190b displays the maximum resolution of 1024*800,
the integrated frame buffer 134-1 allocates the storage area to be
the frame buffer size capable of displaying the resolution of
1024*1600. In the first area of the integrated frame buffer 134-1,
the first image frame where the first display 190a is displayed is
stored, and in the second area of the integrated frame buffer
134-1, a second image frame displayed on the second display 190b is
stored.
[0290] The display driver 134-2 uses the first image frame stored
in the integrated frame buffer 134-1 and the address of the second
frame to provide the first and second image frames to the first
display 190a and the second display 190b, to drive each
display.
[0291] As aforementioned, the sensor 150 of the multi display
apparatus 100 may include proximate sensing means, for example, the
proximity sensor 155. Hereinbelow is explanation on the
configuration and operations of the proximity sensor 155.
[0292] FIG. 34 is a view illustrating a proximate touch according
to an exemplary embodiment of the present general inventive
concept. The construction illustrated in FIG. 34 may be applied to
either or both of first display 190a and second display 190b.
[0293] According to the exemplary embodiment illustrated in FIG.
34, a user may input a control command into the multi display
apparatus 100 merely by bringing a user object 50 such as a finger
or other object into proximity of the touch screen 192, without
directly touching the touch screen 192. The multi display apparatus
100 may sense the proximate touch using the proximity sensor 155.
The proximate touch denotes recognizing as one touch gesture when a
movement is recognized within a certain spatial effective
recognition range 5, without directly touching the touch screen
192.
[0294] With reference to FIG. 34, in the upper portion of a display
290, an infrared source 291 to recognize proximate touches is
located, and in the lower portion of the display 290, an infrared
sensor 292 is included.
[0295] The infrared source 291 irradiates an infrared ray in a
surface direction of the display 290. More specifically, the
infrared source 291 is arranged in the lower portion of the display
290 where the image is displayed, and may irradiate the infrared
ray in the surface direction of the display 290. There exists a
certain area which may recognize the approach of the user object 50
above the surface of the display 290. This area is the effective
recognition area 5 in which the proximate touch may be
recognized.
[0296] The meaning of the user object 50 is the means to input a
command into the multi display apparatus 100, for example, a body
part such as the hand.
[0297] When the user object 50 approaches towards inside the
effective recognition area 5, the sensor 292 senses the infrared
reflected by the approach of the user object 50 and generates the
infrared scan image. More specifically, the infrared sensor 292
uses a plurality of infrared sensing elements arranged in an array
format to generate an infrared scan image corresponding to the
infrared ray reflected by the approach of the user object 50. The
multi display apparatus 100 may sense the proximate touch input
using the generated infrared scan image.
[0298] Hereinbelow is explanation on the exterior configuration of
the multi display apparatus 100 according to an exemplary
embodiment of the present general inventive concept with reference
to the perspective view attached.
[0299] FIG. 35 is a detailed perspective view of the multi display
apparatus 100 according to an exemplary embodiment of the present
general inventive concept. Herein, it is illustrated that the multi
display apparatus 100 has two displays 190a and 190b, combined by
the hinge 185.
[0300] With reference to FIG. 35, the multi display apparatus 100
includes the first body 2 and the second body 4, the first body 2
and the second body 4 configured to be connected by the hinge 185
and thus movable relative to one another. On one surface of the
first body 2, there is provided the first display 190a, while on
one side of the display 190a, at least one physical button 161 may
be arranged. On one surface of the second body 4, the second
display 190b is provided, while on one side of the second display
190b, at least one physical button 161' may be arranged. The
physical buttons 161 and 161' may include at least one of the push
button and touch button. In an exemplary embodiment of the present
general inventive concept, the first display 190a arranged in the
first body 2 having the speaker 163 and microphone 162 may operate
as a main screen, while the second display 190b arranged in the
second body 4 may operate as a sub screen.
[0301] In an exemplary embodiment of the present general inventive
concept, the body 2 has the first camera 141, and the second body 4
has the second camera 142.
[0302] With the multi display apparatus 100 arranged as illustrated
in FIG. 35, each of the displays 190a and 190b may display the
screen in a landscape mode.
[0303] As long as the multi display apparatus 100 has the first
body 2 and second body 4 connected by the hinge 185 to be
relatively moveable, it may be any apparatus such as a mobile
phone, notebook, tablet PC, PMP, etc. The first display 190a and
the second display 190b are explained as having the first body 2
and second body 4, respectively, but the following explanation may
be applied to any apparatus which provides a touch screen display
as only one panel. In addition, at least one of the function
buttons 161 on the side of each of the displays 190a, 190b may be
omitted. Furthermore, although it is explained that a connection
between the first body 2 and second body 4 is made by the hinge
185, the hinge 185 may be substituted for any other configurative
element as long as the first body 2 and second body 4 are
configured to move relative to one another.
[0304] Meanwhile, the hinge 185 may include two hinges arranged in
the upper and lower sides of the connecting portion of the first
body 2 and second body 4. In addition, the hinge 185 may include
one hinge arranged in the entirety of the connecting portion of the
first body 2 and second body 4.
[0305] The multi display apparatus 100 has the display apparatus
including the first display 190a and second display 190b which are
separated from each other physically or graphically. The multi
display apparatus 100 may utilize two displays 190a and 190b to
support various screen modes such as illustrated in FIGS. 11 to
15.
[0306] FIGS. 35 to 40 are views illustrating various screen modes
according to the relative angle between the first body 2 and the
second body 4 of the multi display apparatus 100 according to an
exemplary embodiment of the present general inventive concept.
[0307] The relative angle .theta. is a rotation angle between the
second body 4 rotated in a predetermined direction (for example
counterclockwise direction) regarding the first body 2.
[0308] More specifically, the relative angle .theta. may be
detected using the hinge sensor 154 mounted inside the hinge 185
itself. The hinge sensor 154 may include one of a hall sensor,
pressure sensor, induction detection sensor, electricity contact
sensor and optical sensor, and detect the movement of the hinge and
relative location to recognize the relative angle .theta..
[0309] In addition, besides the hinge sensor 154, the geomagnetic
sensor 152 and acceleration sensor 153 may detect each location of
the first body 2 and the second body 4 to recognize the relative
angle 8.
[0310] FIG. 36 is a perspective view illustrating a folded multi
display apparatus 100 according to an exemplary embodiment of the
present general inventive concept. As illustrated in FIG. 36, with
the first display 190a on the first body 2 and the second display
190b on the second body 4 facing each other, the first body 2 and
the second body 4 touch each other. That is, the second display
190b is located on the opposite side of the first display 190a.
When the user sees the first display 190a, the second display 190b
is located at the opposite side, and thus the user cannot see the
second display 190b. As such, the user may only see one
display.
[0311] In the configuration illustrated in FIG. 36, the relative
angle 8 is defined as 0 degrees. For example, when the relative
angle between the first body 2 and the second body 4 is 0 to 60
degrees, the multi display apparatus 100 may recognize as a single
display mode. The single display mode may be used usefully for
applications in which only one display is necessary, for example in
a telephone call application, or when the multi display apparatus
100 is in a locked state and is not used. In the single display
mode, the first display 190a of the front surface displays the
operation screen by at least one application, and the second
display 190b on the rear surface may be turned off. Some
applications may use an option menu (not illustrated) to turn on
the second display 190b of the rear surface. In the case of the
single display mode, the user can see only one display screen, and
thus controller 130 controls one display only. That is, the
controller 130 only transmits a control signal regarding the
display that the user is currently seeing. Additional control
signals and data signals are not transmitted regarding the display
which is not controlled by the controller 130, thereby saving power
consumption.
[0312] Herein, the display which is not provided with the control
signal and data signal operates in a sleep mode, and when the
signal to be released from the sleep mode is applied by the
controller 130, such as when the user rotates the multi display
apparatus 100 to see the display in the sleep mode, the display may
be released from the sleep mode, and therefore receive control
signals from the controller 130.
[0313] FIG. 37 illustrates a state where the relative angle .theta.
is or approximates 180 degrees within a predetermined range, with
the first body 2 and second body 4 substantially parallel to each
other. Hereinbelow this mode is called an expanding mode. For
example, when the relative angle between the first body 2 and the
second body 4 is in the range of 175 to 185 degrees, the multi
display apparatus 100 may determine that the first body 2 and the
second body 4 are unfolded. The expanding mode may display two
operation screens regarding two applications on each of the
displays 190a and 190b, or display one screen regarding one
application on both of the displays 190a and 190b, or display two
operation screens regarding one application on the two displays
190a and 190b. No application may be displayed on either one of the
displays, but a defaulted home screen may be displayed. The
expanding mode may be useful in applications best performed with
multiple displays or a single large display, such as for example in
E-books, video player applications, and web browsing screens.
[0314] For example, the controller 130 may display a first web
browsing screen (not illustrated) on the first display 190a, and
display a second web browsing screen (not illustrated) on the
second display 190b. Herein, the first web browsing screen and
second web browsing screen are related to each other, and a third
web browsing screen (not illustrated) related to the second web
browsing screen may be displayed on the first display 190a, and a
fourth web browsing screen (not illustrated) related to the third
browsing screen may be displayed on the second display 190b.
[0315] The controller 130 may display the web browsing screens
alternately on the first display 190a and second display 190b, and
may perform screen conversion between the plurality of web browsing
screens displayed on each display 190a and 190b, and may display
the web browsing screens displayed on the first display 190a and
the web browsing screens displayed on the second display 190b in
different layouts from each other.
[0316] FIG. 38 illustrates a state of the multi display apparatus
100 in which the relative angle .theta. of the second body 4 to the
first body 2 exceeds 180 degrees, that is, when the two displays
190a and 190b are folded slightly towards an inside direction,
i.e., folded slightly towards each other. This is referred to
herein as a tool kit mode. For example, when the relative angle
between the first body 2 and second body 4 is 180 to 265 degrees,
it is recognized as the tool kit mode. The tool kit mode is useful
when the multi display apparatus 100 is used in a similar format as
a notebook computer. For example, various operation environments
may be provided such as displaying an operation screen on the first
display 190a and displaying a tool such as a keyboard on the second
display 190b.
[0317] FIG. 39 illustrates a state where the relative angle 8 of
the first body 2 and second body 4 is less than 180 degrees, that
is, when the two displays 190a and 190b are folded towards an
outside, i.e., folded so that the displays 190a and 190b face
almost opposite directions. This mode is referred to herein as a
standing mode. For example, when the relative angle between the
first body 2 and the second body 4 is between 30 to 90 degrees, the
multi display apparatus 100 may recognize the standing mode. Since
the standing mode is a structure where the two displays 190a and
190b are folded towards outside, it is possible to stand the multi
display apparatus 100 on the bottom in a triangular shape, i.e.,
balanced on an edge portion of the first body 2 and the second body
4. The standing mode of the multi display apparatus 100 may be
useful when charging, viewing an image (for example, using the
multi display apparatus 100 as a digital clock or frame), or
watching personal broadcast, film, and video etc. In another
exemplary embodiment of the present general inventive concept, the
standing mode may be applied to the application which requires
cooperation or interaction by two or more users, for example, a
conference call, a collaborative game, etc. Some applications
display operations screens on only the first display 190a in the
standing mode, and may turn off the second display 190b. Some
applications may turn on the second display 190b using the option
menu.
[0318] FIG. 40 is another exemplary embodiment of the standing mode
illustrated in FIG. 39, referred to herein as length view mode
where the multi display apparatus 100 is placed in a standing
manner so that a portion of the hinge 185 touches the ground. When
it is determined that the relative angle 8 between the first body 2
and the second body 4 is 30 to 90 degrees and that the multi
display apparatus 100 is placed in a standing manner by the
acceleration sensor 153, the multi display apparatus 100 may
recognize the length view mode.
[0319] More specifically, the acceleration sensor 153 may sense
rotation of the multi display apparatus 100. The acceleration
sensor 153 may sense the conversion between the length view mode
where the first display 190a and the second display 190b are placed
on the left or right side, as illustrated in FIG. 40, and a width
view mode where the first display 190a and the second display 190b
are placed on an upper or lower side, as illustrated in FIG.
39.
[0320] The length view mode may be applied to an application where
different images need to be provided to two or more users, for
example, conference call and multi video player etc.
[0321] FIGS. 41 and 42 are views illustrating an arrangement of two
cameras 141 and 142 of the multi display apparatus 100 according to
an exemplary embodiment of the present general inventive concept.
FIGS. 41 and 42 are views of the multi display 100 apparatus seen
from the direction of both displays 190a and 190b while in the
expanding mode. These are views to illustrate the arrangement of
the cameras 141 and 142, and thus other elements and configurations
of the multi display apparatus 100 are omitted for convenience of
explanation.
[0322] With reference to FIG. 41, the multi display apparatus 100
includes a first body 2 and second body 4, the first body 2
including the first display 190a and the second body 4 including
the second display 190b. The first body 2 and second body 4 are
connected by the hinge 185 so that they are moveable relative to
one another. The first camera 141 may be arranged on the first body
2 in a central portion of an edge area opposite to the hinge 185.
Similarly, the second camera 142 may be arranged on the second body
4 in a central portion of an edge area opposite to the hinge
185.
[0323] Next, an alternate arrangement of first camera 141' and
second camera 142' is explained with reference to FIG. 42.
Similarly to FIG. 41, the explanation is made with the multi
display apparatus 100 in the expanding mode. The first camera 141'
may be arranged on a central surface of the left edge area of the
first body 2, and the second camera 142' may be arranged on a
central surface of the left edge of the second body 4. In another
exemplary embodiment of the present general inventive concept, each
of the first camera 141' and the second camera 142' may be arranged
in a central surface of the right edge area of the first and second
body 2, 4, respectively. In another exemplary embodiment of the
present general inventive concept, the first and second camera may
be arranged in a corner area of the first body 2 and the second
body 4, respectively. The multi display apparatus 100 with the
first camera 141' and the second camera 142' arranged as in FIG. 42
may not only be used in the width view mode as aforementioned but
also in the length view mode.
[0324] Hereinbelow is explanation on a multi display method
according to various exemplary embodiments of the present general
inventive concept.
[0325] FIGS. 43 to 47 are flowcharts illustrating a multi display
method according to various exemplary embodiments of the present
general inventive concept.
[0326] The operations of the multi display method already explained
above will be omitted.
[0327] The multi display method as in FIG. 43 may be performed in a
multi display apparatus 100 including a first body 2 where a first
display 190a is arranged, a second body 4 where a second display
190b is arranged, and a hinge 185 which connects the first body 2
and the second body 4 and supports the first body 2 and the second
body 4 so that they are rotatable relative to one another.
[0328] More specifically, with the multi display apparatus 100
displaying the unstructured first screen and second screen, through
the first display 190a and the second display 190b, respectively,
the user may perform user operations for each screen. Accordingly,
when a first user manipulation regarding a nondecisive location of
the first screen is sensed (operation S4310), and a second user
manipulation regarding a nondecisive location of the second screen
is sensed (operation S4320), the sensed first user manipulation and
second user manipulation are combined, and an operation
corresponding to that combination results if performed (operation
S4330).
[0329] With reference to FIG. 44, the multi display apparatus 100
according to an exemplary embodiment of the present general
inventive concept may include sensing the first user manipulation
(operation S4410), changing the functions matching the second user
manipulation according to the first user manipulation (operation
S4420), sensing a second user manipulation (operation S4430), and
performing an operation according to the changed functions when the
second user manipulation is sensed (operation S4440).
[0330] Herein, the second user manipulation may be a line input,
and when the first user manipulation is a pinch out input, the
changing the function (operation S4420) may increase the thickness
of the input line, and when the first user manipulation is a pinch
in input, the changing the function (operation S4420) may reduce
the thickness of the input line, and when the first user
manipulation is a touch & rotate input, the changing the
function (operation S4420) may change the texture of the line.
[0331] In addition, in performing the aforementioned operations
(operation S4330) in the exemplary embodiment illustrated in FIG.
43, if the first user manipulation is a touch & hold input, and
when the second user manipulation by the touch input is sensed
together with the first user manipulation, a menu corresponding to
the second user manipulation may be displayed.
[0332] With reference to FIG. 45, the multi display method
according to an exemplary embodiment of the present general
inventive concept includes sensing a first user manipulation
(operation S4510), changing the screen displayed on the second
display according to the first user manipulation when the first
user manipulation is sensed (operation S4520), sensing a second
user manipulation (operation S4530), and changing the screen which
has been changed by the first manipulation and then displayed when
the second user manipulation is sensed (operation S4540).
[0333] Herein, changing the screen (operation S4520) may expand the
screen displayed on the second display when the first user
manipulation is a pinch out input, reduce the screen displayed on
the second display 190b when the first user manipulation is a pinch
in input, and rotate the screen displayed on the second display
when the first user manipulation is a touch & rotate input.
[0334] With reference to FIG. 46, the multi display method
according to an exemplary embodiment of the present general
inventive concept may include sensing a first user manipulation
(operation S4610), extracting contents displayed on the first
display 190a when the first user manipulation is sensed (operation
S4620), sensing a second user manipulation (operation S4630), and
displaying a screen corresponding to the extracted contents on the
second display 190b when the second user manipulation is sensed
(operation S4640).
[0335] Herein, the extracting (operation S4620) may capture the
screen displayed on the first display 190a according to the first
user manipulation, and displaying the screen (operation S4640) may
display the captured screen in the location where the second user
manipulation is made on the second display 190b.
[0336] With reference to FIG. 47, the multi display method of the
multi display apparatus 100 according to an exemplary embodiment of
the present general inventive concept includes sensing a first user
manipulation regarding the nondecisive location of the first screen
when each of the first display 190a and second display 190b
displays the unstructured first screen and second screen,
respectively (operation S4710), sensing the second user
manipulation regarding the nondecisive location of the second
screen (operation S4720), and connecting the sensed first user
manipulation and second user manipulation to recognize as one user
gesture, and displaying the execution screen corresponding to the
user gesture at one of the first display 190a and the second
display 190b (operation S4730).
[0337] In the aforementioned multi display method according to
various exemplary embodiments of the present general inventive
concept, the first user manipulation may be a proximate input or
touch input to the first display 190a using the user's body, and
the second user manipulation may be a proximate or touch input to
the second display 190b using an input pen 200.
[0338] Meanwhile, the aforementioned various multi display methods
were explained to recognize the nonstructured user manipulation
made for each of the plurality of displays in the multi display
apparatus 100 and perform operations accordingly. However, the
present general inventive concept is not limited thereto, but
rather, the nonstructured user manipulation made for each of the
plurality of displays may be recognized as one user manipulation
and an operation corresponding to the one user manipulation may be
performed.
[0339] For example, when the user performs a drag operation from a
point on the first display 190a to the second display 190b, a user
manipulation is sensed in both the first display 190a and second
display 190b, respectively. When the point where the user
manipulation is sensed in the first display 190a and the second
display 190b is within the predetermined time range and when the
dragging track on the first display 190a is continuous with the
dragging track on the second display 190b, the controller 130
recognizes those user manipulations as one dragging operation.
Accordingly, an operation matching the dragging manipulation is
performed. For example, the controller 130 may perform various
operations such as for example an operation of expanding the first
screen displayed on the first display 190a where the initial
dragging began on the second display 190b, an operation of dividing
one screen and displaying the divided screen on one or both of the
first display 190a and the second display 190b, an operation of
displaying one common screen on each of the first display 190a and
the second display 190b, an operation of simultaneously activating
or inactivating the first display 190a and the second display 190b,
and an operation of turning on or off power of the multi display
apparatus 100.
[0340] The aforementioned multi display method may be embodied as a
program including an algorithm executable in a computer, and the
program may be stored in and provided by means of a non-transitory
computer readable medium.
[0341] The computer-readable medium can include a computer-readable
recording medium and a computer-readable transmission medium. The
computer-readable recording medium is any data storage device that
can store data as a program which can be thereafter read by a
computer system. Examples of the computer-readable recording medium
include a semiconductor memory, a read-only memory (ROM), a
random-access memory (RAM), a USB memory, a memory card, a Blu-Ray
disc, CD-ROMs, magnetic tapes, floppy disks, and optical data
storage devices. The computer-readable recording medium can also be
distributed over network coupled computer systems so that the
computer-readable code is stored and executed in a distributed
fashion. The computer-readable transmission medium can transmit
carrier waves or signals (e.g., wired or wireless data transmission
through the Internet). Also, functional programs, codes, and code
segments to accomplish the present general inventive concept can be
easily construed by programmers skilled in the art to which the
present general inventive concept pertains.
[0342] According to various exemplary embodiments of the present
general inventive concept, it is possible to perform unstructured
independent inputting through each display of a multi display
apparatus 100, and combine each input to perform an output
corresponding thereto.
[0343] Although exemplary embodiments of the present general
inventive concept have been described above, it will be understood
that the present general inventive concept is not limited to these
exemplary embodiments. For example, the multi display 190 of the
multi display apparatus 100 does not necessarily include displays
disposed on separate bodies which are physically connected to one
another. For example, the first display 190a and the second display
190b may be disposed on opposite sides of a single body, similarly
to the configuration illustrated in FIG. 36, but without the hinge
185, and with both displays active. This configuration would allow
the multi display apparatus 100 to be made more compact.
Furthermore, since user manipulations may be made in a nondecisive
location on the first display 190a to affect operations of the
second display 190b, the user does not need to see the first
display 190a in order to control the second display 190b. For
example, the first display 190a could be a touch pad that does not
display a screen, and instead receives user manipulations which may
modify the operation of user manipulations on the second display
190b, as described above. This configuration would allow a user to
efficiently perform operations on the second display 190b without
needing to hold a separate first display 190a.
[0344] Furthermore, the displays of the multi display 190 do not
need to be physically connected, and may communicate wirelessly
with each other through communication module 110, thereby allowing
a greater degree of flexibility in the positioning of the displays.
The controller 130 may determine a relative location of each
display with the GPS module 115 or geomagnetic sensor 152 for
example. This would allow a user of one display to perform user
manipulations relative to another display. For example, a user may
perform a drag operation on first display 190b in the direction of
second display 190b, to perform an operation on second display
190b.
[0345] In the case of a plurality of displays, manipulations on one
display can affect multiple other displays. For example, as
illustrated in FIG. 48, a user of first display 190a in wireless
communication (represented by the double arrow) with second display
190b and a third display 190c may perform a manipulation (a touch
and hold operation) to capture content displayed on the first
display 190a (see FIG. 48 view 1). Subsequently, user manipulations
(for example, a double tap with an input pen 200) performed on
second display 190b and third display 190c may insert the content
into each corresponding display as a memo 2110 (see FIG. 48 view
2).
[0346] Although a few embodiments of the present general inventive
concept have been shown and described, it will be appreciated by
those skilled in the art that changes may be made in these
exemplary embodiments without departing from the principles and
spirit of the invention, the scope of which is defined in the
appended claims and their equivalents.
* * * * *