U.S. patent application number 15/214472 was filed with the patent office on 2016-11-10 for controlling method for a sensing system.
The applicant listed for this patent is PixArt Imaging Inc.. Invention is credited to Cho-Yi Lin, Chih-Hung Lu.
Application Number | 20160328088 15/214472 |
Document ID | / |
Family ID | 43068122 |
Filed Date | 2016-11-10 |
United States Patent
Application |
20160328088 |
Kind Code |
A1 |
Lin; Cho-Yi ; et
al. |
November 10, 2016 |
CONTROLLING METHOD FOR A SENSING SYSTEM
Abstract
A controlling method for a sensing system includes the following
step. Whether a specific function is started is determined
according to the amount of pixels of a total pattern acquired by
sensing a first object. In an embodiment of the present invention,
the step of determining whether the specific function is started
according to the amount of pixels of the total pattern is the step
of determining whether the specific function is started according
to the amount of pixels of the total pattern and the length of time
to successively acquire the total pattern. Accordingly, it is more
convenient for a user to use the sensing system having the said
controlling method.
Inventors: |
Lin; Cho-Yi; (HSINCHU,
TW) ; Lu; Chih-Hung; (HSINCHU, TW) |
|
Applicant: |
Name |
City |
State |
Country |
Type |
PixArt Imaging Inc. |
HSINCHU |
|
TW |
|
|
Family ID: |
43068122 |
Appl. No.: |
15/214472 |
Filed: |
July 20, 2016 |
Related U.S. Patent Documents
|
|
|
|
|
|
Application
Number |
Filing Date |
Patent Number |
|
|
13850311 |
Mar 26, 2013 |
|
|
|
15214472 |
|
|
|
|
12557466 |
Sep 10, 2009 |
8629854 |
|
|
13850311 |
|
|
|
|
Current U.S.
Class: |
1/1 |
Current CPC
Class: |
G06F 3/0428 20130101;
G06F 3/042 20130101; H04N 5/2258 20130101; G06F 3/04166 20190501;
G06K 9/00375 20130101; G06F 3/0416 20130101; G06F 3/03547 20130101;
G06F 3/0421 20130101 |
International
Class: |
G06F 3/042 20060101
G06F003/042; G06F 3/041 20060101 G06F003/041; G06F 3/0354 20060101
G06F003/0354; H04N 5/225 20060101 H04N005/225 |
Foreign Application Data
Date |
Code |
Application Number |
May 18, 2009 |
TW |
098116457 |
Claims
1. A controlling method for a sensing system, comprising: capturing
a first image and a second image above a panel from different
locations; extracting a first sub-pattern and a second sub-pattern
respectively from the first image and the second image, wherein the
first sub-pattern and the second sub-pattern are corresponding to
an object above the panel; and starting an erasing function when a
size of a total pattern is greater than or equal to a first
predetermined value, wherein the total pattern includes the first
sub-pattern and the second sub-pattern.
2. The controlling method for a sensing system as in claim 1,
wherein the sensing system comprises the panel, a first image
sensor and a second image sensor, the first image sensor and the
second image sensor are disposed at the panel, a sensing range of
the first image sensor and a sensing range of the second image
sensor cover an area of the panel, and the step of starting the
erasing function is determined according to an amount of pixels of
the total pattern and a length of time to successively acquire the
total pattern and comprises: (a) the first image sensor and the
second image sensor simultaneously sensing the area respectively
for acquiring the first image and the second image, wherein the
first image comprises the first sub-pattern, the second image
comprises the second sub-pattern, the first sub-pattern and the
second sub-pattern correspond to the object, and the total pattern
is comprised of the first sub-pattern and the second sub-pattern;
(b) determining whether the amount of pixels of the total pattern
is greater than or equal to the first predetermined value; (c)
determining whether the length of time to successively acquire the
total pattern is greater than or equal to a second predetermined
value, if the amount of pixels of the total pattern is greater than
or equal to the first predetermined value; and (d) starting the
erasing function if the length of time to successively acquire the
total pattern is greater than or equal to the second predetermined
value.
3. The controlling method for a sensing system as in claim 2
further comprising: (c1) calculating and outputting a first
coordinate of the object and returning to the Step (a) if the
amount of pixels of the total pattern is smaller than the first
predetermined value.
4. The controlling method for a sensing system as in claim 2,
further comprising: (d1) calculating and outputting a first
coordinate of the object and returning to the Step (a) if the
length of time to successively acquire the total pattern is smaller
than the second predetermined value.
5. The controlling method for a sensing system as in claim 2,
wherein the first image sensor has a plurality of first sensing
pixels adapted for sensing the area, the second image sensor has a
plurality of second sensing pixels adapted for sensing the area,
and the amount of pixels of the total pattern is a sum of an amount
of the first sensing pixels corresponding to the first sub-pattern
and an amount of the second sensing pixels corresponding to the
second sub-pattern.
6. The controlling method for a sensing system as in claim 1,
wherein the sensing system comprises the panel, an image sensor,
and a reflective mirror element, the image sensor is disposed at
the panel, the reflective mirror element is disposed at the panel
and mirrors a first area of the panel for forming a second area, a
sensing range of the image sensor covers the first area and the
second area, and the step of starting the erasing function is
determined according to an amount of pixels of the total pattern
and a length of time to successively acquire the total pattern and
comprises: (a) the image sensor sensing the first area and the
second area for acquiring a first image, wherein the first image
comprises a first sub-pattern and a second sub-pattern, the first
sub-pattern and the second sub-pattern correspond to the first
object, and the total pattern is comprised of the first sub-pattern
and the second sub-pattern; (b) determining whether the amount of
pixels of the total pattern is greater than or equal to the first
predetermined value; (c) determining whether the length of time to
successively acquire the total pattern is greater than or equal to
a second predetermined value, if the amount of pixels of the total
pattern is greater than or equal to the first predetermined value;
and (d) starting the erasing function if the length of time to
successively acquire the total pattern is greater than or equal to
the second predetermined value.
7. The controlling method for a sensing system as in claim 6,
further comprising: (c1) calculating and outputting a first
coordinate of the object and returning to the Step (a) if the
amount of pixels of the total pattern is smaller than the first
predetermined value.
8. The controlling method for a sensing system as in claim 6,
further comprising: (d1) calculating and outputting a first
coordinate of the first object and returning to the Step (a) if the
length of time to successively acquire the total pattern is smaller
than the second predetermined value.
9. The controlling method for a sensing system as in claim 6,
wherein the image sensor has a plurality of sensing pixels adapted
for sensing the first area and the second area, and the amount of
pixels of the total pattern is a sum of an amount of the sensing
pixels corresponding to the first sub-pattern and an amount of the
sensing pixels corresponding to the second sub-pattern.
Description
CROSS-REFERENCES
[0001] This is a Continuation Application of U.S. application Ser.
No. 13/850,311 (filed Mar. 26, 2013), which is a division of an
application Ser. No. 12/557,466, filed Sep. 10, 2009. U.S.
application Ser. No. 13/850,311 is based upon and claims the
benefit of priority from a Taiwan application No. 098116457 filed
on May 18, 2009. The entirety of the above-mentioned patent
applications are hereby incorporated by reference herein and made a
part of this specification.
FIELD OF THE INVENTION
[0002] The present invention relates to a controlling method and in
particular, to a controlling method for a sensing system.
BACKGROUND OF THE INVENTION
[0003] A conventional sensing system includes a panel, two image
sensors, and a processor. Each of the image sensors senses an area
of the panel for acquiring an image, and the processer successively
processes for calculating a position of a pointer on the area. The
abovementioned technology has been disclosed in many related
patents, such as U.S. Pat. No. 4,782,328, U.S. Pat. No. 6,803,906,
and U.S. Pat. No. 6,954,197.
[0004] However, the functions of the abovementioned sensing systems
still can not satisfy requirements of designers and therefore, the
conventional sensing systems need to be improved.
SUMMARY OF THE INVENTION
[0005] The present invention is directed to provide a controlling
method for a sensing system in which whether a specific function is
started is determined according to the amount of pixels of a first
total pattern acquired by sensing a first object.
[0006] The present invention provides a controlling method for a
sensing system including the following step. Whether a specific
function is started is determined according to the amount of pixels
of a first total pattern acquired.
[0007] In an embodiment of the present invention, the controlling
method, capture a first image and a second image above a panel from
different locations. Then, extract a first sub-pattern and a second
sub-pattern respectively from the first image and the second image,
wherein the first sub-pattern and the second sub-pattern are
corresponding to an object above the panel. Afterwards, start an
erasing function when a size of a total pattern is greater than or
equal to a first predetermined value, wherein the total pattern
includes the first sub-pattern and the second sub-pattern.
BRIEF DESCRIPTION OF THE DRAWINGS
[0008] The accompanying drawings are included to provide a further
understanding of the invention, and are incorporated in and
constitute part of this specification. The drawings illustrate
embodiments of the invention and, together with the description,
serve to explain the principles of the invention.
[0009] FIG. 1 is a schematic view of a sensing system according to
a first embodiment of the present invention.
[0010] FIG. 2A to FIG. 2B are a flow chart of a controlling method
for the sensing system of FIG. 1 according to the first embodiment
of the present invention.
[0011] FIG. 3A to FIG. 3B are a detailed flow chart of the
controlling method of FIGS. 2A and 2B.
[0012] FIG. 4 is a schematic view of a first image and a second
image respectively sensed by the first image sensor and the second
image sensor of the sensing system of FIG. 1.
[0013] FIG. 5 is a schematic view of a third image and a fourth
image sensed by the first image sensor and the second image sensor
of the sensing system of FIG. 1.
[0014] FIG. 6 is a schematic view of a sensing system according to
a second embodiment of the present invention.
[0015] FIG. 7A to FIG. 7B are a flow chart of a controlling method
for the sensing system of FIG. 6 according to the second embodiment
of the present invention.
[0016] FIG. 8A to FIG. 8B are a detailed flow chart of the
controlling method of FIGS. 7A and 7B.
[0017] FIG. 9 is a schematic view of a first image sensed by the
image sensor of the sensing system of FIG. 6.
[0018] FIG. 10 is a schematic view of a second image sensed by the
image sensor of the sensing system of FIG. 6.
DETAILED DESCRIPTION OF EMBODIMENTS
First Embodiment
[0019] FIG. 1 is a schematic view of a sensing system according to
a first embodiment of the present invention. Referring to FIG. 1,
the sensing system 200 of the embodiment includes a panel 210, a
first image sensor 220, a second image sensor 230 and a processor
240. The panel 210 has a surface 214 and an area 212, and the area
212 is located at the surface 214. The area 212 is a quadrilateral
for example. The first image sensor 220 and the second image sensor
230 are disposed at the surface 214 of the panel 210 and are
electrically connected to the processor 240. A sensing range of the
first image sensor 220 and a sensing range of the second image
sensor 230 cover the area 212, respectively. In this embodiment,
the panel 210 is a whiteboard or a touch screen for example.
[0020] When a user uses an object 20 which is capable of being
sensed by the sensing system 200 such that the object 20 approaches
the area 212, the first image sensor 220 and the second image
sensor 230 sense the area 212 respectively and the processor 240
calculates a position of the object 20 according to images sensed
by the first image sensor 220 and the second image sensor 230. The
abovementioned technologies can be referred to U.S. Pat. No.
4,782,328 and U.S. Pat. No. 6,803,906 and should not be described
in detail herein. In addition, when the object 20 moves on the area
212, the movement path will be displayed on the panel 210 such as a
touch screen or a display device (not shown) electrically connected
to the panel 210.
[0021] FIG. 2A to FIG. 2B are a flow chart of a controlling method
for the sensing system of FIG. 1 according to the first embodiment
of the present invention. FIG. 3A to FIG. 3B are a detailed flow
chart of the controlling method of FIGS. 2A and 2B. FIG. 4 is a
schematic view of a first image and a second image respectively
sensed by the first image sensor and the second image sensor of the
sensing system of FIG. 1. Referring to FIGS. 1, 2A, 2B, 3A, 3B and
4, generally speaking, the controlling method for the sensing
system includes the step of determining whether a specific function
is started according to the size of a first total pattern P1
acquired. In this embodiment, the abovementioned step is the step
of determining whether the specific function is started according
to the size of the first total pattern P1 acquired and the length
of time to successively acquire the first total pattern P1. The
description will be given in detail below.
[0022] First, a step (a) is executed. The step (a) includes that
the first image sensor 220 and the second image sensor 230
simultaneously sense the area 212 for respectively acquiring the
first image 222 and the second image 232. The first image 222
includes a first sub-pattern 222a and the second image 232 includes
a second sub-pattern 232a. The first total pattern P1 is composed
of the first sub-pattern 222a and the second sub-pattern 232a.
[0023] In this embodiment, the first image sensor 220 has a
plurality of first sensing pixels (not shown) suitable for sensing
the area 212, and the second image sensor 230 has a plurality of
second sensing pixels (not shown) suitable for sensing the area
212. When a first object (not shown) approaches the area 212 of the
panel 210, the first sensing pixels of the first image sensor 220
and the second sensing pixels of the second image sensor 230
simultaneously sense the area 212, and the first sensing pixels of
the first image sensor 220 acquire the first image 222 and the
second sensing pixels of the second image sensor 230 acquire the
second image 232. Meanwhile, at least a part of the first sensing
pixels sense the first object to acquire the first sub-pattern 222a
and at least a part of the second sensing pixels sense the first
object to acquire the second sub-pattern 232a. That is, at this
moment, the first object is adjacent to the area 212 of the panel
210 and sensed by the first image sensor 220 and the second image
sensor 230 to be acquired as the first sub-pattern 222a of the
first image 222 and the second sub-pattern 232a of the second image
232.
[0024] The first object may be a stylus, a finger, a board eraser
or a palm. If the first object is a stylus or a finger, the size of
the first total pattern P1 is relatively small. The size of the
first total pattern P1 is, for example, the sum total of the amount
of the corresponding first sensing pixels which acquire the first
sub-pattern 222a and the amount of the corresponding second sensing
pixels which acquire the second sub-pattern 232a. In addition, if
the first object is a board eraser or a palm, the size of the first
total pattern P1 is relatively large. It should be noted that the
first object corresponding to the first sub-pattern 222a and the
second sub-pattern 232a of FIG. 4 is, for example, a board
eraser.
[0025] It should be noted that when a plurality of first objects
approach the area 212 of the panel 210, for example, when two
fingers touch the area 212 of the panel 210, the first image 222
may include a plurality of first sub-patterns 222a and the second
image 232 may include a plurality of second sub-patterns 232a.
Meanwhile, the first total pattern P1 is composed of the first
sub-patterns 222a and the second sub-patterns 232a, and the size of
the first total pattern P1 is the sum total of the amount of the
corresponding first sensing pixels which acquire the first
sub-patterns 222a and the amount of the corresponding second
sensing pixels which acquire the second sub-patterns 232a.
[0026] Next, a step (b) is executed. The step (b) includes that the
processor 240 determines whether the size of the first total
pattern P1 is greater than or equal to a first predetermined value.
In this embodiment, the first predetermined value may be a
predetermined amount of pixels which may be between 1/2 of the sum
total of the amount of the first sensing pixels of the first image
sensor 220 and the amount of the second sensing pixels of the
second image sensor 230 and 2/3 of the sum total of the amount of
the first sensing pixels of the first image sensor 220 and the
amount of the second sensing pixels of the second image sensor
230.
[0027] Next, a step (c) is executed. The step (c) includes that if
the size of the first total pattern P1 is greater than or equal to
the first predetermined value, whether the length of time to
successively acquire the first total pattern P1 is greater than or
equal to a second predetermined value (e.g. 1.5 seconds) is further
determined. It should be noted that the step (c) of FIG. 2A may be
accomplished by means of the counting manner of FIG. 3A. For
example, the first image sensor 220 and the second image senor 230
are designed to sense the area 212 once every time interval (e.g.
1/240 seconds). Whenever the first image sensor 220 and the second
image sensor 230 sense the area 212, the processor 240 determines
whether the size of the first total pattern P1 is greater than or
equal to the first predetermined value. If the determination result
is positive, the counting value is increased by one and the
processor 240 further determines whether the counting value is
greater than or equal to a third predetermined value, as shown in
sub-steps (c2) and (c3) of FIGS. 3A and 3B. In this embodiment, the
second predetermined value (e.g. 1.5 seconds) is converted to the
third predetermined value (e.g. 360 times) by means of the
calculation (e.g. 1.5 divided by 1/240).
[0028] Next, a step (d) is executed. The step (d) includes that if
the length of time to successively acquire the first total pattern
P1 is greater than or equal to the second predetermined value, that
is, if the counting value is greater than or equal to the third
predetermined value, the processor 240 starts the specific
function, calculates and outputs a first coordinate of the first
object which corresponds at that time to the first sub-pattern 222a
and the second sub-pattern 232a. In this embodiment, at this
moment, the first object such as a board eraser or a palm is
regarded by the sensing system 200 as an erasing instrument, and
the specific function may erase the movement path displayed by the
panel 210 such as a touch screen or by the additional display
device. That is, the specific function may be an erasing
function.
[0029] In addition, after the step (b) is executed, the controlling
method for the sensing system further includes executing step (c1).
The step (c1) includes that if the size of the first total pattern
P1 is smaller than the first predetermined value, the processor 240
calculates and outputs the first coordinate of the first object
which corresponds at that time to the first sub-pattern 222a and
the second sub-pattern 232a and that the controlling process
returns to the step (a). In this embodiment, at this moment, the
first object such as a stylus or a finger is regarded by the
sensing system 200 as a pointer.
[0030] In addition, after the step (c) is executed, the controlling
method for the sensing system further includes executing step (d1).
The step (d1) includes that if the length of time to successively
acquire the first total pattern P1 is smaller than the second
predetermined value, that is, if the counting value is smaller the
third predetermined value, the processor 240 calculates and outputs
the first coordinate of the first object which corresponds at that
time to the first sub-pattern 222a and the second sub-pattern 232a
and that the controlling process returns to the step (a). In this
embodiment, at this moment, the first object such as a board eraser
or a palm is regarded by the sensing system 200 as a pointer
because the length of time to be successively acquired is not long
enough. For example, at this moment, the user may make a board
eraser or a palm approach the area 212 by mistake and immediately
remove the board eraser or the palm away from the area 212, and the
sensing system 200 is prevented from starting the specific function
because of misjudgment.
[0031] It should be noted that a first controlling period is
composed of the abovementioned steps (a), (b), (b1), (c), (d) and
(d1). During the first controlling period, each of things capable
of being sensed by the first image sensor 220 and the second image
sensor 230 is named as the first object. That is, during the first
controlling period, the first object is a common name of each of
the things capable of being sensed by the first image sensor 220
and the second image sensor 230.
[0032] In the controlling method for the sensing system, whether
the specific function is started is determined according to the
size of the first total pattern P1 acquired. Accordingly, it is
more convenient for a user to use the sensing system 200 having the
abovementioned controlling method. In addition, in the controlling
method for the sensing system, the abovementioned step may be the
step of determining whether the specific function is started
according to the size of the first total pattern P1 acquired and
the length of time to successively acquire the first total pattern
P1. Therefore, the possibility that the sensing system starts the
specific function because of misjudgment can be reduced.
[0033] FIG. 5 is a schematic view of a third image and a fourth
image sensed by the first image sensor and the second image sensor
of the sensing system of FIG. 1. In this embodiment, after the step
(d) is executed, the controlling method for the sensing system
further includes the following steps. Referring to FIGS. 1, 2A, 2B,
3A, 3B and 5, step (e) is executed. The step (e) includes that the
first image sensor 220 and the second image sensor 230
simultaneously sense the area 212 for respectively acquiring the
third image 224 and the fourth image 234. The third image 224
includes a third sub-pattern 224a, and the fourth image 234
includes a fourth sub-pattern 234a. A second total pattern P2 is
composed of the third sub-pattern 224a and the fourth sub-pattern
234a. It should be noted that the third image 224 and the fourth
image 234 correspond at that time to a second object adjacent to
the area 212. That is, at this moment, the second object is
adjacent to the area 212 of the panel 210 and sensed by the first
image sensor 220 and the second image sensor 230 to be acquired as
the third sub-pattern 224a of the third image sensor 224 and the
fourth sub-pattern 234a of the fourth image sensor 234.
[0034] The second object may be a stylus, a finger, a board eraser
or a palm. It should be noted that the second object corresponding
to the third sub-patter 224a and the fourth sub-pattern 234a of
FIG. 5 is, for example, a stylus.
[0035] Next, a step (f) is executed. The step (f) includes that
whether the size of the second total pattern P2 is smaller than the
first predetermined value is determined. The size of the second
total pattern P2 is, for example, the sum total of the amount of
the corresponding first sensing pixels which acquire the third
sub-pattern 224a and the amount of the corresponding second sensing
pixels which acquire the fourth sub-pattern 234a.
[0036] Next, a step (g) is executed. The step (g) includes that if
the size of the second total pattern P2 is smaller than the first
predetermined value, whether the length of time to successively
acquire the second total pattern P2 is greater than or equal to the
second predetermined value is further determined. It should be
noted that the step (g) of FIG. 2B may be accomplished by means of
the counting manner of FIG. 3B. For example, whenever the first
image sensor 220 and the second image sensor 230 sense the area
212, the processor 240 determines whether the size of the second
total pattern P2 is smaller than the first predetermined value. If
the determination result is positive, the counting value is
increased by one and the processor 240 further determines whether
the counting value is greater than or equal to the third
predetermined value, as shown in sub-steps (g2) and (g3) of FIGS.
3A and 3B.
[0037] Next, a step (h) is executed. The step (h) includes that if
the length of time to successively acquire the second total pattern
P2 is greater than or equal to the second predetermined value, that
is, if the counting value is greater than or equal to the third
predetermined value, the processor 240 ends the specific function,
calculates and outputs a second coordinate of the second object
which corresponds at that time to the third sub-pattern 224a and
the fourth sub-pattern 234a. In this embodiment, at this moment,
the second object such as a stylus or a finger is regarded by the
sensing system 200 as a pointer.
[0038] In addition, after the step (f) is executed, the controlling
method for the sensing system further includes executing step (g1).
The step (g1) includes that if the size of the second total pattern
P2 is greater than or equal to the first predetermined value, the
processor 240 successively starts the specific function, calculates
and outputs the second coordinate of the second object which
corresponds at that time to the third sub-pattern 224a and the
fourth sub-pattern 234a and that the controlling process returns to
the step (e). In this embodiment, at this moment, the second object
such as a board eraser or a palm is regarded by the sensing system
200 as an erasing instrument.
[0039] In addition, after the step (g) is executed, the controlling
method for the sensing system further includes executing step (h1).
The step (h1) includes that if the length of time to successively
acquire the first total pattern P2 is smaller than the second
predetermined value, that is, if the counting value is smaller the
third predetermined value, the processor 240 successively starts
the specific function, calculates and outputs the second coordinate
of the second object which corresponds at that time to the third
sub-pattern 224a and the fourth sub-pattern 234a and that the
controlling process returns to the step (e). In this embodiment, at
this moment, the second object is such as a stylus is regarded by
the sensing system 200 as an erasing instrument because the length
of time to be successively acquired is not long enough. For
example, at this moment, the user may make a stylus or a finger
approach the area 212 by mistake and immediately remove the stylus
or the finger away from the area 212, and the sensing system 200 is
prevented from ending the specific function because of
misjudgment.
[0040] It should be noted that the abovementioned a second
controlling period is composed of steps (e), (f), (g1), (g), (h)
and (h1). During the second controlling period, each of things
capable of being sensed by the first image sensor 220 and the
second image sensor 230 is named as the second object. That is,
during the second controlling period, the second object is a common
name of each of the things capable of being sensed by the first
image sensor 220 and the second image sensor 230.
[0041] It should be noted that in practice, before the step (a) of
the first controlling period of the controlling method for the
sensing system 200 is executed, the user may use stylus such
another specific function (regarded as the second specific
function) of the sensing system 200 is performed. The second
specific function is, for example, the handwriting function or the
drawing function. Next, after the specific function such as the
erasing function (regarded as the first specific function which is
different from the second specific function) is started for a
period of time (e.g. the user uses a board eraser) and then is
ended (e.g., the user removes the board eraser and picks up the
stylus), the second specific function is started again. That is,
the second specific function is a predetermined function of the
sensing system 200.
Second Embodiment
[0042] FIG. 6 is a schematic view of a sensing system according to
a second embodiment of the present invention. The sensing system
300 includes a panel 310, an image sensor 320, a reflective mirror
element 330 and a processor 340. The image sensor 320 is disposed
at the panel 310. The reflective mirror element 330 is disposed at
the panel 310 and a reflective mirror plane 332 of the reflective
mirror element 330 mirrors a first area 312 of the panel 312 for
forming a second area 312'. The image sensor 320 faces the
reflective mirror element 330 and is electrically connected to the
processor 340. A sensing range of the image sensor 320 covers the
first area 312 and the second area 312'. That is, the second image
sensor 230 of the first embodiment is omitted in the sensing system
300 of this embodiment and the reflective mirror element 330 is
added in the sensing system 300 of this embodiment.
[0043] When a user uses an object 30 which is capable of being
sensed by the sensing system 300 such that the object 20 approaches
to the area 312, the object 30 is mirrored by the reflective mirror
element 330 to form a mirror image 30'. The image sensor 320 senses
the first area 312 and the second area 312' and the processor 340
calculates a position of the object 30 according to images sensed
by the image sensor 320. The abovementioned technologies can be
referred to Taiwan patent application No. 097126033, Taiwan patent
application No. 097142355 and Taiwan patent application No.
098100969 and U.S. patent application Ser. No. 12/249,222, U.S.
patent application Ser. No. 12/334,449 and U.S. patent application
Ser. No. 12/422,191, which respectively correspond to the
abovementioned Taiwan patent applications, and should not be
described in detail herein.
[0044] FIG. 7A to FIG. 7B are a flow chart of a controlling method
for the sensing system of FIG. 6 according to the second embodiment
of the present invention. FIG. 8A to FIG. 8B are a detailed flow
chart of the controlling method of FIGS. 7A and 7B. FIG. 9 is a
schematic view of a first image sensed by the image sensor of the
sensing system of FIG. 6. FIG. 10 is a schematic view of a second
image sensed by the image sensor of the sensing system of FIG. 6.
Referring to FIGS. 6 to 10, the difference between the controlling
method for the sensing system of this embodiment and the
controlling method for the sensing system of the first embodiment
is that the steps (a) and (e) of this embodiment are different from
the steps (a) and (e) of the first embodiment.
[0045] In the step (a) of this embodiment, the image sensor 320
senses the first area 312 and the second area 312' for acquiring a
first image 322. The first image 322 includes a first sub-pattern
322a and a second sub-pattern 322b, and a first total pattern P3 is
composed of the first sub-pattern 322a and the second sub-pattern
322b.
[0046] In this embodiment, the first image sensor 320 has a
plurality of sensing pixels (not shown) suitable for sensing the
first area 312 and the second area 312'. When a first object (not
shown) approaches the first area 312 of the panel 310, the sensing
pixels of the image sensor 320 sense the first area 312 and the
second area 312', and the sensing pixels of the image sensor 320
acquire the first image 322. Meanwhile, at least a part of the
sensing pixels sense the first object to acquire the first
sub-pattern 322a, and at least a part of the sensing pixels sense
the mirror image of the first object to acquire the second
sub-pattern 322b. That is, at this moment, the first object is
adjacent to the first area 312 of the panel 310, and the first
object and its mirror image are sensed by the image sensor 320 to
be acquired as the first sub-pattern 322a and the second
sub-pattern 322b of the first image 322.
[0047] The size of the first total pattern P3 is, for example, the
sum total of the amount of the corresponding sensing pixels which
acquire the first sub-pattern 322a and the amount of the
corresponding sensing pixels which acquire the second sub-pattern
322b. It should be noted that the first object corresponding to the
first sub-pattern 322a of FIG. 9 is, for example, a board
eraser.
[0048] In the step (e) of this embodiment, the image sensor 320
senses the first area 312 and the second area 312' for acquiring a
second image 324. The first image 324 includes a third sub-pattern
324a and a fourth sub-pattern 324b, and a second total pattern P4
is composed of the third sub-pattern 324a and the fourth
sub-pattern 324b. The third sub-pattern 324a and the fourth
sub-pattern 324b correspond at that time to a second object
adjacent to the first area 312 and a mirror image of the second
object, respectively. That is, at this moment, the second object is
adjacent to the first area 312 of the panel 310, and the second
object and its mirror image are sensed by the image sensor 320 to
be acquired as the third sub-pattern 324a and the fourth
sub-pattern 324b of the second image 324.
[0049] The size of the second total pattern P4 is, for example, the
sum total of the amount of the corresponding sensing pixels which
acquire the third sub-pattern 324a and the amount of the
corresponding sensing pixels which acquire the fourth sub-pattern
324b. It should be noted that the second object corresponding to
the third sub-patter 324a of FIG. 10 is, for example, a stylus.
[0050] In addition, the first predetermined value of this
embodiment may be a predetermined amount of pixels which may be
between 1/2 of the amount of the sensing pixels of the image sensor
320 and 2/3 of the amount of the sensing pixels of the image sensor
320.
[0051] According to the mentioned above, the controlling method for
the sensing system of the embodiment of the present invention has
at least one of the following or other advantages:
[0052] 1. In the controlling method for the sensing system, whether
the specific function is started is determined according to the
size of the first total pattern acquired. Accordingly, it is more
convenient for a user to use the sensing system having the
abovementioned controlling method.
[0053] 2. In the controlling method for the sensing system, the
abovementioned step is the step of determining whether the specific
function is started according to the size of the first total
pattern acquired and the length of time to successively acquire the
first total pattern. Therefore, the possibility that the sensing
system starts the specific function because of misjudgment can be
reduced.
[0054] The above description is given by way of example, and not
limitation. Given the above disclosure, one skilled in the art
could devise variations that are within the scope and spirit of the
invention disclosed herein, including configurations ways of the
recessed portions and materials and/or designs of the attaching
structures. Further, the various features of the embodiments
disclosed herein can be used alone, or in varying combinations with
each other and are not intended to be limited to the specific
combination described herein. Thus, the scope of the claims is not
to be limited by the illustrated embodiments.
* * * * *