U.S. patent application number 14/696698 was filed with the patent office on 2016-02-25 for depth detecting apparatus and method, and gesture detecting apparatus and gesture detecting method.
This patent application is currently assigned to SAMSUNG ELECTRONICS CO., LTD.. The applicant listed for this patent is SAMSUNG ELECTRONICS CO., LTD.. Invention is credited to Myungjae JEON, Yonghwa PARK, Heesun YOON, Jangwoo YOU.
Application Number | 20160057340 14/696698 |
Document ID | / |
Family ID | 55349385 |
Filed Date | 2016-02-25 |
United States Patent
Application |
20160057340 |
Kind Code |
A1 |
YOU; Jangwoo ; et
al. |
February 25, 2016 |
DEPTH DETECTING APPARATUS AND METHOD, AND GESTURE DETECTING
APPARATUS AND GESTURE DETECTING METHOD
Abstract
A gesture detecting apparatus including a light emitter
configure to emit light towards an object, a camera configured to
capture light emitted from the light emitter and reflected by the
object, and a signal controller configured to control the light
emitter and the camera, in which the light emitter comprises a
first light and second light, at least one of which is configured
to emit light having non-monotonic intensity characteristics.
Inventors: |
YOU; Jangwoo; (Seoul,
KR) ; PARK; Yonghwa; (Yongin-si, KR) ; YOON;
Heesun; (Seoul, KR) ; JEON; Myungjae;
(Yongin-si, KR) |
|
Applicant: |
Name |
City |
State |
Country |
Type |
SAMSUNG ELECTRONICS CO., LTD. |
Suwon-si |
|
KR |
|
|
Assignee: |
SAMSUNG ELECTRONICS CO.,
LTD.
Suwon-si
KR
|
Family ID: |
55349385 |
Appl. No.: |
14/696698 |
Filed: |
April 27, 2015 |
Current U.S.
Class: |
348/222.1 |
Current CPC
Class: |
H04N 5/2226 20130101;
G06T 7/521 20170101; H04N 5/2256 20130101; G06T 2207/10152
20130101; G06T 2207/30196 20130101; H04N 5/23229 20130101; G06K
9/2036 20130101; G06T 2207/10016 20130101; G06K 9/00355
20130101 |
International
Class: |
H04N 5/232 20060101
H04N005/232; G06K 9/00 20060101 G06K009/00; H04N 5/225 20060101
H04N005/225; G06T 7/00 20060101 G06T007/00 |
Foreign Application Data
Date |
Code |
Application Number |
Aug 22, 2014 |
KR |
10-2014-0109967 |
Claims
1. A gesture detecting apparatus comprising: a light emitter
configure to emit light towards an object; a camera configured to
capture light emitted from the light emitter and reflected by the
object; and a signal controller configured to control the light
emitter and the camera, wherein the light emitter comprises a first
light and second light, at least one of which is configured to emit
light having non-monotonic intensity characteristics.
2. The gesture detecting apparatus of claim 1, further comprising
an image processor connected to the signal controller, the image
processor being configured to generate an image representing light
captured by the camera.
3. The gesture detecting apparatus of claim 1, further comprising a
storage connected to the signal controller configured to store
object information.
4. The gesture detecting apparatus of claim 1, wherein the light
emitter comprises a plurality of light sources and a plurality of
beam shapers.
5. The gesture detecting apparatus of claim 4, wherein the first
light comprises a first light source from among the plurality of
light sources and a first beam shaper from among the plurality of
beam shapers and the second light comprises a second light source
from among the plurality of light sources and a second beam shaper
from among the plurality of beam shapers.
6. The gesture detecting apparatus of claim 1, wherein the light
emitter comprises one or more light sources configured to emit an
infrared ray or a near-infrared ray.
7. The gesture detecting apparatus of claim 1, wherein the light
emitter comprises a laser diode (LD) or a light emitting diode
(LED).
8. The gesture detecting apparatus of claim 1, wherein the camera
comprises an image sensor.
9. The gesture detecting apparatus of claim 8, wherein the image
sensor comprises a charge coupled device (CCD) or a complementary
metal oxide semiconductor (CMOS) image sensor (CIS).
10. The gesture detecting apparatus of claim 1, wherein the camera
unit further comprises a band pass filter configured to allow light
in a band corresponding to a wavelength of the light emitted by the
light emitter to pass through, and to block light in another
band.
11. A gesture detecting method comprising: emitting light having
non-monotonic intensity characteristics toward an object from a
lighting unit; capturing light reflected by the object after being
emitted toward the object by using a camera unit; and analyzing
image information captured by the camera unit by using a signal
processing unit.
12. The gesture detecting method of claim 11, wherein the lighting
unit comprises a first light and a second light, at least one of
which emits light having non-monotonic intensity
characteristics.
13. The gesture detecting method of claim 12, further comprising:
emitting light toward the object from the first light and capturing
light reflected by the object by using the camera unit; emitting
light toward the object from the second light and capturing light
reflected by the object by using the camera unit; and capturing an
image by using the camera unit when a light is not emitted from the
lighting unit.
14. The gesture detecting method of claim 11, wherein the lighting
unit is controlled by a signal control unit, the signal control
unit comprising a switching unit that controls an on/off state and
an intensity of the lighting unit.
15. The gesture detecting method of claim 13, wherein the signal
control unit detects the object and a gesture based on an intensity
ratio, an intensity difference, or a combination thereof by using
first intensity information of the first light and second intensity
information of the second light, the first intensity information
and the second intensity information being captured by the camera
unit.
16. A depth detecting apparatus comprising: a first light
configured to emit light having non-monotonic intensity
characteristics towards an object; a camera configured to capture
light emitted by the first light and reflected by the object; and a
signal controller configured to determine a depth of the object
based on the captured reflected light.
17. The depth detecting apparatus according to claim 16, wherein
the camera is further configured to capture an image of the object
when light is not emitted by the first light, and wherein the
signal controller is further configured to associate the determined
depth of the object with the representation of the object in the
captured image.
18. The depth detecting apparatus according to claim 16, wherein
the signal controller is further configured to control an intensity
of the light emitted by the first light.
19. The depth detecting apparatus according to claim 16, further
comprising: a second light configured to emit light towards the
object, wherein the camera is further configured to capture light
emitted by the second light and reflected by the object, and
wherein the signal controller is configured to determine the depth
of the object based on a comparison of a first intensity of the
captured reflected light emitted by the first light and a second
intensity of the captured reflected light emitted by the second
light.
20. The depth detecting apparatus according to claim 19, wherein
the signal controller is configured to determine the depth of the
object based on at least one of an intensity ratio of the first
intensity and the second intensity, and an intensity difference of
the first intensity and the second intensity.
Description
RELATED APPLICATION
[0001] This application claims the benefit of Korean Patent
Application No. 10-2014-0109967, filed on Aug. 22, 2014, in the
Korean Intellectual Property Office, the disclosure of which is
incorporated herein in its entirety by reference.
BACKGROUND
[0002] 1. Field
[0003] The present disclosure relates to a depth detecting
apparatus and method for detecting a position of an object, and a
gesture detecting apparatus and gesture detecting method for
accurately detecting a gesture of an object.
[0004] 2. Description of the Related Art
[0005] To acquire a depth image of an object, time-of-flight (TOF)
or triangulation may be used.
[0006] When TOF is used, an infrared (IR) light is emitted from a
source toward an object and a depth image of the object is
generated based on the TOF taken by the IR light reflected back to
the source. This method requires the use of a dedicated
complementary metal oxide semiconductor (CMOS) image sensor (CIS)
or a high-speed intensity modulator, which is generally referred to
as an optical shutter. When a dedicated CIS is used, a function
that modulates the intensity of light in each pixel must be
provided, which requires an increased chip size and which increases
power consumption.
[0007] When triangulation is used, pattern-emitting lighting and a
dedicated high-speed image processing for an application specific
integrated circuit (ASIC) for extracting depth information from a
pattern change are required. Conventionally, during triangulation,
a position of the lighting pattern may change proportionally to a
distance to an object and increased computing power may be needed
to extract a position change of the lighting pattern.
SUMMARY
[0008] Provided is a depth detecting apparatus capable of detecting
a depth of an object.
[0009] Provided is a depth detecting method for detecting a depth
of an object by using a depth detecting sensor.
[0010] Provided is a gesture detecting apparatus capable of
detecting a gesture of an object.
[0011] Provided is a gesture detecting method for detecting a
gesture of an object by using a gesture detecting sensor.
[0012] Additional aspects will be set forth in part in the
description which follows and, in part, will be apparent from the
description, or may be learned by practice of the presented
embodiments.
[0013] According to an aspect of the present invention, a gesture
detecting apparatus is provided. The gesture detecting apparatus
includes a light emitter configure to emit light towards an object,
a camera configured to capture light emitted from the light emitter
and reflected by the object, and a signal controller configured to
control the light emitter and the camera, in which the light
emitter includes a first light and a second light, at least one of
which is configured to emit light having non-monotonic
intensity.
[0014] The gesture detecting apparatus may further include an image
processor connected to the signal controller, the image processor
being configured to generate an image representing light captured
by the camera.
[0015] The gesture detecting apparatus may further include a
storage connected to the signal controller configured to store
object information.
[0016] The light emitter may include a plurality of light sources
and a plurality of beam shapers.
[0017] The first light may include a first light source from among
the plurality of light sources and a first beam shaper from among
the plurality of beam shapers and the second light may include a
second light source from among the plurality of light sources and a
second beam shaper from among the plurality of beam shapers.
[0018] The light emitter may include one or more light sources
configured to emit an infrared ray or a near-infrared ray.
[0019] The light emitter may include a laser diode (LD) or a light
emitting diode (LED).
[0020] The camera may include an image sensor.
[0021] The image sensor may include a charge coupled device (CCD)
or a complementary metal oxide semiconductor (CMOS) image sensor
(CIS).
[0022] The camera unit may further include a band pass filter
configured to allow light in a band corresponding to a wavelength
of the light emitted by the light emitter to pass through, and to
block light in another band.
[0023] According to another aspect of the present invention, there
is provided a gesture detecting method. The gesture detecting
method includes emitting light having non-monotonic intensity
characteristics toward an object from a lighting unit, capturing
light reflected from the object after being emitted toward the
object by using a camera unit, and analyzing image information
captured by the camera unit by using a signal processing unit.
[0024] The lighting unit may include a first light and a second
light, at least one of which emits light having non-monotonic
intensity characteristics.
[0025] The gesture detecting method may further include emitting
light toward the object from the first light and capturing light
reflected from the object by using the camera unit, emitting light
toward the object from the second light and capturing light
reflected from the object by using the camera unit, and capturing
an image by using the camera unit when a light is not emitted from
the lighting unit.
[0026] The lighting unit may be a signal control unit, and the
signal control unit may include a switching unit that controls an
on/off state and an intensity of the lighting unit.
[0027] The signal control unit may detect the object and a gesture
based on an intensity ratio, an intensity difference, or a
combination thereof by using first intensity information of the
first light and second intensity information of the second light,
the first intensity information and the second intensity
information being captured by the camera unit.
[0028] According to another aspect of the present invention, a
depth detecting apparatus is provided. The depth detecting
apparatus includes a first light configured to emit light having
non-monotonic intensity characteristics towards an object; a camera
configured to capture light emitted by the first light and
reflected by the object; and a signal controller configured to
determine a depth of the object based on the captured reflected
light.
[0029] The camera may be further configured to capture an image of
the object when light is not emitted by the first light, and the
signal controller may be further configured to associate the
determined depth of the object with the representation of the
object in the captured image.
[0030] The signal controller may be further configured to control
an intensity of the light emitted by the first light.
[0031] The depth detecting apparatus may further include a second
light configured to emit light towards the object. The camera may
be further configured to capture light emitted by the second light
and reflected by the object, and the signal controller may be
configured to determine the depth of the object based on a
comparison of a first intensity of the captured light emitted by
the first light and a second intensity of the captured light
emitted by the second light.
[0032] The signal controller may be configured to determine the
depth of the object based on at least one of an intensity ratio of
the first intensity and the second intensity, and an intensity
difference of the first intensity and the second intensity.
[0033] According to another aspect of the present invention, a
depth detecting method is provided. The depth detecting method
includes: emitting a first light having non-monotonic intensity
characteristics toward an object; capturing the first light
reflected by the object; and determining a depth of the object
based on the captured reflected first light.
[0034] The depth detecting method may further include: emitting a
second light toward the object; and capturing the second light
reflected by the object. The determining may include determining
the depth of the object based on a comparison of a first intensity
of the captured reflected first light and a second intensity of the
captured reflected second light.
[0035] The determining may include determining the depth of the
object based on at least one of an intensity ratio of the first
intensity and the second intensity, and an intensity difference of
the first intensity and the second intensity.
BRIEF DESCRIPTION OF THE DRAWINGS
[0036] These and/or other aspects will become apparent and more
readily appreciated from the following description of one or more
exemplary embodiments, taken in conjunction with the accompanying
drawings in which:
[0037] FIG. 1 is a block diagram schematically illustrating a
gesture or depth detecting apparatus according to an exemplary
embodiment;
[0038] FIG. 2 illustrates detection of a gesture of an object by
using a wearable device including a gesture detecting apparatus
according to an exemplary embodiment;
[0039] FIG. 3A illustrates a basic example of a gesture detecting
method using a gesture detecting apparatus according to an
exemplary embodiment;
[0040] FIG. 3B illustrates an example of determining a distance to
an actual object from a distance measured via a gesture or depth
detecting method using a gesture or depth detecting apparatus
according to an exemplary embodiment;
[0041] FIG. 4A is a graph illustrating a light intensity with
respect to a lighting angle of a gesture or depth detecting
apparatus according to an exemplary embodiment;
[0042] FIG. 4B is a graph illustrating a light intensity ratio with
respect to a lighting angle of a gesture or depth detecting
apparatus according to an exemplary embodiment;
[0043] FIG. 5A is a graph illustrating a light intensity with
respect to a lighting angle of a gesture or depth detecting
apparatus according to an exemplary embodiment;
[0044] FIG. 5B is a graph illustrating a light intensity ratio with
respect to a lighting angle of a gesture or depth detecting
apparatus according to an exemplary embodiment; and
[0045] FIG. 5C is a graph illustrating a light intensity difference
with respect to a lighting angle of a gesture or depth detecting
apparatus according to an exemplary embodiment.
DETAILED DESCRIPTION
[0046] Hereinafter, a depth detecting apparatus and a depth
detecting method using the same, and a gesture detecting apparatus
and a gesture detecting method using the same according to one or
more exemplary embodiments will be described in detail with
reference to the accompanying drawings. Throughout the drawings,
like reference numerals refer to like elements, and the size of
each element may be exaggerated for clarity and convenience.
[0047] FIG. 1 is a block diagram schematically illustrating a
gesture or depth detecting apparatus according to an exemplary
embodiment.
[0048] Referring to FIG. 1, the gesture or depth detecting
apparatus may include a light emitter 110, i.e. a lighting unit,
including first light (light 1) 114 and second light (light 2) 112,
and a camera 120, i.e. a camera unit, capable of capturing light
reflected from an object after being emitted from the light emitter
110. The gesture or depth detecting apparatus may also include a
signal controller 130, i.e. a signal controller unit, that controls
the light emitter 110 and the camera 120, an image processor 140
that is connected to the signal controller 130 and generates an
image of the object for a display or the like to allow a user to
check the image, and a storage 150 capable of storing object
information detected by the gesture or depth detecting
apparatus.
[0049] Each of the first light 114 and the second light 112 of the
light emitter 110 may include a light source and a beam shaper. The
light source may emit, for example, an infrared ray or a near
infrared ray. The light source may be a laser diode (LD) or a light
emitting diode (LED). The beam shaper may be disposed in a
light-emitting unit of the light source, and light emitted from the
light source may be emitted to an object through the beam shaper.
At least one of the first light 114 or the second light 112 of the
light emitter 110 emits light having non-monotonic intensity
characteristics. For example, when a gesture or depth of an object
is to be detected using the gesture or depth detecting apparatus,
the intensity of light emitted from the first light 114 may
gradually increase and the intensity of light emitted from the
second light 112 may gradually decrease. Alternatively, the
intensity of light emitted from the first light 114 may decease and
the intensity of light emitted from the second light 112 may
gradually increase. The light emitted from the first light 114 may
be non-monotonic and the light emitted from the second light 112
may be either monotonic or non-monotonic. The light emitted from
the second light 112 may be non-monotonic and the light emitted
from the first light 114 may be either monotonic or
non-monotonic.
[0050] The camera 120 may include an image sensor that may be, for
example, a charge coupled device (CCD) or a complementary metal
oxide semiconductor (CMOS) image sensor (CIS). The camera 120 may
include a band pass filter through which light in a band
corresponding to a wavelength of light emitted from the first light
114 and the second light 112 may pass, and light in other bands may
be blocked. The camera 120 may include an image sensor having a
one-dimensional (1D) or two-dimensional (2D) array structure.
[0051] The gesture or depth detecting apparatus emits light to an
external object from the first light 114 and captures the light
reflected from the object through the camera 120. Thereafter, the
gesture or depth detecting apparatus emits light from the second
light 112 and captures light reflected from the object through the
camera 120. Then, an image is captured by the camera 120 without
emitting light from the light emitter 110. Image information
captured by the camera 120 is analyzed by the signal controller
130. The signal controller 130 may include a switch capable of
controlling an on/off state and an intensity of the light emitter
110. If ambient light is strong in an area where the object is
placed, the signal control unit 130 may increase the power of the
light emitter 110 and may control the gesture or depth detecting
apparatus to operate in a burst mode, reducing an on-operation
time.
[0052] FIG. 2 illustrates detection of a gesture of the object by
using a wearable device including the gesture detecting apparatus
according to an exemplary embodiment.
[0053] Referring to FIG. 2, a wearable device 20 may include a
light emitter 22 that emits light L1 towards an object 200 and a
camera 24 onto which light L2 reflected from the object 200 is
incident. The wearable device 20 may include a display 26, i.e. a
display unit, and may represent a gesture 220 of the object 200 as
an image 28 corresponding to position information detected by the
gesture detecting apparatus. The wearable device 20 may have
various forms, and may be, for example, a wrist-watch type or
glasses-type wearable device. The gesture detecting apparatus may
be mounted on various electronic devices such as a cellular phone,
a tablet personal computer (PC), a laptop computer, or a television
(TV).
[0054] FIG. 3A illustrates a basic example of a gesture detecting
method using a gesture detecting apparatus.
[0055] Referring to FIG. 3A, a line connecting a lighting unit I
with a camera unit C is set as a reference line. An object P
located outside the gesture detecting is within a range A11 in
which light L11 emitted from the light source I is emitted and also
within a viewing angle range A12 of the camera C. In the gesture
detecting apparatus, a distance d1 between the light emitter I and
the camera C may be fixed. The light L11 is emitted toward the
object P from the light emitter I and light L12 reflected from the
object P is incident onto the camera C. If an angle between the
object P and the light emitter I with respect to the reference line
is 8 and an angle between the object P and the camera unit C with
respect to the reference line is a, a distance R between the object
P and the camera unit C may be calculated as follows:
R=d.times.tan .theta./(sin(.alpha.)+tan(.theta.)cos(.alpha.))
(1)
Since the angle .theta. between the object P and the light emitter
I with respect to the reference line and the angle .alpha. between
the object P and the camera C with respect to the reference line
are known, position information of the object P in a space may be
obtained. By continuously obtaining such position information,
gesture information regarding the object P in the space may also be
obtained.
[0056] As an alternative, a basic example of a depth detecting
method using a depth detecting apparatus would detect the position
information once or occasionally. As a non-limiting example, if the
depth detecting apparatus were combined with a photographic camera,
the depth detecting method may be executed once corresponding to a
photograph being taken.
[0057] FIG. 3B illustrates an example of determining a distance to
an actual object from a distance measured by using a gesture or
depth detecting method using a gesture or depth detecting apparatus
according to an exemplary embodiment.
[0058] Referring to FIG. 3B, a line connecting a light emitter 11
with a camera C1 is set as a reference line d11, and a position of
an object P1 located in a range A21 in which light L21 emitted from
the light emitter 11 is emitted and in a range of a viewing angle
range A22 of the camera C1 is measured. According to the gesture or
depth detecting apparatus, a distance d11 between the light emitter
11 and the camera C1 may be fixed. Light L21 is emitted from the
light emitter 11 toward the object P1, and light L22 reflected from
the object P1 is incident onto the camera C1. As described with
reference to FIG. 1, the gesture or depth detecting method
performed with regard to the object P1 by using the gesture or
depth detecting apparatus includes emitting the light L12 from the
first light 114 of the light emitter 11 and then capturing the
light L22 reflected from the object P1 by using the camera C1. When
light is not emitted from the light emitter 11, an image is
captured by the camera C1. A method for distinguishing between an
actual object P1 and an illusory object P2 will be discussed
below.
[0059] A lighting intensity value measured from each of the first
light 114 and the second light 112 of the lighting unit 11 is shown
in FIG. 4A, and a light intensity ratio of the measured light
intensity values of the first light 114 and the second light 112 is
shown in FIG. 4B.
[0060] FIG. 4A is a graph illustrating a light intensity with
respect to a light angle of the gesture or depth detecting
apparatus according to an exemplary embodiment. In FIG. 4A, the
light intensity of the first light 114 is indicated by a solid line
and the light intensity of the second light 112 is indicated by a
dotted line. It can be seen from FIG. 4A that both the light
intensity of the first light 114 and the light intensity of the
second light 112 have non-monotonic characteristics. To measure a
distance to an object, a lighting angle .theta.0 may be extracted
using the light intensities.
[0061] FIG. 4B is a graph illustrating a light intensity ratio with
respect to a lighting angle of a gesture or depth detecting
apparatus according to an exemplary embodiment.
[0062] Referring to FIG. 4B, a quotient obtained by dividing the
light intensity of the first light 114 by the light intensity of
the second light 112 in FIG. 4A, that is, a light intensity ratio
with respect to a lighting angle, may be expressed via two similar
graphs. For example, in FIG. 4B, a lighting angle .theta. with
respect to a light intensity ratio may be about 4 at lighting
angles .theta.1 and .theta.2.
[0063] In this case, as shown in FIG. 3B, two objects P1 and P2
located at the angles .theta.1 and .theta.2 with respect to the
reference line may be indicated. To extract respective distances
between the objects P1 and P2 and the camera C1, an operation of
identifying an actual object between the two objects P1 and P2 may
be needed. To this end, various methods may be used. For example,
as shown in FIG. 3B, the object P1 corresponding to a region A23 of
interest may be selected as an actual object and the object P2
within a first distance D1 from the reference line connecting the
light emitter L1 with the camera C1 may be disregarded as an
illusory object. A region A23 within a particular distance (an area
between D1 and (D1+D2)) may be set as a region of interest. In this
way, the distance between the actual object P1 and the camera C1
may be measured. If reflective properties of the object P1 are
known in advance, the actual object P1 may be easily
determined.
[0064] The light intensity of the light emitter of the gesture or
depth detecting apparatus according to an exemplary embodiment may
have various structures and the phase, period, and size of the
light intensity may be adjusted arbitrarily.
[0065] FIG. 5A is a graph illustrating a light intensity with
respect to a lighting angle of a gesture or depth detecting
apparatus according to an exemplary embodiment.
[0066] Referring to FIG. 5A, the light intensity of the first light
114 of FIG. 1 is indicated by a solid line and the light intensity
of the second light 112 of FIG. 1 is indicated by a dotted line.
The light intensity of the first light 114 gradually increases from
a lighting angle .theta.30 to a lighting angle .theta.31, and the
light intensity of the second light 112 gradually decreases from
.theta.30 to .theta.31. The light intensity of the second light 112
gradually decreases from the lighting angle .theta.31 to a lighting
angle .theta.32, and the light intensity of the first light 114
gradually increases from .theta.31 to .theta.32. The light
intensity of the first light 114 gradually increases from the
lighting angle .theta.32 to a lighting angle .theta.33, and the
light intensity of the second light 112 gradually decreases from
.theta.32 to .theta.33.
[0067] A ratio of a light intensity M1 of the first light 114 to a
light intensity M2 of the second light 112, M1/M2, is shown in FIG.
5B. A ratio of a difference between the light intensity M1 of the
first light 114 and the light intensity M2 of the second light 112
to a sum of the light intensity M1 of the first light 114 and the
light intensity M2 of the second light 112 is shown in FIG. 5C. As
such, according to the gesture or depth detecting apparatus, the
signal controller 130 of FIG. 1 may detect a distance to the object
P1 and a position of the object P1 based on various values, for
example, an intensity ratio (M1/M2), an intensity difference
((M1-M2)/(M1+M2)), or various combinations thereof, by using the
light intensity M1 of the first light 114 and the light intensity
M2 of the second light 112.
[0068] As described above, the gesture or depth detecting apparatus
may detect a distance to an object located outside the gesture or
depth detecting apparatus and a spatial position of the object by
analyzing a non-monotonic intensity of a lighting unit. Moreover,
the gesture or depth detecting method may provide
improved-precision object gesture or depth detection through
analysis of intensities of first lighting and second lighting.
[0069] It should be understood that the exemplary embodiments
described therein should be considered descriptive and are
non-limiting. Descriptions of features or aspects within each
exemplary embodiment should typically be considered as available
for other similar features or aspects in other exemplary
embodiments.
[0070] While one or more exemplary embodiments have been described
with reference to the figures, it will be understood by those of
ordinary skill in the art that various changes in form and details
may be made therein without departing from the spirit and scope of
the present invention as defined by the following claims.
* * * * *