U.S. patent application number 14/478517 was filed with the patent office on 2015-05-14 for eye detecting device and methods of detecting pupil.
The applicant listed for this patent is PIXART IMAGING INC.. Invention is credited to YU-HAO HUANG.
Application Number | 20150131051 14/478517 |
Document ID | / |
Family ID | 53043552 |
Filed Date | 2015-05-14 |
United States Patent
Application |
20150131051 |
Kind Code |
A1 |
HUANG; YU-HAO |
May 14, 2015 |
EYE DETECTING DEVICE AND METHODS OF DETECTING PUPIL
Abstract
An eye detecting device includes an optical assembly, an image
sensor and an arithmetic unit. The optical assembly provides a
plurality of incident lights. The incident lights enter into eyes
to form a plurality of glints. At least part of glints is near a
pupil of the eye. The image sensor is used to capture an eye image.
The eye image includes images of the glints. The arithmetic unit
analyzes the gray scale value of eye image and obtains the
distributions of the glints through the gray scale value. The
arithmetic unit determines the position of the pupil of eye
according to the distributions of the glints.
Inventors: |
HUANG; YU-HAO; (HSIN-CHU,
TW) |
|
Applicant: |
Name |
City |
State |
Country |
Type |
PIXART IMAGING INC. |
HSIN-CHU |
|
TW |
|
|
Family ID: |
53043552 |
Appl. No.: |
14/478517 |
Filed: |
September 5, 2014 |
Current U.S.
Class: |
351/206 |
Current CPC
Class: |
A61B 3/0025 20130101;
A61B 3/1216 20130101; A61B 3/14 20130101; A61B 3/113 20130101 |
Class at
Publication: |
351/206 |
International
Class: |
A61B 3/12 20060101
A61B003/12; A61B 3/14 20060101 A61B003/14; A61B 3/00 20060101
A61B003/00 |
Foreign Application Data
Date |
Code |
Application Number |
Nov 14, 2013 |
TW |
102141440 |
Claims
1. An eye detecting device comprising: an optical assembly
providing at least one incident light entering an eye to generate
at least one glint near a pupil of the eye; an image sensor
capturing an eye image including a glint image and a pupil image;
an arithmetic unit analyzing a gray scale value of the eye image
and obtaining at least one position of the glint according to the
gray scale value, wherein the arithmetic unit determines the
position of the pupil of eye according to the position of the
glint.
2. The eye detecting device according to claim 1, wherein the
incident light is infrared light.
3. The eye detecting device according to claim 1, wherein a gray
scale value of the pupil image is smaller than a threshold gray
scale value, and a gray scale value of the glints of the eye image
is larger than the threshold gray scale value.
4. The eye detecting device according to claim 1, wherein the
optical assembly provides a plurality of incident lights.
5. The eye detecting device according to claim 4, wherein the
optical assembly comprises at least one light source and at least
one dispersing component, the light source provides a light, and
the incident light is formed from the light through the dispersing
component.
6. The eye detecting device according to claim 4, wherein the
optical assembly comprises a plurality of light sources providing
the incident lights.
7. The eye detecting device according to claim 4 further comprising
a control unit, wherein the control unit controls the timing that
the incident lights are emitted into the eye, and the image sensor
captures the eye image at different timing, and the arithmetic unit
analyzes the gray scale value of the eye image at different
timing.
8. The eye detecting device according to claim 7, wherein the
arithmetic unit commands the control unit, so that the control unit
controls the timing that the incident lights are emitted into the
eye.
9. A method of detecting pupil comprising: providing at least one
incident light entering an eye to form at least one first glint,
and the first glint is located near a pupil of the eye; capturing
an first eye image including a glint image and a pupil image;
analyzing a gray scale value of the first eye image to obtain the
distributions of the first glint; and determining a position of the
pupil of the eye according to the distributions of the first
glint.
10. The method of detecting pupil according to claim 9, wherein the
incident lights is provided by an optical assembly.
11. The method of detecting pupil according to claim 9, wherein the
first eye image is captured by an image sensor.
12. The method of detecting pupil according to claim 9, wherein the
gray scale value of the eye image is analyzed by the arithmetic
unit.
13. The method of detecting pupil according to claim 9, wherein the
optical assembly provides a plurality of incident lights to form a
plurality of first glints on the eye.
14. The method of detecting pupil according to claim 13, wherein
the step of determining the position of the pupil of eye according
to the distributions of the first glint comprising: selecting a
threshold gray scale value; analyzing a gray scale value
distribution in a survey area defined by the first glints;
selecting an area in the survey area to be a specific area, wherein
the gray scale value of the specific area is less than the
threshold gray scale value; and determining whether the shape of
the specific area matches with the shape of the pupil of eye.
15. The method of detecting pupil according to claim 14, wherein
the survey area is surrounded by a plurality of first glints.
16. The method of detecting pupil according to claim 14, wherein
the gray scale value of the first glints in the first image is
large than the threshold gray scale value.
17. The method of detecting pupil according to claim 13, wherein
the first eye image is captured at a first timing, and the method
of detecting pupil further comprising: providing the incident
lights entering the eye to form a plurality of second glints on the
eye at a second timing, and at least part of second glints are
located near the pupil of the eye; capturing an second eye image at
the second timing, wherein the second eye image including a glint
image and a pupil image; analyzing a gray scale value of the second
eye image to obtain the distributions of the second glints, wherein
the distributions of the first glints are different from the
distributions of the second glints of the second glints; and
producing a difference image between the first image and the second
image by image subtraction.
18. The method of detecting pupil according to claim 17, wherein
the incident lights entering the eye at the first timing and the
second timing individually through a control unit.
19. A method of identifying iris comprising: when an eye is located
at a reference position, providing a plurality of incident lights
entering the eye to form a plurality of a first reference point, a
second reference point and a third reference point located near a
pupil of the eye as a mark for locating the eye at the reference
position, wherein the positions of the first reference point, the
second reference point and the third reference point are
corresponding to the emission position of the incident lights; when
the eye moves from the reference position to a measuring position,
the incident lights form a first measuring glint, a second
measuring glint, and a third measuring glint near a pupil of the
eye; capturing a eye image of the eye including a first measuring
glint image, a second measuring image, a third measuring glint
image, and an iris image; analyzing a gray scale value of the eye
image to obtain the positions of the first measuring glint, the
second measuring, and the third measuring glint; and Calculating a
displacement amount of the first measuring glint, the second
measuring glint, and the third measuring glint relative to the
first reference point, the second reference point and the third
reference point respectively, so as to obtain a deformation amount
caused by the iris image of the eye at the measuring position
relative to the iris image of the eye at the reference
position.
20. The method of identifying iris according to claim 19, wherein
the incident light is provided by at least one light source and at
least one dispersing component, the light source provides a light,
and the incident light is formed from the light through the
dispersing component.
Description
BACKGROUND
[0001] 1. Technical Field
[0002] The present disclosure relates to an eye detecting device,
in particular an eye detecting device for detecting pupil and
identifying iris.
[0003] 2. Description of Related Art
[0004] Currently, the eye detecting device can be used to detect
gaze direction or identify iris boundary. Most eye detecting
devices detects the eye gaze direction by using the characteristic
that the position of pupil changes with the gaze direction.
[0005] Generally, conventional eye detecting device detects the eye
gaze direction by using the glint formed by emitting the incident
light into the eye, and the glint is used to be a reference point
for locating eye.
[0006] Specifically, after capturing eye image, the conventional
eye detecting device identifies the pupil and glint from the whole
cornea image. In the process of identifying the pupil, the
conventional eye detecting device scans whole eye image. The
conventional eye detecting device analyzes the gray scale value
distribution of whole eye image for identifying the pupil and
glint. The conventional eye detecting device can obtain the
relative position of the pupil and glint, and then determines the
gaze direction according to the relative position.
SUMMARY
[0007] An exemplary embodiment of the present disclosure
illustrates an eye detecting device which determines the position
of the pupil according to at least one glint.
[0008] An exemplary embodiment of the present disclosure
illustrates an eye detecting device. The eye detecting device
includes an optical assembly, an image sensor, and an arithmetic
unit. The optical assembly provides a plurality of incident lights
entering an eye to generate a plurality of light glints near a
pupil of the eye. The image sensor captures an eye image including
a glint image and a pupil image. The arithmetic unit analyzes a
gray scale value of the eye image and obtains at least one position
of the glint according to the gray scale value. The arithmetic unit
determines the position of the pupil of eye through the position of
the glint.
[0009] An exemplary embodiment of the present disclosure
illustrates a method of detecting pupil which determines the
position of the pupil according to one glint or plurality of
glints.
[0010] An exemplary embodiment of the present disclosure
illustrates a method of detecting pupil including providing at
least one incident light entering an eye to form at least one first
glint, and the first glint is located near a pupil of the eye.
Capturing a first eye image including a glint image and a pupil
image. Analyzing a gray scale value of the first eye image to
obtain the distributions of the first glint. Determining a position
of the pupil of the eye according to the distributions of the first
glint.
[0011] An exemplary embodiment of the present disclosure
illustrates a method of identifying iris which obtains a
deformation amount of the iris image of the eye while the eye
moving.
[0012] An exemplary embodiment of the present disclosure
illustrates a method of identifying iris including when an eye is
located at a reference position, providing a plurality of incident
lights entering the eye to form a plurality of a first reference
point, a second reference point and a third reference point located
near a pupil of the eye as a mark for locating the eye at the
reference position, wherein the positions of the first reference
point, the second reference point and the third reference point are
corresponding to the emission position of the incident lights. When
the eye moves from the reference position to a measuring position,
the incident lights form a first measuring glint, a second
measuring glint, and a third measuring glint near a pupil of the
eye. Capturing a eye image of the eye including a first measuring
glint image, a second measuring image, a third measuring glint
image, and an iris image. Analyzing a gray scale value of the eye
image to obtain the positions of the first measuring glint, the
second measuring, and the third measuring glint. Calculating a
displacement amount of the first measuring glint, the second
measuring glint, and the third measuring glint relative to the
first reference point, the second reference point and the third
reference point respectively, so as to obtain a deformation amount
caused by the iris image of the eye at the measuring position
relative to the iris image of the eye at the reference
position.
[0013] An exemplary embodiment of the present disclosure
illustrates a method of identifying iris which obtains a resolution
variation of the iris image of the eye.
[0014] An exemplary embodiment of the present disclosure
illustrates a method of identifying iris including providing a
plurality of incident lights entering the eye. Setting a first
reference point, a second reference point and a third reference
point as a mark for locating the eye at the reference position,
wherein the positions of the first reference point, the second
reference point and the third reference point are corresponding to
the emission position of the incident lights. The incident lights
form a first measuring glint, a second measuring glint, and a third
measuring glint near a pupil of the eye, and a plurality of
positions of the first measuring glint, the second measuring glint,
and the third measuring glint are corresponding to those positions
of the first reference point, the second reference point and the
third reference point. Capturing a eye image of the eye including
the first measuring glint image, the second measuring image, the
third measuring glint image, and an iris image. Analyzing a gray
scale value of the eye image to obtain the positions of the first
measuring glint, the second measuring glint, and the third
measuring glint. Calculating a variation of the distance between
the first measuring glint and the second measuring glint with
respect to the distance between the first reference point and the
second reference point, and calculating a variation of the distance
between the second measuring glint and the third measuring glint
with respect to the distance between the second reference point and
the third reference point, so as to obtain an resolution variation
of an iris image when the eye is located at the measuring
position.
[0015] In summary, the present disclosure provides eye detecting
device, methods of detecting pupil and identifying iris. The eye
detecting device includes an optical assembly, an image sensor, and
an arithmetic unit. The arithmetic unit can analyze the gray scale
value distribution of the survey area near the arrangement of the
glint in the first eye image so as to reduce searching scope of the
pupil. Hence, the position of the pupil can be searched quickly.
Therefore, compared with conventional technology, the arithmetic
unit does not analyze the gray scale value distribution of whole
first eye image for searching scope of the pupil.
[0016] The present disclosure provides eye detecting device,
methods of detecting pupil. The eye detecting device includes an
optical assembly, an image sensor, an arithmetic unit, and the
control unit. Since the control unit controls the different
incident lights to emit into the different positions of the eye at
the different timing, the arrangement of the first and second glint
at the different timing can be arranged. The arrangement of the
first and second glint can be more confirmed through the gray scale
value and the special pattern after image subtraction. Hence, the
possibility of the misjudgments of the glints position can be more
reduced. The arithmetic unit can analyze the gray scale value
distribution of the survey area near the arrangement of the first
glints and/or second glints in the difference image so as to search
the position of the pupil quickly. Therefore, compared with
conventional technology, the arithmetic unit does not analyze the
gray scale value distribution of whole first or second eye image
for searching scope of the pupil.
[0017] The arithmetic unit can calculate the major axis and minor
axis of the said ellipse according to the first variation, the
second variation, and the third variation. Hence, the boundary of
the pupil P1 can be estimated so that the boundary of the pupil P1
can be searched quickly.
[0018] The present disclosure provides methods of identifying iris.
The arithmetic unit can calculate the major axis and minor axis of
the said ellipse according to the first, second, and third
variation. Hence, the boundary of the pupil can be estimated so
that the boundary of the pupil can be searched quickly.
[0019] The present disclosure provides methods of identifying iris.
The arithmetic unit can calculate the first, second, and third
variation. Hence, the boundary of the pupil can be estimated so
that the boundary of the pupil can be searched quickly.
[0020] In order to further understand the techniques, means and
effects of the present disclosure, the following detailed
descriptions and appended drawings are hereby referred, such that
through which, the purposes, features and aspects of the present
disclosure can be thoroughly and concretely appreciated; however,
the appended drawings are merely provided for reference and
illustration, without any intention to be used for limiting the
present disclosure.
BRIEF DESCRIPTION OF THE DRAWINGS
[0021] The accompanying drawings are included to provide a further
understanding of the present disclosure, and are incorporated in
and constitute a part of this specification. The drawings
illustrate exemplary embodiments of the present disclosure and,
together with the description, serve to explain the principles of
the present disclosure.
[0022] FIG. 1A depicts a side view of the eye detecting device in
accordance with the first embodiment of the present invention.
[0023] FIG. 1B is a front view of the eye detecting device shown in
FIG. 1A.
[0024] FIG. 1C is a function block diagram of the eye detecting
device in accordance with the first embodiment of the present
invention.
[0025] FIG. 1D depicts a flow diagram of a method of detecting
pupil in accordance with the first exemplary embodiment of the
present disclosure.
[0026] FIG. 2A depicts a side view of the eye detecting device in
accordance with the second embodiment of the present invention.
[0027] FIG. 2B is a function block diagram of the eye detecting
device in accordance with the second embodiment of the present
invention.
[0028] FIG. 2C depicts a flow diagram of a method of detecting
pupil in accordance with the second exemplary embodiment of the
present disclosure.
[0029] FIG. 3A depicts a function block diagram of the eye
detecting device in accordance with the third embodiment of the
present invention.
[0030] FIG. 3B depicts a flow diagram of a method of detecting
pupil in accordance with the second exemplary embodiment of the
present disclosure.
[0031] FIG. 4 depicts a flow diagram of a method of identifying
iris in accordance with the third exemplary embodiment of the
present disclosure.
[0032] FIG. 5 depicts a flow diagram of a method of identifying
iris in accordance with the fourth exemplary embodiment of the
present disclosure.
DESCRIPTION OF THE EXEMPLARY EMBODIMENTS
[0033] FIG. 1A is a side view of the eye detecting device in
accordance with the first embodiment of the present invention. FIG.
1B is a front view of the eye detecting device shown in FIG. 1A.
FIG. 1C is a function block diagram of the eye detecting device in
accordance with the first embodiment of the present invention.
Please refer to FIG. 1A to 1C, the eye detecting device 100
includes an optical assembly 110, an image sensor 120, and an
arithmetic unit 130. The optical assembly 110 provides at least one
incident light L1 to form at least one glint G1 located near a
pupil P1 of the eye E1. Specifically, the eye E1 has the pupil P1
and a periphery surrounding the pupil P1, and the glint G1 is
formed on the periphery. The periphery includes an iris I1 and a
sclera. The image sensor 120 is used to capture an eye image, and
the eye E1 image includes the glint G1 image. The arithmetic unit
130 analyzes a gray scale value of the eye E1 image and obtains at
least one position of the glint G1 according to the gray scale
value. Hence, the arithmetic unit 130 can determine the position of
the pupil P1 of eye E1 according to the position of the glint
G1.
[0034] The eye detecting device 100 can be disposed on the
eyeglasses frame, and the eye detecting device 100 also can be
disposed on the laptop or the screen of the smartphone. In this
embodiment, the eye detecting device 100 may be wearable, like
eyeglasses. The optical assembly 110 and the image sensor 120 are
disposed on the supporting frame 150. User can wear the supporting
frame 150, and the optical assembly 110 and the image sensor 120
are in front of the user. However, in other embodiment, the eye
detecting device 100 can be disposed on mobile device, for example,
laptop, the front camera lens or the screen of the smartphone.
However, the present disclosure does not limit the disposition of
the eye detecting device 100.
[0035] Practically, the supporting frame 150 can be an eyeglasses
frame. The supporting frame 150 includes two rims 152 and two
temples 154 connected to rims 152 respectively. User can put the
temples 154 on ears, and the rims 152 are in front of the eye E1.
However, the present disclosure does not limit the supporting frame
150.
[0036] The optical assembly 110 can emit at least one incident
light L1 entering the eye E1. The incident light L1 falls on the
eye E1 to form at least one glint by reflecting at the iris I1 of
the eye E1. The glint is located near a pupil P1 of the eye E1.
Specifically, the glint may be formed on the periphery surrounding
the pupil P1, namely iris I1 or sclera. In this embodiment,
providing one incident light L1 entering the eye E1 so that the
number of the glint is one. It is worth to mention that the
incident light L1 is the invisible light, such as infrared light or
near infrared light. The cornea covered on the iris I1 has a smooth
surface so that the incident light L1 emitted in many directions
can form the glint G1 through the path between the cornea and the
image sensor 120.
[0037] Specifically, the optical assembly 110 includes at least one
light source 112 and at least one dispersing component 114 so that
the optical assembly 110 provides at least one incident light L1.
Practically, the light source 112 can be light emitting diode
(LED), and the dispersing component 114 can guide light and has a
plurality of optical microstructures. The optical microstructures
can be optical microstructures, trenches or ribs. The trenches may
be V-cut grooves When the light provided by the light source 112 is
emitted into the dispersing component 114, the light can be
reflected, refracted, or scattered by the optical microstructures
so as to be transmitted from an outgoing surface of the dispersing
component 114.
[0038] The image sensor 120 is used to capture the eye E1 image. It
is worth to mention that the wavelength range of the light captured
by the image sensor 120 covers the wavelength range of the incident
light L1. The eye E1 image appears in the eye region of user, for
example, the eye white area (not shown), the iris I1 area, and the
pupil P1 area. Besides, the eye E1 image shows the glint G1 image.
Specifically, the image sensor 120 senses the incident light L1
through photo-sensitive elements. The photo-sensitive elements can
be complementary metal-oxide-semiconductor sensors (CMOS) or
charge-coupled devices (CCD).
[0039] The arithmetic unit 130 can be a digital signal processor
(DSP) or a central processing unit (CPU). The arithmetic unit 130
analyzes a gray scale value of the eye image and obtains the
distribution of the glint G1 through the gray scale value. The
arithmetic unit 130 determines the position of the pupil P1 of eye
E1 according to at least one distribution of the glint G1.
[0040] FIG. 1D depicts a flow diagram of a method of detecting
pupil in accordance with the first exemplary embodiment of the
present disclosure. Please refer to FIGS. 1B, FIG. 1C and FIG.
1D.
[0041] Implementing the step S101, when the user uses the eye
detecting device 100, such as the user wearing the supporting frame
150 of the eye detecting device 100, the optical assembly 110
provides one incident light L1 entering into the eye E1. The
incident light L1 is located at the eye E1 and reflects to form one
glint G1 near the pupil P1, such as the iris I1.
[0042] It is worth to notice that the position where the incident
light L1 enters the iris I1 near the pupil P1 can be adjusted by
changing the arrangement of the light source 112 or the disposition
of the light source 112 and the dispersing component 114. Namely,
the position of the glint G1 can be changed by the emission
position of the incident light L1. Hence, the position of the glint
G1 depends on the emission position of the incident light L1.
[0043] Implementing the step S102, the image sensor 120 captures a
first eye image by photographing the eye E1. The first eye image
photographed by image sensor 120 shows the image of the eye E1
region and the image of the said glint G1. Then, the image sensor
120 transmits the data of the first eye image to the arithmetic
unit 130.
[0044] Implementing the step S103, the arithmetic unit 130 analyzes
a gray scale value of the first eye image to obtain the
distributions of the glint G1. The 8-bit color image, namely
256-grayscale image is used as an example. The grayscale value is
quantified as 256 colors from the pure black, through gray to
white, and the grayscale value ranges from 0 to 255. It is worth to
notice that the gray scale value of the glint G1 is near to or
equal to 255, whereas the gray scale value of the pupil P1 is near
to 0. The arithmetic unit 130 can obtain the arrangement, shape and
range of the pixels which is close to the maximum gray scale value
in all pixels through the gray scale value distribution of the
first eye image. Further, the arithmetic unit 130 speculates the
arrangement of the pixels corresponding to the arrangement of the
glint G1 in the first image.
[0045] Implementing the step S104, the arithmetic unit 130
determines the position of the pupil P1 according to the position
of the glint G1. Specifically, the arithmetic unit 130 selects an
appropriate threshold gray scale value first. The gray scale value
of the pupil P1 is less than the said threshold gray scale value,
whereas the gray scale value of the glint G1 in the first eye image
is greater than the said threshold gray scale value.
[0046] After confirming the position of the glint G1, the
arithmetic unit 130 scans the survey area M1 near the arrangement
of the glint G1 (shown in FIG. 1B), and analyzes the gray scale
value distribution of the survey area M1. The arithmetic unit 130
determines the part of the survey area M1, whose gray scale value
is less than the threshold gray scale value. The survey area M1 can
be defined by at least one glint G1. The positions of the glint G1
and the pupil P1 are in the survey area M1. It is worth to mention
that the position of the glint G1 can be at the boundary of the
survey area M1 or in the survey area M1. User can set the range of
the survey area M1 according to the pupil P1 size through the
arithmetic unit 130. The present disclosure does not limit the
range of the survey area M1.
[0047] The arithmetic unit 130 determines an area from the survey
area M1 to be a specific area, and the gray scale value of the
specific area is less than the threshold gray scale value. Further,
the arithmetic unit 130 determines whether the shape of the
specific area matches with the shape of the pupil P1 to reduce the
possibility of the misjudgment of the pupil P1 position. For
instance, the arithmetic unit 130 selects two specific areas
satisfied by the condition that the gray scale value of the
specific areas are less than the threshold gray scale value. When
one specific area is rectangle, and the other specific area is
circular, the arithmetic unit 130 then determines one of the
circular specific areas, which is circular, matches with the shape
of the pupil P1. Besides, in order to reduce the possibility of the
misjudgment of the pupil P1 position more, user can set the range
of the pupil P1 area in the first image. The arithmetic unit 130
determines whether the proportion of the specific area is within
the range of the pupil P1 to reduce the possibility of the
misjudgment of the pupil P1 position more.
[0048] It is worth to mention that the arithmetic unit 130 can
analyze the gray scale value distribution of the survey area M1
near the glint G1 in the first eye image so as to reduce searching
scope of the pupil P1. Hence, the position of the pupil P1 can be
found quickly. Therefore, compared with conventional technology,
the arithmetic unit 130 does not analyze the gray scale value
distribution of whole first eye image for searching for the pupil P
1.
[0049] FIG. 2A is a side view of the eye detecting device in
accordance with the second embodiment of the present invention.
FIG. 2B is a function block diagram of the eye detecting device in
accordance with the second embodiment of the present invention.
Please refer to FIG. 2A and 2B. The structure of an eye detecting
device 200 in accordance with second exemplary embodiment is
similar to the eye detecting device 100 in accordance with first
exemplary embodiment. For example, the eye detecting device 100 and
200 include the image sensor 120. However, there are some
differences between the eye detecting devices 100 and 200. The
following detailed description explains the difference between the
eye detecting devices 100 and 200, and the same features are
basically not described again.
[0050] The eye detecting device 200 in accordance with the second
embodiment includes an optical assembly 210, an image sensor 120,
and an arithmetic unit 130. The optical assembly 210 provides a
plurality of incident lights L1 to form a plurality of glints G1
located near a pupil P1 of the eye E1. The image sensor 120 is used
to capture an eye image, and the eye image includes these glints G1
image. The arithmetic unit 130 analyzes a gray scale value of the
eye E1 image and obtains distribution of the glints G1 according to
the gray scale value. Hence, the arithmetic unit 130 determines the
position of the pupil P1 of eye E1 according to the distribution of
the glints G1.
[0051] The optical assembly 210 can emit a plurality of incident
lights L1 enter into the eye E1. The incident lights L1 fall on the
eye E1 to form a plurality of glints by reflecting at an iris I1 of
the eye E1, and at least part of glints are located near a pupil P1
of the eye E1.
[0052] In this embodiment, the optical assembly 210 includes only
one or less light source 212 and a dispersing component 214. The
incident lights can be formed by dividing at least one light
through the optical assembly 210. In other embodiment, the optical
assembly 210 may include a plurality of light sources 212 and
exclude any dispersing component 214. The present disclosure does
not limit the number of the light source 212 and the structure of
dispersing component 214.
[0053] FIG. 2C depicts a flow diagram of a method of detecting
pupil in accordance with the second exemplary embodiment of the
present disclosure. Please refer to FIG. 2A, FIG. 2B and FIG.
2C.
[0054] Implementing the step S201, when the user uses the eye
detecting device 200, the optical assembly 210 provides a plurality
of incident lights L1 enter the eye E1. The incident lights L1
reflect to form a plurality of glints G1 near the pupil P1, such as
the iris I1.
[0055] It is worth to notice that the positions of the glints G1
can be changed with the emission positions of the incident lights
L1. For instance, there are four emission positions of the incident
lights L1 approximately arranged in a rectangle, and the aspect
ratio of the rectangle is 2:1. Then, four glints G1 are formed and
arranged in a rectangle with the aspect ratio of 2:1.
[0056] Implementing the step S202, the image sensor 120 captures a
first eye image by photographing the eye E1. The first eye image
photographed by image sensor 120 shows the image of the eye E1
region and the image of the said glints G1. Then, the image sensor
120 transmits the data of the first eye image to the arithmetic
unit 130.
[0057] Implementing the step S203, the arithmetic unit 130 can
obtain the arrangement, shape and range of the pixels which each
have close to the maximum gray scale value through the gray scale
value. Further, the arithmetic unit 130 speculates the arrangement
of the pixels corresponding to the arrangement of the glints G1 in
the first image.
[0058] Implementing the step S204, the arithmetic unit 130
determines the position of the pupil P1 according to the
distributions of the glints G1. Specifically, the arithmetic unit
130 selects an appropriate threshold gray scale value first. The
gray scale value of the glints G1 in the first eye image are
greater than the said threshold gray scale value. After confirming
the arrangement of the glints G1, the arithmetic unit 130 scans the
survey area M1 near the arrangement of the glints G1 (shown in FIG.
2A), and analyzes the gray scale value distribution of the survey
area M1.
[0059] It worth to mention that the survey area M1 can be defined
by those glints G1. The survey area M1 contains the arrangement of
the glints G1 and the pupil P1, and can be equal to or slightly
larger than the area surrounded by the glints G 1.
[0060] In the same way, in order to reduce the possibility of the
misjudgment of the pupil P1 position, after the specific area which
has gray scale value less than the threshold gray scale value is
determined by the arithmetic unit 130, the arithmetic unit 130
determines whether the shape of the specific area matches the shape
of the pupil P1, and the proportion of the specific area is within
the range of the pupil P 1.
[0061] It is worth to mention that the arithmetic unit 130 can
define the shape or range of the survey area M1 through
distribution of the glints G1 so as to reduce the seeking range.
Hence, the position of the pupil P1 can be searched quickly.
[0062] FIG. 3A is a function block diagram of the eye detecting
device in accordance with the third embodiment of the present
invention. The structure of an eye detecting device 300 in
accordance with third exemplary embodiment is similar to the eye
detecting device 200 in accordance with second exemplary
embodiment. For example, the eye detecting devices 300 and 200 each
include the optical assembly 210 and the image sensor 120. However,
there are some differences between the eye detecting devices 100
and 200. The following detailed description explains the difference
between the eye detecting device 100 and 200, and the same features
are basically not described again.
[0063] The eye detecting device 300 in accordance with the third
embodiment includes an optical assembly 210, an image sensor 120,
an arithmetic unit 230, and the control unit 340. The optical
assembly 210 provides a plurality of incident lights L1 to form a
plurality of glints G1 located near a pupil P1 of the eye E1. The
control unit 340 controls the timing that the incident lights are
emitted into the eyes, namely, the control unit 340 can control
that the optical assembly 210 providing different incident lights
L1 into the eyes E1 at different timing separately. The image
sensor 120 captures the eye images at different timing, and the eye
images include the glint G1a and glint G1b image. Namely, The glint
G1a and glint G1b image appear in each eye images captured at
different timing. The arithmetic unit 230 analyzes a gray scale
value of the eye images captured at different timing and obtains
positions of glint G1a and glint G1b according to the gray scale
value. Hence, the arithmetic unit 230 determines the position of
the pupil P1 of eye E1 according to the position of glint G1a and
glint G1b.
[0064] Specifically, the image sensor 120 is used to capture the
eye images at different timing, and each the eye image shows these
glint G1a and glint G1b image. The arithmetic unit 230 analyzes
gray scale value of the eye image captured at different timing, in
addition, the arithmetic unit 230 commands the control unit 340 so
that the control unit 340 controls the timing that the optical
assembly 210 provides the incident lights L1.
[0065] FIG. 3B depicts a flow diagram of a method of detecting
pupil in accordance with the second exemplary embodiment of the
present disclosure. Please refer to FIG. 3A and FIG. 3B.
[0066] Implementing the step S301, the control unit 340 controls
the optical assembly 210 provides a plurality of incident lights L1
at a first timing. The incident lights L1 enter into the iris I1
area near the pupil P1 and than reflect to form a plurality of
first glints G1a. The arrangement of the first glints G1a is
corresponding to the emission arrangement of the incident lights
L1. It is worth to notice that the optical assembly 210 includes a
plurality of light sources without including any dispersing
component.
[0067] Implementing the step S302, the image sensor 120 captures a
first eye image by photographing the eye E1 at the first timing.
The first eye image photographed by image sensor 120 at the first
timing and shows the image of the eye E1 region and the image of
the said first glints G1a. Then, the image sensor 120 transmits the
data of the first eye image to the arithmetic unit 230.
[0068] Implementing the step S303, the control unit 340 controls
the optical assembly 210 provides a plurality of incident lights L1
at a second timing. The incident lights L1 enters the iris I1 area
near the pupil P1 and reflect to form a plurality of second glints
G1b. The arrangement of the second glints G1b is corresponding to
the emission arrangement of the incident lights L. It is worth to
notice that the first timing is not equal to the second timing, and
the arrangement of the first glints G1a formed at the first timing
is not equal to the arrangement of the second glints G1b formed at
the second timing. Specifically, part of the light source 112
provides some incident lights L1 at the first timing, the other
light source 112 provides some incident lights L1 at the second
timing.
[0069] For example, the amount of the light sources 112 is four,
and the light sources 112 are arranged approximately in the
rectangular array. The aspect ratio of said rectangular array is
2:1. The control unit 340 controls the optical assembly 210 to
provide two light sources 112 arranged in diagonally opposite
corners of the rectangular array at the first timing, and then the
control unit 340 controls the optical assembly 210 to provide the
other light sources 112 arranged in diagonally opposite corners of
the rectangular array at the second timing. The present disclosure
does not limit the number and arrangement of the light sources 112
provided by the optical assembly 210 at different timing. The
present disclosure does not limit the emission sequence of the
light sources 112.
[0070] Implementing the step S304, the image sensor 120 captures a
second eye image by photographing the eye E1 at the second timing.
The second eye image photographed by image sensor 120 at the second
timing and shows the image of the eye E1 region and the image of
the said first glints G1b. Then, the image sensor 120 transmits the
data of the second eye image to the arithmetic unit 230.
[0071] It is worth to notice that the aforementioned first timing
is namely the timing that the user started using the eye detecting
device 300, and the second timing is the another timing different
from the first timing. The first eye image photographed by image
sensor 120 at the first timing, and the second eye image
photographed by image sensor 120 at the second timing.
[0072] Implementing the step S305, the arithmetic unit 230 analyzes
a gray scale value distribution of the first eye image and the
second eye image to obtain the distributions of the first glints
G1a and the second glints G1b. Specifically, the arithmetic unit
230 can obtain the arrangement, shape and range of the pixels which
is close to the maximum gray scale value in all pixels through the
gray scale value distribution of the first eye image and the second
eye image. Further, the arithmetic unit 130 speculates the
arrangement of the pixels corresponding to the arrangement of the
first glint G1a in the first image and the second glint G1b in the
second image.
[0073] Implementing the step S306, a difference image between the
first image and the second image is produced by image subtraction.
In this embodiment, the amount of the light sources 112 is four,
and the first glints G1a in the first image are provided by two
light sources 112 arranged in diagonally opposite corners of the
rectangular array, and the second glints G1b in the second image
are provided by the other light sources 112 arranged in other
diagonally opposite corners of the rectangular array. The
difference image is generated by subtracting the second image from
the first image, and the difference gray scale value of the
difference image range from -255 to 255.
[0074] Since the arrangement of the first glints G1a and the
arrangement of the second glints G1b do not overlap, the gray scale
values corresponding to the arrangement of the first glint G1a and
the second glint G1b in the difference image between the first
image and the second image are proximate to critical value. For
example, in the difference image, the gray scale value
corresponding to the arrangement of the first glint G1a is
proximate to a maximum value, whereas the gray scale value
corresponding to the arrangement of the second glints G1b is
proximate to a minimum value (negative gray scale value).
[0075] Thus, in the difference image, the gray scale value
corresponding to the arrangement of the first glint G1a and the
second glint G1b show a special pattern. In this embodiment, the
special pattern is defined by two brightest spots and two darkest
spots. However, in the difference image, the difference image can
be generated by subtracting the first image from the second image.
Hence, the gray scale value corresponding to the arrangement of the
second glint G1b is proximate to a maximum value, whereas the gray
scale value corresponding to the arrangement of the first glints
G1a is proximate to a minimum value and is not limited to the
examples provided herein.
[0076] In addition, the arrangement of the first glints G1a and the
second glint G1b can be further determined through the arithmetic
unit 230. Specifically, in the process of determining the
arrangement of the first glints G1a and the second glint G1b
through the arithmetic unit 230, the arithmetic unit 230 analyzes
the arrangement, shape and range of the pixels which are close to
the maximum (255) and minimum (-255) gray scale value in all pixels
to speculate a possibility arrangement of the first glints G1a and
the second glint G1b. Then, the arithmetic unit 230 speculates
whether the possibility arrangement of the first glints G1a and the
second glint G1b corresponding to the above-mentioned special
pattern.
[0077] Since the control unit 340 controls the different incident
lights L1 to emit into the different positions of the eye E1 at the
different timing, the arrangement of the first glints G1a and the
second glint G1b at the different timing can be arranged. The
arrangement of the first glints G1a and the second glint G1b can be
more confirmed through the gray scale value and the above-mentioned
special pattern after image subtraction. Hence, the possibility of
the misjudgments of the glints G1 position can be more reduced.
[0078] Implementing the step S307, the arithmetic unit 230
determines the position of the pupil P1 according to the
arrangement of the first glints G1a and the second glints G1b.
Specifically, the arithmetic unit 230 selects an appropriate
threshold gray scale value first. The gray scale value of the pupil
P1 is less than the said threshold gray scale value, whereas the
gray scale values of the first glints G1a and the second glint G1b
in the difference image are greater than the said threshold gray
scale value. The arithmetic unit 230 confirms the arrangement of
the first glints G1a and the second glint G1b through the gray
scale value. After confirming the arrangement of the first glints
G1a and the second glint G1b, the arithmetic unit 230 scans the
survey area M1 near the arrangement of the first glints G1a or the
second glint G1b, and analyzes the gray scale value distribution of
the survey area M1. The arithmetic unit 230 determines the part of
the survey area M1, whose gray scale value is less than the
threshold gray scale value.
[0079] For example, the gray scale values corresponding to the
arrangement of the first glint G1a and the second glint G1b in the
difference image are proximate to critical value, and the gray
scale value corresponding to the arrangement of the first glint G1a
is proximate to a maximum value, whereas the gray scale value
corresponding to the arrangement of the second glints G1b is
proximate to a minimum value. Thus, the arithmetic unit 230
confirms the arrangement of the first glints G1a through the gray
scale value., and then scans the survey area M1 near the
arrangement of the first glint G1a to determines the position of
the pupil P1.
[0080] In particular, the survey area M1 can be defined by these
first glints G1a and/or second glints G1b. The range of the survey
area M1 contains the arrangement of the first glints G1a and/or
second glints G1b, and can be equal to or slightly larger than the
area surrounded by the arrangement of the first glints G1a and/or
second glints G1b. Specifically, the position of the first glints
G1a and/or second glints G1b can be at the boundary of the survey
area M1 or in the survey area M1. User can set the range of the
survey area M1 according to the pupil P1 size through the
arithmetic unit 230. The present disclosure is not limited to the
range of the survey area M1.
[0081] The arithmetic unit 230 selects the specific area from the
survey area, and the gray scale value of the specific area is less
than the threshold gray scale value, and then determines whether
the shape and proportion of the specific area matches the pupil P1
in the difference image to reduce the possibility of the
misjudgment of the pupil P1 position.
[0082] The arithmetic unit 230 can analyze the gray scale value
distribution of the survey area M1 near the arrangement of the
first glints G1a and/or second glints G1b in the difference image
so as to search the position of the pupil P1 quickly. Therefore,
compared with conventional technology, the arithmetic unit 230 does
not analyze the gray scale value distribution of whole first or
second eye image for searching scope of the pupil P1.
[0083] FIG. 4 depicts a flow diagram of a method of identifying
iris in accordance with the third exemplary embodiment of the
present disclosure. The method of identifying iris in accordance
with the third exemplary embodiment can be implemented through eye
detecting device 200 (shown in FIG. 2A). Please refer to FIG. 2A
and FIG. 4.
[0084] Implementing the step S401, when the eye E1 is located at a
reference position, and the reference position is corresponding to
a position where the eye gazes straight ahead in this embodiment.
the optical assembly 210 provides a plurality of incident lights L1
entering the eye E1. The incident lights L1 reflect to form a
plurality of glints G1 near the pupil P1, and the arrangement of
those glints G1 is defined to a first reference point, a second
reference point and a third reference point.
[0085] Specifically, the incident lights L1 can be provided by the
light source 212 and the dispersing component 214 so that the
emission positions of the incident lights L1 are the illuminated
position of the dispersing component 214. Or, the incident lights
L1 can be provided by at least three light sources 212 without any
dispersing component 214 so that the emission positions of the
incident lights L1 are the position where the light sources 212 are
placed. The position that the incident lights L1 enter in the iris
I1 area near the pupil P1 can be adjusted by adjusting the
arrangement of the light source 212 or the disposition of the light
sources 212 and the dispersing component 214.
[0086] The first reference point, the second reference point and
the third reference point are located near a pupil P1 of the eye E1
as a mark for locating the eye E1 at the reference position. The
positions of the first reference point, the second reference point
and the third reference point are corresponding to the emission
position of the incident lights L1. In this embodiment, when the
user looks straight ahead, namely, the eye gazes straight ahead,
the user presets those glints positions corresponding to the
emission arrangement of the incident lights L1 to be regarded as
the positions of the first reference point, the second reference
point and the third reference point. Specifically, a first
reference axis is formed between the first reference point and the
second reference point. A second reference axis is formed between
the second reference point and the third reference point. A
reference angle is formed between the first reference axis and the
second reference axis. Besides, in order to mark the reference
position clearly, the method of identifying iris can further
include presets the fourth reference point or more other reference
point, but not limited to the examples provided herein.
[0087] In this embodiment, three emission positions of the incident
lights L1 are provided and are arranged approximately as the right
angled triangle. The ratio of two sides of said right angled
triangle is 2:1. Hence, those glints are located near a pupil P1 of
the eye E1 and arranged approximately as the right angled triangle.
Namely, the ratio between the first reference axis and the second
reference axis is 2:1, and the reference angle is approximate to 90
degrees.
[0088] Implementing the step S402, when the eye E1 moves from the
reference position to a measuring position, the incident lights L1
form a first measuring glint, a second measuring glint, and a third
measuring glint near a pupil P1 of the eye E1. A first axis is
formed between the first measuring glint and the second measuring
glint. A second axis is formed between the second measuring glint
and the third measuring glint. An angle is formed between the first
axis and the second axis.
[0089] Specifically, the eye E1 is substantially spherical, and the
iris I1 is the portion rising slightly above the surface of the
sphere. The arrangement of those glints G1 is changed while the eye
E1 moves corresponding to the reference position, whereas the
glints G1 are the first reference point, the second reference point
and the third reference point. Namely, when the eye E1 gaze
direction moves from the front direction to lateral direction, the
arrangement of those glints G1 is changed from the first reference
point, the second reference point and the third reference point to
the first measuring glint, the second measuring glint, and the
third measuring glint.
[0090] Implementing the step S403, the image sensor 120 captures an
eye image by photographing the eye E1. The eye image photographed
by image sensor 120 shows the image of the eye E1 region and the
image of the said first measuring glint, the second measuring
glint, and the third measuring glint. Then, the image sensor 120
transmits the data of the eye image to the arithmetic unit 130 or
230.
[0091] Implementing the step S404, the arithmetic unit 130 or 230
analyzes a gray scale value distribution of the eye image to obtain
the arrangement of the first measuring glint, the second measuring,
and the third measuring glint. Specifically, the arithmetic unit
130 or 230 can obtain the arrangement, shape and range of the
pixels which is close to the maximum gray scale value (255) in all
pixels through the gray scale value distribution of the eye image.
Further, the arithmetic unit 130 or 230 speculates the arrangement
of the pixels corresponding to the arrangement of the first
measuring glint, the second measuring, and the third measuring
glint in the image.
[0092] Implementing the step S405, the displacement amounts of the
first measuring glint, the second measuring glint, and the third
measuring glint relative to the first reference point, the second
reference point and the third reference point are calculated
respectively. Therefore, a deformation amount caused by the iris
image of the eye at the measuring position relative to the iris
image of the eye at the reference position is obtained.
Specifically, the arithmetic unit 130 or 230 calculates the first
variation, which is a length and angular variation of the first
axis relatives to the first reference axis. The arithmetic unit 130
or 230 calculates the second variation, which is a length and
angular variation of the second axis relatives to the second
reference axis. Equally, the third variation which is an angular
variation of the angle relatives to the reference angle is
calculated. Hence, the arithmetic unit 130 or 230 calculates the
iris image deformation amount according to the first variation, the
second variation, and the third variation. Furthermore, the
proportion of the iris image deformation amount can be estimated
according to the relative proportion of the first axis relatives to
the first reference axis and the relative proportion of the second
axis relatives to the second reference axis. Besides, the distance
between the image sensor 120 and eye E1 can be estimated according
to the length of the first axis and the second axis. Hence, the
size of the pupil P1 can be estimated so that the position of the
pupil P1 can be searched quickly.
[0093] It is worth to notice that when the user looks straight
ahead, the shape of the pupil P1 image photographed by image sensor
120 is similar to a circle. While the measuring position is equal
to the reference position, namely, the user keeps looking straight
ahead, the shape of the pupil P1 image photographed by image sensor
120 keeps being similar to circle. While the measuring position is
not equal to the reference position, namely, the eye E1 gaze
direction moves from the front direction to lateral direction, the
shape of the pupil P1 image photographed by image sensor 120 is
similar to an ellipse.
[0094] The arithmetic unit 130 or 230 can calculate the major axis
and minor axis of the said ellipse according to the first
variation, the second variation, and the third variation. Hence,
the boundary of the pupil P1 can be estimated so that the boundary
of the pupil P1 can be searched quickly.
[0095] FIG. 5 depicts a flow diagram of a method of identifying
iris in accordance with the fourth exemplary embodiment of the
present disclosure. Please refer to FIG. 5 and FIG. 2A. The method
of identifying iris shown in FIG. 5 is similar to the method of
identifying iris shown in FIG. 4. The differences between these
methods of identifying iris s are further discloses as follows.
[0096] Implementing the step S501, in this embodiment, the optical
assembly 210 provides a plurality of incident lights L1 entering
the eye E1. The incident lights L1 reflect to form a plurality of
glints G1 near the pupil P1. Specifically, the incident lights L1
can be provided by the light source 212 and the dispersing
component 214 so that the emission positions of the incident lights
L1 are the illuminated position of the dispersing component 214.
Or, the incident lights L1 can be provided by at least three light
sources 212 without any dispersing component 214 so that the
emission positions of the incident lights L1 are the position where
the light sources 212 are placed. The position that the incident
lights L1 enter in the iris I1 area near the pupil P1 can be
adjusted by adjusting the arrangement of the light source 212 or
the disposition of the light sources 212 and the dispersing
component 214.
[0097] Implementing the step S502, the user sets a first reference
point, a second reference point and a third reference point as a
mark for locating the eye E1 at the reference position. The
positions of the first reference point, the second reference point
and the third reference point are corresponding to the emission
position of the incident lights L1. In this embodiment, when the
user looks straight ahead and there is a reference distance between
the optical assembly 210 and the eye E1, the user presets those
glints positions corresponding to the emission arrangement of the
incident lights L1 to be regarded as the positions of the first
reference point, the second reference point and the third reference
point. However, the reference position cannot be the position right
front the eye E1 when the eye E1 gaze direction deviates from the
front direction of the eye E1, and is not limited to the examples
provided herein.
[0098] Specifically, a first reference axis is formed between the
first reference point and the second reference point. A second
reference axis is formed between the second reference point and the
third reference point. A reference angle is formed between the
first reference axis and the second reference axis. In addition, in
order to mark the reference position clearly, the method of
identifying iris can further include presets the fourth reference
point or more other reference point, but not limited to the
examples provided herein.
[0099] In this embodiment, three emission positions of the incident
lights L1 are provided and are arranged approximately as the right
angled triangle. The ratio of two sides of said right angled
triangle is 2:1. Hence, those glints are located near a pupil P1 of
the eye E1 and arranged approximately as the right angled triangle.
Namely, the ratio between the first reference axis and the second
reference axis is 2:1, and the reference angle is approximate to 90
degrees. It is worth to note that the user presets those glints
positions corresponding to the emission arrangement of the incident
lights L1 to be regarded as the positions of the first reference
point, the second reference point and the third reference point
while there exists the reference distance between the optical
assembly 210 and the eye E1.
[0100] Implementing the step S503, when the eye E1 is located at a
measuring position, there is a measuring distance between the
optical assembly 210 and the eye E1. The incident lights L1 form a
first measuring glint, a second measuring glint, and a third
measuring glint near a pupil P1 of the eye E1. The positions of the
first measuring glint, the second measuring glint, and the third
measuring glint are corresponding to the positions of the first
reference point, the second reference point and the third reference
point. A first axis is formed between the first measuring glint and
the second measuring glint. A second axis is formed between the
second measuring glint and the third measuring glint. An angle is
formed between the first axis and the second axis.
[0101] Specifically, since different user has different face and
nose height, there exists different distance between the optical
assembly 210 and the eye E1 while different user wear eye detecting
device 200 or 300. Hence, the glint G1 position formed by emitting
the incident light L1 into the eye E1 can be changed to the first,
second and the third measuring glint. That is, when the eye gaze
direction remains without moving, the glint G1 position formed by
emitting the incident light L1 into the eye E1 can be changed
proportionally from the aforementioned the first, second and third
reference point to the first, second and the third measuring glint.
The angle is equal to the reference angle.
[0102] In this embodiment, since the ratio between the first and
second reference axis is 2:1, the ratio between the first and
second axis is 2:1. Since the reference angle is approximate to 90
degrees, the angle is approximate to 90 degrees.
[0103] Implementing the step S504, the image sensor 120 captures an
eye image by photographing the eye E1. The eye image photographed
by image sensor 120 shows the image of the eye E1 region and the
image of the said first, second, and the third measuring glint, and
an iris I1 image. Then, the image sensor 120 transmits the data of
the eye image to the arithmetic unit 130 or 230.
[0104] Implementing the step S505, the arithmetic unit 130 or 230
analyzes a gray scale value distribution of the eye image to obtain
the arrangement of the first, second, and third measuring glint.
Specifically, the arithmetic unit 130 or 230 can obtain the
arrangement, shape and range of the pixels which each have close to
the maximum gray scale value (255) through the gray scale value
distribution of the eye image. Further, the arithmetic unit 130 or
230 speculates the arrangement of the pixels corresponding to the
arrangement of the first, second, and third measuring glint in the
image.
[0105] Implementing the step S506, the variation of the distance
between the first measuring glint and the second measuring glint
with respect to the distance between the first reference point and
the second reference point is calculated. The variation of the
distance between the second measuring glint and the third measuring
glint with respect to the distance between the second reference
point and the third reference point is calculated. Accordingly, the
resolution variation of an iris image when the eye is located at
the measuring position is obtained. Specifically, the arithmetic
unit 130 or 230 calculates the first variation, which is a length
variation of the first axis relatives to the first reference axis.
The arithmetic unit 130 or 230 calculates the second variation,
which is a length variation of the second axis relatives to the
second reference axis. Hence, the arithmetic unit 130 or 230
calculates the resolution variation of an iris image according to
the first and second variation.
[0106] For example, the first reference axis has 20 pixels, whereas
the second reference axis has 10 pixels. The ratio between pixel of
the first and second reference axis is 2:1. The arithmetic unit 130
or 230 calculates that the first axis has 10 pixels and the second
axis has 5 pixels. Hence, the arithmetic unit 130 or 230 calculates
that the first axis is decrease by two times compare to the first
reference axis and the second axis is decrease by two times compare
to the second reference axis. Namely, the first and second
variation is two. Hence, the boundary of the pupil P1 can be
estimated so that the boundary of the pupil P1 can be searched
quickly.
[0107] In summary, the present disclosure provides eye detecting
device, methods of detecting pupil and identifying iris. The eye
detecting device includes an optical assembly, an image sensor, and
an arithmetic unit. The arithmetic unit can analyze the gray scale
value distribution of the survey area near the arrangement of the
glint in the first eye image so as to reduce searching scope of the
pupil. Hence, the position of the pupil can be searched quickly.
Therefore, compared with conventional technology, the arithmetic
unit does not analyze the gray scale value distribution of whole
first eye image for searching scope of the pupil.
[0108] The present disclosure provides eye detecting device,
methods of detecting pupil. The eye detecting device includes an
optical assembly, an image sensor, an arithmetic unit, and the
control unit. Since the control unit controls the different
incident lights to emit into the different positions of the eye at
the different timing, the arrangement of the first and second glint
at the different timing can be arranged. The arrangement of the
first and second glint can be more confirmed through the gray scale
value and the special pattern after image subtraction. Hence, the
possibility of the misjudgments of the glints position can be more
reduced. The arithmetic unit can analyze the gray scale value
distribution of the survey area near the arrangement of the first
glints and/or second glints in the difference image so as to search
the position of the pupil quickly. Therefore, compared with
conventional technology, the arithmetic unit does not analyze the
gray scale value distribution of whole first or second eye image
for searching scope of the pupil.
[0109] The arithmetic unit can calculate the major axis and minor
axis of the said ellipse according to the first variation, the
second variation, and the third variation. Hence, the boundary of
the pupil P1 can be estimated so that the boundary of the pupil P1
can be searched quickly.
[0110] The present disclosure provides methods of identifying iris.
The arithmetic unit can calculate the major axis and minor axis of
the said ellipse according to the first, second, and third
variation. Hence, the boundary of the pupil can be estimated so
that the boundary of the pupil can be searched quickly.
[0111] The present disclosure provides methods of identifying iris.
The arithmetic unit can calculate the first, second, and third
variation. Hence, the boundary of the pupil can be estimated so
that the boundary of the pupil can be searched quickly.
[0112] The above-mentioned descriptions represent merely the
exemplary embodiment of the present disclosure, without any
intention to limit the scope of the present disclosure thereto.
Various equivalent changes, alternations or modifications based on
the claims of present disclosure are all consequently viewed as
being embraced by the scope of the present disclosure.
* * * * *