Technical Field
[0001] The present disclosure relates to an information processing system, an eye state
measurement system, an information processing method, and a non-transitory computer
readable medium, more particularly to an information processing system, an eye state
measurement system, an information processing method, and a non-transitory computer
readable medium that evaluate a change in state of an eye of a subject.
Background Art
[0002] There has been proposed a technique of evaluating a change in state of an eye of
a subject, based on a captured image of the eye of the subject. For example, Patent
Literature 1 discloses a method of calculating a pupil motion from a change in a pupil
position detected based on a captured image of a face of a driver. However, in the
above-mentioned method described in Patent Literature 1, when a head moves and an
eyeball is not within a visual field range of a camera, a pupil position cannot be
detected. Therefore, the subject is required to fix the head in such a way that the
eyeball stays within the visual field range of the camera.
[0003] Further, Patent Literature 2 discloses a method of detecting a movement amount of
an eyeball, based on a difference between a reference position indicating a position
of a blood vessel figure in a white region of an eye in a reference image and a position
of the blood vessel figure in the white region of the eye in an image for detection.
In this method, an eyeball image in the image for detection is corrected to be at
a position in the reference image through Affine conversion from three positions including
positions of an inner eye corner and an outer eye corner, and hence image blurring
can be prevented. However, even in this method, a subject is also required to fix
the head to a certain extent in such a way that the eyeball stays within a visual
field range of a camera.
[0004] Meanwhile, there has been known a method of capturing an image of a designated region
in an enlarged manner through use of cameras with different angles of view. Patent
Literature 3 discloses an information processing system in which a second narrow-angle
camera captures an image of a target region designated in an image captured by a first
wide-angle camera. However, the second camera is limited for capturing an image of
a human face, and Patent Literature 3 described above does not disclose capturing
an image of an eye region being a micro region. Further, Patent Literature 3 described
above does not disclose a technique of measuring and evaluating a microscopic change
in state of an eye.
Citation List
Patent Literature
Summary of Invention
Technical Problem
[0006] The present disclosure has been made in order to solve the problems described above,
and an object thereof is to provide an information processing system, an eye state
measurement system, an information processing method, and a non-transitory computer
readable medium that suitably evaluate a change in state of an eye of a subject while
a head is in a relaxed posture without being fixed.
Solution to Problem
[0007] An information processing system according to one aspect of the present disclosure
includes a first acquisition means, a second acquisition means, a movement control
means, and a state evaluation means. The first acquisition means acquires, from a
first imaging means, image data relating to a first image of a head of a subject,
the first image being captured at a first angle of view. The second acquisition means
acquires, from a second imaging means, image data relating to a second image of an
eye region of the subject, the second image being captured at a second angle of view
narrower than the first angle of view. The movement control means moves a visual field
range of the second imaging means, based on position information relating to the head
of the subject, the position information being acquired based on the first image.
The state evaluation means evaluates a change in state of an eye of the subject, based
on chronological data relating to the second image.
[0008] An eye state measurement system according to one aspect of the present disclosure
includes a first imaging means, a second imaging means, and an information processing
device. The first imaging means captures an image of a head of a subject at a first
angle of view. The second imaging means captures an image of an eye region of the
subject at a second angle of view narrower than the first angle of view. The information
processing device includes a first acquisition means, a second acquisition means,
a movement control means, and a state evaluation means. The first acquisition means
acquires, from a first imaging means, image data relating to a first image of a head
of a subject, the first image being captured at a first angle of view. The second
acquisition means acquires, from a second imaging means, image data relating to a
second image of an eye region of the subject, the second image being captured at a
second angle of view narrower than the first angle of view. The movement control means
moves a visual field range of the second imaging means, based on position information
relating to the head of the subject, the position information being acquired based
on the first image. The state evaluation means evaluates a change in state of an eye
of the subject, based on chronological data relating to the second image.
[0009] An information processing method according to one aspect of the present disclosure
includes: acquiring, from a first imaging means, image data relating to a first image
of a head of a subject, the first image being captured at a first angle of view; acquiring,
from a second imaging means, image data relating to a second image of an eye region
of the subject, the second image being captured at a second angle of view narrower
than the first angle of view; moving a visual field range of the second imaging means,
based on position information relating to the head of the subject, the position information
being acquired based on the first image; and evaluating a change in state of an eye
of the subject, based on chronological data relating to the second image.
[0010] A non-transitory computer readable medium according to one aspect of the present
disclosure stores a program for causing a computer to execute first acquisition processing,
second acquisition processing, movement control processing, and state evaluation processing.
The first acquisition processing is processing of acquiring, from a first imaging
means, image data relating to a first image of a head of a subject, the first image
being captured at a first angle of view. The second acquisition processing is processing
of acquiring, from a second imaging means, image data relating to a second image of
an eye region of the subject, the second image being captured at a second angle of
view narrower than the first angle of view. The movement control processing is processing
of moving a visual field range of the second imaging means, based on position information
relating to the head of the subject, the position information being acquired based
on the first image. The state evaluation processing is processing of evaluating a
change in state of an eye of the subject, based on chronological data relating to
the second image.
Advantageous Effects of Invention
[0011] According to the present disclosure, it is possible to provide an information processing
system, an eye state measurement system, an information processing method, and a non-transitory
computer readable medium that suitably evaluate a change in state of an eye of a subject
while a head is in a relaxed posture without being fixed.
Brief Description of Drawings
[0012]
Fig. 1 is a block diagram illustrating a functional configuration of an information
processing system according to a first example embodiment;
Fig. 2 is a block diagram illustrating a hardware configuration of the information
processing system according to the first example embodiment;
Fig. 3 is a system configuration diagram of an eye state measurement system to which
an information processing system according to a second example embodiment is applicable;
Fig. 4 is a block diagram illustrating a functional configuration of an information
processing device according to the second example embodiment;
Fig. 5 is a diagram for describing movement control processing executed by a movement
control unit according to the second example embodiment;
Fig. 6 is a diagram for describing focal position control processing executed by the
movement control unit according to the second example embodiment;
Fig. 7 is a diagram for describing state evaluation processing executed by a state
evaluation unit according to the second example embodiment;
Fig. 8 is a flowchart illustrating a procedure of information processing of the information
processing device according to the second example embodiment;
Fig. 9 is a system configuration diagram of an eye state measurement system according
to a third example embodiment;
Fig. 10 is a system configuration diagram of an eye state measurement system according
to a fourth example embodiment; and
Fig. 11 is a block diagram illustrating a functional configuration of an information
processing device according to the fourth example embodiment.
Example Embodiment
[0013] The present disclosure is described below based on example embodiments, but the invention
within the scope of the claims is not limited to the embodiments given below. Further,
not all the configurations described in the example embodiments are necessary as means
for solving the problems. For clarification of the description, the description and
the drawings given below are omitted and simplified as appropriate. Note that, in
each of the drawings, the same elements are denoted with the same reference symbols.
<First Example Embodiment>
[0014] First, with reference to Fig. 1 and Fig. 2, a first example embodiment of the present
disclosure is described. Fig. 1 is a block diagram illustrating a functional configuration
of an information processing system 10 according to the first example embodiment.
The information processing system 10 is a computer device that evaluates a change
in state of an eye of a subject. The information processing system 10 includes a first
acquisition unit 11, a second acquisition unit 12, a movement control unit 13, and
a state evaluation unit 15.
[0015] The first acquisition unit 11 is also referred to as a first acquisition means. The
first acquisition unit 11 is connected to a first imaging unit 30 in a communicable
manner, and acquires, from the first imaging unit 30, image data (first image data)
relating to a first image of a head of subject that is captured at a first angle of
view. Herein, the first imaging unit 30 is a camera that captures an image of the
head of the subject at the first angle of view. The first imaging unit 30 is also
referred to as a first imaging means.
[0016] The second acquisition unit 12 is also referred to as a second acquisition means.
The second acquisition unit 12 is connected to a second imaging unit 40 in a communicable
manner, and acquires, from the second imaging unit 40, image data (second image data)
relating to a second image of an eye region of the subject that is captured at a second
angle of view. Herein, the second imaging unit 40 is a camera that captures an image
of the eye region of the subject at the second angle of view. The second imaging unit
40 is also referred to as a second imaging means. The second angle of view is narrower
than the first angle of view. Further, the eye region may be an eyeball or a peripheral
region including an eyeball.
[0017] The movement control unit 13 is also referred to as a movement control means. The
movement control unit 13 moves a visual field range of the second imaging unit 40,
based on position information relating to the head of the subject that is acquired
based on the first image. The visual field range of the second imaging unit 40 is
a range that is captured as an image by the second imaging unit 40, and is also referred
to as a capture volume. The visual field range is defined based on an angle of view
and a camera optical axis, and is wider as the angle of view is larger.
[0018] The state evaluation unit 15 is also referred to as a state evaluation means. The
state evaluation unit 15 evaluates a change in state of the eye of the subject, based
on chronological data relating to the second image.
[0019] Fig. 2 is a block diagram illustrating a hardware configuration of the information
processing system 10 according to the first example embodiment.
[0020] The information processing system 10 includes a processor 100, a read only memory
(ROM) 101, a random access memory (RAM) 102, and an interface (IF) unit 103 as main
hardware configurations. The processor 100, the ROM 101, the RAM 102, and the interface
unit 103 are connected mutually to one another via a data bus or the like.
[0021] The processor 100 has a function as an arithmetic device that executes control processing
and arithmetic processing. The processor 100 may be a central processing unit (CPU),
a graphics processing unit (GPU), a field-programmable gate array (FPGA), a digital
signal processor (DSP), an application specific integrated circuit (ASIC), or a combination
thereof. The ROM 101 has a function of storing a control program, an arithmetic program,
and the like that are executed by the processor 100. The RAM 102 has a function of
temporarily storing processing data and the like. The interface unit 103 performs
input and output of a signal with the outside in a wired or wireless manner. Further,
the interface unit 103 receives an input operation of data from a user, and displays
information for the user. For example, the interface unit 103 communicates with the
first imaging unit 30 and the second imaging unit 40.
[0022] As described above, with the information processing system 10 according to the first
example embodiment, the visual field range of the narrow-angle camera is moved based
on the position information relating to the head of the subject that is acquired based
on the first wide-angle image, the eye region is captured as an enlarged image, and
thus a change in state of the eye is evaluated. Therefore, even when the head of the
subject is not fixed, the eye region can be prevented from deviating from the visual
field range of the narrow-angle camera. With this, the information processing system
10 can suitably evaluate a change in state of the eye of the subject while the head
is in a relaxed posture without being fixed.
<Second Example Embodiment>
[0023] Fig. 3 is a system configuration diagram of an eye state measurement system 1a to
which an information processing system (hereinafter, referred to as an information
processing device) according to a second example embodiment is applicable. Note that,
in this drawing, it is assumed that a right-and-left direction of a subject P corresponds
to an X-axis direction, a front-and-rear direction of the subject P corresponds to
a Y-axis direction, and an up-and-down direction of the subject P corresponds to the
Z-axis direction.
[0024] Herein, the eye state measurement system 1a is a computer system that measures and
evaluates a change in state of the eye of the subject. In the second example embodiment,
the eye state measurement system 1a measures and evaluates a vibration state of a
pupil that is caused by a saccade phenomenon. The eye state measurement system 1a
includes an information processing device 10a, a first imaging unit 30a, the second
imaging unit 40, and a movable mirror 50, a driving unit 51, and a light source 61.
Note that a distance between the first imaging unit 30a and the subject P in the Y-axis
direction is denoted as D. D is, for example, 2±0.5 [m].
[0025] The information processing device 10a is associated with the information processing
system 10 in Fig. 1. The information processing device 10a is connected to the first
imaging unit 30a, the second imaging unit 40, and the driving unit 51. When the first
image data is received from the first imaging unit 30a, the information processing
device 10a generates a control signal required for causing the eye region of the subject
P to fall within the visual field range of the second imaging unit 40 and a control
signal required for causing the second imaging unit 40 to focus on the eye region
of the subject P. Then, the information processing device 10a transmits the control
signal for the visual field range in the X-axis direction and the Z-axis direction
to the driving unit 51, and transmits the control signal for the focal position to
the second imaging unit 40.
[0026] The first imaging unit 30a is a camera having a function similar to that of the first
imaging unit 30 in Fig. 1. In the second example embodiment, the first imaging unit
30a includes a wide-angle camera 31. The wide-angle camera 31 captures an image of
at least a face of the subject P, and generates the first image data. A focal distance
of the wide-angle camera 31 is set in advance in such a way that the face of the subject
P at a distance of D [m] can be captured as an image. For example, the focal distance
of the wide-angle camera 31 may be less than 200 [mm], preferably less than 70 [mm],
and is 28 [mm] in the second example embodiment. Note that the focal distance of the
wide-angle camera 31 may be 12 [mm].
[0027] A frame rate of the wide-angle camera 31 is set in advance in such a way as to follow
vibration of the eye region due to vibration of the head of the subject P and suppress
an excessive increase of the data amount. For example, the frame rate of the wide-angle
camera 31 may be 120 [fps] to 1,200 [fps], preferably 240 [fps] to 1,000 [fps], more
preferably 480 [fps] to 1,000 [fps], and is 500 [fps] in the second example embodiment.
Herein, the wide-angle camera 31 may capture an image of an object at a frame rate
equal to or greater than a frame rate of the second imaging unit 40 described later.
With this, the visual field range of the second imaging unit 40 can easily be controlled
in a suitable manner, according to vibration of the eye region.
[0028] When the first image data is generated, the wide-angle camera 31 transmits the first
image data to the information processing device 10a.
[0029] The second imaging unit 40 is a camera that captures an image of the eye region of
the subject at the second angle of view narrower than the first angle of view and
generates the second image data. For facilitation of pupil detection described later,
the second imaging unit 40 may be a near-infrared ray camera. A detection wavelength
of a photodetector of the second imaging unit 40 is referred to as a wavelength for
pupil detection, and the wavelength for pupil detection is a wavelength of, for example,
940 [nm]. Further, the second imaging unit 40 includes a telephoto lens and a liquid
lens. The liquid lens is a lens for determining a focal position (focus point), and
may be controlled based on the control signal for the focal position from the information
processing device 10a. The focal distance of the second imaging unit 40 is set in
such a way as to capture an image of the eye region of the subject P at the distance
D [m] from the movable mirror 50, and is longer than the focal distance of the wide-angle
camera 31. For example, the focal distance of the second imaging unit 40 may be 100
[mm] or longer, preferably 150 [mm] or longer, and is 200±50 [mm] in the second example
embodiment. Herein, the second imaging unit 40 has a long focal distance, and thus
has an extremely narrow depth of field. Therefore, the second imaging unit 40 controls
the focal position, based on the distance information, and thus focuses on the eye
region (focusing). Further, the frame rate of the second imaging unit 40 is set in
advance in such a way as to observe a change in state of the eye, in this example,
a saccade phenomenon, and suppress an excessive increase of the data amount. For example,
the frame rate of the second imaging unit 40 may be 120 [fps] to 1,000 [fps], preferably
240 [fps] to 1,000 [fps], more preferably 500 [fps] to 1,000 [fps], and is 500 [fps]
in the second example embodiment. When the second image data is generated, the second
imaging unit 40 transmits the second image data to the information processing device
10a.
[0030] The movable mirror 50 is a pair of mirrors that move an optical axis of the second
imaging unit 40. The movable mirror 50 includes a first movable mirror 50-1 and a
second movable mirror 50-2. In the following description, when there is no need to
discriminate the first movable mirror 50-1 and the second movable mirror 50-2 from
each other, the movable mirror 50 is simply used. Each of the movable mirrors 50 is
connected to the driving unit 51 via a support portion (not illustrated) in a fixed
manner in such a way as to form a predetermined inclination angle, and is configured
in such a way as to change the inclination angle along with rotation of the support
portion. For example, the first movable mirror 50-1 is configured in such a way that
the support portion being the connection destination rotates about the Z-axis, and
the second movable mirror 50-2 is configured in such a way that the support portion
being the connection destination rotates about the X-axis. With this, the movable
mirror 50 is capable of moving the optical axis of the second imaging unit 40 in the
X-axis direction and the Z-axis direction, that is, moving the visual field range
of the second imaging unit 40 in the X-axis direction and the Z-axis direction.
[0031] Note that, in the first example embodiment, the movable mirror 50 is a galvanometer
mirror having a relatively small mass and high responsiveness. With this, the eye
state measurement system 1a easily captures an image of the eye region while associating
the visual field range of the second imaging unit 40 with microscopic and high-speed
movement of the head of the subject P.
[0032] The driving unit 51 is also referred to as a driving means. The driving unit 51 is
a driving motor that rotates each of the movable mirrors 50 via the support portion.
The driving unit 51 includes a first driving unit 51-1 associated with the first movable
mirror 50-1 and a second driving unit 51-2 associated with the second movable mirror
50-2. Note that the first driving unit 51-1 and the second driving unit 51-2 are also
simply referred to as the driving unit 51. Herein, the driving unit 51 rotates the
movable mirror 50, based on the control signal for the visual field range in the X-axis
direction and the Z-axis direction from the information processing device 10a.
[0033] The light source 61 is a light source that irradiates the face of the subject P.
The light source 61 is a light source having a wavelength region associated with the
wavelength for pupil detection, and is a near-infrared light source of 940 [nm] in
the second example embodiment.
[0034] In other words, the second imaging unit 40 captures an image of the eye region of
the subject P with incident light through a path including the eye region of the subject
P, the first movable mirror 50-1, the second movable mirror 50-2, and the second imaging
unit 40 in the stated order.
[0035] Fig. 4 is a block diagram illustrating a functional configuration of the information
processing device 10a according to the second example embodiment. The information
processing device 10a in Fig. 4 includes an element position estimation unit 14 and
an output unit 16 in addition to the configuration elements of the information processing
device 10 in Fig. 1.
[0036] The first acquisition unit 11 is connected to the first imaging unit 30a (the wide-angle
camera 31), and receives and acquires the first image data from the first imaging
unit 30a. The first acquisition unit 11 may acquire the first image data at the frame
rate equal to or greater than the frame rate at which the second acquisition unit
12 acquires the second image data. The first acquisition unit 11 supplies the first
image data thus acquired to the movement control unit 13.
[0037] The second acquisition unit 12 is connected to the second imaging unit 40, and receives
and acquires the second image data from the second imaging unit 40. The second acquisition
unit 12 supplies the second image data thus acquired to the element position estimation
unit 14. Further, the second acquisition unit 12 may supply the second image data
thus acquired to the movement control unit 13.
[0038] The movement control unit 13 generates change information relating to a position
of the face of the subject P in the X-axis direction and the Z-axis direction, based
on chronological data relating to the first image. The movement control unit 13 calculates
each of rotation amounts of the first movable mirror 50-1 and the second movable mirror
50-2, based on the change information. The movement control unit 13 may use the second
image data in addition to the change information for calculating the rotation amounts.
Then, the movement control unit 13 generates the control signal for the visual field
range in the X-axis direction and the Z-axis direction, based on each of the rotation
amounts, transmits the control signal for the visual field range in the Z-axis direction
to the first driving unit 51-1, and transmits the control signal for the visual field
range in the X-axis direction to the second driving unit 51-2. Further, the movement
control unit 13 executes focal position control processing, based on the chronological
data relating to the first image and the second image, generates the change information
relating to a position of the face of the subject P in the Y-axis direction, and generates
the control signal for the focal position. Further, the movement control unit 13 transmits
the control signal to the second imaging unit 40, to control the liquid lens. In this
manner, the second imaging unit 40 moves the visual field range and the focal position
(focus point), based on the change information relating to the position of the face.
[0039] The element position estimation unit 14 is also referred to as an element position
estimation means. The element position estimation unit 14 estimates a position of
an element of an eye in the second image. In the second example embodiment, the position
of the element of the eye that is estimated by the element position estimation unit
14 includes a gravity center position of the pupil of the subject P and positions
of an outer eye corner and an inner eye corner. Note that any one of the positions
of the outer eye corner and the inner eye corner may be used, or a position of another
freely selected point may be used instead. The element position estimation unit 14
supplies the position information relating to the element of the eye to the state
evaluation unit 15.
[0040] The state evaluation unit 15 generates difference information relating to the position
of the element of the eye in the second image, and evaluates the vibration state of
the pupil of the subject P, based on the difference information. The vibration state
of the pupil may be at least one of a vibration amount, a vibration direction, a vibration
frequency, and a vibration duration of the pupil. The state evaluation unit 15 supplies
the information relating to the vibration state of the pupil to the output unit 16.
[0041] The output unit 16 is also referred to as an output means. The output unit 16 outputs
the information relating to the vibration state of the pupil, as an evaluation result.
The output unit 16 may include a display unit (not illustrated) that displays the
evaluation result. Further, the output unit 16 may include a transmission unit (not
illustrated) that transmits the evaluation result to an external device (not illustrated).
[0042] Herein, with reference to Fig. 5, movement control processing is described in detail.
Fig. 5 is a diagram for describing the movement control processing executed by the
movement control unit 13 according to the second example embodiment. In this drawing,
a second image IMG_b is illustrated. An eye region including an inner eye corner Ib
and an outer eye corner Ob of the subject P is captured as the second image IMG_b.
[0043] First, the movement control unit 13 calculates position coordinates of the eye region
in the first image (referred to as eye position coordinates). In the second example
embodiment, the eye position coordinates are, but not limited to, position coordinates
of the gravity center of the eye region when the eye region in the first image is
similar to a substantially elliptic region, and may be a position coordinate range
of the eye region in the first image. Subsequently, the movement control unit 13 calculates
position coordinates of a gravity center VC of an image region (projection region)
VA in the second image, the position coordinates being associated with the eye position
coordinates in the first image when the eye region in the first image is virtually
projected onto the second image IMG_b. In this case, the movement control unit 13
may calculate the position coordinates of the gravity center VC of the projection
region VA, based on the difference information relating to the eye position coordinates
in the first image between the previous image-capturing timing and the current image-capturing
timing and the second image at the previous image-capturing timing. Note that the
difference information relating to the eye position coordinates is an example of the
change information relating to the position of the face that is described above. Then,
the movement control unit 13 calculates the rotation amounts of the first movable
mirror 50-1 and the second movable mirror 50-2 in such a way that the gravity center
VC of the projection region VA is arranged at the center C of the second image. With
this, the movement control unit 13 is capable of moving the visual field range of
the second imaging unit 40 in the X-axis direction and the Z-axis direction in such
a way that an eye region A is always arranged at the center of the second image IMG_b.
Note that, in this drawing, the projection region VA and the eye region A are substantially
elliptic regions that are symmetric with respect to the major axis and the minor axis,
and an intermediate point between the outer eye corner Ob and the inner eye corner
Ib that are positioned on both ends of each of the projection region VA and the eye
region A matches with the gravity center.
[0044] Note that the number of pixels in the width direction of the eye region included
in the second image IMG_b is set in such a way as to fall within a predetermined range
with respect to the number of pixels in the width direction of the second image IMG_b.
The number of pixels in the width direction of the second image IMG_b is denoted with
xb, and the number of pixels in the height direction is denoted with zb. As an example,
xb×zb=640×480 is satisfied. Further, the number of pixels from the inner eye corner
Ib to the outer eye corner Ob of the subject P captured in the second image IMG_b
is denoted with x1. In other words, x1/xb is maintained to fall within a predetermined
range. For example, x1/xb is 0.5 or greater and less than 1, preferably 0.8 or greater
and less than 1. With this, the information processing device 10a is capable of detecting
a pupil from the second image IMG_b at high accuracy.
[0045] Herein, when the movement control unit 13 controls movement of the first movable
mirror 50-1 and the second movable mirror 50-2 in such a way that the eye region A
is arranged at the center of the second image IMG_b, the rotation amounts differ according
to the distance D between the camera and the subject P in the Y-axis direction. Therefore,
the movement control unit 13 is required to adjust the focus of the second image IMG_b.
[0046] Herein, Fig. 6 is a diagram for describing the focal position control processing
executed by the movement control unit 13 according to the second example embodiment.
This drawing illustrates a top schematic diagram the subject P, the first imaging
unit 30a, and the second imaging unit 40 in the upper part, and illustrates the first
imageIMG_a of the first imaging unit 30a and the second image IMG_b of the second
imaging unit 40 in the lower part.
[0047] First, the movement control unit 13 calculates position coordinates (xai, zai) of
an inner eye corner Ia of the subject P in the first image IMG_a. Then, the movement
control unit 13 controls rotation of the movable mirror 50 to acquire the second image
IMG_b. At this state, an angle of the movable mirror 50 is denoted with G (xg, zg).
The movement control unit 13 calculates position coordinates (xbi, zbi) of the inner
eye corner Ib in the second image IMG_b. The movement control unit 13 uses the position
coordinates (xai, zai) of the inner eye corner Ia in the first image IMG_a, the position
coordinates (xbi, zbi) of the inner eye corner Ib in the second image IMG_b, and the
angel G (xg, zg) of the movable mirror 50 to calculate the distance D with a stereo
image method. Further, the movement control unit 13 determines the focal position
of the second imaging unit 40, based on the distance D, and generates the control
information for moving the focus to the focal position. In other words, the movement
control unit 13 uses the first imaging unit 30a and the second imaging unit 40 as
stereo cameras, and adjusts the focus of the second imaging unit 40, which has narrow
depth of field, to match with the eye position of the subject P.
[0048] Note that, in the example described above, the movement control unit 13 uses the
position coordinates of the inner eye corners Ia and Ib, and may use position coordinates
of outer eye corners Oa and Ob instead.
[0049] Next, with reference to Fig. 7, state evaluation processing is described in detail.
Fig. 7 is a diagram for illustrating the state evaluation processing executed by the
state evaluation unit 15 according to the second example embodiment. In this drawing,
the second image IMG_b is also illustrated.
[0050] First, the state evaluation unit 15 calculates a relative position of a gravity center
G of a pupil being a moving point with respect to a reference point in the second
image IMG_b at each image-capturing timing. The reference point is preferably a fixed
point at a position that is hardly changed by an opening degree of an eyelid or movement
of a visual line. Herein, the state evaluation unit 15 may use a position of a cornea
reflection image as the reference point, but uses a point based on a position being
at least one of the outer eye corner Ob and the inner eye corner Ib in the second
example embodiment. In other words, the state evaluation unit 15 calculates the relative
position based on the position information relating to at least one of the outer eye
corner Ob and the inner eye corner Ib and the position information relating to the
gravity center G of the pupil. For example, the state evaluation unit 15 calculates
the relative position between a linear line L connecting the position of the outer
eye corner Ob and the position of the inner eye corner Ib to each other and the gravity
center G of the pupil. The state evaluation unit 15 calculates distances Δx and Δz
from the intermediate point between the outer eye corner Ob and the inner eye corner
Ib (which may match with the center C of the second image IMG_b) to the gravity center
G in the X-axis direction and the Z-axis direction, as the relative position of the
gravity center G of the pupil. The state evaluation unit 15 may calculate a value
acquired by standardizing the distances Δx and Δz in the X-axis direction and the
Z-axis direction with the distance x1, as the relative position of the gravity center
G of the pupil in the second image IMG_b. Note that the state evaluation unit 15 may
only use the position of any one of the outer eye corner Ob and the inner eye corner
Ib, as the reference point. Then, the state evaluation unit 15 evaluates the vibration
state of the pupil, based on the difference information relating to the relative position
of the gravity center G of the pupil between the adjacent image-capturing timings.
[0051] As described above, the state evaluation unit 15 uses the position of at least one
of the outer eye corner Ob and the inner eye corner Ib, as the reference point. Therefore,
as compared to a case of using a cornea reflection image, the system configuration
is more simplified because the eye state measurement system 1a is not required to
include an infrared light source for cornea reflection image formation. Further, the
state evaluation unit 15 enables high speed processing and downsizing of the device
because relative position calculation is facilitated and a calculation amount is reduced.
Therefore, in the second example embodiment, description is made on a case in which
the information processing device 10a is an independent computer device, but the information
processing device 10a may be implemented in the first imaging unit 30a or the second
imaging unit 40.
[0052] Fig. 8 is a flowchart illustrating a procedure of information processing executed
by the information processing device 10a according to the second example embodiment.
[0053] First, the first acquisition unit 11 acquires the first image data at t=ti from the
first imaging unit 30a at a predetermined frame rate (Step S10).
[0054] Subsequently, the movement control unit 13 detects the face of the subject P from
the first image (Step S11). For example, the movement control unit 13 may detect the
face of the subject P through use of a convolutional neural network (CNN) that is
leant with the first image as an input. Then, the movement control unit 13 may generate
a regularized image acquired by extracting an image region associated with the face
of the subject P.
[0055] Subsequently, the movement control unit 13 calculates the eye position coordinates
in the first image (Step S12). For example, the movement control unit 13 may detect
the eye region of the subject P from the regularized image thus generated through
template matching, and may calculate the position coordinates of the gravity center
of the eye region in the first image, as the eye position coordinates.
[0056] Then, the movement control unit 13 determines whether the eye position coordinates
in the first image at t=ti change as compared to the eye position coordinates in the
first image at t=ti-1 (Step S13). When the eye position coordinates do not change
(No in Step S13), the movement control unit 13 determines whether a series of image-capturing
is terminated (Step S16), and the processing returns to Step S10 when a series of
image-capturing is not terminated (No in Step S16). Meanwhile, when the eye position
coordinates change (Yes in Step S13), the movement control unit 13 calculates the
rotation amount of the movable mirror 50 at a subsequent image-capturing timing (for
example, t=ti+1) with the above-mentioned method illustrated in Fig. 5 (Step S14).
At this state, the movement control unit 13 may acquire the second image data from
the second imaging unit 40 at the current or previous image-capturing timing (for
example, the second image data at t=ti-1), and may calculate the rotation amount of
the movable mirror 50, based on the eye position coordinates and the second image.
Then, the movement control unit 13 generates the control signal for the visual field
range in the X-axis direction and the Z-axis direction, based on the calculated rotation
amount. Note that, in addition to this, the movement control unit 13 calculates the
movement amount of the focal position of the second imaging unit 40 with the method
illustrated in Fig. 6, and generates the control signal for the focal position, based
on the movement amount.
[0057] Subsequently, the movement control unit 13 controls movement of the visual field
range and the focal position of the second imaging unit 40 (Step S15). Specifically,
the movement control unit 13 transmits, to the driving unit 51, the control signal
for the visual field range in the X-axis direction and the Z-axis direction. Further,
the movement control unit 13 transmits the control signal for the focal position to
the liquid lens.
[0058] Subsequently, the movement control unit 13 determines whether a series of image-capturing
is terminated (Step S16). When a series of image-capturing is not terminated (No in
Step S16), the movement control unit 13 returns the processing to Step S10. When a
series of image-capturing is terminated (Yes in Step S16), the processing proceeds
to Step S17.
[0059] In Step S17, the second acquisition unit 12 acquires the second image data from the
second imaging unit 40.
[0060] Herein, the information processing device 10a repeats the processing in Step S18
and Step S19 for the number of times that is equivalent to the number of frames of
the second image.
[0061] In Step S18, the element position estimation unit 14 detects the eye region from
the second image through, for example, template matching, and estimates position coordinates
of the elements (the outer eye corner, the inner eye corner, and the gravity center
of the pupil) of the eye. The element position estimation unit 14 may estimate the
position coordinates of the gravity center of the pupil by detecting an image region
of the pupil through, for example, binarization, edge detection, and Hough transformation,
and calculating the position coordinates of the gravity center of the image region.
[0062] Then, in Step S19, the state evaluation unit 15 evaluates the vibration state by
calculating the relative position of the gravity center of the pupil with the above-mentioned
method illustrated in Fig. 7 and generating the difference information relating to
the relative position.
[0063] Subsequently, in Step S20, the output unit 16 outputs the information relating to
the evaluated vibration state.
[0064] Note that the processing in Step S17 to Step S19 may be executed in parallel to the
processing in Step S10 to Step S16. Further, specific processing in each of the steps
is not limited to the description given above.
[0065] As described above, the information processing device 10a according to the second
example embodiment can exert effects similar to those in the first example embodiment.
In particular, when microscopic vibration of an eyeball is to be measured, vibration
of a head needs to be strictly suppressed, and hence the head needs to be firmly fixed.
Therefore, it is difficult to fix a head immediately after a subject have the head
operated due to a brain disease, and thus an eye state such as a saccade phenomenon
cannot be examined. Further, the current technology only enables examination on a
response of a brain in a tensed state due to fixation of a head. The same holds true
for determination on a degree of interest in an image when a subject sees the image,
as well as examination on a recovery state from a brain disease.
[0066] However, with the information processing system 10a according to the second example
embodiment, such a problem can be solved. With this, for example, an evaluation value
acquired from the information processing system 10a through evaluation can be utilized
as evidence for deciding a transitional phase of rehabilitation from a brain disease,
and a medical worker can easily diagnose a state of recovery from illness while alleviating
a burden on a subject being a patient. Further, an evaluation value acquired from
the information processing system 10a through evaluation is collected as a degree
of interest of a subject being a consumer in an advertisement image, and thus a company
can measure an effect of the advertisement quantitatively.
<Third Example Embodiment>
[0067] Next, with reference to Fig. 9, a third example embodiment of the present disclosure
is described. The third example embodiment is characterized in that the visual field
range of the first imaging unit moves according to movement of the head of the subject
and the visual field range of the second imaging unit also moves accordingly.
[0068] Fig. 9 is a system configuration diagram of an eye state measurement system 1b according
to the third example embodiment.
[0069] The eye state measurement system 1b according to the third example embodiment basically
includes configurations and functions that are similar to those of the eye state measurement
system 1a according to the second example embodiment. However, the eye state measurement
system 1b is different from the eye state measurement system 1a in that a first imaging
unit 30b and an information processing device 10b are included in place of the first
imaging unit 30a and the information processing device 10a, and a light source 63,
a long-pass filter 70, and a half mirror 71 are further included.
[0070] The first imaging unit 30b includes a tracking camera 32 that has a visual field
range moving according to movement of the head of the subject P. In the third example
embodiment, the visual field range of the tracking camera 32 moves according to movement
of the eye region of the subject P in such a way that the eye region of the subject
P is captured at the center of the first image. Specifically, similarly to the second
imaging unit 40, the visual field range of the tracking camera 32 moves by rotation
of the movable mirror 50 by the driving unit 51 following the control signal for the
visual field range of the information processing device 10b. In other words, the movable
mirror 50 moves an optical axis of the first imaging unit 30b in addition to the optical
axis of the second imaging unit 40. As a result, the visual field range of the second
imaging unit 40 moves in association with movement of the visual field range of the
tracking camera 32. Herein, a detection wavelength of a photodetector of the tracking
camera 32 is referred to as a wavelength for tracking, and the wavelength for tracking
is smaller than the wavelength for pupil detection being the detection wavelength
of the photodetector of the second imaging unit 40. As an example, the wavelength
for tracking is 850 [nm]. Note that the angle of view, the focal distance, and the
frame rate of the tracking camera 32 are similar to those of the wide-angle camera
31 in the second example embodiment.
[0071] The light source 63 is a light source that irradiates the eye region of the subject
P. The light source 63 is a light source having a wavelength region associated with
the wavelength for tracking, and is a near-infrared light source of 850 [nm] in the
third example embodiment.
[0072] The half mirror 71 is a half mirror that reflects part of 850-nm incident light from
the light source 63 toward the long-pass filter 70. Further, the half mirror 71 causes
part of 850-nm incident light from the long-pass filter 70 to pass therethrough toward
the tracking camera 32. Note that the half mirror 71 may be a beam splitter having
a freely selected ratio of transmission and reflection, in place of a half mirror.
[0073] The long-pass filter 70 is an optical filter that causes light having the wavelength
for pupil detection to pass therethrough and reflects light having the wavelength
for tracking. The long-pass filter 70 is provided between the second imaging unit
40 and the movable mirror 50, and causes incident light having the wavelength for
pupil detection from the movable mirror 50 to pass therethrough toward the second
imaging unit 40. Further, the long-pass filter 70 reflects, toward the half mirror
71, incident light having the wavelength for tracking from the movable mirror 50.
[0074] In other words, part of light from the light source 63 arrives at the eye region
of the subject P via a path including the half mirror 71, the long-pass filter 70,
the second movable mirror 50-2, and the first movable mirror 50-1 in the stated order.
[0075] Further, the tracking camera 32 of the first imaging unit 30b captures an image of
the eye region (or the face) of the subject P at the first angle of view with incident
light through a path including the eye region (or the face) of the subject P, the
first movable mirror 50-1, the second movable mirror 50-2, the long-pass filter 70,
the half mirror 71, and the tracking camera 32 in the stated order.
[0076] Further, the second imaging unit 40 captures an image of the eye region of the subject
P at the second angle of view with incident light through a path including the eye
region of the subject P, the first movable mirror 50-1, the second movable mirror
50-2, the long-pass filter 70, and the second imaging unit 40 in the stated order.
[0077] The information processing device 10b basically includes configurations and functions
that are similar to those of the information processing device 10a, but is different
from the second example embodiment in movement control processing executed by the
movement control unit 13 in the X-axis direction and the Z-axis direction (processing
shown in Step S14 in Fig. 5 and Fig. 8).
[0078] In the third example embodiment, first, the movement control unit 13 calculates the
eye position coordinates in the first image. Then, the movement control unit 13 calculates
the rotation amounts of the first movable mirror 50-1 and the second movable mirror
50-2 in such a way that the gravity center of the eye region in the first image is
arranged at the center of the first image at a subsequent image-capturing timing.
Herein, the tracking camera 32 of the first imaging unit 30b has the frame rate equal
to or greater than the frame rate of the second imaging unit 40, and hence the eye
region of the subject P is also captured at the image center of the second image captured
by the second imaging unit 40.
[0079] As described above, according to the third example embodiment, the information processing
device 10b is capable of moving the visual field range of the second imaging unit
40 following movement of the head of the subject P at higher accuracy. With this,
the information processing device 10b can suitably evaluate a change of a state of
the eye of the subject P while the head is in a relaxed posture without being fixed.
<Fourth Example Embodiment>
[0080] Next, with reference to Fig. 10 and Fig. 11, a fourth example embodiment of the present
disclosure is described. The fourth example embodiment is characterized in that a
position of a visual field range of a narrowest-angle camera is roughly adjusted with
a widest-angle image and a position of a visual field range is finely adjusted with
an intermediate-angle image.
[0081] Fig. 10 is a system configuration diagram of an eye state measurement system 1c according
to the fourth example embodiment. The eye state measurement system 1c according to
the fourth example embodiment basically includes configurations and functions that
are similar to those of the eye state measurement system 1b according to the third
example embodiment. However, the eye state measurement system 1c is different from
the eye state measurement system 1b in that a third imaging unit 33 and an information
processing device 10c in place of the information processing device 10b are included.
[0082] The third imaging unit 33 is a camera that captures an image of at least the head
of the subject P at a third angle of view and generates third image data relating
to a third image. The third angle of view is larger than the first angle of view of
the tracking camera 32, and is larger than the second angle of view of the second
imaging unit 40. When the third image data is generated, the third imaging unit 33
transmits the third image data to the information processing device 10c.
[0083] The information processing device 10c basically includes functions that are similar
to those of the information processing device 10b, but is different from the information
processing device 10b in that movement control for the visual field ranges of the
tracking camera 32 and the second imaging unit 40 is roughly adjusted based on the
third image data.
[0084] Fig. 11 is a block diagram illustrating a functional configuration of the information
processing device 10c according to the fourth example embodiment. In addition to the
configuration of the information processing device 10b, the information processing
device 10c includes a third acquisition unit 17. The third acquisition unit 17 receives
and acquires the third image data from the third imaging unit 33, and supplies the
third image data to the movement control unit 13.
[0085] The movement control unit 13 detects the face of the subject P from the third image,
calculates the eye position coordinates in the third image, and calculates the rotation
amount of the movable mirror 50, based on the eye position coordinates, in such a
way that the eye region falls within the first image. Then, the movement control unit
13 transmits, to the driving unit 51, the control signal of the visual field range
based on the rotation amount, and roughly adjusts the inclination angle of the movable
mirror 50.
[0086] Such rough adjustment processing may be executed before Step S10 in Fig. 8. Further,
for preventing an increase of a processing time, the rough adjustment processing may
be executed only when the face of the subject P is not detected from the first image
in Step S11 in Fig. 8 or when the eye region of the subject P is not detected from
the first image in Step S12, for example.
[0087] Note that the movement control processing in which the first image at the first angle
of view narrower than the third image is used is processing associated with Step S10
to Step S15 in Fig. 8, but may also be referred to fine adjustment processing. The
first angle of view in the fourth example embodiment is only required to be an angle
of view for capturing an image of at least the eye region of the subject P, and may
be narrower than the first angle of view in the second example embodiment and the
third example embodiment. In this case, in the fine adjustment processing, face detection
processing associated with Step S11 in Fig. 8 may be omitted. With this, even when
the rough adjustment processing is added, an increase of a time for a series of processing
can be prevented.
[0088] As described above, even when the head of the subject P largely moves, the information
processing device 10c according to the fourth example embodiment is capable of moving
the visual field range of the second imaging unit 40 in such a way as to prevent the
eye region of the subject P from deviating from the visual field range of the second
imaging unit 40. With this, the subject P can freely move during image-capturing,
and hence the information processing device 10c is capable of evaluating a change
in state of an eye in a more relaxed state.
[0089] With reference to the example embodiments, the present disclosure is described above,
but the invention of the present application is not limited thereto. Various modifications
that can be understood by a person skilled in the art may be made to the configurations
and the details of the invention of the present application within the scope of the
invention. For example, the state of the eye being an evaluation target is a contraction
amount of the pupil, and the state evaluation unit 15 may evaluate a contraction amount
of the pupil of the subject P, based on the chronological data relating to the second
image. Further, when an image such as an advertisement image is viewed, the state
evaluation unit 15 may evaluate a degree of interest, based on a contraction amount
of the pupil.
[0090] In the example embodiment described above, the present disclosure is described as
a hardware configuration, but the present disclosure is not limited thereto. In the
present disclosure, various processing relating to the state evaluation method can
be executed by causing a processor to execute a computer processing.
[0091] In the example described above, a program can be stored through use of a non-transitory
computer readable medium of various types, and can be supplied to a computer. The
non-transitory computer readable medium includes a tangible storage medium of various
types. Examples of the non-transitory computer readable medium include a magnetic
recording medium (for example, a flexible disk, a magnetic tape, a hard disk drive),
a magneto-optic recording medium (for example, a magneto-optic disk), a CD-read only
memory (ROM), CD-R, CD-R/W, and a semiconductor memory (for example, a mask ROM, a
programmable ROM (PROM), an erasable PROM (EPROM), a flash ROM, a random access memory
(RAM)). Further, the program may be supplied to the computer from a transitory computer
readable medium of various types. Examples of the transitory computer readable medium
include an electric signal, an optical signal, and an electromagnetic wave. The transitory
computer readable medium is capable of supplying the program to the computer via a
wired communication path such as an electric cable and an optical fiber or via a wireless
communication path.
[0092] In the example embodiments described above, the computer is configured as a computer
system including a personal computer, a word processor, or the like. However, the
computer may be configured as, but not limited to, a server of a local area network
(LAN), a host of a computer (personal computer) communication, a computer system connected
to the Internet, or the like. Further, the computer may be configured as a network
as whole by functionally distributing devices via the network.
[0093] A part or the entirety of the example embodiments described above may be described
as in the following supplementary notes, but is not limited to the followings.
(Supplementary Note 1)
[0094] An information processing system including:
a first acquisition unit configured to acquire, from a first imaging unit, image data
relating to a first image of a head of a subject, the first image being captured at
a first angle of view;
a second acquisition unit configured to acquire, from a second imaging unit, image
data relating to a second image of an eye region of the subject, the second image
being captured at a second angle of view narrower than the first angle of view;
a movement control unit configured to move a visual field range of the second imaging
unit, based on position information relating to the head of the subject, the position
information being acquired based on the first image; and
a state evaluation unit configured to evaluate a change in state of an eye of the
subject, based on chronological data relating to the second image.
(Supplementary Note 2)
[0095] The information processing system according to Supplementary Note 1, wherein
the movement control unit generates change information relating to a position of the
head of the subject, based on chronological data relating to the first image, and
moves the visual field range of the second imaging unit, based on the change information.
(Supplementary Note 3)
[0096] The information processing system according to Supplementary Note 2, wherein
the first acquisition unit acquires image data relating to the first image at a frame
rate equal to or greater than a frame rate at which the second acquisition unit acquires
image data relating to the second image.
(Supplementary Note 4)
[0097] The information processing system according to any one of Supplementary Notes 1 to
3, further including:
an element position estimation unit configured to estimate a position of a gravity
center of a pupil of the subject and a position of at least one of an outer eye corner
and an inner eye corner in the second image, wherein
the state evaluation unit evaluates a vibration state of the pupil of the subject,
based on position information relating to at least one of the outer eye corner and
the inner eye corner and position information relating to the gravity center of the
pupil.
(Supplementary Note 5)
[0098] The information processing system according to Supplementary Note 4, wherein
the state evaluation unit calculates a relative position of the gravity center of
the pupil with respect to a linear line connecting a position of the outer eye corner
and a position of the inner eye corner, and evaluates a vibration state of the pupil
of the subject, based on the relative position.
(Supplementary Note 6)
[0099] The information processing system according to any one of Supplementary Notes 1 to
3, wherein
the state evaluation unit evaluates a contraction amount of the pupil of the subject,
based on the chronological data relating to the second image.
(Supplementary Note 7)
[0100] An eye state measurement system including:
a first imaging unit configured to capture an image of a head of a subject at a first
angle of view;
a second imaging unit configured to capture an image of an eye region of the subject
at a second angle of view narrower than the first angle of view; and
an information processing device, wherein
the information processing device includes:
a first acquisition unit configured to acquire, from the first imaging unit, image
data relating to a first image;
a second acquisition unit configured to acquire, from the second imaging unit, image
data relating to a second image;
a movement control unit configured to move a visual field range of the second imaging
unit, based on position information relating to the head of the subject, the position
information being acquired based on the first image; and
a state evaluation unit configured to evaluate a change in state of an eye of the
subject, based on chronological data relating to the second image.
(Supplementary Note 8)
[0101] The eye state measurement system according to Supplementary Note 7, further including:
a movable mirror configured to move an optical axis of the second imaging unit; and
a driving unit configured to drive the movable mirror, wherein
the movement control unit generates change information relating to a position of the
head of the subject, based on chronological data relating to the first image, and
calculates a rotation amount of the movable mirror, based on the change information.
(Supplementary Note 9)
[0102] The eye state measurement system according to Supplementary Note 8, wherein
the movable mirror move optical axes of the first imaging unit and the second imaging
unit.
(Supplementary Note 10)
[0103] The eye state measurement system according to any one of Supplementary Notes 7 to
9, wherein
the first imaging unit captures an image of an object at a frame rate equal to or
greater than a frame rate of the second imaging unit.
(Supplementary Note 11)
[0104] An information processing method including:
acquiring, from a first imaging unit, image data relating to a first image of a head
of a subject, the first image being captured at a first angle of view;
acquiring, from a second imaging unit, image data relating to a second image of an
eye region of the subject, the second image being captured at a second angle of view
narrower than the first angle of view;
moving a visual field range of the second imaging unit, based on position information
relating to the head of the subject, the position information being acquired based
on the first image; and
evaluating a change in state of an eye of the subject, based on chronological data
relating to the second image.
(Supplementary Note 12)
[0105] A program for causing a computer to execute:
first acquisition processing of acquiring, from a first imaging unit, image data relating
to a first image of a head of a subject, the first image being captured at a first
angle of view;
second acquisition processing of acquiring, from a second imaging unit, image data
relating to a second image of an eye region of the subject, the second image being
captured at a second angle of view narrower than the first angle of view;
movement control processing of moving a visual field range of the second imaging unit,
based on position information relating to the head of the subject, the position information
being acquired based on the first image; and
state evaluation processing of evaluating a change in state of an eye of the subject,
based on chronological data relating to the second image.
[0106] This application is based upon and claims the benefit of priority from
Japanese patent application No. 2020-176066, filed on October 20, 2020, the disclosure of which is incorporated herein in its
entirety by reference.
Industrial Applicability
[0107] The information processing system according to the present disclosure is applicable
for evaluating a change in state of an eye of a subject.
Reference Signs List
[0108]
1a, 1b, 1c EYE STATE MEASUREMENT SYSTEM
10, 10a, 10b, 10c INFORMATION PROCESSING SYSTEM (INFORMATION PROCESSING DEVICE)
11 FIRST ACQUISITION UNIT
12 SECOND ACQUISITION UNIT
13 MOVEMENT CONTROL UNIT
14 ELEMENT POSITION ESTIMATION UNIT
15 STATE EVALUATION UNIT
16 OUTPUT UNIT
17 THIRD ACQUISITION UNIT
30, 30a, 30b FIRST IMAGING UNIT
31 WIDE-ANGLE CAMERA
32 TRACKING CAMERA
33 THIRD IMAGING UNIT
40 SECOND IMAGING UNIT
50 MOVABLE MIRROR
51 DRIVING UNIT
61 LIGHT SOURCE
63 LIGHT SOURCE
70 LONG-PASS FILTER
71 HALF MIRROR
100 PROCESSOR
101 ROM
102 RAM
103 INTERFACE (IF) UNIT
P SUBJECT
IMG_a FIRST IMAGE
IMG_b SECOND IMAGE