BACKGROUND OF THE INVENTION
1. Field of the Invention
[0001] A technique of the present disclosure relates to a diagnosis support device, an operation
method of a diagnosis support device, an operation program of a diagnosis support
device, and a dementia diagnosis support method.
2. Description of the Related Art
[0002] In diagnosis of a disease, for example, dementia represented by Alzheimer's disease,
a doctor refers to a medical image such as a head magnetic resonance imaging (MRI)
image. The doctor obtains a dementia opinion by observing, for example, a degree of
atrophy of a hippocampus, a parahippocampal gyrus, an amygdala, and the like, a degree
of a vascular disorder of a white matter, the presence or absence of a decrease in
blood flow metabolism in a frontal lobe, a temporal lobe, and an occipital lobe.
[0003] JP6483890B describes a diagnosis support device that derives a dementia opinion on a head MRI
image by a machine learning model and provides the dementia opinion to a doctor. The
diagnosis support device described in
JP6483890B extracts a plurality of anatomical regions according to a Brodmann's brain map or
the like from the head MRI image, and calculates a Z value indicating a degree of
atrophy of each of the anatomical regions. In addition, the calculated Z value of
each of the anatomical regions is input to a machine learning model, and a dementia
opinion is output from the machine learning model.
SUMMARY OF THE INVENTION
[0004] As described above, in order to obtain an opinion of a disease such as dementia,
it is necessary to thoroughly examine each of anatomical regions of an organ such
as a brain from various viewpoints. However, in
JP6483890B, only one index value such as the Z value which is statistically obtained is used.
For this reason, there is a limit to prediction accuracy of a disease opinion that
is obtained with only such limited information.
[0005] In addition, even though a disease opinion with high prediction accuracy is obtained,
in a case where the doctor does not recognize a degree of contribution of each of
the anatomical regions to output of the opinion, the doctor cannot determine a plausibility
of the opinion.
[0006] In one embodiment according to the technique of the present disclosure, there are
provided a diagnosis support device, an operation method of a diagnosis support device,
an operation program of a diagnosis support device, and a dementia diagnosis support
method capable of obtaining a more accurate disease opinion and recognizing a degree
of contribution of each of anatomical regions of an organ to output of the opinion.
[0007] According to the present disclosure, there is provided a diagnosis support device
including: a processor; and a memory connected to or built in the processor, in which
the processor is configured to: acquire a medical image; extract a plurality of anatomical
regions of an organ from the medical image; input images of the plurality of anatomical
regions to a plurality of feature amount derivation models prepared for each of the
plurality of anatomical regions and output a plurality of feature amounts for each
of the plurality of anatomical regions from the feature amount derivation models;
input the plurality of feature amounts which are output for each of the plurality
of anatomical regions to a disease opinion derivation model and output a disease opinion
from the disease opinion derivation model; derive a first contribution which represents
a degree of contribution to output of the opinion for each of the anatomical regions;
and present the opinion and a derivation result of the first contribution for each
of the anatomical regions.
[0008] Preferably, the processor is configured to: present the derivation result in descending
order of the first contribution.
[0009] Preferably, the processor is configured to: input disease-related information related
to the disease to the disease opinion derivation model in addition to the plurality
of feature amounts.
[0010] Preferably, the disease-related information includes a plurality of items, and the
processor is configured to: derive a second contribution which represents a degree
of contribution to output of the opinion for each of the items; and present a derivation
result of the second contribution for each of the items.
[0011] Preferably, the feature amount derivation model includes at least one of an auto-encoder,
a single-task convolutional neural network for class determination, or a multi-task
convolutional neural network for class determination.
[0012] Preferably, the processor is configured to: input an image of one anatomical region
of the anatomical regions to the plurality of different feature amount derivation
models, and output the feature amounts from each of the plurality of feature amount
derivation models.
[0013] Preferably, the disease opinion derivation model is configured by any one method
of a neural network, a support vector machine, or boosting.
[0014] Preferably, the processor is configured to: perform normalization processing of matching
the acquired medical image with a reference medical image prior to extraction of the
anatomical regions.
[0015] Preferably, the organ is a brain and the disease is dementia. In this case, preferably,
the plurality of anatomical regions include at least one of a hippocampus or a temporal
lobe. Further, preferably, the disease-related information includes at least one of
a volume of the anatomical region, a score of a dementia test, a test result of a
genetic test, a test result of a spinal fluid test, or a test result of a blood test.
[0016] According to the present disclosure, there is provided an operation method of a diagnosis
support device, the method including: acquiring a medical image; extracting a plurality
of anatomical regions of an organ from the medical image; inputting images of the
plurality of anatomical regions to a plurality of feature amount derivation models
prepared for each of the plurality of anatomical regions, and outputting a plurality
of feature amounts for each of the plurality of anatomical regions from the feature
amount derivation models; inputting the plurality of feature amounts which are output
for each of the plurality of anatomical regions to a disease opinion derivation model,
and outputting a disease opinion from the disease opinion derivation model; deriving
a first contribution which represents a degree of contribution to output of the opinion
for each of the anatomical regions; and presenting the opinion and a derivation result
of the first contribution for each of the anatomical regions.
[0017] According to the present disclosure, there is provided an operation program of a
diagnosis support device, the program causing a computer to execute a process including:
acquiring a medical image; extracting a plurality of anatomical regions of an organ
from the medical image; inputting images of the plurality of anatomical regions to
a plurality of feature amount derivation models prepared for each of the plurality
of anatomical regions, and outputting a plurality of feature amounts for each of the
plurality of anatomical regions from the feature amount derivation models; inputting
the plurality of feature amounts which are output for each of the plurality of anatomical
regions to a disease opinion derivation model, and outputting a disease opinion from
the disease opinion derivation model; deriving a first contribution which represents
a degree of contribution to output of the opinion for each of the anatomical regions;
and presenting the opinion and a derivation result of the first contribution for each
of the anatomical regions.
[0018] According to the present disclosure, there is provided a dementia diagnosis support
method causing a computer that includes a processor and a memory connected to or built
in the processor to execute a process including: acquiring a medical image in which
a brain appears; extracting a plurality of anatomical regions of the brain from the
medical image; inputting images of the plurality of anatomical regions to a plurality
of feature amount derivation models prepared for each of the plurality of anatomical
regions, and outputting a plurality of feature amounts for each of the plurality of
anatomical regions from the feature amount derivation models; inputting the plurality
of feature amounts which are output for each of the plurality of anatomical regions
to a dementia opinion derivation model, and outputting a dementia opinion from the
dementia opinion derivation model; deriving a first contribution which represents
a degree of contribution to output of the opinion for each of the anatomical regions;
and presenting the opinion and a derivation result of the first contribution for each
of the anatomical regions.
[0019] According to the technique of the present disclosure, it is possible to provide a
diagnosis support device, an operation method of a diagnosis support device, an operation
program of a diagnosis support device, and a dementia diagnosis support method capable
of obtaining a more accurate disease opinion and recognizing a degree of contribution
of each of anatomical regions of an organ to output of the opinion.
BRIEF DESCRIPTION OF THE DRAWINGS
[0020]
Fig. 1 is a diagram illustrating a medical system including a diagnosis support device.
Fig. 2 is a block diagram illustrating a computer including the diagnosis support
device.
Fig. 3 is a block diagram illustrating a processing unit of a CPU of the diagnosis
support device.
Fig. 4 is a diagram illustrating processing of a normalization unit.
Fig. 5 is a diagram illustrating processing of an extraction unit.
Fig. 6 is a diagram illustrating processing of a feature amount derivation unit.
Fig. 7 is a diagram illustrating processing of a dementia opinion derivation unit.
Fig. 8 is a diagram illustrating first contribution information.
Fig. 9 is a diagram illustrating a first display screen.
Fig. 10 is a diagram illustrating a second display screen.
Fig. 11 is a diagram illustrating a third display screen.
Fig. 12 is a diagram illustrating switching of display of the first contribution information.
Fig. 13 is a diagram illustrating a configuration of an auto-encoder and a structure
of a feature amount derivation model.
Fig. 14 is a diagram explaining convolution processing.
Fig. 15 is a diagram illustrating a configuration of operation data.
Fig. 16 is a diagram explaining pooling processing.
Fig. 17 is a diagram illustrating an outline of processing in a learning phase of
the auto-encoder.
Fig. 18 is a diagram illustrating an outline of processing in a learning phase of
a dementia opinion derivation model.
Fig. 19 is a flowchart illustrating a processing procedure of the diagnosis support
device.
Fig. 20 is a flowchart illustrating a processing procedure of the diagnosis support
device.
Fig. 21 is a diagram illustrating another example of dementia opinion information.
Fig. 22 is a diagram illustrating still another example of dementia opinion information.
Fig. 23 is a diagram illustrating processing of a dementia opinion derivation unit
according to a second embodiment.
Fig. 24 is a diagram illustrating an outline of processing in a learning phase of
a dementia opinion derivation model according to the second embodiment.
Fig. 25 is a diagram illustrating processing of a contribution derivation unit and
second contribution information according to the second embodiment.
Fig. 26 is a diagram illustrating a third display screen according to the second embodiment.
Fig. 27 is a diagram illustrating a configuration of a single-task convolutional neural
network for class determination and a structure of a feature amount derivation model.
Fig. 28 is a diagram illustrating an outline of processing in a learning phase of
a single-task convolutional neural network for class determination.
Fig. 29 is a diagram illustrating a configuration of a multi-task convolutional neural
network for class determination and a structure of a feature amount derivation model.
Fig. 30 is a diagram illustrating an outline of processing in a learning phase of
a multi-task convolutional neural network for class determination.
Fig. 31 is a diagram illustrating processing of a feature amount derivation unit according
to a fifth embodiment.
Fig. 32 is a diagram illustrating a configuration of an auto-encoder, a configuration
of a single-task convolutional neural network for class determination, and a structure
of a feature amount derivation model.
Fig. 33 is a diagram illustrating a detailed configuration of an output unit.
Fig. 34 is a diagram illustrating an outline of processing in a learning phase of
the auto-encoder and the single-task convolutional neural network for class determination.
Fig. 35 is a graph illustrating a change of a weight given to a loss of the auto encoder.
Fig. 36 is a diagram illustrating processing of a dementia opinion derivation unit
according to a sixth embodiment.
Fig. 37 is a table showing a performance comparison between a method of predicting
progress of dementia that is described in literatures in the related art and a method
of predicting progress of dementia according to the sixth embodiment.
DESCRIPTION OF THE PREFERRED EMBODIMENTS
[First Embodiment]
[0021] As illustrated in Fig. 1 as an example, a medical system 2 includes an MRI apparatus
10, a picture archiving and communication system (PACS) server 11, and a diagnosis
support device 12. The MRI apparatus 10, the PACS server 11, and the diagnosis support
device 12 are connected to a local area network (LAN) 13 provided in a medical facility,
and can communicate with each other via the LAN 13.
[0022] The MRI apparatus 10 images a head of a patient P and outputs a head MRI image 15.
The head MRI image 15 is voxel data representing a three-dimensional shape of the
head of the patient P. In Fig. 1, a head MRI image 15S having a sagittal cross section
is illustrated. The MRI apparatus 10 transmits the head MRI image 15 to the PACS server
11. The PACS server 11 stores and manages the head MRI image 15 from the MRI apparatus
10. The head MRI image 15 is an example of a "medical image" according to the technique
of the present disclosure.
[0023] The diagnosis support device 12 is, for example, a desktop personal computer, and
includes a display 17 and an input device 18. The input device 18 is a keyboard, a
mouse, a touch panel, a microphone, or the like. The doctor transmits a distribution
request of the head MRI image 15 of the patient P to the PACS server 11 by operating
the input device 18. The PACS server 11 searches for the head MRI image 15 of the
patient P that is requested to be distributed, and distributes the head MRI image
15 to the diagnosis support device 12. The diagnosis support device 12 displays the
head MRI image 15 distributed from the PACS server 11 on the display 17. The doctor
diagnoses dementia on the patient P by observing a brain of the patient P appearing
in the head MRI image 15. The brain is an example of an "organ" according to the technique
of the present disclosure, and the dementia is an example of a "disease" according
to the technique of the present disclosure. Further, in Fig. 1, only one MRI apparatus
10 and one diagnosis support device 12 are illustrated. On the other hand, a plurality
of MRI apparatuses 10 and a plurality of diagnosis support devices 12 may be provided.
[0024] As illustrated in Fig. 2 as an example, a computer including the diagnosis support
device 12 includes a storage 20, a memory 21, a central processing unit (CPU) 22,
and a communication unit 23, in addition to the display 17 and the input device 18.
The components are connected to each other via a bus line 24. The CPU 22 is an example
of a "processor" according to the technique of the present disclosure.
[0025] The storage 20 is a hard disk drive that is built in the computer including the diagnosis
support device 12 or is connected via a cable or a network. Alternatively, the storage
20 is a disk array in which a plurality of hard disk drives are connected in series.
The storage 20 stores a control program such as an operating system, various application
programs, and various data associated with the programs. A solid state drive may be
used instead of the hard disk drive.
[0026] The memory 21 is a work memory which is necessary to execute processing by the CPU
22. The CPU 22 loads the program stored in the storage 20 into the memory 21, and
executes processing according to the program. Thereby, the CPU 22 collectively controls
each unit of the computer. The communication unit 23 controls transmission of various
types of information to an external device such as the PACS server 11. The memory
21 may be built in the CPU 22.
[0027] As illustrated in Fig. 3 as an example, an operation program 30 is stored in the
storage 20 of the diagnosis support device 12. The operation program 30 is an application
program for causing the computer to function as the diagnosis support device 12. That
is, the operation program 30 is an example of "the operation program of the diagnosis
support device" according to the technique of the present disclosure. The storage
20 also stores the head MRI image 15, a reference head MRI image 35, a segmentation
model 36, a feature amount derivation model group 38 including a plurality of feature
amount derivation models 37, and a dementia opinion derivation model 39.
[0028] In a case where the operation program 30 is started, the CPU 22 of the computer including
the diagnosis support device 12 functions as a read/write (hereinafter, abbreviated
as RW) control unit 45, a normalization unit 46, an extraction unit 47, a feature
amount derivation unit 48, a dementia opinion derivation unit 49, a contribution derivation
unit 50, and a display control unit 51, in cooperation with the memory 21 and the
like.
[0029] The RW control unit 45 controls storing of various types of data in the storage 20
and reading of various types of data in the storage 20. For example, the RW control
unit 45 receives the head MRI image 15 from the PACS server 11, and stores the received
head MRI image 15 in the storage 20. In Fig. 3, only one head MRI image 15 is stored
in the storage 20. On the other hand, a plurality of head MRI images 15 may be stored
in the storage 20.
[0030] The RW control unit 45 reads the head MRI image 15 of the patient P designated by
the doctor for diagnosing dementia from the storage 20, and outputs the read head
MRI image 15 to the normalization unit 46 and the display control unit 51. The RW
control unit 45 acquires the head MRI image 15 by reading the head MRI image 15 from
the storage 20.
[0031] In addition, the RW control unit 45 reads the reference head MRI image 35 from the
storage 20, and outputs the read reference head MRI image 35 to the normalization
unit 46. The RW control unit 45 reads the segmentation model 36 from the storage 20,
and outputs the read segmentation model 36 to the extraction unit 47. The RW control
unit 45 reads the feature amount derivation model group 38 from the storage 20, and
outputs the read feature amount derivation model group 38 to the feature amount derivation
unit 48. Further, the RW control unit 45 reads the dementia opinion derivation model
39 from the storage 20, and outputs the read dementia opinion derivation model 39
to the dementia opinion derivation unit 49.
[0032] The normalization unit 46 performs normalization processing of matching the head
MRI image 15 with the reference head MRI image 35, and sets the head MRI image 15
as a normalized head MRI image 55. The normalization unit 46 outputs the normalized
head MRI image 55 to the extraction unit 47.
[0033] The reference head MRI image 35 is a head MRI image in which a brain having a reference
shape, a reference size, and a reference shade (pixel value) appears. The reference
head MRI image 35 is, for example, an image generated by averaging head MRI images
15 of a plurality of healthy persons, or an image generated by computer graphics.
The reference head MRI image 35 is an example of a "reference medical image" according
to the technique of the present disclosure.
[0034] The extraction unit 47 inputs the normalized head MRI image 55 to the segmentation
model 36. The segmentation model 36 is a machine learning model that performs so-called
semantic segmentation of assigning a label representing each of anatomical regions
of a brain, such as a hippocampus, an amygdala, and a frontal lobe, to each pixel
of the brain appearing in the normalized head MRI image 55. The extraction unit 47
extracts images 56 of a plurality of anatomical regions of the brain (hereinafter,
referred to as anatomical region images) from the normalized head MRI image 55 based
on the labels assigned by the segmentation model 36. The extraction unit 47 outputs
an anatomical region image group 57 including the plurality of anatomical region images
56 for each of the plurality of anatomical regions to the feature amount derivation
unit 48, the contribution derivation unit 50, and the display control unit 51.
[0035] One feature amount derivation model 37 is prepared for each of the plurality of anatomical
regions of the brain (refer to Fig. 6). The feature amount derivation unit 48 inputs
the anatomical region images 56 to the corresponding feature amount derivation models
37. In addition, a feature amount set 58 including a plurality of types of feature
amounts Z (refer to Fig. 6) is output from the feature amount derivation model 37.
The feature amount derivation unit 48 outputs a feature amount set group 59 including
a plurality of feature amount sets 58 corresponding to the plurality of anatomical
regions, to the dementia opinion derivation unit 49 and the contribution derivation
unit 50.
[0036] The dementia opinion derivation unit 49 inputs the feature amount set group 59 to
the dementia opinion derivation model 39. In addition, dementia opinion information
60 representing a dementia opinion is output from the dementia opinion derivation
model 39. The dementia opinion derivation unit 49 outputs the dementia opinion information
60 to the contribution derivation unit 50 and the display control unit 51. The dementia
opinion derivation model 39 is an example of a "disease opinion derivation model"
according to the technique of the present disclosure.
[0037] The contribution derivation unit 50 derives a first contribution, which represents
a degree of contribution to output of the dementia opinion information 60, for each
of the anatomical regions, based on the anatomical region image group 57 from the
extraction unit 47, the feature amount set group 59 from the feature amount derivation
unit 48, and the dementia opinion information 60 from the dementia opinion derivation
unit 49. For the derivation of the first contribution by the contribution derivation
unit 50, for example, a method by gradient-weighted class activation mapping (Grad-CAM)++
described in the following literature is used.
<
Daniel Omeiza, etc., Smooth Grad-CAM++: An Enhanced Inference Level Visualization
Technique for Deep Convolutional Neural Network Models, arXiv: 1908.01224, Aug 2019.>
[0038] The contribution derivation unit 50 generates first contribution information 61 from
the derived first contribution. The first contribution information 61 is an example
of a "derivation result of the first contribution" according to the technique of the
present disclosure. The contribution derivation unit 50 outputs the first contribution
information 61 to the display control unit 51.
[0039] The display control unit 51 controls a display of various screens on the display
17. The various screens include a first display screen 75 (refer to Fig. 9) for instructing
analysis by the segmentation model 36, the feature amount derivation model 37, and
the dementia opinion derivation model 39, a second display screen 80 (refer to Fig.
10) for displaying the dementia opinion information 60, a third display screen 85
(refer to Fig. 11) for displaying the first contribution information 61, and the like.
[0040] As illustrated in Fig. 4 as an example, the normalization unit 46 performs, as normalization
processing, shape normalization processing 65 and shade normalization processing 66
on the head MRI image 15. The shape normalization processing 65 is processing of extracting,
for example, landmarks serving as references for registration from the head MRI image
15 and the reference head MRI image 35, and performing parallel displacement, rotation,
and/or enlargement/reduction of the head MRI image 15 in accordance with the reference
head MRI image 35 such that a correlation between the landmark of the head MRI image
15 and the landmark of the reference head MRI image 35 is maximized. The shade normalization
processing 66 is, for example, processing of correcting a shade histogram of the head
MRI image 15 in accordance with a shade histogram of the reference head MRI image
35.
[0041] As illustrated in Fig. 5 as an example, the extraction unit 47 extracts, as the anatomical
region images 56, the anatomical region image 56_1 of a hippocampus, the anatomical
region image 56_2 of a parahippocampal gyrus, the anatomical region image 56_3 of
a frontal lobe, the anatomical region image 56_4 of a temporal lobe, the anatomical
region image 56_5 of an occipital lobe, the anatomical region image 56_6 of a thalamus,
the anatomical region image 56_7 of a hypothalamus, the anatomical region image 56_8
of an amygdala, the anatomical region image 56_9 of a pituitary gland, and the like.
In addition to these images, the extraction unit 47 extracts the anatomical region
images 56 of the anatomical regions such as mammillary bodies, corpora callosa, fornices,
and lateral ventricles. The anatomical regions such as a hippocampus, a frontal lobe,
a temporal lobe, and an amygdala come in pairs of a left anatomical region and a right
anatomical region. Although not illustrated in the drawings, the anatomical region
image 56 of each of the left and right anatomical regions is extracted from the pairs
of the left and right anatomical regions. For example, for the hippocampus, the anatomical
region image 56_1 of a left hippocampus and the anatomical region image 56_1 of a
right hippocampus are extracted. Preferably, the anatomical region includes at least
one of a hippocampus or a temporal lobe. More preferably, the anatomical region includes
all of a hippocampus and a temporal lobe. The temporal lobe means a front portion
of a temporal lobe. For the extraction of the anatomical regions by the extraction
unit 47 using the segmentation model 36, for example, a method described in the following
literature is used.
<
Patrick McClure, etc., Knowing What You Know in Brain Segmentation Using Bayesian
Deep Neural Networks, Front. Neuroinform., 17 October 2019.>
[0042] As illustrated in Fig. 6 as an example, the feature amount derivation unit 48 inputs
the anatomical region image 56_1 of the hippocampus to the feature amount derivation
model 37_1 of the hippocampus, and outputs the feature amount set 58_1 of the hippocampus
from the feature amount derivation model 37_1 of the hippocampus. The feature amount
set 58_1 of the hippocampus includes a plurality of feature amounts Z1_1, Z2_1, ···,
ZN_1. N is the number of feature amounts, and is, for example, several tens to hundreds
of thousands.
[0043] Similarly, the feature amount derivation unit 48 inputs the anatomical region image
56_2 of the parahippocampal gyrus to the feature amount derivation model 37_2 of the
parahippocampal gyrus, inputs the anatomical region image 56_3 of the frontal lobe
to the feature amount derivation model 37_3 of the frontal lobe, and inputs the anatomical
region image 56_4 of the temporal lobe to the feature amount derivation model 37_4
of the temporal lobe. In addition, the feature amount set 58_2 of the parahippocampal
gyrus is output from the feature amount derivation model 37_2 of the parahippocampal
gyrus, the feature amount set 58_3 of the frontal lobe is output from the feature
amount derivation model 37_3 of the frontal lobe, and the feature amount set 58_4
of the temporal lobe is output from the feature amount derivation model 37_4 of the
temporal lobe. The feature amount set 58_2 of the parahippocampal gyrus includes a
plurality of feature amounts Z1_2, Z2_2, ···, ZN_2, the feature amount set 58_3 of
the frontal lobe includes a plurality of feature amounts Z1_3, Z2_3, ···, ZN_3, and
the feature amount set 58_4 of the temporal lobe includes a plurality of feature amounts
Z1_4, Z2_4, ···, ZN_4.
[0044] Further, the feature amount derivation unit 48 inputs the anatomical region image
56_5 of the occipital lobe to the feature amount derivation model 37_5 of the occipital
lobe, and inputs the anatomical region image 56_6 of the thalamus to the feature amount
derivation model 37_6 of the thalamus. In addition, the feature amount set 58_5 of
the occipital lobe is output from the feature amount derivation model 37_5 of the
occipital lobe, and the feature amount set 58_6 of the thalamus is output from the
feature amount derivation model 37_6 of the thalamus. The feature amount set 58_5
of the occipital lobe includes a plurality of feature amounts Z1_5, Z2_5, ···, ZN_5,
and the feature amount set 58_6 of the thalamus includes a plurality of feature amounts
Z1_6, Z2_6, ···, ZN_6. In this way, the plurality of anatomical region images 56 are
respectively input to the corresponding feature amount derivation models 37. Thereby,
the plurality of feature amount sets 58 for each of the anatomical region images 56
are output from the feature amount derivation models 37. The number of the feature
amounts Z may be the same in each anatomical region as in a case of N in the example,
or may be different in each anatomical region.
[0045] As illustrated in Fig. 7 as an example, the dementia opinion derivation unit 49 inputs
the feature amount set group 59 to the dementia opinion derivation model 39. As the
dementia opinion information 60, any one of normal control (NC), mild cognitive impairment
(MCI), and Alzheimer's disease (AD) is output from the dementia opinion derivation
model 39.
[0046] As illustrated in Fig. 8 as an example, the first contribution information 61 includes
a first contribution map group 71, which includes a plurality of first contribution
maps 70 for each of the plurality of anatomical regions, and a ranking table 72. The
first contribution maps 70 includes a first contribution map 70_1 of the hippocampus,
a first contribution map 70_3 of the frontal lobe, a first contribution map 70_5 of
the occipital lobe, and the like. The first contribution map 70 is, so to speak, a
heat map of the first contribution, in which colors are assigned to each pixel of
the anatomical region image 56 according to a level of the first contribution. Specifically,
warm colors such as red and orange are assigned to the pixels having a relatively
high first contribution, and cold colors such as purple and blue are assigned to the
pixels having a relatively low first contribution.
[0047] The ranking table 72 is a table in which anatomical regions are arranged in descending
order of the first contribution. Fig. 8 illustrates a case where the hippocampus corresponds
to a first ranking, the parahippocampal gyrus corresponds to a second ranking, the
frontal lobe corresponds to a third ranking, the temporal lobe corresponds to a fourth
ranking, and so on.
[0048] Fig. 9 illustrates an example of the first display screen 75 for instructing the
analysis by the segmentation model 36, the feature amount derivation model 37, and
the dementia opinion derivation model 39. The head MRI images 15 of the patient P
for diagnosing dementia are displayed on the first display screen 75. The head MRI
images 15 include a head MRI image 15S having a sagittal cross section, a head MRI
image 15A having an axial cross section, and a head MRI image 15C having a coronal
cross section. A button group 76 for switching the display is provided in a lower
portion of each of the head MRI images 15S, 15A, and 15C.
[0049] An analysis button 77 is provided on the first display screen 75. The doctor selects
the analysis button 77 in a case where he/she wants to perform analysis using the
segmentation model 36, the feature amount derivation model 37, and the dementia opinion
derivation model 39. In response to the selection, the CPU 22 receives an instruction
for analysis by the segmentation model 36, the feature amount derivation model 37,
and the dementia opinion derivation model 39.
[0050] Fig. 10 illustrates an example of a second display screen 80 for displaying dementia
opinion information 60 obtained as a result of analysis by the segmentation model
36, the feature amount derivation model 37, and the dementia opinion derivation model
39. On the second display screen 80, a message 81 according to the dementia opinion
information 60 is displayed. Fig. 10 illustrates an example in which the dementia
opinion information 60 is mild cognitive impairment (MCI) and "suspected as mild cognitive
impairment" is displayed as the message 81.
[0051] A contribution derivation button 82 is provided on the second display screen 80.
The doctor selects the contribution derivation button 82 in a case where he/she wants
to know a degree of contribution of each anatomical region to the output of the dementia
opinion information 60. In response to the selection, the CPU 22 receives an instruction
for deriving the first contribution information 61. In a case where a confirmation
button 83 is selected, the display control unit 51 turns off the display of the message
81, and returns the second display screen 80 to the first display screen 75.
[0052] Fig. 11 illustrates an example of a third display screen 85 for displaying the first
contribution information 61 output by the contribution derivation unit 50. On the
third display screen 85, in addition to the head MRI image 15 and the message 81 according
to the dementia opinion information 60, a message 86 indicating the ranking of the
first contribution and the anatomical region corresponding to the ranking, the anatomical
region image 56, and the first contribution map 70 are displayed. The anatomical region
image 56 and the first contribution map 70 are displayed side by side. A button group
87 for switching the display is provided in lower portions of the anatomical region
image 56 and the first contribution map 70. In a case where a confirmation button
88 is selected, the display control unit 51 turns off the display of the message 86,
the anatomical region image 56, and the first contribution map 70, and returns the
third display screen 85 to the second display screen 80.
[0053] As illustrated in Fig. 12 as an example, in a case where the button group 87 is operated,
the display control unit 51 switches the display of the message 86, the anatomical
region image 56, and the first contribution map 70. The display control unit 51 switches
the display of the message 86, the anatomical region image 56, and the first contribution
map 70 in descending order of the first contribution. Fig. 12 illustrates an example
in which the display is switched according to the ranking table 72 illustrated in
Fig. 8. Specifically, the display is sequentially switched in order of the display
of the anatomical region image 56_1 and the first contribution map 70_1 of the hippocampus
corresponding to the first ranking, the display of the anatomical region image 56_2
and the first contribution map 70_2 of the parahippocampal gyrus corresponding to
the second ranking, and the display of the anatomical region image 56_3 and the first
contribution map 70_3 of the frontal lobe corresponding to the third ranking.
[0054] As illustrated in Fig. 13 as an example, a compression unit 91 of an auto-encoder
(hereinafter, abbreviated as AE) 90 is used in the feature amount derivation model
37. The AE 90 includes a compression unit 91 and a restoration unit 92. The anatomical
region image 56 is input to the compression unit 91. The compression unit 91 converts
the anatomical region image 56 into the feature amount set 58. The compression unit
91 transmits the feature amount set 58 to the restoration unit 92. The restoration
unit 92 generates a restoration image 93 of the anatomical region image 56 from the
feature amount set 58.
[0055] The compression unit 91 converts the anatomical region image 56 into the feature
amount set 58 by performing a convolution operation as illustrated in Fig. 14 as an
example. Specifically, the compression unit 91 includes a convolutional layer 200
represented by "convolution (abbreviated as conv)". The convolutional layer 200 applies,
for example, a 3 × 3 filter 203 to the target data 202 including a plurality of elements
201 which are two-dimensionally arranged. In addition, the convolutional layer 200
performs convolution of an element value e of an element of interest 2011, which is
one of the elements 201, and element values a, b, c, d, f, g, h, and i of eight elements
201S adjacent to the element of interest 2011. The convolutional layer 200 sequentially
performs a convolution operation on each of the elements 201 of the target data 202
while shifting the element of interest 2011 by one element, and outputs element values
of elements 204 of operation data 205. Thereby, similarly to the target data 202,
the operation data 205 including a plurality of elements 204 which are two-dimensionally
arranged is obtained. The target data 202 that is first input to the convolutional
layer 200 is the anatomical region image 56, and thereafter, reduction operation data
205S (refer to Fig. 16) to be described later is input to the convolutional layer
200 as the target data 202.
[0056] In a case where it is assumed that coefficients of the filter 203 are r, s, t, u,
v, w, x, y, and z, an element value k of an element 204I of the operation data 205
corresponding to the element of interest 2011 is obtained, for example, by calculating
the following equation (1), the element value k being a result of the convolution
operation on the element of interest 2011.

[0057] One piece of the operation data 205 is output for one filter 203. In a case where
a plurality of types of filters 203 are applied to one piece of the target data 202,
the operation data 205 is output for each of the filters 203. That is, as illustrated
in Fig. 15 as an example, pieces of the operation data 205 are generated for the number
of filters 203 applied to the target data 202. In addition, the operation data 205
includes the plurality of elements 204 which are two-dimensionally arranged, and thus
the operation data 205 has a width and a height. The number of pieces of the operation
data 205 is called the number of channels. Fig. 15 illustrates four channels of pieces
of the operation data 205 that are output by applying the four filters 203 to the
target data 202.
[0058] As illustrated in Fig. 16 as an example, the compression unit 91 includes a pooling
layer 210 represented by "pooling (abbreviated as pool)" in addition to the convolutional
layer 200. The pooling layer 210 obtains local statistics of the element values of
the elements 204 of the operation data 205, and generates reduction operation data
205S in which the obtained statistics are used as element values. Here, the pooling
layer 210 performs maximum value pooling processing of obtaining, as the local statistic,
a maximum value of the element values in a 2 × 2 element block 211. By performing
the processing while shifting the block 211 by one element in a width direction and
a height direction, a size of the reduction operation data 205S is reduced to 1/2
of a size of the original operation data 205. Fig. 16 illustrates a case where the
element value b among the element values a, b, e, and f in the block 211A is a maximum
value, the element value b among the element values b, c, f, and g in the block 211B
is a maximum value, and the element value h among the element values c, d, g, and
h in the block 211C is a maximum value. Average value pooling processing of obtaining,
as a local statistic, an average value instead of the maximum value may be performed.
[0059] The compression unit 91 outputs final operation data 205 by repeating the convolution
processing by the convolutional layer 200 and the pooling processing by the pooling
layer 210 a plurality of times. The final operation data 205 is, in other words, the
feature amount set 58, and the element value of each element 204 of the final operation
data 205 is, in other words, the feature amount Z. The feature amount Z obtained in
this way represents a shape of the anatomical region and a feature of a texture, such
as a degree of atrophy of the hippocampus, a degree of a vascular disorder of a white
matter, and the presence or absence of a decrease in blood flow metabolism in the
frontal lobe, the temporal lobe, and the occipital lobe. Here, for the sake of simplicity,
the description is given that the processing is performed in a two-dimensional manner.
On the other hand, the processing is actually performed in a three-dimensional manner.
[0060] As illustrated in Fig. 17 as an example, the AE 90 is trained by inputting learning
anatomical region images 56L in a learning phase before the compression unit 91 is
adapted as the feature amount derivation model 37. The AE 90 outputs learning restoration
images 93L in response to the learning anatomical region images 56L. Loss calculation
of the AE 90 using a loss function is performed based on the learning anatomical region
images 56L and the learning restoration images 93L. In addition, update settings of
various coefficients of the AE 90 (such as coefficients of the filters 203) are performed
according to a result of the loss calculation, and the AE 90 is updated according
to the update settings.
[0061] In the learning phase of the AE 90, while exchanging the learning anatomical region
images 56L, a series of processing including inputting of the learning anatomical
region images 56L to the AE 90, outputting of the learning restoration images 93L
from the AE 90, the loss calculation, the update settings, and updating of the AE
90 is repeatedly performed. The repetition of the series of processing is ended in
a case where accuracy of restoration from the learning anatomical region images 56L
to the learning restoration images 93L reaches a predetermined setting level. The
compression unit 91 of the AE 90 of which the restoration accuracy reaches the setting
level in this manner is used as the trained feature amount derivation model 37 by
being stored in the storage 20.
[0062] In Fig. 18 illustrating an example of an outline of processing in the learning phase
of the dementia opinion derivation model 39, the dementia opinion derivation model
39 is configured by using any one method of a neural network, a support vector machine,
and boosting. In the learning phase, the dementia opinion derivation model 39 is trained
by inputting learning data 100. The learning data 100 is a set of a learning feature
amount set group 59L and correct dementia opinion information 60CA corresponding to
the learning feature amount set group 59L. The learning feature amount set group 59L
is obtained by inputting the anatomical region image 56 of a certain head MRI image
15 to the feature amount derivation model 37. The correct dementia opinion information
60CA is a result obtained by actually diagnosing, by the doctor, the dementia opinion
on the head MRI image 15 from which the learning feature amount set group 59L is obtained.
[0063] In the learning phase, the learning feature amount set group 59L is input to the
dementia opinion derivation model 39. The dementia opinion derivation model 39 outputs
learning dementia opinion information 60L in response to the learning feature amount
set group 59L. A loss calculation of the dementia opinion derivation model 39 using
a loss function is performed based on the learning dementia opinion information 60L
and the correct dementia opinion information 60CA. In addition, update settings of
various coefficients of the dementia opinion derivation model 39 are performed according
to a result of the loss calculation, and the dementia opinion derivation model 39
is updated according to the update settings.
[0064] In the learning phase of the dementia opinion derivation model 39, while exchanging
the learning data 100, a series of processing including inputting of the learning
feature amount set group 59L to the dementia opinion derivation model 39, outputting
of the learning dementia opinion information 60L from the dementia opinion derivation
model 39, the loss calculation, the update settings, and updating of the dementia
opinion derivation model 39 is repeatedly performed. The repetition of the series
of processing is ended in a case where prediction accuracy of the learning dementia
opinion information 60L with respect to the correct dementia opinion information 60CA
reaches a predetermined setting level. The dementia opinion derivation model 39 of
which the prediction accuracy reaches the setting level in this way is stored in the
storage 20, and is used as a trained dementia opinion derivation model in the dementia
opinion derivation unit 49.
[0065] Next, an operation according to the configuration will be described with reference
to flowcharts illustrated in Fig. 19 and Fig. 20. First, in a case where the operation
program 30 is started in the diagnosis support device 12, as illustrated in Fig. 3,
the CPU 22 of the diagnosis support device 12 functions as the RW control unit 45,
the normalization unit 46, the extraction unit 47, the feature amount derivation unit
48, the dementia opinion derivation unit 49, the contribution derivation unit 50,
and the display control unit 51.
[0066] As illustrated in Fig. 19 as an example, in a case where the analysis button 77 is
selected on the first display screen 75 illustrated in Fig. 9, the RW control unit
45 reads the corresponding head MRI image 15 and the reference head MRI image 35 from
the storage 20 (step ST100). The head MRI image 15 and the reference head MRI image
35 are output from the RW control unit 45 to the normalization unit 46.
[0067] As illustrated in Fig. 4, the normalization unit 46 performs normalization processing
(shape normalization processing 65 and shade normalization processing 66) of matching
the head MRI image 15 with the reference head MRI image 35 (step ST110). Thereby,
the head MRI image 15 is set as a normalized head MRI image 55. The normalized head
MRI image 55 is output from the normalization unit 46 to the extraction unit 47.
[0068] As illustrated in Fig. 5, the extraction unit 47 extracts a plurality of anatomical
region images 56 from the normalized head MRI image 55 using the segmentation model
36 (step ST120). The anatomical region image group 57 including the plurality of anatomical
region images 56 is output from the extraction unit 47 to the feature amount derivation
unit 48, the contribution derivation unit 50, and the display control unit 51.
[0069] As illustrated in Fig. 6, the feature amount derivation unit 48 inputs the anatomical
region images 56 to the corresponding feature amount derivation models 37. Thereby,
the feature amount set 58 is output from the feature amount derivation model 37 (step
ST130). The feature amount set group 59 including the plurality of feature amount
sets 58 is output from the feature amount derivation unit 48 to the dementia opinion
derivation unit 49 and the contribution derivation unit 50.
[0070] As illustrated in Fig. 7, the dementia opinion derivation unit 49 inputs the feature
amount set group 59 to the dementia opinion derivation model 39. Thereby, the dementia
opinion information 60 is output from the dementia opinion derivation model 39 (step
ST140). The dementia opinion information 60 is output from the dementia opinion derivation
unit 49 to the contribution derivation unit 50 and the display control unit 51.
[0071] Under a control of the display control unit 51, the second display screen 80 illustrated
in Fig. 10 is displayed on the display 17 (step ST150). The doctor confirms the dementia
opinion information 60 via the message 81 on the second display screen 80.
[0072] As illustrated in Fig. 20 as an example, in a case where the contribution derivation
button 82 is selected on the second display screen 80 and an instruction for deriving
the first contribution information 61 is received by the CPU 22 (YES in step ST200),
the contribution derivation unit 50 derives a first contribution for each of the anatomical
regions based on the anatomical region image group 57, the feature amount set group
59, and the dementia opinion information 60, and generates the first contribution
information 61 illustrated in Fig. 8 from the first contribution (step ST210). The
first contribution information 61 is output from the contribution derivation unit
50 to the display control unit 51.
[0073] Under a control of the display control unit 51, the third display screen 85 illustrated
in Fig. 11 is displayed on the display 17 (step ST220). The doctor confirms the first
contribution information 61 via the third display screen 85.
[0074] As described above, the CPU 22 of the diagnosis support device 12 includes the RW
control unit 45, the extraction unit 47, the feature amount derivation unit 48, the
dementia opinion derivation unit 49, the contribution derivation unit 50, and the
display control unit 51. The RW control unit 45 acquires the head MRI image 15 by
reading the head MRI image 15 of the patient P for diagnosing dementia from the storage
20. The extraction unit 47 extracts the anatomical region images 56 of the plurality
of anatomical regions of the brain from the normalized head MRI image 55. The feature
amount derivation unit 48 inputs the plurality of anatomical region images 56 to the
plurality of feature amount derivation models 37 prepared for each of the plurality
of anatomical regions, and outputs the plurality of feature amount sets 58 for each
of the plurality of anatomical regions from the feature amount derivation models 37.
The dementia opinion derivation unit 49 inputs the feature amount set group 59 including
the plurality of feature amount sets 58 to the dementia opinion derivation model 39,
and outputs the dementia opinion information 60 from the dementia opinion derivation
model 39. The contribution derivation unit 50 derives the first contribution representing
a degree of contribution to output of the dementia opinion information 60 for each
of the anatomical regions. The display control unit 51 presents the dementia opinion
information 60 and the first contribution information 61 to the doctor on the third
display screen 85.
[0075] The number of feature amounts Z is very large, for example, several tens to hundreds
of thousands. For this reason, the feature amount Z does not represent a limited feature
of the anatomical region as in the Z value described in
JP6483890B, but represents a comprehensive feature of the anatomical region. In addition, the
feature amount Z is not a single value which is statistically obtained as in the Z
value described in
JP6483890B, but is obtained by inputting the anatomical region image 56 to the feature amount
derivation model 37. Therefore, according to the method of the present disclosure
for deriving the dementia opinion information 60 based on the feature amounts Z (the
feature amount set group 59 including the plurality of feature amount sets 58), it
is possible to improve the prediction accuracy of the dementia opinion as compared
with the method described in
JP6483890B. Thereby, it is possible to obtain a more accurate dementia opinion.
[0076] In dementia, as compared with other diseases such as cancer, specific lesions that
can be recognized with the naked eye are less likely to appear in the image. In addition,
dementia has an effect on the entire brain, and is not local. Because of this background,
in the related art, it is difficult to obtain an accurate dementia opinion from a
medical image such as a head MRI image 15 by using a machine learning model. On the
other hand, according to the technique of the present disclosure, the brain is subdivided
into the plurality of anatomical regions, feature amounts are derived for each of
the plurality of anatomical regions, and the derived feature amounts are input to
one dementia opinion derivation model 39. Therefore, it is possible to achieve the
object for obtaining a more accurate dementia opinion, as compared with the technique
in the related art in which it is difficult to obtain an accurate dementia opinion.
[0077] In addition, the doctor can recognize a degree of contribution of each of the anatomical
regions of the brain to the output of the dementia opinion information 60 via the
third display screen 85. Thereby, the doctor can determine a plausibility of the dementia
opinion information 60.
[0078] A history of the ranking of the first contribution of each of the anatomical regions
may be stored, and the history may be reflected in the learning of the dementia opinion
derivation model 39. Specifically, the feature amount set 58 of the anatomical region
having a relatively low ranking may be excluded from the learning feature amount set
group 59L to be input to the dementia opinion derivation model 39.
[0079] As illustrated in Fig. 12, the display control unit 51 presents the derivation results
in descending order of the first contribution. The doctor can recognize a degree of
contribution of each of the anatomical regions at a glance. A form for presenting
the derivation results in descending order of the first contribution is not limited
the form illustrated in Fig. 12 in which display is switched between display of the
message 86, display of the anatomical region image 56, and display of the first contribution
map 70 in response to an operation of the button group 87. The messages 86, the anatomical
region images 56, and the first contribution maps 70 for each ranking may be vertically
displayed side by side, and the message 86, the anatomical region image 56, and the
first contribution map 70 corresponding to a higher ranking may be displayed in an
upper portion.
[0080] As illustrated in Fig. 13, the feature amount derivation model 37 is obtained by
adapting the compression unit 91 of the AE 90. The AE 90 is one of neural network
models which are frequently used in the field of machine learning, and is generally
very well known. Therefore, the compression unit 91 of the AE 90 can be relatively
easily adapted as the feature amount derivation model 37.
[0081] As illustrated in Fig. 18, the dementia opinion derivation model 39 is configured
by any method of a neural network, a support vector machine, and boosting. Any method
of a neural network, a support vector machine, and boosting is generally very well
known. Therefore, the dementia opinion derivation model 39 can be relatively easily
configured.
[0082] As illustrated in Fig. 4, the normalization unit 46 performs normalization processing
of matching the head MRI image 15 with the reference head MRI image 35, prior to extraction
of the anatomical regions. Therefore, after an individual difference of the patient
P and an apparatus difference of the MRI apparatus 10 are substantially eliminated,
subsequent processing can be performed. Thereby, it is possible to improve reliability
of the dementia opinion information 60.
[0083] The dementia has become a social problem with the advent of an aging society in recent
years. Therefore, it can be said that the present embodiment of outputting the dementia
opinion information 60 in which a brain is set as an organ and dementia is set as
a disease is a form that matches the current social problem.
[0084] The hippocampus and the temporal lobe are anatomical regions that are particularly
highly correlated with dementia such as Alzheimer's disease. Therefore, as in the
present example, in a case where the plurality of anatomical regions include at least
one of the hippocampus or the temporal lobe, it is possible to obtain a more accurate
dementia opinion.
[0085] The presentation form of the dementia opinion information 60 and the first contribution
information 61 is not limited to the third display screen 85. The dementia opinion
information 60 and the first contribution information 61 may be printed out on a paper
medium, or the dementia opinion information 60 and the first contribution information
61 may be transmitted to a mobile terminal of the doctor as an attachment file of
an e-mail.
[0086] The dementia opinion information is not limited to the content illustrated in Fig.
7 (normal control/mild cognitive impairment/Alzheimer's disease). For example, as
in the dementia opinion information 105 illustrated in Fig. 21, the dementia opinion
information may indicate whether a degree of progression of dementia of the patient
P after one year is fast or slow. Alternatively, as in the dementia opinion information
108 illustrated in Fig. 22, the dementia opinion information may be a type of dementia,
such as Alzheimer's disease, dementia with Lewy body, or vascular dementia.
[Second Embodiment]
[0087] In the second embodiment illustrated in Fig. 23 to Fig. 26, dementia-related information
111 related to dementia is input to the dementia opinion derivation model 112 in addition
to the plurality of feature amounts Z.
[0088] As illustrated in Fig. 23 as an example, the dementia opinion derivation unit 110
according to the present embodiment inputs dementia-related information 111 related
to dementia to the dementia opinion derivation model 112 in addition to the feature
amount set group 59. In addition, dementia opinion information 113 is output from
the dementia opinion derivation model 112. The dementia-related information 111 is
an example of "disease-related information" according to the technique of the present
disclosure.
[0089] The dementia-related information 111 is information on the patient P for diagnosing
dementia. The dementia-related information 111 includes a plurality of items. The
items include, for example, a volume of the hippocampus. In addition, the items include,
for example, a score of revised Hasegawa's dementia scale (HDS-R), a score of mini-mental
state examination (MMSE), a score of a rivermead behavioural memory test (RBMT), clinical
dementia rating (CDR), activities of daily living (ADL), Alzheimer's disease assessment
scale-cognitive subscale (ADAS-Cog), and the like. In addition, the items include
a genotype of an ApoE gene, an amyloid-β measurement value, a tau protein measurement
value, an apolipoprotein measurement value, a complement protein measurement value,
and a transthyretin measurement value. Further, the items include a gender and an
age of the patient P and the like. The HDS-R score, the MMSE score, the RBMT score,
the CDR, the ADL, the ADAS-Cog, the genotype of the ApoE gene, the amyloid-β measurement
value, the tau protein measurement value, the apolipoprotein measurement value, the
complement protein measurement value, the transthyretin measurement value, the gender
and the age of the patient P, and the like are taken from an electronic chart system
that is not illustrated.
[0090] The volume of the hippocampus is, for example, the total number of pixels of the
anatomical region image 56_1 of the hippocampus. The volume of the hippocampus is
an example of a "volume of the anatomical region" according to the technique of the
present disclosure. In addition to or instead of the volume of the hippocampus, a
volume of another anatomical region such as the amygdala may be included in the dementia-related
information 111.
[0091] The HDS-R score, the MMSE score, the RBMT score, the CDR, the ADL, and the ADAS-Cog
are an example of a "dementia test score" according to the technique of the present
disclosure.
[0092] The genotype of the ApoE gene is a combination of two types among three types of
ApoE genes of ε2, ε3, and ε4 (ε2 and ε3, ε3 and ε4, and the like). A risk of development
of the Alzheimer's disease having a genotype including one or two ε4 (ε2 and ε4, ε4
and ε4, and the like) is approximately 3 times to 12 times a risk of development of
the Alzheimer's disease having a genotype without ε4 (ε2 and ε3, ε3 and ε3, and the
like). The genotype of the ApoE gene is converted into a numerical value. For example,
a combination of ε2 and ε3 is converted into 1, and a combination of ε3 and ε3 is
converted into 2. The numerical value is input to the dementia opinion derivation
model 142. The genotype of the ApoE gene is an example of a "test result of a genetic
test" according to the technique of the present disclosure.
[0093] The amyloid-β measurement value and the tau protein measurement value are an example
of a "test result of a spinal fluid test" according to the technique of the present
disclosure. In addition, the apolipoprotein measurement value, the complement protein
measurement value, and the transthyretin measurement value are an example of a "test
result of a blood test" according to the technique of the present disclosure.
[0094] Fig. 24 illustrates an example of an outline of processing in a learning phase of
the dementia opinion derivation model 112. The dementia opinion derivation model 112
is trained by inputting learning data 118. The learning data 118 is a combination
of the learning feature amount set group 59L, the learning dementia-related information
111L, and the correct dementia opinion information 113CA corresponding to the learning
feature amount set group 59L and the learning dementia-related information 111L. The
learning feature amount set group 59L is obtained by inputting the anatomical region
image 56 of a certain head MRI image 15 to the feature amount derivation model 37.
The learning dementia-related information 111L is information of the patient P whose
the head MRI image 15 is imaged, the head MRI image 15 being an image from which the
learning feature amount set group 59L is obtained. The correct dementia opinion information
113CA is a result obtained by actually diagnosing, by the doctor, the dementia opinion
on the head MRI image 15 from which the learning feature amount set group 59L is obtained
in consideration of the learning dementia-related information 111L.
[0095] In the learning phase, the learning feature amount set group 59L and the learning
dementia-related information 111L are input to the dementia opinion derivation model
112. The dementia opinion derivation model 112 outputs the learning dementia opinion
information 113L in response to the learning feature amount set group 59L and the
learning dementia-related information 111L. A loss calculation of the dementia opinion
derivation model 112 using a loss function is performed based on the learning dementia
opinion information 113L and the correct dementia opinion information 113CA. In addition,
update settings of various coefficients of the dementia opinion derivation model 112
are performed according to a result of the loss calculation, and the dementia opinion
derivation model 112 is updated according to the update settings.
[0096] In the learning phase of the dementia opinion derivation model 112, while exchanging
the learning data 118, a series of processing including inputting of the learning
feature amount set group 59L and the learning dementia-related information 111L to
the dementia opinion derivation model 112, outputting of the learning dementia opinion
information 113L from the dementia opinion derivation model 112, the loss calculation,
the update settings, and updating of the dementia opinion derivation model 112 is
repeatedly performed. The repetition of the series of processing is ended in a case
where prediction accuracy of the learning dementia opinion information 113L with respect
to the correct dementia opinion information 113CA reaches a predetermined setting
level. The dementia opinion derivation model 112 of which the prediction accuracy
reaches the setting level in this way is stored in the storage 20, and is used as
a trained dementia opinion derivation model in the dementia opinion derivation unit
110.
[0097] As illustrated in Fig. 25 as an example, the contribution derivation unit 120 according
to the present embodiment generates first contribution information 121 based on the
anatomical region image group 57, the feature amount set group 59, and the dementia
opinion information 113, similarly to the first contribution information 61 according
to the first embodiment, and outputs the generated first contribution information
121 to the display control unit 123. In addition, the contribution derivation unit
120 derives a second contribution representing a degree of contribution to the output
of the dementia opinion information 113 based on the dementia-related information
111 and the dementia opinion information 113 for each item of the dementia-related
information 111. The contribution derivation unit 120 derives, as a second contribution,
for example, a numerical value in 10 steps from 1 to 10. The contribution derivation
unit 120 outputs the second contribution information 122 in which the derived second
contribution is summarized to the display control unit 123. In the second contribution
information 122, the second contribution corresponding to each item of the dementia-related
information 111 is registered. The second contribution information 122 is an example
of a "derivation result of the second contribution" according to the technique of
the present disclosure. For the derivation of the second contribution by the contribution
derivation unit 120, for example, the method described in the following literature
is used.
<
Scott M. Lundberg, etc., Explainable machine-learning predictions for the prevention
of hypoxaemia during surgery, Nature Biomedical Engineering volume 2, pages 749-760
(2018)>
[0098] As illustrated in Fig. 26 as an example, on the third display screen 130 according
to the present embodiment, in addition to the head MRI image 15, the message 81 according
to the dementia opinion information 60, the message 86 indicating the ranking of the
first contribution and the anatomical region corresponding to the ranking, the anatomical
region image 56, and the first contribution map 70, a list 131 of items of the dementia-related
information 111 is displayed. Each item of the list 131 is displayed according to
a level of the second contribution, as illustrated by shade of hatching. The display
according to the level of the second contribution is, for example, display in which
an item having a higher second contribution is displayed in a darker color and an
item having a lower second contribution is displayed in a lighter color. Fig. 26 illustrates
a case where the second contribution for each of the volume of the hippocampus, the
HDS-R score, the genotype of the ApoE gene, and the like is relatively high and the
second contribution for each of the MMSE score, the CDR, the ADL, the ADAS-Cog, and
the like is relatively low. In a case where a confirmation button 132 is selected,
the display control unit 123 turns off the display of the message 86, the anatomical
region image 56, the first contribution map 70, and the list 131, and returns the
third display screen 130 to the second display screen 80.
[0099] As described above, in the second embodiment, the dementia-related information 111
is input to the dementia opinion derivation model 112. The dementia-related information
111 includes the volume of the hippocampus, the HDS-R score, the MMSE score, the CDR,
the ADL, the ADAS-Cog, the genotype of the ApoE gene, the amyloid-β measurement value,
the tau protein measurement value, the apolipoprotein measurement value, the complement
protein measurement value, the transthyretin measurement value, the gender and the
age of the patient P, and the like. Pieces of powerful information useful for prediction
such as various types of dementia-related information 111 related to dementia are
added. Thus, as compared with the case where the dementia opinions are predicted by
using only the feature amount set group 59, it is possible to dramatically improve
the prediction accuracy of the dementia opinion.
[0100] In addition, as illustrated in Fig. 25, the contribution derivation unit 120 derives
a second contribution representing a degree of contribution to the output of the dementia
opinion information 113 for each item of the dementia-related information 111. In
addition, as illustrated in Fig. 26, the display control unit 123 presents the derivation
result of the second contribution for each item on the third display screen 130. Therefore,
the doctor can recognize the degree of contribution of each item of the dementia-related
information 111 to the output of the dementia opinion information 60. Thereby, the
doctor can determine a plausibility of the dementia opinion information 60 with greater
confidence.
[0101] The dementia-related information 111 may include at least one of a volume of the
anatomical region, a score of a dementia test, a test result of a genetic test, a
test result of a spinal fluid test, or a test result of a blood test. The dementia-related
information 111 may include a medical history of the patient P, whether or not the
patient P has a relative who develops dementia, and the like. In addition, as in the
case of the first contribution, a history of the second contribution of each item
may be stored, and the history may be reflected in the learning of the dementia opinion
derivation model 112. Specifically, the item having a relatively low second contribution
may be excluded from the learning dementia-related information 111L to be input to
the dementia opinion derivation model 112.
[Third Embodiment]
[0102] In the third embodiment illustrated in Fig. 27 and Fig. 28, instead of the compression
unit 91 of the AE 90, a compression unit 141 of a single-task convolutional neural
network for class determination (hereinafter, abbreviated as a single-task CNN) 140
is used as a feature amount derivation model 145.
[0103] As illustrated in Fig. 27 as an example, the single-task CNN 140 includes a compression
unit 141 and an output unit 142. The anatomical region image 56 is input to the compression
unit 141. Similar to the compression unit 91, the compression unit 141 converts the
anatomical region image 56 into a feature amount set 143. The compression unit 141
transmits the feature amount set 143 to the output unit 142. The output unit 142 outputs
one class 144 based on the feature amount set 143. In Fig. 27, the output unit 142
outputs, as the class 144, a determination result indicating whether dementia is developed
or not developed. The compression unit 141 of the single-task CNN 140 is used as the
feature amount derivation model 145.
[0104] As illustrated in Fig. 28 as an example, the single-task CNN 140 is trained by inputting
learning data 148 in a learning phase before the compression unit 141 is adapted as
the feature amount derivation model 145. The learning data 148 is a set of the learning
anatomical region image 56L and a correct class 144CA corresponding to the learning
anatomical region image 56L. The correct class 144CA is a result obtained by actually
determining, by the doctor, whether or not dementia is developed on the head MRI image
15 from which the learning anatomical region image 56L is obtained.
[0105] In the learning phase, the learning anatomical region image 56L is input to the single-task
CNN 140. The single-task CNN 140 outputs a learning class 144L in response to the
learning anatomical region image 56L. The loss calculation of the single-task CNN
140 is performed based on the learning class 144L and the correct class 144CA. In
addition, update settings of various coefficients of the single-task CNN 140 are performed
according to a result of the loss calculation, and the single-task CNN 140 is updated
according to the update settings.
[0106] In the learning phase of the single-task CNN 140, while exchanging the learning data
148, a series of processing including inputting of the learning anatomical region
image 56L to the single-task CNN 140, outputting of the learning class 144L from the
single-task CNN 140, the loss calculation, the update settings, and updating of the
single-task CNN 140 is repeatedly performed. The repetition of the series of processing
is ended in a case where prediction accuracy of the learning class 144L with respect
to the correct class 144CA reaches a predetermined setting level. The compression
unit 141 of the single-task CNN 140 of which the prediction accuracy reaches the setting
level is stored in the storage 20 as the trained feature amount derivation model 145,
and is used in the feature amount derivation unit 48.
[0107] As described above, in the third embodiment, the compression unit 141 of the single-task
CNN 140 is used as the feature amount derivation model 145. The single-task CNN 140
is also one of neural network models which are frequently used in the field of machine
learning, and is generally very well known. Therefore, the compression unit 141 of
the single-task CNN 140 can be relatively easily adapted as the feature amount derivation
model 145.
[0108] The class 144 may include, for example, content indicating that the patient P is
younger than 75 years old or content indicating that the patient P is 75 years old
or older, or may include an age group of the patient P such as 60's and 70's.
[Fourth Embodiment]
[0109] In the fourth embodiment illustrated in Fig. 29 and Fig. 30, instead of the compression
unit 91 of the AE 90 and the compression unit 141 of the single-task CNN 140, a compression
unit 151 of a multi-task class determination CNN (hereinafter, abbreviated as a multi-task
CNN) 150 is used as a feature amount derivation model 156.
[0110] As illustrated in Fig. 29 as an example, the multi-task CNN 150 includes a compression
unit 151 and an output unit 152. The anatomical region image 56 is input to the compression
unit 151. The compression unit 151 converts the anatomical region image 56 into a
feature amount set 153 in the same manner as the compression unit 91 and the compression
unit 141. The compression unit 151 transmits the feature amount set 153 to the output
unit 152. The output unit 152 outputs two classes of a first class 154 and a second
class 155 based on the feature amount set 153. In Fig. 29, the output unit 152 outputs,
as the first class 154, a determination result indicating whether dementia is developed
or not developed. Further, in Fig. 29, the output unit 152 outputs, as the second
class 155, the age of the patient P. The compression unit 151 of the multi-task CNN
150 is used as a feature amount derivation model 156.
[0111] As illustrated in Fig. 30 as an example, the multi-task CNN 150 is trained by inputting
learning data 158 in a learning phase before the compression unit 151 is adapted as
the feature amount derivation model 156. The learning data 158 is a set of the learning
anatomical region image 56L and a correct first class 154CA and a correct second class
155CA corresponding to the learning anatomical region image 56L. The correct first
class 154CA is a result obtained by actually determining, by the doctor, whether or
not dementia is developed on the head MRI image 15 from which the learning anatomical
region image 56L is obtained. In addition, the correct second class 155CA is the actual
age of the patient P whose the head MRI image 15 is imaged, the head MRI image 15
being an image from which the learning anatomical region image 56L is obtained.
[0112] In the learning phase, the learning anatomical region image 56L is input to the multi-task
CNN 150. The multi-task CNN 150 outputs a learning first class 154L and a learning
second class 155L in response to the learning anatomical region image 56L. The loss
calculation of the multi-task CNN 150 is performed based on the learning first class
154L and the learning second class 155L, and the correct first class 154CA and the
correct second class 155CA. In addition, update settings of various coefficients of
the multi-task CNN 150 are performed according to a result of the loss calculation,
and the multi-task CNN 150 is updated according to the update settings.
[0113] In the learning phase of the multi-task CNN 150, while exchanging the learning data
158, a series of processing including inputting of the learning anatomical region
image 56L to the multi-task CNN 150, outputting of the learning first class 154L and
the learning second class 155L from the multi-task CNN 150, the loss calculation,
the update settings, and updating of the multi-task CNN 150 is repeatedly performed.
The repetition of the series of processing is ended in a case where prediction accuracy
of the learning first class 154L and the learning second class 155L with respect to
the correct first class 154CA and the correct second class 155CA reaches a predetermined
setting level. The compression unit 151 of the multi-task CNN 150 of which the prediction
accuracy reaches the setting level is stored in the storage 20 as the trained feature
amount derivation model 156, and is used in the feature amount derivation unit 48.
[0114] As described above, in the fourth embodiment, the compression unit 151 of the multi-task
CNN 150 is used as the feature amount derivation model 156. The multi-task CNN 150
performs more complicated processing of outputting a plurality of classes (the first
class 154 and the second class 155) as compared with the AE 90 and the single-task
CNN 140. For this reason, there is a high possibility that the feature amount set
153 output from the compression unit 151 more comprehensively represents a feature
of the anatomical region image 56. Therefore, as a result, it is possible to further
improve the prediction accuracy of the dementia opinion by the dementia opinion derivation
model 39.
[0115] The first class 154 may be, for example, a degree of progression of dementia in five
levels. In addition, the second class 155 may be a determination result of the age
group of the patient P. The multi-task CNN 150 may output three or more classes.
[Fifth Embodiment]
[0116] In the fifth embodiment illustrated in Fig. 31, the anatomical region image 56 of
one anatomical region is input to a plurality of different feature amount derivation
models.
[0117] In Fig. 31, the feature amount derivation unit 160 according to the present embodiment
inputs the anatomical region image 56 of one anatomical region to a first feature
amount derivation model 161, a second feature amount derivation model 162, and a third
feature amount derivation model 163. Thereby, the feature amount derivation unit 160
outputs a first feature amount set 164 from the first feature amount derivation model
161, outputs a second feature amount set 165 from the second feature amount derivation
model 162, and outputs a third feature amount set 166 from the third feature amount
derivation model 163. The first feature amount derivation model 161 is obtained by
adapting the compression unit 91 of the AE 90 according to the first embodiment. The
second feature amount derivation model 162 is obtained by adapting the compression
unit 141 of the single-task CNN 140 according to the third embodiment. The third feature
amount derivation model 163 is obtained by adapting the compression unit 151 of the
multi-task CNN 150 according to the fourth embodiment.
[0118] As described above, in the fifth embodiment, the feature amount derivation unit 160
inputs the anatomical region image 56 of one anatomical region to the first feature
amount derivation model 161, the second feature amount derivation model 162, and the
third feature amount derivation model 163. In addition, the first feature amount set
164, the second feature amount set 165, and the third feature amount set 166 are output
from each of the models 161 to 163. Therefore, a wide variety of feature amounts Z
can be obtained as compared with a case where one kind of feature amount derivation
model 37 is used. As a result, it is possible to further improve the prediction accuracy
of the dementia opinion by the dementia opinion derivation model 39.
[0119] The plurality of different feature amount derivation models may be, for example,
a combination of the first feature amount derivation model 161 obtained by adapting
the compression unit 91 of the AE 90 and the second feature amount derivation model
162 obtained by adapting the compression unit 141 of the single-task CNN 140. Alternatively,
a combination of the second feature amount derivation model 162 obtained by adapting
the compression unit 141 of the single-task CNN 140 and the third feature amount derivation
model 163 obtained by adapting the compression unit 151 of the multi-task CNN 150
may be used. Further, a combination of the second feature amount derivation model
162, which outputs whether or not dementia is developed as the class 144 and is obtained
by adapting the compression unit 141 of the single-task CNN 140, and the second feature
amount derivation model 162, which outputs the age group of the patient P as the class
144 and is obtained by adapting the compression unit 141 of the single-task CNN 140,
may be used.
[Sixth Embodiment]
[0120] In the sixth embodiment illustrated in Fig. 32 to Fig. 37, a model obtained by combining
the AE 250 and the single-task CNN 251 is used as a feature amount derivation model
252.
[0121] As illustrated in Fig. 32 as an example, the AE 250 includes a compression unit 253
and a restoration unit 254, similar to the AE 90 according to the first embodiment.
The anatomical region image 56 is input to the compression unit 253. The compression
unit 253 converts the anatomical region image 56 into the feature amount set 255.
The compression unit 253 transmits the feature amount set 255 to the restoration unit
254. The restoration unit 254 generates a restoration image 256 of the anatomical
region image 56 from the feature amount set 255.
[0122] The single-task CNN 251 includes a compression unit 253 and an output unit 257, similar
to the single-task CNN 140 according to the third embodiment. That is, the compression
unit 253 is shared by the AE 250 and the single-task CNN 251. The compression unit
253 transmits the feature amount set 255 to the output unit 257. The output unit 257
outputs one class 258 based on the feature amount set 255. In Fig. 32, the output
unit 257 outputs, as the class 258, a determination result indicating that the patient
P with mild cognitive impairment remains a state of mild cognitive impairment after
2 years or progresses to Alzheimer's disease after 2 years. In addition, the output
unit 257 outputs aggregated feature amounts ZA obtained by aggregating the plurality
of feature amounts Z included in the feature amount set 255. The aggregated feature
amounts ZA are output for each of the anatomical regions. In the present embodiment,
the aggregated feature amounts ZA are input to the dementia opinion derivation model
282 (refer to Fig. 36) instead of the feature amount set 255.
[0123] As illustrated in Fig. 33 as an example, the output unit 257 includes a self-attention
(hereinafter, abbreviated as SA) mechanism layer 265, a global average pooling (hereinafter,
abbreviated as GAP) layer 266, a fully connected (hereinafter, abbreviated as FC)
layer 267, a softmax function (hereinafter, abbreviated as SMF) layer 268, and a principal
component analysis (hereinafter, abbreviated as PCA) layer 269.
[0124] The SA mechanism layer 265 performs convolution processing illustrated in Fig. 14
on the feature amount set 255 while changing the coefficients of the filter 203 according
to the element value of the element of interest 2011. Hereinafter, the convolution
processing performed by the SA mechanism layer 265 is referred to as SA convolution
processing. The SA mechanism layer 265 outputs the feature amount set 255 after the
SA convolution processing to the GAP layer 266.
[0125] The GAP layer 266 performs global average pooling processing on the feature amount
set 255 after the SA convolution processing. The global average pooling processing
is processing of obtaining average values of the feature amounts Z for each channel
(refer to Fig. 15) of the feature amount set 255. For example, in a case where the
number of channels of the feature amount set 255 is 512, average values of 512 feature
amounts Z are obtained by the global average pooling processing. The GAP layer 266
outputs the obtained average values of the feature amounts Z to the FC layer 267 and
the PCA layer 269.
[0126] The FC layer 267 converts the average values of the feature amounts Z into variables
handled by the SMF of the SMF layer 268. The FC layer 267 includes an input layer
including units corresponding to the number of the average values of the feature amounts
Z (that is, the number of channels of the feature amount set 255) and an output layer
including units corresponding to the number of variables handled by the SMF. Each
unit of the input layer and each unit of the output layer are fully coupled to each
other, and weights are set for each unit. The average values of the feature amounts
Z are input to each unit of the input layer. The product sum of the average value
of the feature amounts Z and the weight which is set for each unit is an output value
of each unit of the output layer. The output value is a variable handled by the SMF.
The FC layer 267 outputs the variable handled by the SMF to the SMF layer 268. The
SMF layer 268 outputs the class 258 by applying the variable to the SMF.
[0127] The PCA layer 269 performs PCA on the average values of the feature amounts Z, and
aggregates the average values of the plurality of feature amounts Z into aggregated
feature amounts ZA of which the number is smaller than the number of the average values.
For example, the PCA layer 269 aggregates the average values of 512 feature amounts
Z into one aggregated feature amount ZA.
[0128] As illustrated in Fig. 34 as an example, the AE 250 is trained by inputting learning
anatomical region images 56L in a learning phase. The AE 250 outputs learning restoration
images 256L in response to the learning anatomical region images 56L. Loss calculation
of the AE 250 using a loss function is performed based on the learning anatomical
region images 56L and the learning restoration images 256L. In addition, update settings
of various coefficients of the AE 250 are performed according to a result of the loss
calculation (hereinafter, referred to as a loss L1), and the AE 250 is updated according
to the update settings.
[0129] In the learning phase of the AE 250, while exchanging the learning anatomical region
images 56L, a series of processing including inputting of the learning anatomical
region images 56L to the AE 250, outputting of the learning restoration images 256L
from the AE 250, the loss calculation, the update settings, and updating of the AE
250 is repeatedly performed.
[0130] The single-task CNN 251 is trained by inputting learning data 275 in a learning phase.
The learning data 275 is a set of the learning anatomical region image 56L and a correct
class 258CA corresponding to the learning anatomical region image 56L. The correct
class 258CA indicates that the patient P whose the head MRI image 15 is imaged and
who has mild cognitive impairment remains a state of mild cognitive impairment after
2 years or progresses to Alzheimer's disease after 2 years, the head MRI image 15
being an image from which the learning anatomical region image 56L is obtained.
[0131] In the learning phase, the learning anatomical region image 56L is input to the single-task
CNN 251. The single-task CNN 251 outputs a learning class 258L in response to the
learning anatomical region image 56L. The loss calculation of the single-task CNN
251 using a cross-entropy function or the like is performed based on the learning
class 258L and the correct class 258CA. In addition, update settings of various coefficients
of the single-task CNN 251 are performed according to a result of the loss calculation
(hereinafter, referred to as a loss L2), and the single-task CNN 251 is updated according
to the update settings.
[0132] In the learning phase of the single-task CNN 251, while exchanging the learning data
275, a series of processing including inputting of the learning anatomical region
image 56L to the single-task CNN 251, outputting of the learning class 258L from the
single-task CNN 251, the loss calculation, the update settings, and updating of the
single-task CNN 251 is repeatedly performed.
[0133] The update setting of the AE 250 and the update setting of the single-task CNN 251
are performed based on a total loss L represented by the following equation (2). α
is a weight.

[0134] That is, the total loss L is a weighted sum of the loss L1 of the AE 250 and the
loss L2 of the single-task CNN 251.
[0135] As illustrated in Fig. 35 as an example, the weight α is set to 1 in an initial stage
of the learning phase. Assuming that the weight α is 1, the total loss L is represented
by L = L1. Therefore, in this case, only the learning of the AE 250 is performed,
and the learning of the single-task CNN 251 is not performed.
[0136] The weight α is gradually decreased from 1 as the learning is progressed, and is
eventually set as a fixed value (0.8 in Fig. 35). In this case, the learning of the
AE 250 and the learning of the single-task CNN 251 are both performed with intensity
corresponding to the weight α. As described above, the weight given to the loss L1
is larger than the weight given to the loss L2. Further, the weight given to the loss
L1 is gradually decreased from a maximum value of 1, and the weight given to the loss
L2 is gradually increased from a minimum value of 0. Both the weight given to the
loss L1 and the weight given to the loss L2 are set as fixed values.
[0137] The learning of the AE 250 and the single-task CNN 251 is ended in a case where accuracy
of restoration from the learning anatomical region image 56L to the learning restoration
image 256L by the AE 250 reaches a predetermined setting level and where prediction
accuracy of the learning class 258L with respect to the correct class 258CA by the
single-task CNN 251 reaches a predetermined setting level. The AE 250 of which the
restoration accuracy reaches the setting level in this way and the single-task CNN
251 of which the prediction accuracy reaches the setting level in this way are stored
in the storage 20, and are used as the trained feature amount derivation model 252.
[0138] As illustrated in Fig. 36 as an example, the dementia opinion derivation unit 280
according to the present embodiment inputs an aggregated feature amount group ZAG
and the dementia-related information 281 to the dementia opinion derivation model
282. The aggregated feature amount group ZAG includes a plurality of aggregated feature
amounts ZA which are output for each of the anatomical regions. Similar to the dementia-related
information 111 according to the second embodiment, the dementia-related information
281 includes a gender and an age of the patient P for diagnosing dementia, a volume
of the anatomical region, a score of a dementia test, a test result of a genetic test,
a test result of a spinal fluid test, a test result of a blood test, and the like.
[0139] The dementia opinion derivation model 282 includes a quantile normalization unit
283 and a linear discriminant analysis unit 284. The aggregated feature amount group
ZAG and the dementia-related information 281 are input to the quantile normalization
unit 283. The quantile normalization unit 283 performs quantile normalization of converting
the plurality of aggregated feature amounts ZA included in the aggregated feature
amount group ZAG and each of parameters of the dementia-related information 281 into
data according to a normal distribution, in order to handle the plurality of aggregated
feature amounts ZA and the parameters in the same sequence. The linear discriminant
analysis unit 284 performs linear discriminant analysis on the aggregated feature
amounts ZA and each of the parameters of the dementia-related information 281 after
the quantile normalization processing, and outputs dementia opinion information 285
as a result of the linear discriminant analysis. The dementia opinion information
285 indicates that the patient P with mild cognitive impairment remains a state of
mild cognitive impairment after 2 years or progresses to Alzheimer's disease after
2 years. The learning of the dementia opinion derivation model 282 is the same as
the learning of the dementia opinion derivation model 112 illustrated in Fig. 24,
except that the learning feature amount set group 59L is changed to the learning aggregated
feature amount group ZAG Thus, illustration and description thereof will be omitted.
[0140] As described above, in the sixth embodiment, the single-task CNN 251 that performs
a main task such as outputting of the class 258 and the AE 250 that is partially common
to the single-task CNN 251 and performs a sub-task such as generation of the restoration
image 256 are used as the feature amount derivation model 252, the sub-task being
a task having a more general purpose as compared with the main task. In addition,
the AE 250 and the single-task CNN 251 are trained at the same time. Therefore, as
compared with a case where the AE 250 and the single-task CNN 251 are separate, the
feature amount set 255 that is more appropriate and the aggregated feature amounts
ZA that are more appropriate can be output. As a result, it is possible to improve
the prediction accuracy of the dementia opinion information 285.
[0141] In the learning phase, the update setting is performed based on the total loss L,
which is a weighted sum of the loss L1 of the AE 250 and the loss L2 of the single-task
CNN 251. Therefore, by setting the weight α to an appropriate value, the AE 250 can
be intensively trained, the single-task CNN 251 can be intensively trained, or the
AE 250 and the single-task CNN 251 can be trained in a well-balanced manner.
[0142] The weight given to the loss L1 is larger than the weight given to the loss L2. Therefore,
the AE 250 can always be intensively trained. In a case where the AE 250 is always
intensively trained, the feature amount set 255 that more represents the shape of
the anatomical region and the feature of the texture can be output from the compression
unit 253. As a result, the aggregated feature amounts ZA having a higher plausibility
can be output from the output unit 257.
[0143] Further, the weight given to the loss L1 is gradually decreased from the maximum
value, and the weight given to the loss L2 is gradually increased from the minimum
value. After the learning is performed a predetermined number of times, both the weight
given to the loss L1 and the weight given to the loss L2 are set as fixed values.
Thus, the AE 250 can be more intensively trained in an initial stage of the learning.
The AE 250 is responsible for a relatively simple sub-task such as generation of the
restoration image 256. Therefore, in a case where the AE 250 is more intensively trained
in the initial stage of the learning, the feature amount set 255 that more represents
the shape of the anatomical region and the feature of the texture can be output from
the compression unit 253 in the initial stage of the learning.
[0144] As an example, a table 300 illustrated in Fig. 37 shows performance comparison between
Nos. 1 to 7 and Nos. 8 and 9, Nos. 1 to 7 being described in the following literatures
A, B, C, D, E, F, and G and being related to a method of predicting progress of dementia,
and Nos. 8 and 9 being related to a method of predicting progress of dementia according
to the present embodiment. In the method of predicting progress of dementia according
to the present embodiment, No. 8 indicates a case where only the aggregated feature
amount group ZAG is input to the dementia opinion derivation model 282 and the dementia-related
information 281 is not input. On the other hand, No. 9 indicates a case where the
aggregated feature amount group ZAG and the dementia-related information 281 are input
to the dementia opinion derivation model 282.
[0145] Literature A <
Tam, A., Dansereau, C., Iturria-Medina, Y, Urchs, S., Orban, P., Sharmarke, H., Breitner,
J., & Alzheimer's Disease Neuroimaging Initiative., "A highly predictive signature
of cognition and brain atrophy for progression to Alzheimer's dementia.", GigaScience,
8 (5), giz055 (2019).>
[0146] Literature B <
Ledig, C., Schuh, A., Guerrero, R., Heckemann, R. A., & Rueckert, D., "Structural
brain imaging in Alzheimer's disease and mild cognitive impairment: biomarker analysis
and shared morphometry database.", Scientific reports, 8 (1), 11258 (2018).>
[0147] Literature C <
Lu, D., Popuri, K., Ding, G W., Balachandar, R., & Beg, M. F., "Multimodal and multiscale
deep neural networks for the early diagnosis of Alzheimer's disease using structural
MR and FDG-PET images", Scientific reports, 8 (1), 5697 (2018).>
[0148] Literature D <
Basaia, S., Agosta, F., Wagner, L., Canu, E., Magnani, G, Santangelo, R., Filippi,
M., Automated classification of Alzheimer's disease and mild cognitive impairment
using a single MRI and deep neural networks, NeuroImage: Clinical 21, 101645 (2019).>
[0149] Literature E <
Nakagawa, T., Ishida, M., Naito, J., Nagai, A., Yamaguchi, S., Onoda, K., "Prediction
of conversion to Alzheimer's disease using deep survival analysis of MRI images",
Brain Communications, Vol. 2 (1) (2020).>
[0150] Literature F <
Lee, G, Nho, K., Kang, B., Sohn, K. A., & Kim, D., "Predicting Alzheimer's disease
progression using multi-modal deep learning approach.", Scientific reports, 9 (1),
1952 (2019).>
[0151] Literature G <
Goto, T., Wang, C., Li, Y, Tsuboshita, Y, Multi-modal deep learning for predicting
progression of Alzheimer's disease using bi-linear shake fusion, Proc. SPIE 11314,
Medical Imaging (2020).>
[0152] The accuracy of No. 8 and the accuracy of No. 9 are 0.84 and 0.90. In particular,
the accuracy of No. 9 is 0.90 and is higher than the accuracy of any one of Nos. 1
to 7. An area under the curve (AUC) of No. 8 and an area under the curve (AUC) of
No. 9 are 0.93 and 0.97. These values are larger than a value in No. 5 that is related
to a method of predicting progress of dementia and is described in Literature E. Therefore,
it can be said that the method of predicting progress of dementia according to the
present embodiment can predict progress of dementia with higher accuracy as compared
with the methods of predicting progress of dementia in the related art that are described
in Literatures A to G
[0153] A sensitivity of No. 8 and a sensitivity of No. 9 are 0.85 and 0.91. These values
are higher than sensitivities in Nos. 1 to 7. In particular, the sensitivity of No.
9 is 0.91, and is a maximum value among the sensitivities. Therefore, it can be said
that the method of predicting progress of dementia according to the present embodiment
can predict that the patient P with mild cognitive impairment will progress to Alzheimer's
disease after a prediction period without overlooking the progress as compared with
the methods of predicting progress of dementia in the related art that are described
in Literatures A to G
[0154] A specificity of No. 8 and a specificity of No. 9 are 0.84 and 0.90. These values
are smaller than 0.97 in No. 1 related to the method of predicting progress of dementia
that is described in Literature A, but are larger than values in other Literatures
B, C, D, and F. Therefore, it can be said that the method of predicting progress of
dementia according to the present embodiment can more accurately predict that the
patient P with mild cognitive impairment remains a state of mild cognitive impairment
even after a prediction period as compared with many other methods of predicting progress
of dementia in the related art.
[0155] In the table 300, ADNI in the items of the learning image is an abbreviation of "Alzheimer's
disease Neuroimaging Initiative". AIBL is an abbreviation of "Australian Imaging Biomarkers
and Lifestyle Study of Ageing". J-ADNI is an abbreviation of "Japanese Alzheimer's
Disease Neuroimaging Intiative". The items indicate a database in which head MRI images
15 and the like of patients P with Alzheimer's disease are accumulated.
[0156] Instead of the single-task CNN 251, the multi-task CNN 150 according to the fourth
embodiment may be used.
[0157] The learning of the AE 90 illustrated in Fig. 17, the learning of the dementia opinion
derivation model 39 illustrated in Fig. 18, the learning of the dementia opinion derivation
model 112 illustrated in Fig. 24, the learning of the single-task CNN 140 illustrated
in Fig. 28, the learning of the multi-task CNN 150 illustrated in Fig. 30, the learning
of the AE 250 and the single-task CNN 251 illustrated in Fig. 34, and the like may
be performed by the diagnosis support device 12 or by a device other than the diagnosis
support device 12. In addition, the learning may be continuously performed after storing
each model in the storage 20 of the diagnosis support device 12.
[0158] The PACS server 11 may function as the diagnosis support device 12.
[0159] The medical image is not limited to the head MRI image 15 in the example. The medical
image may be a positron emission tomography (PET) image, a single photon emission
computed tomography (SPECT) image, a computed tomography (CT) image, an endoscopic
image, an ultrasound image, or the like.
[0160] The organ is not limited to the illustrated brain, and may be a heart, a lung, a
liver, or the like. In a case of a lung, right lungs S1 and S2 and left lungs S1 and
S2 are extracted as the anatomical regions. In a case of a liver, a right lobe, a
left lobe, a gall bladder, and the like are extracted as the anatomical regions. In
addition, the disease is not limited to dementia in the example, and may be a heart
disease, pneumonia, dyshepatia, or the like.
[0161] In each of the embodiments, for example, as a hardware structure of the processing
unit that executes various processing, such as the RW control unit 45, the normalization
unit 46, the extraction unit 47, the feature amount derivation units 48 and 160, the
dementia opinion derivation units 49, 110, and 280, the contribution derivation units
50 and 120, and the display control units 51 and 123, the following various processors
may be used. The various processors include, as described above, the CPU 22 which
is a general-purpose processor that functions as various processing units by executing
software (an operation program 30), a programmable logic device (PLD) such as a field
programmable gate array (FPGA) which is a processor capable of changing a circuit
configuration after manufacture, a dedicated electric circuit such as an application
specific integrated circuit (ASIC) which is a processor having a circuit configuration
specifically designed to execute specific processing, and the like.
[0162] One processing unit may be configured by one of these various processors, or may
be configured by a combination of two or more processors having the same type or different
types (for example, a combination of a plurality of FPGAs and/or a combination of
a CPU and an FPGA). Further, the plurality of processing units may be configured by
one processor.
[0163] As an example in which the plurality of processing units are configured by one processor,
firstly, as represented by a computer such as a client and a server, a form in which
one processor is configured by a combination of one or more CPUs and software and
the processor functions as the plurality of processing units may be adopted. Secondly,
as represented by system on chip (SoC), there is a form in which a processor that
realizes the functions of the entire system including a plurality of processing units
with one integrated circuit (IC) chip is used. As described above, the various processing
units are configured by using one or more various processors as a hardware structure.
[0164] Further, as the hardware structure of the various processors, more specifically,
an electric circuit (circuitry) in which circuit elements such as semiconductor elements
are combined may be used.
[0165] The technique of the present disclosure can also appropriately combine the various
embodiments and/or the various modification examples. In addition, the technique of
the present disclosure is not limited to each embodiment, and various configurations
may be adopted without departing from the scope of the present disclosure. Further,
the technique of the present disclosure extends to a program and a storage medium
for non-temporarily storing the program.
[0166] The described contents and the illustrated contents are detailed explanations of
a part according to the technique of the present disclosure, and are merely examples
of the technique of the present disclosure. For example, the descriptions related
to the configuration, the function, the operation, and the effect are descriptions
related to examples of a configuration, a function, an operation, and an effect of
a part according to the technique of the present disclosure. Therefore, it goes without
saying that, in the described contents and illustrated contents, unnecessary parts
may be deleted, new components may be added, or replacements may be made without departing
from the spirit of the technique of the present disclosure. Further, in order to avoid
complications and facilitate understanding of the part according to the technique
of the present disclosure, in the described contents and illustrated contents, descriptions
of technical knowledge and the like that do not require particular explanations to
enable implementation of the technique of the present disclosure are omitted.
[0167] In this specification, "A and/or B" is synonymous with "at least one of A or B".
That is, "A and/or B" means that only A may be included, that only B may be included,
or that a combination of A and B may be included. Further, in this specification,
even in a case where three or more matters are expressed by being connected using
"and/or", the same concept as "A and/or B" is applied.
[0168] All documents, patent applications, and technical standards mentioned in this specification
are incorporated herein by reference to the same extent as in a case where each document,
each patent application, and each technical standard are specifically and individually
described by being incorporated by reference.