[Technical Field]
[0001] The present invention relates to an information processing apparatus, an information
processing method and an information processing program.
[Background Art]
[0002] There are various services for providing satellite images shot by artificial satellites.
Patent Document 1, for example, discloses a satellite image selection support method,
etc. that generates, if a purchase order for a satellite image is accepted, a masked
image subjected to masking processing for masking a part that does not cover a purchase
desired area and transmits the masked satellite image to the terminal of an orderer.
[Prior Art Document]
[Patent Document]
[Summary of Invention]
[Problems to be Solved by Invention]
[0004] However, data of an image such as a satellite image observed from space has a huge
data size as compared with that of a general photography or the like, which makes
it difficult for the user who acquires such observation data to handle it.
[0005] In one aspect, an object is to provide an information processing apparatus, etc.
capable of readily acquiring easy-to-handle observation data.
[Means for Solving Problems]
[0006] In one aspect, an information processing apparatus comprises: an acquisition unit
that acquires observation information obtained through observation of a target region
from a flying object flying in outer space; a classification unit that inputs the
observation information acquired by the acquisition unit to a classifier so trained
as to output a classification result obtained by classifying a target object present
in the target region if the observation information is input, and classifies the target
object; an acceptance unit that accepts designation input for designating the target
object; and an output unit that outputs the observation information including a classification
result of the target object designated.
[Effect of Invention]
[0007] In one aspect, it is possible to readily acquire easy-to-handle observation data.
[Brief Description of Drawings]
[0008]
FIG. 1 is a schematic view illustrating one example of the configuration of a satellite
image provision system.
FIG. 2 is a block diagram illustrating one example of the configuration of a server.
FIG. 3 illustrates one example of the record layout of a user DB and an image DB.
FIG. 4 illustrates land cover classification processing.
FIG. 5 illustrates object classification processing.
FIG. 6 illustrates one example of a display screen of a terminal.
FIG. 7 illustrates one example of a display screen of the terminal.
FIG. 8 illustrates one example of a display screen of the terminal.
FIG. 9 illustrates one example of a display screen of the terminal.
FIG. 10 illustrates one example of a display screen of the terminal.
FIG. 11A illustrates one example of a display screen of the terminal.
FIG. 11B illustrates one example of a display screen of the terminal.
FIG. 12 illustrates one example of a display screen of the terminal.
FIG. 13 is a flowchart of one example of the processing procedure of classifier generation
processing.
FIG. 14 is a flowchart of one example of the processing procedure of target object
classification processing.
FIG. 15 is a flowchart of one example of the processing procedure of image purchase
processing.
FIG. 16 is a flowchart of one example of the processing procedure of image browsing
processing.
FIG. 17 illustrates one example of a display screen according to Embodiment 2.
FIG. 18 illustrates one example of a display screen according to Embodiment 2.
FIG. 19 is a flowchart of one example of the processing procedure of image purchase
processing according to Embodiment 2.
FIG. 20 is a flowchart of one example of the processing procedure of image browsing
processing according to Embodiment 2.
FIG. 21 illustrates the outline of Embodiment 3.
FIG. 22 is a flowchart of one example of the processing procedure executed by a server
according to Embodiment 3.
FIG. 23 illustrates the outline of Embodiment 4.
FIG. 24 is a flowchart of one example of the processing procedure executed by a server
according to Embodiment 4.
[Mode for Carrying Out Invention]
[0009] The present invention is described below with reference to the drawings depicting
embodiments.
Embodiment 1
[0010] FIG. 1 is a schematic view illustrating one example of the configuration of a satellite
image provision system. In the present embodiment, a satellite image provision system
allowing the user to purchase (use) a desired satellite image is described. The satellite
image provision system includes an information processing apparatus 1, a terminal
2 and a satellite 3. The information processing apparatus 1 and the terminal 2 are
communicatably connected to each other over the network N such as the Internet or
the like.
[0011] The information processing apparatus 1 is an information processing apparatus capable
of performing various information processing as well as transmitting and receiving
information, and is, for example, a server device, a personal computer or the like.
In the present embodiment, the information processing apparatus 1 is assumed as a
server device and referred to as a server 1 for the sake of simplicity. The server
1 is a managing device for continuously acquiring satellite images obtained by shooting
the surface of the earth from the satellite 3 as an artificial satellite to manage
the acquired satellite image on a database and provides a platform allowing the user
to purchase a satellite image. It is noted that the satellite image is image data
of each pixel value within the image associated with the position information (information
on latitude and longitude) of the corresponding location on the surface of the earth.
In the present embodiment, several tens of compact satellites 3 each having a weight
of several tens of kilograms present on a predetermined orbit over the earth. The
server 1 acquires satellite images from the satellites 3 and manages them on the database.
The server 1 accepts designation input for designating a geographical range (region)
and a shooting date and time (shooting time point) of a satellite image desired to
purchase by the user on a Web browser, extracts an image corresponding to the designated
range and date and time from the satellite images on the database and provides the
user with the image.
[0012] Though the following description is made assuming that a satellite image (photograph)
shot by the satellite 3 detecting a visible ray, near infrared rays or the like is
provided to the user in the present embodiment, the mode of the present embodiment
is not limited thereto. For example, the satellite 3 may be provided with an optical
sensor for detecting thermal infrared radiation and may provide data acquired by detecting
infrared rays emitted by radiation (emission) from the surface of the earth. Furthermore,
the satellite 3 may be provided with a microwave sensor (synthetic aperture radar,
for example), not the optical sensor, for radiating microwave (radio wave) and detecting
the microwave reflected from the surface of the earth, and may provide data observed
by the microwave sensor. As such, the server 1 may be configured essentially to provide
the user with observation information obtained by observing the surface of the earth
from the satellite 3, and the observation information to be provided is not limited
to an image based on a visible ray.
[0013] Though the surface of the earth is observed by the artificial satellite (satellite
3) in the following description, that the server 1 may be configured essentially to
provide the user with data observed by any flying object moving in the space, and
the flying object is not limited to the artificial satellite.
[0014] In the present embodiment, when providing the user with a satellite image, the server
1 provides the user with data of a satellite image including a classification result
obtained by classifying a target object (object) within the satellite image using
a classifier generated by machine learning. Specifically, as will be described later,
the server 1 provides a satellite image to which metadata is added that indicates
a land cover within a satellite image and the number of movable objects present at
each area of the satellite image by using a land cover classifier for classifying
a land cover (for example, cloud covering the surface of the earth, water, trees,
bare ground, etc.) covering the surface of the earth and an object classifier for
classifying a specific object (for example, a movable object such as a vehicle, a
ship, etc.) present on the surface of the earth.
[0015] The terminal 2 is a terminal device operated by the user and is a personal computer,
for example. It is noted that the terminal 2 may be a smartphone, a tablet terminal
or other devices. The terminal 2 accesses the server 1 in response to operation input
by the user and performs a purchase request for a satellite image (divided image to
be described later) to download the satellite image from the server 1 and displays
the downloaded image.
[0016] Though the description is made assuming that a satellite image is sold to the user
so as to be downloaded to the terminal 2 in the present embodiment, a satellite image
may be made available for a certain period of time according to a contract with the
user in a subscription form, for example. The server 1 may be configured essentially
to output a satellite image from the terminal 2 in accordance with a request for use
so that the image is available to the user, not necessarily selling each satellite
image for the user to download the image in the terminal 2.
[0017] FIG. 2 is a block diagram illustrating one example of the configuration of the server
1. The server 1 is provided with a control unit 11, a main storage unit 12, a communication
unit 13 and an auxiliary storage unit 14.
[0018] The control unit 11 includes one or more arithmetic processing devices such as a
central processing unit (CPU), a micro processing unit (MPU), a graphics processing
units (GPU) or the like and performs various information processing, control processing,
etc. by reading and executing a program P stored in the auxiliary storage unit 14.
The main storage unit 12 is a temporary storage area such as a static random access
memory (SRAM), a dynamic random access memory (DRAM), a flash memory or the like and
temporarily stores data required for the control unit 11 to execute arithmetic processing.
The communication unit 13 is a communication module for performing communication processing
as well as transmits and receives information with the outside.
[0019] The auxiliary storage unit 14 is a nonvolatile storage area such as a large capacity
memory, a hard disk or the like and stores a program P required for the control unit
11 to execute processing and other data. Moreover, the auxiliary storage unit 14 stores
a user DB 141, an image DB 142, a land cover classifier 143 and an object classifier
144. The user DB 141 is a database storing information on each of the users. The image
DB 142 is a database storing satellite images acquired from the satellite 3. The land
cover classifier 143 is a learned model generated by machine learning and is a classifier
for classifying a land cover that covers the surface of the earth. Likewise, the object
classifier 144 is a learned model generated by machine learning and is a classifier
for classifying an object (movable object) present on the surface of the earth.
[0020] It is noted that the auxiliary storage unit 14 may be an external storage device
connected to the server 1. Moreover, the server 1 may be a multicomputer formed by
multiple computers or may be a virtual machine virtually constructed by software.
[0021] In the present embodiment, the server 1 is not limited to the above-described ones
and may include, for example, an input unit for accepting operation input, a display
unit for displaying an image, etc. Furthermore, the server 1 may include a reading
unit for reading a portable storage medium P1 such as a compact disc (CD)-ROM, a digital
versatile disc (DVD)-ROM or the like and read a program P from the portable storage
medium P1 to execute the program P. Alternatively, the server 1 may read out a program
P from a semiconductor memory P2.
[0022] FIG. 3 illustrates one example of the record layout of the user DB 141 and the image
DB 142.
[0023] The user DB 141 includes a user ID column, a name column, a user information column
and a purchased image column. The user ID column stores the user ID for identifying
each user. The name column, user information column and purchased image column respectively
store a user name, other information on the user and information on the image purchased
by the user in association with the user ID. The user information column stores account
information, information required for making a payment upon purchasing an image and
other information of the user on the platform. The purchased image column stores,
for example, an ID for identifying a divided image, which will be described later.
[0024] The image DB 142 includes an image ID column, a date and time column, a satellite
ID column, an image column, a cell ID column, a cell region column and an object column.
The image ID column stores an image ID for identifying a satellite image acquired
from the satellite 3. The date and time column, the satellite ID column, the image
column, the cell ID column, the cell region column and the object column respectively
store a date and time when a satellite image is shot, the ID of a satellite 3 that
shoots the satellite image, a shot satellite image, a cell ID for identifying a cell
image (divided image), which will be described later, obtained by dividing the satellite
image by predetermined unit, coordinates information of an area within the satellite
image corresponding to the cell image and a classification result of each object (target
object) contained in the cell image in association with the image ID.
[0025] FIG. 4 illustrates land cover classification processing. In the present embodiment,
as described above, the server 1 performs processing of classifying a target object
within a satellite image, specifically, a land cover and a movable object by using
classifiers generated by machine learning. FIG. 4 conceptually illustrates the details
of the processing related to a land cover out of the land cover and the movable object.
The processing details of the land cover classification processing is described with
reference to FIG. 4.
[0026] For example, the server 1 performs machine learning to learn the features of a land
cover by deep learning to thereby generate the land cover classifier 143. The land
cover classifier 143 is a neural network related to a convolution neural network (CNN),
for example and includes an input layer for accepting input of a satellite image (observation
information), an output layer for outputting a classification result of a land cover
contained in the satellite image and an intermediate layer for extracting feature
values of the satellite image.
[0027] The input layer includes multiple neurons accepting pixel values of respective pixels
contained in a satellite image and passes the input pixel values to the intermediate
layer. The intermediate layer includes multiple neurons extracting image feature values
of the satellite image and passes the extracted feature values to the output layer.
If the land cover classifier 143 is a CNN, the intermediate layer includes convolution
layers for convoluting the pixel values of the respective pixel values input from
the input layer and pooling layers for mapping the pixel values convoluted in the
convolution layer, the convolution layers and the pooling layers being alternately
connected, and finally extracts the feature values of the satellite image while compressing
the pixel information of the satellite image. The output layer includes one or more
neurons that output the classification results of the land cover and classifies the
land cover that covers the surface of the earth based on the image feature values
output from the intermediate layer.
[0028] Though the following is described assuming that the land cover classifier 143 is,
but not limited to, a CNN in the present embodiment, the land cover classifier 143
may be a learned model that is constructed by another learning algorithm such as a
neural network other than the CNN, a support vector machine (SVM), a Bayesian network,
a regression tree or the like.
[0029] The server 1 performs learning by using training data including multiple satellite
images obtained by shooting the surface of the earth from the satellite 3 that are
associated with correct answer values as classification results obtained when the
land covers of the satellite images are classified. The server 1 inputs a satellite
image included in the training data to the input layer, followed by arithmetic processing
in the intermediate layer and acquires an output value indicating the classification
result of the land cover from the output layer. It is noted that the output value
is a discrete value (value of "0" or "1," for example) or a continuous probability
value (value in the range of "0" to "1," for example).
[0030] Here, since the satellite image shot by the satellite 3 has a huge data size, the
server 1 inputs, as the image data to be input to the land cover classifier 143, divided
images (for example, cell images to be described later) obtained by dividing a satellite
image by predetermined unit, not the satellite image as it is (raw data), and classifies
them. This makes it possible to reduce an arithmetic load required to perform processing
on individual images and suitably perform the classification processing.
[0031] The server 1 compares the output value from the output layer with the information
labeled on the satellite image in the training data, i.e., the correct answer value
to thereby optimize the parameters used for the arithmetic processing in the intermediate
layer such that the output value approximates the correct answer.
The parameters include a weight between the neurons (coupling coefficient) and a coefficient
of an activation function used in each of the neurons, for example. Though the method
of optimizing parameters is not limited to a particular method, the server 1 optimizes
various parameters by using backpropagation, for example. The server 1 performs the
above-described processing on each of the satellite images included in the training
data to generate the land cover classifier 143.
[0032] The server 1 generates the land cover classifier 143 related to semantic segmentation
which is one type of the CNN, for example. The semantic segmentation is a method of
performing class determination by pixel indicating which object (target object) each
pixel within the image represents. Data used here includes information (correct answer
value) indicating the type of the land cover that is attached as a label to the image
area corresponding to each type of land covers in the satellite image. The server
1 inputs the satellite image included in the training data to the land cover classifier
143, acquires the output value indicating the classification result of each land cover
by pixel and optimizes the parameters by comparing the output value with the correct
answer value to thereby generate the land cover classifier 143 that allows for classification
by pixel indicating which land object each pixel represents.
[0033] If a satellite image is acquired from the satellite 3, the server 1 performs classification
of a land cover by using the land cover classifier 143. The land cover is a predetermined
object covering the surface of the earth and is classified as cloud, woods, a bare
ground, water areas, ice and snow, artificial objects, etc. It is noted that such
classification is exemplification, and the land cover is not limited to the above-described
ones. The server 1 divides a satellite image into images by predetermined unit and
inputs the divided images to the input layer of the land cover classifier 143, performs
computation for extracting image feature values in the intermediate layer and inputs
the extracted feature values to the output layer. The server 1 acquires from the output
layer the classification result by pixel indicating which type of land cover each
pixel represents as an output value.
[0034] Though the following is described assuming that the server 1 generates the land cover
classifier 143 by supervised learning in the present embodiment, the land cover classifier
143 may be generated by semi-supervised learning or unsupervised learning. The same
applies to the object classifier 144 to be described later.
[0035] FIG. 5 illustrates object classification processing. FIG. 5 conceptually illustrates
measurement of the number of specific objects, specifically, movable objects (for
example, vehicles, ships, etc.) contained in a satellite image by using the object
classifier 144. The processing details of the object classification processing are
described with reference to FIG. 5.
[0036] The object classifier 144 is a neural network, which is related to the CNN, generated
by deep learning similarly to the above-mentioned land cover classifier 143. Similarly
to the land cover classifier 143, the object classifier 144, which includes an input
layer, an intermediate layer and an output layer, accepts input of a satellite image
and outputs a classification result obtained by classifying movable objects within
the satellite image.
[0037] The server 1 performs learning by using training data including a correct answer
value for a classification result obtained when a movable object is classified, the
correct answer value being associated with each of the satellite images. Similarly
to learning with the land cover classifier 143, the server 1 divides each of the satellite
images included in the training data by predetermined unit and inputs the divided
ones to the object classifier 144 while acquiring the classification result of the
movable object as an output value. The server 1 compares the output value with the
correct answer value to optimize parameters such as weights or the like, to thereby
generate the object classifier 144.
[0038] For example, the server 1 performs learning by using training data including a correct
answer value indicating the type of the movable object that is attached as a label
to a coordinate point (plot) corresponding to each of the movable objects present
in a satellite image. The resolution of the satellite images is often too low to extract
feature values such as the shape and color of each of the movable objects and detect
an object. Hence, the server 1 learns feature values of each predetermined unit area
(see a rectangular frame at the lower right of FIG. 5) within the image and the number
of movable objects (the number of plots) present in each area based on the plotted
coordinate point of the movable object in the training data to thereby generate the
object classifier 144 that directly estimates the number of movable objects without
detecting (recognizing) individual movable objects. This makes it possible to accurately
estimate the number of movable objects even if the resolution is low.
[0039] It is noted that individual movable objects may be detected to thereby measure the
number of movable objects by the object classifier 144 depending on the resolution
of a satellite image. That is, the object classifier 144 may be configured essentially
to estimate the number of movable objects, and the algorithm therefor is not limited
to a particular one.
[0040] The server 1 inputs a satellite image shot by the satellite 3 to the object classifier
144 while acquiring a classification result that is obtained by classifying the movable
objects contained in the satellite image and that indicates the number of movable
objects as an output value. As illustrated at the lower right of FIG. 5, for example,
the server 1 acquires an estimation result obtained by estimating the number of movable
objects for each area in the image.
[0041] Here, when inputting a satellite image to the object classifier 144, the server 1
narrows down the image area within the satellite image according to the classification
result of the land cover by the land cover classifier 143 and inputs the narrowed
area. This allows the server 1 to reduce a load on the arithmetic processing using
the object classifier 144.
[0042] In the case where the movable object as a target to be classified is a ship, for
example, the image area to be taken into account as an analysis target is water areas
such as sea, river, etc. Here, if making a classification of a ship, the server 1
specifies the image area corresponding to the water areas with reference to the classification
result of the land cover and inputs the specified image area to the object classifier
144. This allows the server 1 to acquire the classification result indicating the
number of ships present at the water areas as an output value as illustrated at the
lower right of FIG. 5.
[0043] It is noted that the movable object such as a ship, a vehicle, etc. are mere examples
of objects to be classified by using the object classifier 144, and the object may
be a static object present at a fixed location.
[0044] As described above, the server 1 performs classification processing for a land cover
and a movable object from a satellite image by using the land cover classifier 143
and the object classifier 144, respectively. The server 1 sequentially acquires satellite
images obtained by shooting each target region at shooting time points (observation
time points) from each of multiple satellites 3, 3, .... and classifies the satellite
images. The server 1 stores each of the satellite images and the classification result
of the satellite image in association with the shooting time point and the target
region in the image DB 142.
[0045] FIGs. 6 to 12 each illustrate one example of a display screen of the terminal 2.
The outline of the present embodiment is described with reference to FIGs. 6 to 12.
[0046] The server 1 in the present embodiment provides a platform to sell a satellite image
acquired from each of the satellites 3 to the user. FIG. 6, which is one example of
a display screen displayed on the terminal 2, illustrates a display example of a Web
browser screen related to the present platform. The server 1 provides the user with
a satellite image labeled with the classification results obtained by using the land
cover classifier 143 and the object classifier 144 in response to operation input
on the screen.
[0047] For example, the terminal 2 searches for a target region to be purchased based on
operation input performed on the screen (not illustrated) and displays the screen
illustrated in FIG. 6. More specifically, as illustrated in FIG. 6, the terminal 2
displays a map image representing the target region.
[0048] The terminal 2 displays a grid 41 (dividing line) for dividing the target region
into multiple areas with a predetermined width in row and a predetermined width in
column while being superimposed on the map image. Though the width (distance) of each
area divided by the grid 41 is not limited to a particular value, the width of 5 km
in all directions is set in the present embodiment.
[0049] In the present embodiment, though the grid 41 is set to extend along a north-south
direction and an east-west direction in accordance with latitude and longitude so
as to divide the map image to rectangles, the way of drawing (direction) of the grid
41 is not particularly limited thereto. Furthermore, the user may arbitrarily set
a dividing line to set any area available as a purchase target region.
[0050] Moreover, the grid 41 is a rough approximation set when a target region is divided
into multiple areas, and cell images to be described later do not necessarily precisely
match the areas on the map image divided by grid 41. For example, the neighboring
cell images may be overlapped with each other.
[0051] The terminal 2 accepts designation input for designating multiple areas as a target
to be purchased (target to be used) from the areas divided by the grid 41. The terminal
2 accepts designation input for designating multiple areas by drag operation, click
operation or the like performed on the screen, for example. If accepting designation
input of the areas as a target to be purchased, the terminal 2 makes communication
with the server 1 to shift to the screen illustrated in FIG. 7.
[0052] If accepting designation input for designating the areas to be purchased from the
terminal 2, the server 1 searches the image DB 142 for a satellite image including
the designated areas. The server 1 then extracts cell images (area observation information)
obtained by cutting out the image area corresponding to the areas designated on the
map image from the searched satellite image. The server 1 outputs the extracted cell
images to the terminal 2 and causes the terminal 2 to display the images.
[0053] More specifically, as illustrated in FIG. 7, the server 1 displays a preview image
in which the cell images 51 extracted from the satellite image (observation information)
of the target region are superimposed on the corresponding areas on the map image.
It is noted that in the drawings of the present embodiment, the areas on which the
satellite image (photograph) is to be displayed are hatched for convenience of drawing.
At the time of displaying the preview image in FIG. 7, the server 1 first displays
cell images 51 extracted from the satellite image shot at the latest shooting time
point (observation time point) out of the multiple satellite images obtained when
the target region is shot at the past shooting time points. The cell images 51 as
a target to be purchased are displayed so as to be superimposed on the map image,
whereby the user can easily perceive the location of the cell images while confirming
the preview image.
[0054] Here, the terminal 2 displays a cloud amount setting section 52 and an angle setting
section 53 for respectively narrowing down the amount of cloud and an off-nadir angle
(shooting angle) at a predetermined position of the preview image. The terminal 2
accepts input of set values for the cloud amount and/or off-nadir angle at the respective
setting sections and switches a display manner of the cell images 51 depending on
whether or not the set values are satisfied.
[0055] The cloud amount is a ratio of the area of the cloud to the shot range (observation
range) of the satellite image. If the user purchases a satellite image (cell images
51) and due to a large amount of cloud within the image, the user can hardly confirm
the surface of the earth, it is difficult to say that this image is useful for the
user. Hence, in the present embodiment, narrowing down the cell images depending on
the amount of cloud is made possible at the time of displaying the preview image.
For example, the terminal 2 accepts a setting input for setting the upper limit value
of the amount of cloud in the area of each of the cell images 51 at the time of shooting
(observation time point) by the cloud amount setting section 52. This makes it possible
for the terminal 2 to narrow down the cell images 51.
[0056] The off-nadir angle is a shooting angle obtained when the satellite 3 shoots a target
region (area of each cell image 51) and is an angle formed by a normal line connecting
the satellite 3 at the time of shooting (observation) and the surface of the earth
and a straight line connecting the satellite 3 and the target region (area). Though
the "off-nadir angle" naturally means an illuminating angle of a microwave in the
synthetic aperture radar, it is temporarily used to express a shooting angle obtained
when a target region is shot from the satellite 3 for the sake of convenience in the
present embodiment.
[0057] In the case of a high off-nadir angle, a satellite image (cell image 51) is an image
obtained by shooting the target region obliquely, not directly overhead. Accordingly,
regardless of the desire for an image obtained by shooting a building directly overhead,
the user purchases an image obtained by shooting a building obliquely as a result.
Hence, in the present embodiment, narrowing the cell images depending on the off-nadir
angle is made possible at the time of displaying a preview image.
[0058] For example, the server 1 calculates the angle formed by a normal connecting the
satellite 3 and the surface of the earth and a direct line connecting the satellite
3 and the center point of the area of each cell image 51 as an off-nadir angle related
to each cell image 51. The terminal 2 accepts setting input for setting the upper
limit value of the off-nadir angle at the time of shooting (observation time point)
by the angle setting section 53. According to the upper limit values for the cloud
amount and/or the off-nadir angle set above, the terminal 2 displays the cell images
51 in a different manner between the cell image 51 having a value equal to or less
than the upper limit value and the cell image 51 having a value above the upper limit
value. For example, the terminal 2 displays the cell image 51 having a value above
the upper limit value so as to be darker than the cell image 51 having a value equal
to or less than the upper limit value. In FIG. 7, display in dark color is represented
by shades of the color of the hatching for the sake of convenience. By the above-described
switching display, the user can purchase a satellite image (cell image 51) for which
the cloud amount and/or the off-nadir angle satisfy a predetermined condition.
[0059] Furthermore, the terminal 2 accepts designation input for changing the cell images
51 to be purchased from the images shot at the latest time point to the images shot
at another time point in the past. For example, the terminal 2 displays in list form
the shooting time points of the satellite images that are stored in the image DB 142
and that are obtained by shooting a target region in an operation menu at the right
of the screen. The terminal 2 accepts designation input for designating any shooting
time point in the past by the operation menu.
[0060] If another time point is designated, the terminal 2 accesses the server 1 to acquire
cell images 51 obtained by extracting the areas designated by the user from the satellite
image shot at the designated time point. The terminal 2 switches the cell images 51
superimposed on the map image to the cell images 51 newly acquired and displays the
switched image. If change to another time point is further designated, the terminal
2 switches the screen display to the image at the designated time point and displays
the switched image. By repeating the above-described processing, the user can select
an image to be purchased while simply switching images shot at respective time points
in the past for confirmation.
[0061] In the case where the cell image 51 is to be purchased, the terminal 2 registers
a cell image 51 in a cart in response to operation input from the user. The cart is
a list for tentatively registering a cell image 51 to be purchased. For example, the
terminal 2 accepts click operation for any of the cell images 51 that are being displayed
and registers the operated cell image 51 in the cart.
[0062] FIG. 8 illustrates one example of a display screen after registration in the cart.
If a cell image 51 is registered in the cart, the terminal 2 displays each registered
cell image 51 labeled with an icon 54. Displaying the icon 54 allows the user to easily
grasp that the cell image 51 that is being displayed has already been registered in
the cart. Furthermore, the terminal 2 displays at the operation menu a dot mark 55
associated with the shooting time point (date and time) of the cell images 51 that
are being displayed and a bar 56 having a length corresponding to the amount (the
number) of the cell images 51 registered in the cart if some cell images 51 are registered
in the cart. This allows the user to easily grasp the time point when the registered
cell images 51 were shot and the number of registered cell images 51.
[0063] If the user proceeds to apply for purchase of some cell images 51, the terminal 2
shifts to a purchase application screen. FIG. 9 illustrates one example of a display
screen displayed at the time of purchase application. As illustrated in FIG. 9, the
terminal 2 displays the thumbnail images of the cell images 51 that are being registered
in the cart in list form and as well as a purchase amount (not illustrated).
[0064] In this case, the terminal 2 accepts designation input for designating a target object
desired by the user as a target to be classified for the cell images 51 to be purchased.
For example, the terminal 2 accepts designation input for designating the type of
the land cover and the movable object to be targeted as illustrated at the lower right
of FIG. 9.
[0065] The terminal 2 transmits a purchase request (use request) for cell images 51 including
the details of the designation of the target object to the server 1 in response to
operation input by the user. If accepting the purchase request from the terminal 2,
the server 1 performs processing of settling the purchase price according to the number
of cell images 51 to be purchased. It is noted that the server 1 may vary the purchase
amount depending on the number of types of the target object designated above.
[0066] After completion of the purchase processing, the server 1 outputs the purchased cell
images 51 to the terminal 2 and causes the terminal 2 to download them. In this case,
the server 1 outputs data of the cell images 51 including the classification result
of the target object designated by the user to the terminal 2. For example, the server
1 outputs image data of the cell images 51 to which metadata indicating the classification
result of the target object is added to the terminal 2. This allows the user to not
merely receive a satellite image but also receive data of the analyzed satellite image.
[0067] It is noted that though a satellite image to which a classification result (metadata)
of the target object is added is provided in the present embodiment, the image to
be provided is not limited thereto. Alternatively, data of an image itself that is
so processed as to show the position and the range of the target object within the
image may be provided. For example, the server 1 may generate a satellite image (cell
image 51) in which the area of the land cover to be targeted is sorted by color like
color-coding display to be describe later and output the colored image to the terminal
2. Hence, the server 1 may be configured essentially to output a satellite image including
the classification results of the target object, and the output manner of the image
data to be output is not limited to a particular one.
[0068] FIG. 10 illustrates one example of a display screen at the time of browsing purchased
cell images 51. In the present embodiment, the server 1 makes the purchased cell images
51, i.e., the cell images 51 available to the user browsable on the same platform
(Web browser screen) as they are purchased.
[0069] The terminal 2 accepts selection input for selecting cell images 51 as a target to
be browsed out of the cell images 51 that have been downloaded from the server 1 on
a screen (not illustrated). The terminal 2 displays the selected cell images 51.
In this case, the terminal 2 displays the selected cell images 51 superimposed on
the corresponding areas of the map image.
[0070] More specifically, the terminal 2 displays the cell images 51 superimposed on the
map image while displaying the dot mark 55 applied to the shooting time corresponding
to the cell images 51 that are being displayed at the operation menu. This allows
the user to easily grasp where and when the images are shot similarly to purchasing
time.
[0071] Next, the details of the processing at the time of browsing the classification result
of target objects contained in the purchased cell image 51 is described. FIGs. 11A
and 11B each illustrate one example of a display screen of the classification result.
For example, if accepting designation input for designating any one of the cell images
51 superimposed on the map image shown in FIGs. 11A and 11B, the terminal 2 displays
the classification result of the target objects related to the designated cell image
51.
[0072] FIG. 11A illustrates one example of the display screen related to a land cover while
FIG. 11B illustrates one example of the display screen related to a movable object.
As illustrated in FIG. 11A, for example, in the case where the classification result
related to a land cover is displayed, the terminal 2 displays the image area corresponding
to the land cover designated by the user in a manner different from the other image
areas. For example, the terminal 2 highlights the pixels corresponding to the land
cover designated by the user by color-coding or the like.
[0073] As illustrated in FIG. 11B, for example, in the case where the classification result
related to a movable object is displayed, the terminal 2 displays an area in which
one or more movable objects are present by a rectangular frame while displaying the
number of movable objects within the area in graph form.
[0074] Next, the details of the processing at the time of accepting additional purchase
of a cell image 51 is described. The terminal 2 accepts additional purchase of some
cell images 51 shot at another time point from the display screen shown in FIG. 10.
For example, the terminal 2 accepts designation input for designating another time
point by the operation menu. If another time point is designated, the terminal 2 accesses
the server 1 and requests the server 1 to output the cell images 51 that are related
to the satellite image shot at the designated time point and that are in the same
area as the cell images 51 that are being displayed. If acquiring the cell images
51 from the server 1, the terminal 2 switches the screen to a preview image of the
acquired cell images 51 and displays the switched preview image.
[0075] FIG. 12 illustrates one example of the switched screen.
[0076] Similarly to Embodiment 1, the terminal 2 displays the cell images 51 superimposed
on the map image. The terminal 2 accepts setting input of the cloud amount and the
off-nadir angle by the cloud amount setting section 52 and the angle setting section
53, respectively, accepts registration operation in a cart on this screen, and then
proceeds to purchase application. The cart registration processing, the purchase processing
for the cell images 51, etc. are similar to those in the above description, and thus
the detailed description is not repeated.
[0077] As described above, the user can designate the geographical range (area), the shooting
time point, etc. of the cell images 51 to be purchased on the map image with simple
operation. Furthermore, cell images 51 (area observation information) of an arbitrary
area are cut out and provided from the raw image data acquired from the satellite
3, that is, the satellite image (observation information) having a huge data size.
This allows the user to acquire an easy-to-handle image with a small data size. According
to the present embodiment, it is possible to easily acquire a satellite image that
permits ease in handling.
[0078] Additionally, according to the present embodiment, the user can acquire cell images
51 including the classification result of a desired target object. This makes it possible
to utilize satellite images for various applications such as grasping of the situation
of a land, the traffic volume or the like.
[0079] FIG. 13 is a flowchart of one example of the processing procedure of classifier generation
processing. The details of the processing of generating the land cover classifier
143 and the object classifier 144 by conducting machine learning are described with
reference to FIG. 13.
[0080] The control unit 11 of the server 1 acquires training data of a satellite image associated
with a correct answer value as a classification result obtained when each target object
contained in the satellite image is classified (step S11). As described above, the
target object to be classified includes a land cover that covers the surface of the
earth, an object (movable object, for example) present on the surface of the earth,
etc. The control unit 11 acquires training data of the satellite image labeled with
information (correct answer value) of various objects.
[0081] The control unit 11 generates the land cover classifier 143 for outputting, when
a satellite image is input by using the training data, a classification result obtained
by classifying a land cover within the satellite image (step S12). That is, the control
unit 11 inputs the satellite image included in the training data to the land cover
classifier 143 and acquires the classification result obtained when a land cover is
classified as output. The control unit 11 compares the acquired classification result
with the correct answer value and optimizes parameters such as weights between the
neurons such that the output classification result approximates the correct answer
value.
[0082] Furthermore, the control unit 11 generates the object classifier 144 for outputting,
when a satellite image is input by using the training data, a classification result
obtained by classifying an object within the satellite image (step S13). More specifically,
the control unit 11 inputs the satellite image included in the training data to the
object classifier 144 and acquires an estimation result (classification result) obtained
by estimating the number of objects present at each area within the satellite image.
The control unit 11 compares the acquired classification result with the correct answer
value and optimizes parameters such as weights between the neurons such that the output
estimation result approximates the correct answer value. The control unit 11 ends
the series of processing.
[0083] FIG. 14 is a flowchart of one example of the processing procedure of target object
classification processing. The details of the processing of classifying various objects
contained in the satellite image by using the land cover classifier 143 and the object
classifier 144 are described.
[0084] The control unit 11 of the server 1 acquires a satellite image obtained by the satellite
3 shooting a target region (step S31). The control unit 11 divides the acquired satellite
image into cell images 51, 51, 51... corresponding to areas into which the target
region is sorted (step S32).
[0085] The control unit 11 inputs the cell images 51 to the land cover classifier 143 to
classify the land cover within the cell image 51 (step S33). For example, the control
unit 11 acquires a classification result indicating the type of the land cover corresponding
to each of the pixel values within the cell image 51 from the land cover classifier
143.
[0086] The control unit 11 extracts image areas corresponding to various types of the land
covers from the cell image 51 according to the classification result acquired at step
S33 (step S34). The control unit 11 inputs the extracted image areas to the object
classifier 144 to acquire a classification result indicating the number of specific
objects contained in the image areas from the object classifier 144 (step S35).
[0087] The control unit 11 stores the satellite image acquired at step S31 and the classification
results of the object related to each cell image 51 acquired at step S33 and S35 in
association with each other in the image DB 142 (step S36). The control unit 11 ends
the series of processing.
[0088] FIG. 15 is a flowchart of one example of the processing procedure of image purchase
processing. The details of the processing executed by the server 1 is described with
reference to FIG. 15.
[0089] The control unit 11 of the server 1 accepts designation input for designating a target
region representing a map image from the terminal 2 (step S51). The control unit 11
outputs the map image of the designated region to the terminal 2 and causes the terminal
2 to display the map image (step S52). More specifically, as described above, the
control unit 11 displays a map image on which the grid 41 (dividing line) for dividing
the target region into multiple areas is superimposed.
[0090] The control unit 11 accepts designation input for designating multiple areas out
of the respective areas on the map image divided by the grid 41 (step S53). The control
unit 11 extracts the cell images 51 (area observation information) corresponding to
the designated areas from the satellite image (observation information) of the target
region, and outputs a preview image on which the extracted cell images 51 are superimposed
at the corresponding areas on the map image to the terminal 2 and causes the terminal
2 to display the preview image (step S54). At step S54, the control unit 11 extracts
the cell images 51 from the satellite image shot at the latest time point out of the
satellite images shot at the target region and superimposes the extracted cell images
51 on the map image.
[0091] The control unit 11 determines whether or not designation input for designating change
of the shooting time point to another time point is accepted (step S55). If determining
that designation input for designating the shooting time point is accepted (S55: YES),
the control unit 11 extracts cell images 51 from the satellite image shot at the designated
shooting time point, outputs the cell images 51 to the terminal 2 and causes the terminal
2 to switch the cell images 51 on the map image to the acquired cell images (step
S56).
[0092] After execution of the processing at step S56 or if NO at step S55, the control unit
11 determines whether or not input of set values related to the cloud amount and/or
the off-nadir angle for each of the cell images 51 is accepted (step S57). For example,
the control unit 11 accepts setting input of the upper limit values for the cloud
amount and/or the off-nadir angle. If determining that input of set values of the
cloud amount and/or the off-nadir angle is accepted (step S57: YES), the control unit
11 switches the display of the cell images 51 such that a cell image 51 satisfying
the set value and a cell image 51 not satisfying the set value are displayed in a
different manner (step S58).
[0093] After execution of the processing at step S58 or if NO at step S57, the control unit
11 accepts registration operation of registering cell images 51 in the cart (purchase
candidate list) from the terminal 2 (step S59). The control unit 11 determines whether
or not a purchase request (use request) for the cell images 51 registered in the cart
is accepted (step S60). If determining that a purchase request is not accepted (S60:
NO), the control unit 11 returns the processing to step S55.
[0094] If determining that a purchase request is accepted (step S60:YES), the control unit
11 accepts designation input for designating a target object desired as a target to
be classified by the user (step S61). The target object designated at step S61 includes
the type of a land cover, the type of an object (movable object) or the like as described
above. The control unit 11 outputs to the terminal 2 the cell image 51 to which data
indicating the classification result of the target object is added for each of the
cell images 51 for which a purchase request is accepted with reference to the image
DB 42, and causes the terminal 2 to download the cell image 51 (step S62). The control
unit 11 ends the series of processing.
[0095] FIG. 16 is a flowchart of one example of the processing procedure of image browsing
processing. The details of processing performed at the time of browsing the purchased
satellite image is described with reference to FIG. 16.
[0096] The control unit 11 of the server 1 accepts designation input for designating purchased
cell images 51, that is, user available cell images 51 from the terminal 2 (step S71).
The control unit 11 generates a map image on which the designated cell images 51 are
superimposed at the corresponding locations on the map image and causes the terminal
2 to display the superimposed image (step S72).
[0097] The control unit 11 determines whether or not designation input for designating the
cell image 51 for which the classification result of the target object is to be displayed
is accepted from the terminal 2 (step S73). If determining that the designation input
is accepted (S73:YES), the control unit 11 causes the terminal 2 to display the cell
image 51 indicating the classification result of the target object contained in the
designated cell image 51 and the classification result of the target object designated
by the user upon purchasing (step S74). For example, if a land cover is assumed to
be a target, the control unit 11 displays the cell image 51 in which the image area
(pixels) corresponding to each land cover is displayed in a different manner by color-coding
or the like. Alternatively, if the number of objects is assumed to be a target, for
example, the control unit 11 displays the cell image 51 in which the number of object
present in each image area is shown in a graph form or the like.
[0098] After execution of the processing at step S74 or if "NO" at step S73, the control
unit 11 determines whether or not designation input for designating switching of the
shooting time point of the cell images 51 to another time point is accepted (step
S75). If determining that designation input for designating the shooting time point
is accepted (S75: YES), the control unit 11 switches the screen to the cell images
51 at the designated time point and displays the switched cell images 51 (step S76).
After execution of the processing at step S76 or if "NO" is determined at step S75,
the control unit 11 determines whether or not input of set values of the cloud amount
and/or the off-nadir angle are accepted (step S77). If determining that input of set
values of the cloud amount and/or the off-nadir angle are accepted (step S77: YES),
the control unit 11 switches the display of the cell images 51 such that a cell image
51 satisfying the set value and a cell image 51 not satisfying the set value are displayed
in a different manner (step S78).
[0099] After execution of the processing at step S78 or if "NO" is determined at step S77,
the control unit 11 determines whether or not a purchase request for the cell images
51 is accepted from the terminal 2 (step S79). If determining that a purchase request
is not accepted (S79: NO), the control unit 11 returns the processing to step S73.
If determining that a purchase request is accepted (step S79: YES), the control unit
11 accepts designation input for designating a target object desired by the user as
a target to be classified (step S80). The control unit 11 outputs the cell images
51 containing the classification result of the designated target object to the terminal
2, causes the terminal 2 to download the cell images 51 (step S81) and end the series
of processing.
[0100] Hence, according to Embodiment 1, it is possible to easily acquire easy-to-deal satellite
images (observation information).
[0101] Furthermore, according to Embodiment 1, a preview of the cell image at the region
designated by the user is displayed on the map image. This makes it possible to easily
grasp the location of the cell image 51 as a target to be purchased (used).
[0102] Moreover, according to Embodiment 1, it is possible to purchase (use) a satellite
image at an arbitrary time point from the respective satellite images shot at the
past time points stored in the image DB 142.
[0103] Additionally, according to Embodiment 1, the purchased cell images 51 can easily
be browsed while additional purchase of a cell image 51 can be easily performed.
[0104] In addition, according to Embodiment 1, cell images 51 can be narrowed down depending
on the cloud amount and/or off-nadir angle, whereby a desired cell image 51 can easily
be acquired.
[0105] Furthermore, according to Embodiment 1, by using the classifiers constructed by machine
learning, a desired target object can be extracted and provided (presented) from the
satellite image.
[0106] Moreover, according to Embodiment 1, classification of the land cover using the land
cover classifier 143 is performed to thereby provide the user with information on
the land cover covering the target region.
[0107] Additionally, according to Embodiment 1, using the object classifier 144, information
on the number of specific objects (movable object, for example) contained in the satellite
image can be provided to the user. Especially, in the present embodiment, the image
area to be input is narrowed down by using the classification result of the land cover
performed by the land cover classifier 143, resulting in reduction in a processing
load.
Embodiment 2
[0108] In the present embodiment, a mode allowing the user to purchase in bulk satellite
images shot during the time period designated by the user is described.
[0109] FIGs. 17 and 18 each illustrate one example of a display screen according to Embodiment
2. The outline of the present embodiment will be described with reference to FIGs.
17 and 18.
[0110] FIG. 17 shows one example of a display screen obtained when the user designates multiple
areas to be purchased from the map image. If multiple areas to be purchased are designated,
the terminal 2 displays an operation menu at the right of the screen to accept operation
input as to whether or not cell images 51 are to be purchased in bulk while the shooting
time period is designated, for example. If accepting designation input for designating
the shooting time period at the operation menu, the terminal 2 registers in the cart
the cell images 51 obtained by extracting the respective image areas corresponding
to the designated multiple areas from the respective satellite images shot at multiple
shooting time points included in the designated period. Naturally, it may be possible
to display a preview image before registration in the cart.
[0111] If the cell images 51 related to the respective shooting time points are registered
in the cart, the terminal 2 shifts to purchase application in response to operation
input by the user and accesses the server 1 to perform purchase processing. In this
case, the terminal 2 accepts designation input for designating a target object to
be classified and transmits a purchase request for the cell images 51 including the
details of the designation to the server 1 similarly to Embodiment 1. If accepting
the purchase request from the terminal 2, the server 1 outputs to the terminal 2 an
image group (observation information group) that consists of cell images 51, 51, 51...at
the time points to each of which a classification result of the designated target
object is added and causes the terminal 2 to download the group. According to the
processing described above, the user can purchase the cell images 51 shot at arbitrary
period of time in bulk by the above-described processing, resulting in improved convenience
of the user.
[0112] FIG. 18 displays one example of a display screen at the time of displaying the classification
result of the target object related to the purchased cell images 51. For example,
if accepting designation input of the cell image 51 on the display screen in FIG.
10 similarly to Embodiment 1, the terminal 2 displays the classification result related
to the target object of the designated cell image 51. In this case, the terminal 2
displays in time series the image group that includes the designated cell image 51
and consists of multiple cell images 51 obtained by shooting the same region as the
designated cell image 51 at different shooting time points.
[0113] For example, the terminal 2 reproduces video of the respective cell images 51 shot
at the multiple shooting time points in time series frame by frame.
[0114] The terminal 2 here displays a land cover by color-coding and the number of moving
objects in graph form for the cell image 51 shot at each of the time points. Thus,
the terminal 2 can display time-series change of the target object such as a land
cover, a moving object, etc. As such, the user can grasp the change of the state and
situation of the designated target object as well as can purchase the cell images
51 in bulk.
[0115] FIG. 19 is a flowchart of one example of the processing procedure of image purchase
processing according to Embodiment 2.
[0116] After accepting designation input for designating multiple areas on the map image
(step S53), the control unit 11 of the server 1 executes the following processing.
The control unit 11 determines whether or not designation input for designating a
shooting time period to be purchased is accepted for the cell images 51 corresponding
to the designated areas (step S201). If determining that designation input for a time
period is not accepted (S201: NO), the control unit 11 shifts the processing to step
S54.
[0117] If determining that designation input for a time period is accepted (S201: YES),
the control unit 11 registers in the cart of the user the image group that consists
of respective cell images 51 shot at multiple time points included in the designated
time period and that corresponds to the areas designated at step S53 (step S202).
The control unit 11 determines whether or not a purchase request for purchasing the
cell images 51 registered in the cart is accepted (step S203). If determining that
a purchase request is not accepted (S203: NO), the control unit 11 returns the processing
to step S201. If determining that a purchase request is accepted (S203: YES), the
control unit 11 shifts the processing to step S60.
[0118] FIG. 20 is a flowchart of one example of the processing procedure of image browsing
processing according to Embodiment 2.
[0119] If determining that designation input for the cell image 51 for which the classification
result related to the target object is to be displayed is accepted from the terminal
2 (S73: YES), the control unit 11 of the server 1 executes the following processing.
The control unit 11 displays in time series the image group that includes the designated
cell image 51 and consists of multiple cell images 51 obtained by shooting the same
region as the designated cell image 51 at different shooting time points (step S221).
For example, the control unit 11 reproduces video of the cell images 51 as described
above. Here, the control unit 11 can display video showing time-series change of the
target object by displaying the land cover by color-coding and the number of moving
objects in graph form for the cell image 51 shot at each of the shooting time points.
The control unit 11 shifts the processing to step S75.
[0120] As such, according to Embodiment 2, it is possible to purchase in bulk the cell images
51 shot at multiple shooting time points, resulting in improved convenience of the
user.
[0121] Moreover, according to Embodiment 2, it becomes possible for the user to grasp the
time-series change of a target object by provision of the cell images 51 shot at multiple
shooting time points to each of which the classification result of the target object
is added.
Embodiment 3
[0122] In the present embodiment, a mode allowing the user to search for a desired image
from the satellite images stored in the image DB 142 is described.
[0123] FIG. 21 illustrates the outline of Embodiment 3. As described above, the server 1
stores each of the satellite images acquired from the satellite 3 and classification
results obtained by classifying a target object contained in each of the satellite
image by the land cover classifier 143 and the movable object classifier in association
with the target region and the shooting time point of the satellite image in the image
DB 142. In the present embodiment, the user can search the image DB 142 for a desired
image while using a target object as a search query. The outline of the present embodiment
will be described with reference to FIG. 21.
[0124] For example, the terminal 2 accepts designation input for designating the type of
a target object (land cover or movable object) as a search query and transmits the
search request to the server 1.
It is noted that the terminal 2 may issue a search request while designating more
detailed search criteria such as the size (area) of a land cover, the threshold for
the number of movable objects, etc. concerning the target object.
[0125] If accepting the search request from the terminal 2, the server 1 searches the image
DB 142 for cell images 51 containing the target object designated by the user with
reference to the classification result of the target object associated with each of
the satellite images (cell images 51). The server 1 outputs the search result to the
terminal 2 and causes the terminal 2 to display the search result. As illustrated
in FIG. 21, for example, the server 1 displays the thumbnail images of the searched
cell images 51 in list form.
[0126] For example, the terminal 2 accepts selection input for selecting any one of the
cell images 51 displayed in list form as a search result and displays a preview image
showing the selected cell image 51 superimposed on the map image. Here, the terminal
2 may display a preview image of all or a part of the information on the target object
as a search query such as a land cover, a movable object, etc. The server 1 accepts
registration operation of the cell image 51 in the cart, then accepts a purchase request
for the cell image 51 and outputs the cell image 51 including the classification result
of the target object to be searched to the terminal 2.
[0127] FIG. 22 is a flowchart of one example of a processing procedure executed by a server
1 according to Embodiment 3.
[0128] The control unit 11 of the server 1 accepts designation input for designating a target
object to be searched (step S301). The control unit 11 searches the image DB 142 for
satellite images (cell images 51) including the designated target (step S302). The
control unit 11 outputs the search result to the terminal 2 (step S303). For example,
the control unit 11 displays the thumbnail images of the searched cell images 51.
In this case, the control unit 11 may output all or a part of the information on the
target object as a search query for the searched cell images 51 to the terminal 2
and causes the terminal 2 to display it.
[0129] The control unit 11 accepts selection input for selecting any one of the cell images
51 displayed as a search result and causes the terminal 2 to display a preview screen
showing that the selected cell image 51 is superimposed on the map (step S304). The
control unit 11 accepts registration operation of the previewed cell image 51 in the
cart (step S305). The control unit 11 accepts a purchase request for the cell image
51 registered in the cart from the terminal 2 to perform purchase processing and causes
the terminal 2 to download the cell image 51 (step S306). The control unit 11 ends
the series of processing.
[0130] As such, according to Embodiment 3, it is possible to search the image DB 142 for
a satellite image including the target object designated by the user, resulting in
improved convenience of the user.
Embodiment 4
[0131] In the present embodiment, a mode is described in which a target object is monitored
based on the satellite images to determine whether or not a predetermined change occurs
in the target object.
[0132] FIG. 23 illustrates the outline of Embodiment 4. In the present embodiment, the server
1 detects a predetermined phenomenon occurring in a target region from the satellite
images continuously shot by the satellite 3 based on a time-series change of a classification
result of the target object by the land cover classifier 143 or the like and reports
the phenomenon to the user. The phenomenon as a target to be monitored is a phenomenon
related to a disaster such as a landslide, a volcanic eruption or the like.
[0133] The present embodiment will be described with reference to FIG. 23.
[0134] For example, the terminal 2 first pre-registers a target region to be monitored and
a phenomenon to be monitored. For example, the terminal 2 accepts designation input
for designating a target region from the user by the user designating the region to
be monitored on the map image similarly to FIG. 6. Furthermore, the terminal 2 accepts
designation input as to the type of the disastrous phenomenon to be monitored such
as a landslide, a volcanic eruption, etc. The terminal 2 transmits the designated
various types of information to the server 1 and pre-registers them.
[0135] Every time the server 1 acquires a satellite image obtained by shooting the target
region from the satellite 3, it inputs the satellite image to the land cover classifier
143 and obtains the classification result of the land cover. The server 1 here determines
whether or not a change of the target object corresponding to the pre-registered phenomenon
occurs, comparing the classification result of the land cover classified based on
the satellite image previously acquired and the classification result of the land
cover classified based on the satellite image currently acquired for the satellite
image of the target region. For example, if a landslide is designated as a phenomenon
to be monitored, the server 1 determines whether or not "woods" change to "bare ground."
[0136] If determining that a change in the target object occurs, the server 1 reports a
determination result to the terminal 2. For example, the server 1 reports the terminal
2 that the pre-registered phenomenon occurs at the target region and outputs to the
terminal 2 the satellite image representing the target region and labeled with the
image area estimated that the phenomenon occurs.
Describing with reference to the above-described example, the server 1 outputs to
the terminal 2 the satellite image labeled with the area in which "woods" changes
to "bare land."
[0137] In this case, the server 1 may output a satellite image as a preview display, for
example, and may then accept a purchase request for the satellite image from the terminal
2. Alternatively, the server 1 may automatically make a payment for the purchase of
the satellite image and cause the terminal 2 to download the satellite image. Hence,
if determining that a change in the target object occurs, the server 1 may by hand
or automatically execute the purchase processing of the satellite image labeled with
the area in which change occurs.
[0138] FIG. 24 is a flowchart of one example of the processing procedure executed by the
server 1 according to Embodiment 4.
[0139] The control unit 11 of the server 1 accepts designation input of designating a target
region to be monitored from the terminal 2 (step S401). The control unit 11 further
accepts designation input of designating a phenomenon to be monitored (change in the
target object) (step S402).
[0140] The control unit 11 acquires a classification result of the target object obtained
by classifying the satellite image corresponding to the target region designated at
step 401 from the image DB 142 (step S403). The control unit 11 determines whether
or not a change of the target object designated at step S402 occurs as compared with
the classification result of the target object shot at the previous shooting time
point (step S404).
[0141] If determining that no change occurs (S404: NO), the control unit 11 returns the
processing to step S403. If determining that a change occurs (404: YES), the control
unit 11 reports the determination result to the terminal 2 (step S405) and ends the
series of processing.
[0142] As such, according to Embodiment 4, it is possible to automatically report a predetermined
phenomenon occurring at a target region to the user and to utilize the present system
for various applications such as disaster observation, etc.
[0143] It is to be understood that the embodiments disclosed here are illustrative in all
respects and not restrictive. The scope of the present invention is defined by the
appended claims, and all changes that fall within the meanings and the bounds of the
claims, or equivalence of such meanings and bounds are intended to be embraced by
the claims.
[0144]
- 1
- server (information processing apparatus)
- 11
- control unit
- 12
- main storage unit
- 13
- communication unit
- 14
- auxiliary storage unit
- P
- program
- 141
- user DB
- 142
- image DB
- 2
- terminal
- 3
- satellite