TECHNICAL FIELD
[0001] The disclosure relates to an electronic device for generating a depth map, and an
operation method thereof. Particularly, the disclosure relates to an electronic device
for generating a depth map corresponding to an object, based on images of the object,
and an operation method of the electronic device.
BACKGROUND ART
[0002] Image-based depth estimation technology may be used in various fields such as robotics,
virtual reality, augmented reality, navigation, or autonomous driving. The image-based
depth estimation technology may be for measuring a distance based on image information,
and generating a depth map based on the measured distance. A distance in an image
may be obtained through sensors such as a light-detection- and-ranging (LiDAR) sensor,
a structured-light (SL) sensor, or a time-of-flight (ToF) sensor, and may also be
generated from binocular images through depth estimation based on a stereo matching
algorithm.
DESCRIPTION OF EMBODIMENTS
TECHNICAL PROBLEM
[0003] Recently, with the development of deep learning, it may be possible to measure a
distance, from an image of an object photographed by a single camera, and generate
a depth map, based on deep learning. However, a depth map generated from an image
captured by a single camera may have a low resolution.
SOLUTION TO PROBLEM
[0004] According to an embodiment of the disclosure, provided is a method of generating
a depth map corresponding to input data. The method according to an embodiment of
the disclosure may include providing a target object with light in a plurality of
patterns that change with time, obtaining a plurality of captured images respectively
corresponding to the plurality of patterns, by photographing the target object provided
with the light, obtaining the input data by preprocessing the plurality of captured
images, and generating the depth map based on the input data.
[0005] In an embodiment of the disclosure, each of the plurality of patterns may include
a light-emitting region to which the light is provided, and the position of the light-emitting
region may change with time.
[0006] In an embodiment of the disclosure, each of the plurality of patterns may include
a light-emitting region to which the light is provided, and an area of at least a
partial region of the light-emitting region may change with time.
[0007] In an embodiment of the disclosure, the illumination of the light in each of the
plurality of patterns may change with time.
[0008] In an embodiment of the disclosure, the method may further include obtaining a plurality
of illumination images respectively corresponding to the plurality of patterns, by
photographing the light-emitting device configured to provide the light, and the obtaining
of the input data may include obtaining the input data by preprocessing the plurality
of captured images and the plurality of illumination images.
[0009] In an embodiment of the disclosure, the method may further include obtaining characteristic
information of the light-emitting device, and the obtaining of the input data may
include obtaining the input data by preprocessing the plurality of captured images,
the plurality of illumination images, and the characteristic information.
[0010] In an embodiment of the disclosure, the characteristic information may include information
about the size of the light-emitting device.
[0011] In an embodiment of the disclosure, the method may further include providing the
target object with sub-light, and obtaining sub-characteristic information of a sub-light-emitting
device configured to provide the sub-light, the obtaining of the plurality of captured
images may include obtaining the plurality of captured images by photographing the
target object provided with the light and the sub-light, and the obtaining of the
input data may include obtaining the input data by preprocessing the plurality of
captured images, the plurality of illumination images, the characteristic information,
and the sub-characteristic information.
[0012] In an embodiment of the disclosure, the method may further include comparing the
intensity of an ambient illumination of the target object with the intensity of a
threshold illumination, wherein, based on the intensity of the ambient illumination
being less than or equal to the intensity of the threshold illumination, the light
may be provided to the target object, and the threshold illumination may be a maximum
illumination at which the plurality of captured images reflecting changes in the plurality
of patterns that change with time are able to be obtained.
[0013] In an embodiment of the disclosure, the generating of the depth map may include generating
the depth map by providing a depth map generation module with the input data, and
the depth map generation module may include an autoencoder.
[0014] According to an embodiment of the disclosure, provided is an electronic device for
generating a depth map corresponding to input data. The electronic device according
to an embodiment of the disclosure may include a light-emitting device configured
to provide a target object with light in a plurality of patterns that change with
time, a measuring device configured to obtain a plurality of captured images respectively
corresponding to the plurality of patterns, by photographing the target object provided
with the light, a memory storing one or more instructions, and at least one processor
configured to execute the one or more instructions stored in the memory to obtain
the input data by preprocessing the plurality of captured images, and generate the
depth map based on the input data.
[0015] According to an embodiment of the disclosure, provided a computer-readable recording
medium having recorded thereon a program for causing a computer to perform the method
of an embodiment of the disclosure.
BRIEF DESCRIPTION OF DRAWINGS
[0016] The disclosure may be readily understood from a combination of the following detailed
descriptions and the accompanying drawings, wherein reference numbers refer to structural
elements.
FIG. 1A is a diagram of a method of generating a depth map using an electronic device
including a light-emitting device, according to an embodiment of the disclosure:
FIG. 1B is a diagram of an electronic device including a light-emitting device according
to an embodiment of the disclosure:
FIG. 2A is a diagram of a method of generating a depth map using an electronic device
including a light-emitting device and a sub-light-emitting device, according to an
embodiment of the disclosure:
FIG. 2B is a diagram of an electronic device including a light-emitting device and
a sub-light-emitting device according to an embodiment of the disclosure;
FIG. 3A is a diagram illustrating a plurality of patterns in which the position of
a light-emitting region changes over time, according to an embodiment of the disclosure;
FIG. 3B is a diagram illustrating a plurality of patterns in which the area of a light-emitting
region changes over time, according to an embodiment of the disclosure;
FIG. 3C is a diagram illustrating a plurality of patterns in which the illumination
of light changes over time, according to an embodiment of the disclosure;
FIG. 3D is a diagram illustrating a plurality of patterns in which the shape of a
light-emitting region changes over time, according to an embodiment of the disclosure;
FIG. 4 is a diagram illustrating a plurality of captured images respectively corresponding
to a plurality of patterns according to an embodiment of the disclosure;
FIG. 5A is a diagram of an operation, performed by an image preprocessing module,
of obtaining input data by preprocessing a plurality of captured images, according
to an embodiment of the disclosure;
FIG. 5B is a diagram of an image preprocessing module configured to obtain input data
by preprocessing a plurality of captured images and a plurality of illumination images,
according to an embodiment of the disclosure;
FIG. 5C is a diagram of an image preprocessing module configured to obtain input data
by preprocessing a plurality of captured images, a plurality of illumination images,
and characteristic information, according to an embodiment of the disclosure;
FIG. 5D is a diagram of an image preprocessing module configured to obtain input data
by preprocessing a plurality of captured images, a plurality of illumination images,
characteristic information, and sub-characteristic information, according to an embodiment
of the disclosure;
FIG. 6 is a diagram of an operation of a depth map generation module according to
an embodiment of the disclosure;
FIG. 7A is a diagram of a depth map generation module configured to generate a depth
map based on input data generated by preprocessing a plurality of captured images
and a plurality of illumination images, according to an embodiment of the disclosure;
FIG. 7B is a diagram of a depth map generation module configured to generate a depth
map based on input data generated by preprocessing a plurality of captured images,
a plurality of illumination images, and characteristic information, according to an
embodiment of the disclosure;
FIG. 8 is a diagram of a network structure of a depth map generation module according
to an embodiment of the disclosure;
FIG. 9 is a diagram of a training process of a depth map generation module according
to an embodiment of the disclosure;
FIG. 10 is a flowchart of a method of generating a depth map using an electronic device
including a light-emitting device, according to an embodiment of the disclosure;
FIG. 11 is a flowchart of a method of generating a depth map based on input data obtained
by preprocessing a plurality of captured images, a plurality of illumination images,
and characteristic information, according to an embodiment of the disclosure;
FIG. 12 is a flowchart of a method of generating a depth map using an electronic device
including a light-emitting device and a sub-light-emitting device, according to an
embodiment of the disclosure; and
FIG. 13 is a flowchart of an operation, performed by an electronic device, of generating
a depth map corresponding to input data, according to an embodiment of the disclosure.
MODE OF DISCLOSURE
[0017] Throughout the disclosure, the expression "at least one of a, b, or c" indicates
only a, only b, only c, both a and b, both a and c, both b and c, all of a, b, and
c, or variations thereof.
[0018] Hereinafter, an embodiment of the disclosure is described in detail with reference
to the accompanying drawings for those of skill in the art to be able to perform the
disclosure without any difficulty. The disclosure may, however, be embodied in many
different forms and should not be construed as being limited to an embodiment of the
disclosure set forth herein. In addition, in order to clearly describe the disclosure,
portions that are not relevant to the description of the disclosure are omitted, and
similar reference numerals are assigned to similar elements throughout the specification.
[0019] Although the terms used herein are selected from among common terms that are currently
widely used in consideration of their function in the disclosure, the terms may be
different according to an intention of those of ordinary skill in the art, a precedent,
or the advent of new technology. Also, in particular cases, the terms are discretionally
selected by the applicant of the disclosure, in which case, the meaning of those terms
may be described in detail in the corresponding embodiment of the disclosure. Therefore,
the terms used herein are not merely designations of the terms, but the terms are
defined based on the meaning of the terms and content throughout the disclosure.
[0020] The singular expression may also include the plural meaning as long as it is not
inconsistent with the context. All the terms used herein, including technical and
scientific terms, may have the same meanings as those generally understood by those
of skill in the art.
[0021] Throughout the specification, when a part "includes" a component, it means that the
part may additionally include other components rather than excluding other components
as long as there is no particular opposing recitation. In addition, as used herein,
terms such as "...er (or)", "... unit", "... module", etc., denote a unit that performs
at least one function or operation, which may be implemented as hardware or software
or a combination thereof.
[0022] As used herein, the expression "configured to" may be interchangeably used with,
for example, "suitable for", "having the capacity to", "designed to", "adapted to",
"made to", or "capable of", according to a situation. The expression "configured to"
may not imply only "specially designed to" in a hardware manner. Instead, in a certain
circumstance, the expression "a system configured to" may indicate the system "capable
of" together with another device or components. For example, "a processor configured
(or set) to perform A, B, and C" may imply a dedicated processor (e.g., an embedded
processor) for performing a corresponding operation or a generic-purpose processor
(e.g., central processing unit (CPU) or an application processor) capable of performing
corresponding operations by executing one or more software programs stored in a memory.
[0023] Also, in the disclosure, it should be understood that when components are "connected"
or "coupled" to each other, the components may be directly connected or coupled to
each other, but may alternatively be connected or coupled to each other with a component
therebetween, unless specified otherwise.
[0024] In addition, the term 'augmented reality device' may denote a device capable of creating
'augmented reality', and includes not only augmented reality glasses resembling eyeglasses
that are typically worn on a user's face but also head-mounted display (HMD) apparatuses
and augmented reality helmets that are worn on the user's head, or the like. However,
the disclosure is not limited thereto, and the augmented reality device may be implemented
as various electronic devices such as a mobile device, a smart phone, a laptop computer,
a desktop computer, a tablet personal computer (PC), an electronic book terminal,
a digital broadcasting terminal, a personal digital assistant (PDA), a portable multimedia
player (PMP), a navigation system, an MP3 player, a camcorder, an internet protocol
television (TV) (IPTV), a digital television (DTV), a wearable device, and the like.
[0025] In the disclosure, a 'target object' may refer to an object for which a depth map
is to be generated using an electronic device. The target object may vary depending
on environments surrounding a user using the electronic device. The target object
may be an object in an environment in which a light-emitting device provides light
(e.g., emits light toward) and a measuring device captures images, among the environments
surrounding the electronic device. The target object may include one or more objects.
Other variations or implantations of the target object may be realized by one of skill
in the art from the disclosure herein, and the 'target object' is not necessarily
limited to the objects that are described herein.
[0026] Hereinafter, an embodiment of the disclosure will be described in detail with reference
to the accompanying drawings to allow those of skill in the art to easily carry out
the embodiment. The disclosure may, however, be embodied in many different forms and
should not be construed as being limited to an embodiment of the disclosure set forth
herein.
[0027] Hereinafter, an embodiment of the disclosure will be described in detail with reference
to the drawings.
[0028] FIG. 1A is a diagram of a method of generating a depth map using an electronic device
including a light-emitting device, according to an embodiment of the disclosure.
[0029] Referring to FIG. 1A, an electronic device 100 according to an embodiment of the
disclosure may be a device capable of providing light to a target object 200 and photographing
the target object 200. As an example of the disclosure, the electronic device 100
may be a device such as a smart phone, a TV, a notebook computer, or an augmented
reality device, but is not limited thereto. As described herein, the electronic device
100 may include a light-emitting device and/or other components may be configured
to provide an object with light or sub-light, and such providing of light may also
be referred to as emitting light toward or to the object, projecting light toward
the object, illuminating the object, etc.
[0030] In an embodiment of the disclosure, the electronic device 100 includes a light-emitting
device 111 configured to provide light to the target object 200, a measuring device
120 configured to photograph the target object 200, a memory 130 storing one or more
instructions, and a processor 140 configured to execute the one or more instructions
stored in the memory 130. However, FIG. 1A illustrates only some components for describing
an operation of the electronic device 100, and components included in the electronic
device 100 are not limited those illustrated in FIG. 1A. As an example of the disclosure,
the target object 200 may be an object for which a depth map is generated using the
electronic device 100. As an example of the disclosure, the target object 200 may
include one or more objects. The arrangements and shapes of the objects included in
the target object 200 may vary.
[0031] In an embodiment of the disclosure, the light-emitting device 111 may provide the
target object 200 with light in a plurality of patterns 300 (see FIG. 3A) that change
over time. Although FIG. 1A illustrates that the light-emitting device 111 is a TV
and the electronic device 100 includes the TV, the disclosure is not limited thereto.
The light-emitting device 111 may be a device configured to provide light, such as
a projector, a monitor, an illuminator, or a light-emitting element. In addition,
in a case in which the electronic device 100 is a TV, the light-emitting device 111
may correspond to a display device included in the TV. In addition, in a case in which
the electronic device 100 is a smart phone, the light-emitting device 111 may correspond
to a screen included in the smart phone. The light provided to the target object 200
by the light-emitting device 111 has the plurality of patterns 300 that change over
time. Accordingly, the shape of a shadow of the target object 200 formed by the light
provided from the light-emitting device 111 may change over time in response to the
plurality of patterns 300 described above.
[0032] In an embodiment of the disclosure, each of the patterns 300 may include a light-emitting
region DA that provides light, and a non-light-emitting region NDA that does not provide
light. When the position of the light-emitting region DA that provides the light and
is included in each of the plurality of patterns 300 changes over time, the shape
of the shadow of the target object 200 may change in response to the change in the
position of the light-emitting region DA.
[0033] In an embodiment of the disclosure, when the area of the light-emitting region DA
that provides the light and is included in each of the plurality of patterns 300 changes
over time, the shape of the shadow of the target object 200 may change in response
to the change in the area of the light-emitting region DA. In addition, the illumination
of the light provided to the target object 200 from the light-emitting device 111
may also change over time in response to the plurality of patterns 300 described above.
[0034] In an embodiment of the disclosure, as the illumination of the light included in
each of the plurality of patterns 300 changes over time, the shape of the shadow of
the target object 200 may change in response to the change in the illumination of
the light. In an embodiment of the disclosure, when the wavelength of the light included
in each of the plurality of patterns 300 changes over time, the shape of the shadow
of the target object 200 may change in response to the change in the wavelength of
the light. The shape of the shadow may include a position where the shadow is formed,
the size of the shadow, the shape of the shadow, and the like. In an embodiment of
the disclosure, when the illumination of light provided to the target object 200 from
the light-emitting device 111 changes over time, the luminance of the target object
200 may change in response to the change in the illumination of the light.
[0035] In an embodiment of the disclosure, the target object 200 may include one or more
objects. The shape of a shadow of each of the objects included in the target object
200 corresponding to the plurality of patterns 300 of light and the luminance of the
object may vary depending on the distance between the object and the light-emitting
device 111. As an example of the disclosure, a change in the shape of a shadow and
the luminance of an object at a short distance from the light-emitting device 111
may be greater than a change in the shape of a shadow and the luminance of an object
at a long distance from the light-emitting device 111.
[0036] In an embodiment of the disclosure, the measuring device 120 may obtain a plurality
of captured images 400 (see FIG. 4) by photographing the target object 200. The measuring
device 120 may include a first camera configured to obtain a red-green-blue (RGB)
image of the target object 200. In an embodiment of the disclosure, the first camera
may include an RGB camera. In addition, the measuring device 120 may include a second
camera configured to obtain a depth image of the target object 200. The depth image
may include data related to depth information representing the distance between the
measuring device 120 and the target object 200. In an embodiment of the disclosure,
the second camera may include a time-of-flight camera. In an embodiment of the disclosure,
the measuring device 120 may include a third camera configured to obtain an RGB image
and a depth image. The third camera may include an RGB-depth camera. However, in an
embodiment of the disclosure, the measuring device 120 may include only the first
camera.
[0037] The measuring device 120 may obtain the plurality of captured images 400 respectively
corresponding to the plurality of patterns 300, by photographing the target object
200 receiving, from the light-emitting device 111, light of the plurality of patterns
300 that change over time.
[0038] In an embodiment of the disclosure, the shapes of shadows of the target object 200
included in the plurality of captured images 400 may be different from each other.
For example, the shape of a shadow of the target object 200 may change in response
to the light provided in the plurality of patterns 300 that change over time, and
the plurality of captured images 400 may respectively include shadows of the target
object 200, which are formed by light having different patterns. In addition, in an
embodiment of the disclosure, the illumination of the light provided to the target
object 200 may change in response to the light provided in the plurality of patterns
300 that change over time, and the luminances of the target object 200 included in
the plurality of captured images 400 may be different from each other.
[0039] In an embodiment of the disclosure, when a time period required for the light-emitting
device 111 to provide light having 'n' patterns to the target object 200 is referred
to as one period, the measuring device 120 may obtain 'n' captured images respectively
corresponding to the 'n' patterns, by photographing the target object 200 for one
period. The shapes of shadows and luminances of the target object 200 respectively
included in the 'n' captured images may correspond to the light of the respective
patterns. 'n' may be a natural number.
[0040] In an embodiment of the disclosure, the electronic device 100 may further include
an illumination measuring device configured to photograph the light-emitting device
111. The illumination measuring device may obtain a plurality of illumination images
320 (see FIG. 5B) respectively corresponding to the plurality of patterns 300, by
photographing the light-emitting device 111 providing light.
[0041] In an embodiment of the disclosure, the illumination measuring device may be a separate
device from the electronic device 100. For example, the illumination images 320 obtained
by photographing the light-emitting device 111 included in the electronic device 100
may be provided to the electronic device 100 from the illumination measuring device.
In this case, an operation of photographing the light-emitting device 111 using the
illumination measuring device may be performed before an operation of photographing
the target object 200 using the measuring device 120. However, the disclosure is not
limited thereto, and the operation of photographing the light-emitting device 111
using the illumination measuring device and the operation of photographing the target
object 200 using the measuring device 120 may be performed simultaneously. In addition,
in an embodiment of the disclosure, a plurality of previously captured illumination
images 320 may be provided from an external server or nearby electronic devices through
a communication interface 150 (see FIG. 1B).
[0042] In an embodiment of the disclosure, the memory 130 may store at least one of instructions,
algorithms, data structures, program code, or application programs, which are readable
by the processor 140. In embodiments of the disclosure, operations performed by the
processor 140 may be implemented by executing the instructions, data structures, or
program code stored in the memory 130. The memory 130 may store at least one of instructions,
algorithms, data structures, program code, or application programs for obtaining input
data 500 (FIG. 6) based on the plurality of captured images 400 obtained by the measuring
device 120, and generating a depth map 600 (FIG. 6) based on the input data 500.
[0043] In an embodiment of the disclosure, the processor 140 may control the overall operation
of the electronic device 100. In an embodiment of the disclosure, the processor 140
may execute one or more instructions or program code stored in the memory 130, and
perform a function and/or an operation corresponding to the instructions/program code.
The processor 140 may control the overall operation of the light-emitting device 111
and the measuring device 120 by executing one or more instructions or programs stored
in the memory 130. The processor 140 may obtain the input data 500 by preprocessing
the plurality of captured images 400 obtained through the measuring device 120. The
processor 140 may generate the depth map 600 based on the input data 500. Although
FIG. 1A illustrates that the electronic device 100 includes one processor 140, the
disclosure is not limited thereto. The electronic device 100 may include two or more
processors, and may generate a depth map using at least one of the processors.
[0044] FIG. 1B is a diagram of an electronic device including a light-emitting device according
to an embodiment of the disclosure.
[0045] Referring to FIG. 1B, the electronic device 100 according to an embodiment of the
disclosure may include a light-emitting device 111, a measuring device 120, a memory
130, a processor 140, and a communication interface 150. The light-emitting device
111, the measuring device 120, the memory 130, the processor 140, and the communication
interface 150 may be electrically and/or physically connected to each other. Hereinafter,
the same reference numerals are assigned to the same components as those described
above with reference to FIG. 1A, and descriptions thereof may be omitted.
[0046] The memory 130 may include, for example, at least one of flash memory-type memory,
hard disk-type memory, multimedia card micro-type memory, card-type memory (e.g.,
secured digital (SD) or XD memory), random-access memory (RAM), static RAM (SRAM),
read-only memory (ROM), electrically erasable programmable ROM (EEPROM), programmable
ROM (PROM), mask ROM, flash ROM, a hard disk drive (HDD), or a solid-state drive (SSD).
The memory 130 may store instructions or program code for performing a function or
operation of the electronic device 100. Instructions, algorithms, data structures,
program code, and application programs stored in the memory 130 may be implemented
in a programming or scripting language (e.g., C, C++, Java, or an assembler).
[0047] In an embodiment of the disclosure, the memory 130 may include various types of modules
that may be used to generate the depth map 600 (see FIG. 6) based on the input data
500 (see FIG. 6). The memory 130 may include a light emission control module 131,
a measurement control module 132, an image preprocessing module 133, and a depth map
generation module 134. The 'module' included in the memory 130 may refer to a unit
for processing a function or operation performed by the processor 140, and may be
implemented as software such as instructions, an algorithm, a data structure, or program
code.
[0048] In an embodiment of the disclosure, the light emission control module 131 is configured
with instructions or program code related to an operation or function of the light-emitting
device 111 providing the target object 200 (see FIG. 1A) with light in the plurality
of patterns 300 (see FIG. 3A) that change over time. The light emission control module
131 will be described below with reference to FIGS. 3A to 3D.
[0049] In an embodiment of the disclosure, the measurement control module 132 is configured
with instructions or program code related to an operation or function, performed by
the measuring device 120, of obtaining the plurality of captured images 400 (see FIG.
4) respectively corresponding to the plurality of patterns 300 by photographing the
target object 200 receiving light from the light-emitting device 111. The measurement
control module 132 will be described below with reference to FIG. 4.
[0050] In an embodiment of the disclosure, the image preprocessing module 133 is configured
with instructions or program code related to an operation or function of obtaining
input data 500a (see FIG. 5A) by preprocessing the plurality of captured images 400.
In an embodiment of the disclosure, the image preprocessing module 133 may be configured
with instructions or program code related to an operation or function of obtaining
input data 500b (see FIG. 5B) by preprocessing the plurality of captured images 400
and the plurality of illumination images 320 (see FIG. 5B). In an embodiment of the
disclosure, the image preprocessing module 133 may be configured with instructions
or program code related to an operation or function of obtaining input data 500c (see
FIG. 5C) by preprocessing the plurality of captured images 400, the plurality of illumination
images 320 (see FIG. 5B), and characteristic information CI (see FIG. 5C) of the light-emitting
device 111. The image preprocessing module 133 will be described below with reference
to FIGS. 5A to 5C.
[0051] In an embodiment of the disclosure, the depth map generation module 134 is configured
with instructions or program code related to an operation or function of generating
the depth map 600 based on the input data 500. In an embodiment of the disclosure,
the depth map generation module 134 may be an artificial intelligence model. In an
embodiment of the disclosure, the depth map generation module 134 may include an autoencoder
model. The operation and training of the depth map generation module 134 will be described
below with reference to FIGS. 6 to 9. In the following embodiment of the disclosure,
the operation of the electronic device 100 may be implemented by the processor 140
executing modules included in the memory 130.
[0052] For example, the processor 140 may include at least one of a CPU, a microprocessor,
a graphics processing unit (GPU), an application processor (AP), an application-specific
integrated circuit (ASIC), a digital signal processor (DSP), a digital signal processing
device (DSPD), a programmable logic device (PLD), a field-programmable gate array
(FPGA), or a neural processing unit or a dedicated artificial intelligence processor
designed with a hardware structure specialized for training and processing of an artificial
intelligence model, but is not limited thereto.
[0053] In an embodiment of the disclosure, the communication interface 150 may perform data
communication with an external server, under control by the processor 140. Also, the
communication interface 150 may perform data communication with other nearby electronic
devices, in addition to the external server. The communication interface 150 may perform
data communication with a server or other nearby electronic devices using at least
one of data communication schemes including, for example, wired local area network
(LAN), wireless LAN, Wi-Fi, Bluetooth, ZigBee, Wi-Fi Direct (WFD), Infrared Data Association
(IrDA), Bluetooth Low Energy (BLE), near-field communication (NFC), wireless broadband
internet (WiBro), Worldwide Interoperability for Microwave Access (WiMAX), Shared
Wireless Access Protocol (SWAP), Wireless Gigabit Alliance (WiGig), and radio-frequency
(RF) communication. The communication interface 150 according to an embodiment of
the disclosure may transmit and receive data for generating the depth map 600 based
on the input data 500, to and from an external server or nearby electronic devices.
In an embodiment of the disclosure, the communication interface 150 may receive, from
an external server or nearby electronic devices, the depth map generation module 134
trained to generate the depth map 600 based on the input data 500.
[0054] FIG. 2A is a diagram of a method of generating a depth map using an electronic device
including a light-emitting device and a sub-light-emitting device, according to an
embodiment of the disclosure.
[0055] Referring to FIG. 2A, an electronic device 100a includes a light-emitting module
110, a measuring device 120 configured to photograph the target object 200, a memory
130a storing one or more instructions, and a processor 140a configured to execute
the one or more instructions stored in the memory 130a. Hereinafter, the same reference
numerals are assigned to the same components as those described above with reference
to FIG. 1A, and descriptions thereof may be omitted.
[0056] In an embodiment of the disclosure, the light-emitting module 110 may include the
light-emitting device 111 configured to provide light to the target object 200 and
a sub-light-emitting device 112 configured to provide sub-light to the target object
200. In an embodiment of the disclosure, the light-emitting device 111 may include
a TV, a monitor, a mobile phone display, and the like.
[0057] In an embodiment of the disclosure, the sub-light-emitting device 112 may be arranged
in a space in which the light-emitting device 111 and the target object 200 are present.
The sub-light-emitting device 112 may provide sub-light to the target object 200.
In an embodiment of the disclosure, the sub-light-emitting device 112 may include
an illuminator 112a, a lamp 112b, and the like. In addition, the sub-light-emitting
device 112 may include a home appliance including a light source, such as a refrigerator,
an oven, or a microwave oven. In an embodiment of the disclosure, the sub-light-emitting
device 112 may perform an operation of providing or not providing sub-light to the
target object 200. Unlike the light-emitting device 111, the sub-light-emitting device
112 may provide the target object 200 with sub-light in a constant pattern regardless
of time. In this case, the shape of a shadow of the target object 200 and the luminance
of the target object 200 may change over time due to the light provided in the plurality
of patterns 300 (see FIG. 3A) and the sub-light provided in the constant pattern.
[0058] In an embodiment of the disclosure, the sub-light-emitting device 112 may be an illuminator,
a lamp, or a home appliance connected to Internet of Things (IoT). The sub-light provided
by the sub-light-emitting device 112 may also be provided to the target object 200
in a plurality of patterns that change over time. In this case, the shape of the shadow
of the target object 200 or the luminance of the target object 200 may change over
time due to the light and the sub-light each provided in a plurality of patterns.
In this case, the plurality of patterns 300 in which the light-emitting device 111
provides light may be different from the plurality of patterns in which the sub-light-emitting
device 112 provides sub-light. Although FIG. 2A illustrates that the light-emitting
module 110 includes two sub-light-emitting devices 112a and 112b, the disclosure is
not limited thereto. The light-emitting module 110 may include one sub-light-emitting
device or may include three or more sub-light-emitting devices. In a case in which
the light-emitting module 110 includes the sub-light-emitting device 112 in addition
to the light-emitting device 111, changes in the shape of the shadow and the luminance
of the target object 200 may be diverse due to the light and the sub-light provided
to the target object 200. Accordingly, the resolution of the depth map 600 (see FIG.
6) generated by the depth map generation module 134 (see FIG. 1A) to be described
below may increase.
[0059] In an embodiment of the disclosure, the measuring device 120 may obtain the plurality
of captured images 400 (see FIG. 4) by photographing the target object 200 receiving
the light from the light-emitting device 111 and the sub-light from the sub-light-emitting
device 112.
[0060] FIG. 2B is a diagram of an electronic device including a light-emitting device and
a sub-light-emitting device according to an embodiment of the disclosure.
[0061] Referring to FIG. 2B, the electronic device 100a according to an embodiment of the
disclosure may include a light-emitting module 110, a measuring device 120, a memory
130a, a processor 140a, and a communication interface 150. The light-emitting module
110, the measuring device 120, the memory 130a, the processor 140a, and the communication
interface 150 may be electrically and/or physically connected to each other. Hereinafter,
the same reference numerals are assigned to the same components as those described
above with reference to FIG. 1B, and descriptions thereof may be omitted.
[0062] In an embodiment of the disclosure, the light-emitting module 110 includes the light-emitting
device 111 configured to provide light to the target object 200 (see FIG. 2A) and
the sub-light-emitting device 112 configured to provide sub-light to the target object
200.
[0063] In an embodiment of the disclosure, the memory 130a may include a light emission
control module 131a, a measurement control module 132a, an image preprocessing module
133a, and a depth map generation module 134a.
[0064] In an embodiment of the disclosure, the light emission control module 131a is configured
with instructions or program code related to an operation or function of the light-emitting
module 110. The light emission control module 131a may be configured with instructions
or program code related to operations or functions of the light-emitting device 111
and the sub-light-emitting device 112.
[0065] In an embodiment of the disclosure, the measurement control module 132a may be configured
with instructions or program code related to an operation or function, performed by
the measuring device 120, of obtaining the plurality of captured images 400 (see FIG.
4) by photographing the target object 200 receiving light and sub-light.
[0066] In an embodiment of the disclosure, the image preprocessing module 133a may be configured
with instructions or program code related to an operation or function of obtaining
the input data 500a (see FIG. 5A) by preprocessing the captured images 400. In an
embodiment of the disclosure, the image preprocessing module 133a may be configured
with instructions or program code related to an operation of function of obtaining
input data by preprocessing the plurality of captured images 400, the plurality of
illumination images 320 (see FIG. 5B), and the sub-illumination images obtained by
photographing the sub-light-emitting device 112. In an embodiment of the disclosure,
the image preprocessing module 133a is configured with instructions or program code
related to an operation of function of obtaining input data 500d (see FIG. 5D) by
preprocessing the plurality of captured images 400, the plurality of illumination
images 320, the characteristic information CI (see FIG. 5D) of the light-emitting
device 111, and sub-characteristic information SCI (see FIG. 5D) of the sub-light-emitting
device 112.
[0067] In an embodiment of the disclosure, the depth map generation module 134a is configured
with instructions or program code related to an operation or function of generating
the depth map 600 based on the input data 500.
[0068] FIG. 3A is a diagram illustrating a plurality of patterns in which the position of
a light-emitting region changes over time, according to an embodiment of the disclosure.
[0069] Referring to FIGS. 1A and 3A, the light-emitting device 111 may provide the target
object 200 with light in the plurality of patterns 300 in which the position of the
light-emitting region DA changes over time. In an embodiment of the disclosure, the
plurality of patterns 300 include first patterns 310a in which the position of the
light-emitting region DA moves in a horizontal direction over time, and second patterns
310b in which the position of the light-emitting region DA moves in a vertical direction
over time.
[0070] In an embodiment of the disclosure, one first pattern 311a includes a first light-emitting
region DA_a extending in the vertical direction, and a first non-light-emitting region
NDA_a adjacent to the first light-emitting region DA_a. The first light-emitting region
DA_a moves in the horizontal direction over time. Therefore, when the light-emitting
device 111 provides the target object 200 with light in the first patterns 310a, a
shadow of the target object 200 may be formed to move in the horizontal direction
over time in response to a change in the position of the first light-emitting region
DA_a.
[0071] In an embodiment of the disclosure, in a case in which the target object 200 includes
at least one object, the degree to which a shadow of each object moves in the horizontal
direction over time may vary depending on the distance between the light-emitting
device 111 and the object. The degree to which a shadow of an object close to the
light-emitting device 111 moves in the horizontal direction over time may be greater
than the degree to which a shadow of an object far from the light-emitting device
111 moves in the horizontal direction over time.
[0072] In an embodiment of the disclosure, one second pattern 311b includes a second light-emitting
region DA_b extending in the horizontal direction, and a second non-light-emitting
region NDA_b adjacent to the second light-emitting region DA_b. The second light-emitting
region DA_b included in one second pattern 311b moves in the vertical direction over
time. Therefore, when the light-emitting device 111 provides the target object 200
with light in the second patterns 310b, a shadow of the target object 200 may be formed
to move in the vertical direction over time in response to a change in the position
of the second light-emitting region DA_b.
[0073] In an embodiment of the disclosure, in a case in which the target object 200 includes
at least one object, the degree to which a shadow of each object moves in the vertical
direction over time may vary depending on the distance between the light-emitting
device 111 and the object. The degree to which a shadow of an object close to the
light-emitting device 111 moves in the vertical direction over time may be greater
than the degree to which a shadow of an object far from the light-emitting device
111 moves in the vertical direction over time.
[0074] Although FIG. 3A illustrates that the plurality of patterns 300 includes the first
patterns 310a and the second patterns 310b, the disclosure is not limited thereto.
The plurality of patterns 300 may include only the first patterns 310a or only the
second patterns 310b. Also, the plurality of patterns 300 may include patterns in
which the position of a light-emitting region moves in a direction intersecting a
horizontal direction or a vertical direction over time.
[0075] FIG. 3B is a diagram illustrating a plurality of patterns in which the area of a
light-emitting region changes over time, according to an embodiment of the disclosure.
[0076] Referring to FIGS. 1A and 3B, the light-emitting device 111 may provide the target
object 200 with light in the plurality of patterns 350 in which the area of the light-emitting
region DA changes over time. The plurality of patterns 350 include third patterns
in which an area of at least a partial region of the light-emitting region DA changes
over time. In an embodiment of the disclosure, one third pattern 312 includes a third
light-emitting region DA_c extending in the horizontal direction, and a third non-light-emitting
region NDA_c adjacent to the third light-emitting region DA_c. An area of at least
a partial region of the third light-emitting region DA_c changes over time. For example,
the area of at least a partial region of the third light-emitting region DA_c increases
over time. Although FIG. 3B illustrates that the third light-emitting region DA_c
has a shape extending in the horizontal direction and the shape of the third light-emitting
region DA_c increases in the vertical direction over time, the disclosure is limited
thereto. Each of the plurality of patterns 300 may include the light-emitting region
DA having a shape that extends in the vertical direction and increases in the horizontal
direction over time.
[0077] In an embodiment of the disclosure, the plurality of patterns 350 may include patterns
in which the shape of the light-emitting region DA changes in the vertical direction
over time, and patterns in which the shape of the light-emitting region DA changes
in the horizontal direction over time. In an embodiment of the disclosure, the light-emitting
region DA included in the plurality of patterns 350 may have a particular shape, for
example, a quadrangle, a triangle, a trapezoid, or a circle, the size of which may
change over time. When the light-emitting device 111 provides the target object 200
with light in the plurality of patterns 350 in which the area of the light-emitting
region DA changes over time, the shape of a shadow of the target object 200 may change
in response to a change in the area of the light-emitting region DA.
[0078] In an embodiment of the disclosure, in a case in which the target object 200 includes
at least one object, the degree to which the shape of a shadow of each object changes
over time may vary depending on the distance between the light-emitting device 111
and the object. The degree to which the shape of a shadow of an object close to the
light-emitting device 111 changes over time may be greater than the degree to which
the shape of a shadow of an object far from the light-emitting device 111 changes
over time.
[0079] FIG. 3C is a diagram illustrating a plurality of patterns in which the illumination
of light changes over time, according to an embodiment of the disclosure.
[0080] Referring to FIGS. 1A and 3C, the light-emitting device 111 may provide the target
object 200 with light in the plurality of patterns 360 in which the illumination of
the light changes over time. The plurality of patterns 360 include fourth patterns
in which the illumination of the light changes over time.
[0081] In an embodiment of the disclosure, one fourth pattern 313 includes a fourth light-emitting
region DA_d and a fourth non-light-emitting region NDA_d adjacent to the fourth light-emitting
region. Although FIG. 3C illustrates that the fourth light-emitting region DA_d has
a circular shape, the disclosure is not limited thereto. The fourth light-emitting
region DA_d may have a polygonal shape such as a quadrangle. In addition, one fourth
pattern 313 may not include the fourth non-light-emitting region NDA_d. The entire
fourth pattern 313 may be a light-emitting region that provides light. The illumination
of light provided from the fourth light-emitting region DA_d may change over time.
[0082] In an embodiment of the disclosure, the illumination of the light provided from the
fourth light-emitting region DA_d may increase over time. When the light-emitting
device 111 provides the target object 200 with light in the plurality of patterns
360 in which the illumination of the light increases over time, the shape of a shadow
of the target object 200 may change in response to a change in the illumination of
the light.
[0083] In an embodiment of the disclosure, when the illumination of the light provided to
the target object 200 by the light-emitting device 111 increases over time, the shape
of a shadow of the target object 200 may also increase over time. In an embodiment
of the disclosure, in a case in which the target object 200 includes at least one
object, the degree to which the shape of a shadow of each object changes over time
may vary depending on the distance between the light-emitting device 111 and the object.
The degree to which the shape of a shadow of an object close to the light-emitting
device 111 changes over time may be greater than the degree to which the shape of
a shadow of an object far from the light-emitting device 111 changes over time. In
addition, when the light-emitting device 111 provides the target object 200 with light
in the plurality of patterns 360 in which the illumination of the light increases
over time, the luminance of the target object 200 may change in response to a change
in the illumination of the light.
[0084] In an embodiment of the disclosure, when the illumination of the light provided to
the target object 200 by the light-emitting device 111 increases over time, the luminance
of the target object 200 may also increase over time. In an embodiment of the disclosure,
in a case in which the target object 200 includes at least one object, the degree
to which the luminance of each object changes over time may vary depending on the
distance between the light-emitting device 111 and the object. The degree to which
the luminance of an object close to the light-emitting device 111 changes over time
may be greater than the degree to which the luminance of an object far from the light-emitting
device 111 changes over time.
[0085] FIG. 3D is a diagram illustrating a plurality of patterns in which the shape of a
light-emitting region changes over time, according to an embodiment of the disclosure.
[0086] Referring to FIGS. 1A and 3D, the light-emitting device 111 may provide the target
object 200 with light in the plurality of patterns 370 in which the shape of the light-emitting
region DA changes over time. The plurality of patterns 370 include fifth patterns
in which the shape of the light-emitting region DA changes over time.
[0087] In an embodiment of the disclosure, one fifth pattern 314 includes a fifth light-emitting
region DA_e and a fifth non-light-emitting region NDA_e adjacent to the fifth light-emitting
region. The shape of the fifth light-emitting region DA_e changes over time. Referring
to FIG. 3D, the shape of the fifth light-emitting region DA_e includes a circular
shape that changes in size over time, and a quadrangular shape that changes in size
over time. However, the disclosure is not limited thereto. The plurality of patterns
370 may include the light-emitting region DA that changes over time into various shapes
such as a circle, a quadrangle, a triangle, or a trapezoid, the size of which may
also change.
[0088] In an embodiment of the disclosure, when the light-emitting device 111 provides the
target object 200 with light in the plurality of patterns 370 in which the shape of
the light-emitting region DA changes over time, the shape of a shadow of the target
object 200 may change in response to a change in the shape of the light-emitting region
DA.
[0089] In an embodiment of the disclosure, in a case in which the target object 200 includes
at least one object, the degree to which the shape of a shadow of each object changes
over time may vary depending on the distance between the light-emitting device 111
and the object. The degree to which the shape of a shadow of an object close to the
light-emitting device 111 changes over time may be greater than the degree to which
the shape of a shadow of an object far from the light-emitting device 111 changes
over time.
[0090] In an embodiment of the disclosure, when the light-emitting device 111 provides the
target object 200 with light in the plurality of patterns 370 in which the shape of
the light-emitting region DA changes over time, the luminance of a shadow of the target
object 200 may change in response to a change in the shape of the light-emitting region
DA. In an embodiment of the disclosure, in a case in which the target object 200 includes
at least one object, the degree to which the luminance of each object changes over
time may vary depending on the distance between the light-emitting device 111 and
the object. The degree to which the luminance of an object close to the light-emitting
device 111 changes over time may be greater than the degree to which the luminance
of an object far from the light-emitting device 111 changes over time.
[0091] In an embodiment of the disclosure, the plurality of illumination images 320 (see
FIG. 5B) respectively corresponding to the plurality of patterns may be obtained using
the illumination measuring device to photograph the light-emitting device 111. Patterns
of light provided to the target object 200 by the light-emitting device 111 may be
identified from the plurality of illumination images 320. In an embodiment of the
disclosure, the electronic device 100 may receive a plurality of previously captured
illumination images 320, from an external server or nearby electronic devices through
the communication interface 150 (see FIG. 1B). In this case, the light emission control
module 131 may be configured with instructions or program code related to an operation
or function, performed by the light-emitting device 111, of providing light to the
target object 200 based on the received plurality of illumination images 320.
[0092] In an embodiment of the disclosure, the light emission control module 131 (see FIG.
1B) is configured with instructions or program code related to an operation of function,
performed by the light-emitting device 111, of providing the target object 200 with
light in the plurality of patterns that change over time, as illustrated in FIGS.
3A to 3D. However, the disclosure is not limited thereto, and the light emission control
module 131 may cause the light-emitting device 111 to provide the target object 200
with light in the plurality of patterns that may cause a change in the shape of a
shadow of the target object 200 or a change in the luminance of the target object
200 over time.
[0093] FIG. 4 is a diagram illustrating a plurality of captured images respectively corresponding
to a plurality of patterns according to an embodiment of the disclosure.
[0094] Referring to FIGS. 1A, 3A and 4, the measuring device 120 may obtain the plurality
of captured images 400 respectively corresponding to a plurality of patterns, by photographing
the target object 200 receiving light having the plurality of patterns from the light-emitting
device 111. Hereinafter, the same reference numerals are assigned to the same components
as those described above with reference to FIGS. 1A and 3A, and descriptions thereof
may be omitted.
[0095] FIG. 4 illustrates that the light-emitting device 111 provides the target object
200 with light in the plurality of patterns 300 in which the position of the light-emitting
region DA changes over time. In an embodiment of the disclosure, the plurality of
patterns 300 include the first patterns 310a and the second patterns 310b. When a
time period required for the light-emitting device 111 to provide the target object
200 with light having the first patterns 310a and the second patterns 310b is referred
to as one period, the measuring device 120 may obtain the plurality of captured images
400 respectively corresponding to the first patterns 310a and the second patterns
310b, by photographing the target object 200 for one period.
[0096] In an embodiment of the disclosure, the plurality of captured images 400 include
first captured images 410a respectively corresponding to the first patterns 310a,
and second captured images 410b respectively corresponding to the second patterns
310b.
[0097] In an embodiment of the disclosure, the shape of a shadow of the target object 200
and the luminance of the target object 200 included in each of the first captured
images 410a may change in response to a change in the position of the first light-emitting
region DA_a included in each of the first patterns 310a.
[0098] In an embodiment of the disclosure, among the first patterns 310a, a pattern in which
the first light-emitting region DA_a is at the left end is referred to as a first
sub-pattern 311a_1. Among the first patterns 310a, a pattern in which the first light-emitting
region DA_a is at the center is referred to as a second sub-pattern 311a_2. Among
the first patterns 310a, a pattern in which the first light-emitting region DA_a is
at the right end is referred to as a third sub-pattern 311a_3. Among the first captured
images 410a, an image corresponding to the first sub-pattern 311a_1 is referred to
as a first sub-captured image 411a_1. Among the first captured images 410a, an image
corresponding to the second sub-pattern 311a_2 is referred to as a second sub-captured
image 411a_2. Among the first captured images 410a, an image corresponding to the
third sub-pattern 311a_3 is referred to as a third sub-captured image 411a_3.
[0099] In an embodiment of the disclosure, when the light-emitting device 111 provides the
target object 200 with light in the first sub-pattern 311a_1, a shadow of the target
object 200 included in the first sub-captured image 411a_1 has a shape extending to
the right. In addition, the luminance of a left portion of the target object 200 included
in the first sub-captured image 411a_1 is greater than the luminance of a central
portion or a right portion of the target object 200. When the light-emitting device
111 provides the target object 200 with light in the second sub-pattern 311a_2, a
shadow of the target object 200 included in the second sub-captured image 411a_2 has
a shape extending to the center. In addition, the luminance of a central portion of
the target object 200 included in the second sub-captured image 411a_2 is greater
than the luminance of a left portion or a right portion of the target object 200.
When the light-emitting device 111 provides the target object 200 with light in the
third sub-pattern 311a_3, a shadow of the target object 200 included in the third
sub-captured image 411a_3 has a shape extending to the left. In addition, the luminance
of a right portion of the target object 200 included in the third sub-captured image
411a_3 is greater than the luminance of a left portion or a central portion of the
target object 200.
[0100] In an embodiment of the disclosure, the shape of a shadow of the target object 200
and the luminance of the target object 200 included in each of the second captured
images 410b may change in response to a change in the position of the second light-emitting
region DA_b included in each of the second patterns 310b.
[0101] In an embodiment of the disclosure, among the second patterns 310b, a pattern in
which the second light-emitting region DA_b is at the lower end is referred to as
a fourth sub-pattern 311b_1. Among the second patterns 310b, a pattern in which the
second light-emitting region DA_b is at the center is referred to as a fifth sub-pattern
311b_2. Among the second patterns 310b, a pattern in which the second light-emitting
region DA_b is at the upper end is referred to as a sixth sub-pattern 311b_3. Among
the second captured images 410b, an image corresponding to the fourth sub-pattern
311b_1 is referred to as a fourth sub-captured image 411b_1. Among the second captured
images 410b, an image corresponding to the fifth sub-pattern 311b_2 is referred to
as a fifth sub-captured image 411b_2. Among the second captured images 410b, an image
corresponding to the sixth sub-pattern 311b_3 is referred to as a sixth sub-captured
image 411b_3.
[0102] In an embodiment of the disclosure, when the light-emitting device 111 provides the
target object 200 with light in the fourth sub-pattern 311b_1, a shadow of the target
object 200 included in the fourth sub-captured image 411b_1 has a shape extending
upward. In addition, the luminance of a lower portion of the target object 200 included
in the fourth sub-captured image 411b_1 is greater than the luminance of a central
portion or an upper portion of the target object 200. When the light-emitting device
111 provides the target object 200 with light in the fifth sub-pattern 311b_2, a shadow
of the target object 200 included in the fifth sub-captured image 411b_2 has a shape
extending to the center. In addition, the luminance of a central portion of the target
object 200 included in the fifth sub-captured image 411b_2 is greater than the luminance
of a lower portion or an upper portion of the target object 200. When the light-emitting
device 111 provides the target object 200 with light in the sixth sub-pattern 311b_3,
a shadow of the target object 200 included in the sixth sub-captured image 411b_3
has a shape extending downward. In addition, the luminance of an upper portion of
the target object 200 included in the sixth sub-captured image 411b_3 is greater than
the luminance of a lower portion or a central portion of the target object 200.
[0103] Although FIG. 4 illustrates only the plurality of captured images 400 when the light-emitting
device 111 provides the target object 200 with light in the first patterns 310a and
the second patterns 310b, the disclosure is not limited thereto. In an embodiment
of the disclosure, when the light-emitting device 111 provides the target object 200
with light in the third to fifth patterns or other patterns that change over time,
the measuring device 120 may also obtain the plurality of captured images 400 respective
corresponding to the patterns.
[0104] In an embodiment of the disclosure, the measurement control module 132 (see FIG.
1B) is configured with instructions or program code related to an operation or function,
performed by the measuring device 120, of obtaining the plurality of captured images
400 respectively corresponding to the plurality of patterns 300 as illustrated in
FIG. 4. However, the disclosure is not limited thereto, and the measurement control
module 132 may be implemented with instructions or program code to cause the measuring
device 120 to obtain the plurality of captured images 400 including a change in the
shape of a shadow of the target object 200 or a change in the luminance of the target
object 200 that occurs in response to a plurality of various patterns 300 in which
the light-emitting device 111 provides light.
[0105] FIG. 5A is a diagram of an operation, performed by an image preprocessing module,
of obtaining input data by preprocessing a plurality of captured images, according
to an embodiment of the disclosure.
[0106] Referring to FIGS. 1B and 5A, the image preprocessing module 133 obtains the input
data 500a by preprocessing the plurality of captured images 400. In an embodiment
of the disclosure, the image preprocessing module 133 preprocesses the plurality of
captured images 400 to obtain the input data 500a to be used by the depth map generation
module 134 to generate the depth map 600 (see FIG. 6).
[0107] In an embodiment of the disclosure, the image preprocessing module 133 obtains the
input data 500a through processes of performing wrangling, transformation, integration,
and the like on the plurality of captured images 400. In an embodiment of the disclosure,
in a case in which the depth map generation module 134 is configured as an artificial
intelligence model, the image preprocessing module 133a may obtain the input data
500a in which epochs and batches for training the artificial intelligence model are
set, by preprocessing the plurality of captured images 400. FIG. 5A illustrates that
the image preprocessing module 133 preprocesses the plurality of captured images 400
to obtain the input data 500a including a plurality of preprocessed captured images
510.
[0108] FIG. 5B is a diagram of an image preprocessing module configured to obtain input
data by preprocessing a plurality of captured images and a plurality of illumination
images, according to an embodiment of the disclosure.
[0109] Referring to FIGS. 1B, 4, and 5B, the plurality of captured images 400 and the plurality
of illumination images 320 may be provided to the image preprocessing module 133.
In an embodiment of the disclosure, the plurality of captured images 400 may correspond
to the plurality of illumination images 320, respectively. In an embodiment of the
disclosure, the image preprocessing module 133 obtains the input data 500b by preprocessing
the plurality of captured images 400 and the plurality of illumination images 320.
[0110] In an embodiment of the disclosure, the input data 500b includes the plurality of
preprocessed captured images 510 and a plurality of preprocessed illumination images
520. The plurality of preprocessed captured images 510 may correspond to the plurality
of preprocessed illumination images 520, respectively.
[0111] FIG. 5C is a diagram of an image preprocessing module configured to obtain input
data by preprocessing a plurality of captured images, a plurality of illumination
images, and characteristic information, according to an embodiment of the disclosure.
[0112] Referring to FIGS. 1A, 1B, and 5C, the plurality of captured images 400, the plurality
of illumination images 320, and the characteristic information CI including information
about the light-emitting device 111 may be provided to the image preprocessing module
133. In an embodiment of the disclosure, the characteristic information CI may include
information about the size of the light-emitting device 111. However, the disclosure
is not limited thereto. The characteristic information CI may include information
about a position where the light-emitting device 111 is arranged, information about
the distance between the light-emitting device 111 and the target object 200, information
about the shape of the light-emitting device 111, information about the illumination
of light provided by the light-emitting device 111, information about the direction
of light provided by the light-emitting device 111, and the like.
[0113] In an embodiment of the disclosure, in a case in which the characteristic information
CI includes information about the size of the light-emitting device 111, and the light-emitting
device 111 is a TV 531, the characteristic information CI may include information
about a horizontal length WD1 and a vertical length WD2 of the light-emitting device
111. However, the disclosure is not limited thereto. In a case in which the light-emitting
device 111 is a device having a quadrangular shape other than a TV 531, the characteristic
information CI may include information about the horizontal length WD1 and the vertical
length WD2 of the light-emitting device 111. In an embodiment of the disclosure, in
a case in which the light-emitting device 111 has a round shape, the characteristic
information CI may include information such as the radius of the light-emitting device
111. The image preprocessing module 133 obtains the input data 500c by preprocessing
the plurality of captured images 400, the plurality of illumination images 320, and
the characteristic information CI.
[0114] In an embodiment of the disclosure, the input data 500c may include the plurality
of preprocessed captured images 510, the plurality of preprocessed illumination images
520, and preprocessed characteristic information 530 (e.g., information about the
lengths WD1 and WD2 of the TV 531).
[0115] FIG. 5D is a diagram of an image preprocessing module configured to obtain input
data by preprocessing a plurality of captured images, a plurality of illumination
images, characteristic information, and sub-characteristic information, according
to an embodiment of the disclosure.
[0116] Referring to FIGS. 1A, 1B, and 5D, the plurality of captured images 400, the plurality
of illumination images 320, the characteristic information CI, and the sub-characteristic
information SCI including information about the sub-light-emitting device 112 may
be provided to the image preprocessing module 133. Hereinafter, the same reference
numerals are assigned to the same components as those described above with reference
to FIG. 5C, and descriptions thereof may be omitted.
[0117] In an embodiment of the disclosure, the characteristic information CI may include
information about the size of the light-emitting device 111.
[0118] In an embodiment of the disclosure, the sub-characteristic information SCI may include
information about a position where the sub-light-emitting device 112 is arranged,
information about the distance between the sub-light-emitting device 112 and the target
object 200, information about the shape of the sub-light-emitting device 112, information
about the illumination of light provided by the sub-light-emitting device 112, information
about the direction of light provided by the sub-light-emitting devices 112, information
about the number of sub-light-emitting devices 112, and the like.
[0119] In an embodiment of the disclosure, the image preprocessing module 133 obtain the
input data 500d by preprocessing the plurality of captured images 400, the plurality
of illumination images 320, the characteristic information CI, and the sub-characteristic
information SCI.
[0120] In an embodiment of the disclosure, the input data 500d may include the plurality
of preprocessed captured images 510, the plurality of preprocessed illumination images
520, the preprocessed characteristic information 530, and preprocessed sub-characteristic
information 540.
[0121] FIG. 6 is a diagram of an operation of a depth map generation module according to
an embodiment of the disclosure.
[0122] Referring to FIGS. 5A, 5B, 5C, 5D and 6, the depth map generation module 134 generates
the depth map 600 based on the input data 500. The depth map generation module 134
generates, based on the input data 500, the depth map 600 including information related
to the distance between the measuring device 120 (see FIG. 1A) and a surface of the
target object 200 (see FIG. 1A).
[0123] In an embodiment of the disclosure, in a case in which the measuring device 120 is
collinear with the light-emitting device 111 (see FIG. 1A), the depth map 600 may
include information related to the distance between the light-emitting device 111
and the surface of the target object 200.
[0124] In an embodiment of the disclosure, the depth map generation module 134 compares
the distances between pixels present in the plurality of preprocessed captured images
510 included in the input data 500 and the measuring device 120 with each other, and
generate the depth map 600 representing relative distances between the pixels and
the measuring device 120. The plurality of preprocessed captured images 510 include
information about shadows and luminances of the target object 200 that change in response
to light provided in the plurality of patterns 300 (see FIG. 3A) that change over
time.
[0125] In an embodiment of the disclosure, the target object 200 may include a plurality
of objects. In a case in which the distances between the objects and the light-emitting
device 111 and the shapes of the objects are different from each other, the degree
to which the shape of a shadow and the luminance of the target object 200 change in
response to a change in the pattern of light provided from the light-emitting device
111 may vary. Accordingly, accuracy of calculating the distance between each pixel
present in the plurality of preprocessed captured images 510 and the measuring device
120 may increase. Therefore, the resolution of a depth map generated using the preprocessed
captured images 510 of the disclosure may be higher than the resolution of a related-art
depth map generated based on images captured without changing the pattern of light
provided to the target object 200.
[0126] In an embodiment of the disclosure, the depth map generation module 134 may generate
the depth map 600 based on the plurality of preprocessed captured images 510 and the
plurality of preprocessed illumination images 520. In this case, in generating the
depth map 600, the depth map generation module 134 may accurately reflect changes
in the pattern of light provided to the target object 200, thereby generating the
depth map 600 with high resolution.
[0127] In an embodiment of the disclosure, the depth map generation module 134 may generate
the depth map 600 based on the plurality of preprocessed captured images 510, the
plurality of preprocessed illumination images 520, and the preprocessed characteristic
information 530. In this case, in generating the depth map 600, the depth map generation
module 134 may additionally reflect the characteristics of the light-emitting device
111, for example, the size of the light-emitting device 111, the position of the light-emitting
device 111, the shape of the light-emitting device 111, the illumination of light
provided by the light-emitting device 111, and the distance between the light-emitting
device 111 and the target object 200, thereby generating the depth map 600 with high
resolution.
[0128] In an embodiment of the disclosure, the depth map generation module 134 may generate
the depth map 600 based on the plurality of preprocessed captured images 510, the
plurality of preprocessed illumination images 520, the preprocessed characteristic
information 530, and the preprocessed sub-characteristic information 540. In this
case, in generating the depth map 600, the depth map generation module 134 may additionally
reflect the characteristics of the sub-light-emitting device 112, for example, the
size of the sub-light-emitting device 112, the position of the sub-light-emitting
device 112, the shape of the sub-light-emitting device 112, the illumination of light
provided by the sub-light-emitting device 112, and the distance between the sub-light-emitting
device 112 and the target object 200, thereby generating the depth map 600 with high
resolution.
[0129] FIG. 7A is a diagram of a depth map generation module configured to generate a depth
map based on input data generated by preprocessing a plurality of captured images
and a plurality of illumination images, according to an embodiment of the disclosure.
[0130] Referring to FIGS. 5B, 6, and 7A, the depth map generation module 134 may generate
a depth map 600a based on the input data 500b including the plurality of preprocessed
captured images 510 and the plurality of preprocessed illumination images 520.
[0131] In an embodiment of the disclosure, the plurality of preprocessed captured images
510 correspond to the plurality of preprocessed illumination images 520, respectively.
The depth map generation module 134 may generate the depth map 600a with high resolution,
based on the plurality of preprocessed captured images 510 including changes in the
shapes of shadows and luminances of the target object 200 (see FIG. 1A) that occur
in response to changes in the position of the light-emitting region DA_a (see FIG.
4) included in the plurality of preprocessed illumination images 520.
[0132] FIG. 7B is a diagram of a depth map generation module configured to generate a depth
map based on input data generated by preprocessing a plurality of captured images,
a plurality of illumination images, and characteristic information, according to an
embodiment of the disclosure.
[0133] Referring to FIGS. 5C, 6, and 7B, the depth map generation module 134 may generate
a depth map 600b based on the input data 500c including the plurality of preprocessed
captured images 510, the plurality of preprocessed illumination images 520, and the
preprocessed characteristic information 530. Although FIG. 7B illustrates that the
preprocessed characteristic information 530 includes information about the size of
the light-emitting device 111 (see FIG. 1A), the disclosure is not limited thereto.
[0134] In an embodiment of the disclosure, the depth map generation module 134 may generate
the depth map 600b with high resolution, based on changes in the shapes of shadows
and luminances of the target object 200 (see FIG. 1A) according to changes in the
position of the light-emitting region DA_a (see FIG. 4), and characteristic information
of the light-emitting device 111.
[0135] FIG. 8 is a diagram of a structure of a depth map generation module according to
an embodiment of the disclosure.
[0136] Referring to FIG. 8, the depth map generation module 134 may include an artificial
intelligence model. In an embodiment of the disclosure, the depth map generation module
134 may include a machine learning or deep learning model. FIG. 8 illustrates that
the depth map generation module 134 is configured with an autoencoder. However, the
disclosure is not limited thereto, and the depth map generation module 134 may be
configured with a generative model capable of generating the depth map 600b based
on the input data 500c, for example, a generative adversarial network (GAN) or a variational
autoencoder (VAE).
[0137] In an embodiment of the disclosure, the depth map generation module 134 includes
an encoder 135 and a decoder 136.
[0138] In an embodiment of the disclosure, encoder 135 may include at least one neural network
layer. Each neural network layer may include at least one convolutional layer 135_a
configured to perform a convolution operation, at least one activation function 135_b
for determining activation based on input data, and at least one pooling layer 135_c
configured to extract feature values.
[0139] In an embodiment of the disclosure, the decoder 136 may include a plurality of neural
network layers. Each neural network layer may include at least one unpooling layer
136_a configured to obtain original data based on the extracted feature values, at
least one deconvolutional layer 136_b configured to perform a deconvolution operation,
and at least one activation function 136_c for determining activation based on the
input data.
[0140] In an embodiment of the disclosure, the depth map generation module 134 may receive
the input data 500c and generate the depth map 600b. In an embodiment of the disclosure,
the input data 500c may include the plurality of preprocessed captured images 510,
the plurality of preprocessed illumination images 520, and the preprocessed characteristic
information 530. The depth map generation module 134 may obtain, based on the plurality
of preprocessed illumination images 520, information about light provided to the target
object 200 (see FIG. 1A) in a pattern that changes over time. In an embodiment of
the disclosure, the depth map generation module 134 may obtain, based on the plurality
of preprocessed captured images 510, information about the shape of a shadow and the
luminance of the target object 200 that changes in response to the light provided
in the pattern that changes over time.
[0141] In an embodiment of the disclosure, the depth map generation module 134 may obtain
information about the size of the light-emitting device 111 (see FIG. 1A), based on
the preprocessed characteristic information 530. The depth map generation module 134
may generate the depth map 600b of the target object 200 based on the fact that a
change in the shape of a shadow of an object in response to a change in the pattern
of light provided from a light source increases as the distance between the light
source and the object decreases. In addition, the depth map generation module 134
may generate the depth map 600b of the target object 200 based on the fact that a
change in the luminance of the object in response to a change in the pattern of light
provided from the light source increases as the distance between the light source
and the object decreases.
[0142] In an embodiment of the disclosure, in a case in which the target object 200 includes
a plurality of objects, the depth map generation module 134 may obtain information
that an object with a large change in the shape of a shadow in response to a change
in provided light is closer to the measuring device 120 than an object with a small
change in the shape of a shadow. The depth map generation module 134 may obtain information
that an object with high luminance is closer to the measuring device 120 than an object
with low luminance. Also, the depth map generation module 134 may obtain information
that an object with a large change in the luminance in response to a change in provided
light is closer to the measuring device 120 than an object with a small change in
the luminance.
[0143] In an embodiment of the disclosure, the depth map generation module 134 may extract
the above-described information using the encoder 135, and generate the depth map
600b of the target object 200 with high resolution using the decoder 136 based on
the extracted information.
[0144] In an embodiment of the disclosure, in a case in which the measuring device 120 includes
an RGB camera, the plurality of preprocessed captured images 510 may include RGB images
of the target object 200. In this case, each of the plurality of preprocessed captured
images 510 includes data of horizontal and vertical R, G, and B pixels corresponding
to the size of the target object 200.
[0145] In an embodiment of the disclosure, the plurality of preprocessed illumination images
520 may include data of horizontal and vertical pixels corresponding to the gray-based
light-emitting region DA (see FIG. 1A) and non-light-emitting region NDA (see FIG.
1A) representing luminance. The preprocessed characteristic information 530 may include
data of the horizontal and vertical sizes of the light-emitting device 111 (e.g.,
TV 531).
[0146] In an embodiment of the disclosure, the depth map generation module 134 generate,
based on the input data 500c, the depth map 600b including depth information and data
of horizontal and vertical pixels corresponding to the size of the target object 200.
[0147] In an embodiment of the disclosure, the depth map generation module 134 may be an
autoencoder that is trained, based on a training dataset, to generate the depth map
600b based on captured images of the target object 200.
[0148] In an embodiment of the disclosure, the training dataset may be a dataset including
the plurality of captured images 400 obtained by the electronic device 100 of the
disclosure. The training dataset may be a dataset including the plurality of captured
images 400 (see FIG. 5B) and the plurality of illumination images 320 (see FIG. 5B)
obtained by the electronic device 100 of the disclosure. The training dataset may
be a dataset including the plurality of captured images 400, the plurality of illumination
images 320, and the characteristic information CI of the disclosure. In addition,
in an embodiment of the disclosure, the training dataset may be a dataset including
the input data 500c obtained by the electronic device 100 of the disclosure.
[0149] In an embodiment of the disclosure, the depth map generation module 134 may be trained
to generate the depth map 600b from captured images of the target object 200, using
the training dataset. The depth map generation module 134 may update weights of the
encoder 135 and the decoder 136 using the training dataset. In an embodiment of the
disclosure, the depth map generation module 134 may repeat a process of updating the
weights of the encoder 135 and the decoder 136 multiple times using training data.
[0150] In an embodiment of the disclosure, each batch for training the depth map generation
module 134 may include the plurality of preprocessed illumination images 520 and the
plurality of preprocessed captured images 510. In an embodiment of the disclosure,
the depth map generation module 134 may perform transfer learning using a pre-trained
model, so as to learn to generate the depth map 600b using a training dataset. In
an embodiment of the disclosure, DenseNet169, ResNet50, or the like may be used for
the pre-trained model. The depth map generation module 134 may calculate a loss function
of the artificial intelligence model included in the depth map generation module 134,
using the pre-trained model to analyze the training data.
[0151] In an embodiment of the disclosure, the depth map generation module 134 may update
the weights of the encoder 135 and the decoder 136, based on a loss value of the loss
function. However, the disclosure is not limited thereto, and the depth map generation
module 134 may receive, from an external server or nearby electronic devices, an artificial
intelligence model previously trained using the training dataset.
[0152] FIG. 9 is a diagram of a training process of a depth map generation module according
to an embodiment of the disclosure.
[0153] FIG. 9 illustrates a captured image 700 of the target object 200, a first depth map
600_1, a second depth map 600_2, a third depth map 600_3, and a fourth depth map 600_4
for describing a learning level of an artificial intelligence model. Referring to
the captured image 700, the target object 200 includes a plurality of objects arranged
at different positions from the measuring device 120 and having different shapes.
[0154] In an embodiment of the disclosure, the captured image 700 includes a first region
AA including objects arranged adjacent to each other, and a second area BB including
small objects that are difficult to be distinguished in a picture.
[0155] In an embodiment of the disclosure, the first depth map 600_1 is a depth map generated
by the depth map generation module 134 that has repeated, for a `j' number of times,
learning based on the plurality of captured images 400 obtained through a process
in which the light-emitting device 111 (see FIG. 1A) provides light in a constant
pattern regardless of time, and the measuring device 120 photographs the target object
200 receiving the light.
[0156] In an embodiment of the disclosure, the second depth map 600_2 is a depth map generated
by the depth map generation module 134 that has repeated, for a `j' number of times,
learning based on the plurality of captured images 400 obtained through a process
in which the light-emitting device 111 provides light in a `k' number of patterns
that change over time, and the measuring device 120 photographs the target object
200 receiving the light.
[0157] In an embodiment of the disclosure, the third depth map 600_3 is a depth map generated
by the depth map generation module 134 that has repeated, for a `j' number of times,
learning based on the plurality of captured images 400 obtained through a process
in which the light-emitting device 111 provides light in an 'I' number of patterns
that change over time, and the measuring device 120 photographs the target object
200 receiving the light.
[0158] In an embodiment of the disclosure, the fourth depth map 600_4 is a depth map generated
by the depth map generation module 134 that has repeated, for an 'm' number of times,
learning based on the plurality of captured images 400 obtained through a process
in which the light-emitting device 111 provides light in an 'I' number of patterns
that change over time, and the measuring device 120 photographs the target object
200 receiving the light. `j', 'k', 'I', and 'm' may each be natural numbers, where
'm' is greater than `j', and 'I' is greater than `k'.
[0159] In an embodiment of the disclosure, it may be confirmed, by comparing the first depth
map 600_1 with the second depth map 600_2, that when the depth map generation module
134 is trained based on the plurality of captured images 400 obtained by photographing
the target object 200 while providing light to the target object 200 in patterns that
change over time, the learning level of the artificial intelligence model increases.
[0160] For example, the resolution of a first region AA_b included in the second depth map
600_2 is higher than the resolution of a first area AA_a included in the first depth
map 600_1, and thus, objects arranged adjacent to each other may be distinguished.
In addition, the resolution of a second region BB_b included in the second depth map
600_2 is higher than the resolution of a second region BB_a included in the first
depth map 600_1, and thus, small objects may be distinguished.
[0161] In an embodiment of the disclosure, it may be confirmed, by comparing the second
depth map 600_2 with the third depth map 600_3, that as the number of patterns in
which light is provided to the target object 200 increases, the learning level of
the artificial intelligence model trained based on the plurality of captured images
400 increases.
[0162] For example, the resolution of a first region AA_c included in the third depth map
600_3 is higher than the resolution of the first area AA_b included in the second
depth map 600_2, and thus, objects arranged adjacent to each other may be distinguished.
In addition, the resolution of a second region BB_c included in the third depth map
600_3 is higher than the resolution of the second region BB_b included in the second
depth map 600_2, and thus, small objects may be distinguished.
[0163] Thus, the resolution of the depth map 600b (see FIG. 8) generated by the artificial
intelligence model trained based on the input data 500c (see FIG. 8) of the present
disclosure is higher than the resolution of a depth map generated by an artificial
intelligence model that trained based on images obtained by simply photographing the
target object 200 without changing the pattern of light provided to the target object
200.
[0164] For example, the input data 500c of the disclosure includes information about changes
in the shape of the shadows and the luminance of the target object 200 according to
the arrangements, shapes, distances from the measuring device 120, and the like of
objects included in the target object 200, and thus, the learning level of the artificial
intelligence model trained based on the input data 500c may be higher. In addition,
by training, based on the input data 500c of the disclosure, an artificial intelligence
model for generating the depth map 600b, the artificial intelligence model may be
trained with a smaller amount of training dataset than that in a case in which the
artificial intelligence model is trained based on images obtained by simply photographing
the target object 200 without changing the pattern of light provided to the target
object 200.
[0165] In an embodiment of the disclosure, it may be confirmed, by comparing the third depth
map 600_3 with the fourth depth map 600_4, that as the number of epochs, which is
the number of times the depth map generation module 134 (see FIG. 8) is trained based
on the input data 500c, increases, the learning level of the trained artificial intelligence
model increases.
[0166] For example, the resolution of a first region AA_d included in the fourth depth map
600_4 is higher than the resolution of the first area AA_c included in the third depth
map 600_3, and thus, objects arranged adjacent to each other may be distinguished.
In addition, the resolution of a second region BB_d included in the fourth depth map
600_4 is higher than the resolution of the second region BB_c included in the third
depth map 600_3, and thus, small objects may be distinguished.
[0167] FIG. 10 is a flowchart of a method of generating a depth map using an electronic
device including a light-emitting device, according to an embodiment of the disclosure.
[0168] Referring to FIGS. 1A, 3A and 10, in operation S1000, the electronic device 100 may
provide the target object 200 with light in the plurality of patterns (e.g., patterns
300) that change over time.
[0169] In an embodiment of the disclosure, the electronic device 100 provides the light
to the target object 200 using the light-emitting device 111. The light-emitting device
111 provides the target object 200 with the light in the plurality of patterns 300
that change over time. In an embodiment of the disclosure, depending on the type of
the electronic device 100, the light-emitting device 111 may perform a function of
displaying an image or providing light to an environment surrounding the electronic
device 100. When the electronic device 100 generates the depth map 600 (see FIG. 6)
of the target object 200, the light-emitting device 111 provides the target object
200 with light in the plurality of patterns 300.
[0170] Referring to FIGS. 1A, 4, and 10, in operation S2000, the electronic device 100 obtains
the plurality of captured images 400 respectively corresponding to the plurality of
patterns 300, by photographing the target object 200 receiving the light.
[0171] In an embodiment of the disclosure, the electronic device 100 may obtain the plurality
of captured images 400 using the measuring device 120 to photograph the target object
200. The measuring device 120 may photograph the target object 200 in synchronization
with a change in the pattern of the light provided by the light-emitting device 111.
[0172] In an embodiment of the disclosure, the measuring device 120 may obtain the plurality
of captured images 400 by photographing the target object 200, a shadow of which changes
in response to changes in the pattern of the light received by the target object 200.
In addition, the measuring device 120 may obtain the plurality of captured images
400 by photographing the target object 200, the luminance of which changes in response
to the changes in the pattern of the light received by the target object 200.
[0173] In an embodiment of the disclosure, each of the plurality of captured images 400
includes changes in the shape of a shadow and the luminance of the target object 200
according to the changes in the pattern of the light.
[0174] In an embodiment of the disclosure, the measuring device 120 may be an RGB-depth
camera configured to obtain an RGB image and a depth image. In an embodiment of the
disclosure, measuring device 120 may use an Intel RealSense D455 camera. In this case,
the plurality of captured images 400 may include RGB images and simple depth map images
of the target object 200. However, the disclosure is not limited thereto, and the
measuring device 120 may be an RGB camera configured to obtain an RGB image. In this
case, the plurality of captured images 400 may include RGB images of the target object
200. Even in a case in which the plurality of captured images 400 do not include simple
depth map images, the depth map 600 of the target object 200 may be generated through
the image preprocessing module 133 and the depth map generation module 134.
[0175] Referring to FIGS. 1A, 5A, and 10, in operation S3000, the electronic device 100
obtains the input data 500a by preprocessing the plurality of captured images 400.
[0176] In an embodiment of the disclosure, the electronic device 100 may obtain the input
data 500a using the image preprocessing module 133 (see FIG. 1B) to preprocess the
plurality of captured images 400. In an embodiment of the disclosure, the input data
500a may include the plurality of preprocessed captured images 510.
[0177] In an embodiment of the disclosure, in a case in which the depth map generation module
134 includes an artificial intelligence model, the electronic device 100 may obtain
the input data 500a using the image preprocessing module 133 to preprocess the plurality
of captured images 400 such that the input data 500a includes information about epochs
and batches.
[0178] Referring to FIGS. 1A, 6, and 10, in operation S4000, the electronic device 100 generates
the depth map 600 based on the input data 500a.
[0179] In an embodiment of the disclosure, the electronic device 100 may generate the depth
map 600 using the depth map generation module 134 (see FIG. 1B). The depth map generation
module 134 generates the depth map 600 based on the input data 500a. In an embodiment
of the disclosure, as the number of preprocessed captured images 510 included in the
input data 500a increases, the resolution of the depth map 600 generated by the depth
map generation module 134 may increase. In addition, in an embodiment of the disclosure,
as the number of times the depth map generation module 134 refers to the input data
500a increases, the resolution of the generated depth map 600 may increase.
[0180] FIG. 11 is a flowchart of a method of generating a depth map based on input data
obtained by preprocessing a plurality of captured images, a plurality of illumination
images, and characteristic information, according to an embodiment of the disclosure.
Hereinafter, the same reference numerals are assigned to the same operations as those
described above with reference to FIG. 10, and descriptions thereof may be omitted.
[0181] Referring to FIGS. 1A, 3A and 11, in operation S1100, the electronic device 100 may
obtain the plurality of illumination images 320 (see FIG. 5C) respectively corresponding
to the plurality of patterns 300
[0182] In an embodiment of the disclosure, the electronic device 100 may obtain the plurality
of illumination images 320 using an illumination measuring device to photograph the
light-emitting device 111.
[0183] In an embodiment of the disclosure, the illumination measuring device may be installed
on the opposite side of the light-emitting device 111, to photograph the plurality
of patterns in which the light-emitting device 111 provides light. The illumination
measuring device may obtain the plurality of illumination images 320 by photographing
changes in the position, area, shape of the light-emitting region DA, the illumination
of light, and the like according to changes in the pattern of the light provided by
the light-emitting device 111. However, the disclosure is not limited thereto. In
operation S1100, a plurality of previously captured illumination images 320 may be
provided from an external server or nearby electronic devices, without measuring the
light-emitting device 111 using the illumination measuring device.
[0184] Referring to FIGS. 1A, 3A, and 11, in operation S1200, the electronic device 100
may provide the target object 200 with light in the plurality of patterns that change
over time.
[0185] Referring to FIGS. 1A, 4, and 11, in operation S2100, the electronic device 100 obtains
the plurality of captured images 400 respectively corresponding to the plurality of
patterns 300, by photographing the target object 200 receiving the light.
[0186] Referring to FIGS. 1A, 5C, and 11, in operation S2200, the electronic device 100
obtains the characteristic information CI of the light-emitting device 111 providing
the light.
[0187] In an embodiment of the disclosure, the characteristic information CI may include
information about the size of the light-emitting device 111, information about a position
where the light-emitting device 111 is arranged, information about the distance between
the light-emitting device 111 and the target object 200, information about the shape
of the light-emitting device 111, information about the illumination of light provided
by the light-emitting device 111, information about the direction of light provided
by the light-emitting device 111, and the like. In an embodiment of the disclosure,
the electronic device 100 may obtain the characteristic information CI from the light-emitting
device 111 through the communication interface 150.
[0188] Referring to FIGS. 1A, 5C, and 11, in operation S3000a, the electronic device 100
obtains the input data 500c by preprocessing the plurality of captured images 400,
the plurality of illumination images 320, and the characteristic information CI.
[0189] In an embodiment of the disclosure, the electronic device 100 may obtain the input
data 500c using the image preprocessing module 133 (see FIG. 1B) to preprocess the
plurality of captured images 400, the plurality of illumination images 320, and the
characteristic information CI. In an embodiment of the disclosure, the input data
500a may include the plurality of preprocessed captured images 510, the plurality
of preprocessed illumination images 520, and the preprocessed characteristic information
530.
[0190] In an embodiment of the disclosure, in a case in which the depth map generation module
134 includes an artificial intelligence model, the electronic device 100 may obtain
the input data 500c using the image preprocessing module 133 to preprocess the plurality
of captured images 400, the plurality of illumination images 320, and the characteristic
information CI such that the input data 500c includes information about epochs and
batches.
[0191] Referring to FIGS. 1A, 7B, and 11, in operation S4000a, the electronic device 100
generates the depth map 600 based on the input data 500c.
[0192] In an embodiment of the disclosure, the electronic device 100 may generate the depth
map 600 using the depth map generation module 134 (see FIG. 1B). In an embodiment
of the disclosure, as the number of preprocessed illumination images 520 and preprocessed
captured images 510 included in the input data 500c increases, the resolution of the
depth map 600 generated by the depth map generation module 134 may increase. In addition,
the depth map generation module 134 may generate the depth map 600 with high resolution,
based on the preprocessed characteristic information 530 including information about
the light-emitting device 111.
[0193] In an embodiment of the disclosure, in a case in which the preprocessed characteristic
information 530 includes information about the size of the light-emitting device 111,
the depth map generation module 134 may obtain the incident angle of light provided
to the target object 200, the luminance of the light, the area of the light-emitting
region DA, and the like, based on the plurality of preprocessed illumination images
520 and the preprocessed characteristic information 530.
[0194] In an embodiment of the disclosure, the depth map generation module 134 may generate
the depth map 600 based on obtained information of the light and the plurality of
preprocessed captured images 510. In this case, the target object 200 may include
one or more objects. Changes in the shape of a shadow and the luminance of the target
object 200 according to changes in the pattern of the light provided to the target
object 200 are determined according to the distance between each object and the light-emitting
device 111, the shape of each object, and the like. However, the disclosure is not
limited thereto, and in a case in which the preprocessed characteristic information
530 further includes information about other characteristics of the light-emitting
device 111, the number of pieces of information about the pattern of the light provided
to the target object 200 identified by the depth map generation module 134 may increase.
[0195] FIG. 12 is a flowchart of a method of generating a depth map using an electronic
device including a light-emitting device and a sub-light-emitting device, according
to an embodiment of the disclosure. Hereinafter, the same reference numerals are assigned
to the same operations as those described above with reference to FIGS. 10 and 11,
and descriptions thereof may be omitted.
[0196] Referring to FIGS. 2A, 3A and 12, in operation S1100, the electronic device 100 may
obtain the plurality of illumination images 320 (see FIG. 5C) respectively corresponding
to the plurality of patterns 300
[0197] Referring to FIGS. 2A, 3A, and 12, in operation S1200, the electronic device 100
may provide the target object 200 with light in the plurality of patterns that change
over time.
[0198] Referring to FIGS. 2A and 12, in operation S1300, the electronic device 100 may provide
sub-light to the target object 200.
[0199] In an embodiment of the disclosure, the electronic device 100 may provide the sub-light
to the target object 200 using the sub-light-emitting device 112. In an embodiment
of the disclosure, the sub-light-emitting device 112 may provide the target object
200 with sub-light in the plurality of patterns 300 that change over time. In this
case, the patterns in which the sub-light-emitting device 112 provides the sub-light
may be different from the patterns in which the light-emitting device 111 provides
the light. In addition, the sub-light-emitting device 112 may provide the target object
200 with sub-light in a constant pattern regardless of time.
[0200] In an embodiment of the disclosure, depending on the type of the electronic device
100, in normal times, the sub-light-emitting device 112 may perform a function of
providing light necessary for the electronic device 100 to perform its function or
providing light to an environment surrounding the electronic device 100. The sub-light-emitting
device 112 provides sub-light to the target object 200 when the electronic device
100 generates the depth map 600 (see FIG. 6) of the target object 200.
[0201] In an embodiment of the disclosure, the electronic device 100 may obtain a plurality
of sub-illumination images using the illumination measuring device to photograph the
sub-light-emitting device 112. The plurality of sub-illumination images may correspond
to a plurality of patterns of the sub-light, respectively.
[0202] Referring to FIGS. 2A, 4, and 12, in operation S2100a, the electronic device 100
obtains the plurality of captured images 400 respectively corresponding to the plurality
of patterns 300, by photographing the target object 200 receiving the light and the
sub-light.
[0203] Referring to FIGS. 2A, 5D, and 12, in operation S2200, the electronic device 100
obtains the characteristic information CI of the light-emitting device 111 providing
the light, and the sub-characteristic information SCI of the sub-light-emitting device
112 providing the sub-light.
[0204] In an embodiment of the disclosure, the characteristic information CI may include
information about the size of the light-emitting device 111 and the like. The sub-characteristic
information SCI may include information about the size of the sub-light-emitting device
112, information about a position where the sub-light-emitting device 112 is arranged,
information about the distance between the sub-light-emitting device 112 and the target
object 200, information about the shape of the sub-light-emitting device 112, information
about the illumination of light provided by the sub-light-emitting device 112, information
about the direction of light provided by the sub-light-emitting devices 112, information
about the number of sub-light-emitting devices 112, and the like.
[0205] In an embodiment of the disclosure, the electronic device 100 may obtain the characteristic
information CI from the light-emitting device 111, and obtain the sub-characteristic
information SCI from the sub-light-emitting device 112, through the communication
interface 150.
[0206] Referring to FIGS. 2A, 5D and 12, in operation S3000b, the electronic device 100
obtains the input data 500d by preprocessing the plurality of captured images 400,
the plurality of illumination images 320, the characteristic information CI, and the
sub-characteristic information SCI.
[0207] In an embodiment of the disclosure, the electronic device 100 may obtain the input
data 500d using the image preprocessing module 133a (see FIG. 2B) to preprocess the
plurality of captured images 400, the plurality of illumination images 320, the characteristic
information CI, and the sub-characteristic information SCI.
[0208] In an embodiment of the disclosure, the input data 500d may include the plurality
of preprocessed captured images 510, the plurality of preprocessed illumination images
520, the preprocessed characteristic information 530, and the preprocessed sub-characteristic
information 540. In an embodiment of the disclosure, the image preprocessing module
133a may further obtain a plurality of sub-illumination images. The image preprocessing
module 133a may also obtain input data by preprocessing the plurality of captured
images 400, the plurality of illumination images 320, the plurality of sub-illumination
images, the characteristic information CI, and the sub-characteristic information
SCI.
[0209] Referring to FIGS. 2A and 12, in operation S4000b, the electronic device 100 generates
the depth map 600 based on the input data 500d.
[0210] In an embodiment of the disclosure, the electronic device 100 may generate the depth
map 600 using the depth map generation module 134a (see FIG. 2B). In an embodiment
of the disclosure, the depth map generation module 134a may generate the depth map
600 with high resolution, based on the preprocessed characteristic information 530
and the preprocessed sub-characteristic information 540.
[0211] In an embodiment of the disclosure, in a case in which the preprocessed characteristic
information 530 includes information about the size of the light-emitting device 111
and the preprocessed sub-characteristic information 540 includes information about
the position of the sub-light-emitting device 112, the depth map generation module
134a may obtain information of light and sub-light provided to the target object 200,
based on the plurality of preprocessed illumination images 520, the preprocessed characteristic
information 530, and the preprocessed sub-characteristic information 540. The depth
map generation module 134a may generate the depth map 600 with high resolution, based
on the obtained information of the light and the sub-light, and the plurality of preprocessed
captured images 510.
[0212] FIG. 13 is a flowchart of an operation, performed by an electronic device, of generating
a depth map corresponding to input data, according to an embodiment of the disclosure.
[0213] Referring to FIGS. 1A, 4, and 13, when a threshold illumination is the maximum illumination
at which the plurality of captured images 400 reflecting changes in the plurality
of patterns 300 that change over time may be obtained, in operation S5000, the electronic
device 100 compares the intensity of the ambient illumination of the target object
200, with the intensity of the threshold illumination.
[0214] In an embodiment of the disclosure, in operation S5000, the electronic device 100
determines whether the intensity of the ambient illumination of the target object
200 is less than or equal to the intensity of the threshold illumination. In an embodiment
of the disclosure, in a case in which the intensity of the ambient illumination of
the target object 200 is greater than the intensity of the threshold illumination,
the measuring device 120 is unable to obtain the plurality of captured images 400
including changes in the shape of a shadow or the luminance of the target object 200
according to changes in the pattern of light provided by the light-emitting device
111. Accordingly, because the electronic device 100 is unable to generate the depth
map 600 according to the disclosure, the electronic device 100 terminates the operation
of generating the depth map 600 in operation S5001.
[0215] In an embodiment of the disclosure, in a case in which the intensity of the ambient
illumination of the target object 200 is less than or equal to the intensity of the
threshold illumination, the electronic device 100 performs the operation of generating
the depth map 600. In an embodiment of the disclosure, the intensity of the threshold
illumination may vary depending on the performance of the light-emitting device 111
or the performance of the measuring device 120. As the illumination of light that
may be provided by the light-emitting device 111 increases, the intensity of the threshold
illumination may also increase. In addition, as the performance of the measuring device
120 for photographing a change in the shape of a shadow or the luminance of the target
object 200 improves, the intensity of the threshold illumination may also increase.
[0216] Referring to FIGS. 1A, 3A, and 13, in an embodiment of the disclosure, in a case
in which the plurality of patterns 300 in which the light-emitting device 111 provides
light include 'n' patterns, in operation S6000, the electronic device 100 provides
the target object 200 with light having an i-th pattern using the light-emitting device
111. 'i' may be a natural number less than or equal to 'n'.
[0217] In an embodiment of the disclosure, at least one of the positions, areas, or shapes
of the light-emitting regions DA included in the respective 'n' patterns may be different
from each other. In an embodiment of the disclosure, the illuminations of light of
the 'n' patterns may be different from each other.
[0218] Referring to FIGS. 1A, 4, and 13, in an embodiment of the disclosure, in operation
S7000, the electronic device 100 may obtain an i-th captured image by photographing
the target object 200 provided with the light having the i-th pattern. The i-th captured
image is a captured image corresponding to the light having the i-th pattern. In an
embodiment of the disclosure, in operation S7000, the electronic device 100 may obtain
the i-th captured image using the measuring device 120. In an embodiment of the disclosure,
in a case in which the plurality of patterns includes 'n' patterns, the electronic
device 100 may obtain 'n' captured images 400.
[0219] In operation S8000, the electronic device 100 determines whether 'i' is less than
'n'. In an embodiment of the disclosure, in a case in which '" is less than 'n', the
electronic device 100 returns to operation S6000 to provide light to the target object
200 again. In this case, the electronic device 100 may provide the target object 200
with light in an (i+1)-th pattern. Thereafter, in operation S7000, the electronic
device 100 may obtain an (i+1)-th captured image by photographing the target object
200 provided with the light having the (i+1)-th pattern. Thereafter, the electronic
device 100 determines whether i+1 is less than 'n', and in a case in which i+1 is
less than 'n', the electronic device 100 repeats the above operations.
[0220] In an embodiment of the disclosure, when 'i' is equal to 'n', the measuring device
120 may obtain 'n' captured images 400 respectively corresponding to 'n' patterns
300 provided to the target object 200 through the light-emitting device 111.
[0221] Referring to FIGS. 1A, 5C, and 13, in operation S9000, the electronic device 100
is provided with 'n' illumination images 320 corresponding to the 'n' patterns 300
in which the light-emitting device 111 provides the light. The 'n' illumination images
320 corresponds to the 'n' patterns 300, respectively.
[0222] In an embodiment of the disclosure, the 'n' illumination images 320 may be images
previously captured before the operation, performed by the electronic device 100,
of generating the depth map 600. The electronic device 100 may receive a plurality
of previously captured illumination images 320, from an external server or nearby
electronic devices through the communication interface 150 (see FIG. 1B).
[0223] In an embodiment of the disclosure, the 'n' illumination images 320 may be images
captured in the operation, performed by the electronic device 100, of providing the
light in the 'n' patterns in order to generate the depth map 600. The electronic device
100 is provided with the characteristic information CI of the light-emitting device
111.
[0224] In an embodiment of the disclosure, the characteristic information CI may include
information about the size of the light-emitting device 111. The electronic device
100 may obtain the input data 500c by preprocessing the 'n' captured images 400, the
'n' illumination images 320, and the characteristic information CI.
[0225] Referring to FIGS. 1A, 7B, and 13, in operation S10000, the electronic device 100
may generate the depth map 600 based on the input data 500c.
[0226] In an embodiment of the disclosure, using the generated depth map 600, the electronic
device 100 may perform operations of, for example, providing an augmented reality
(AR) experience using the electronic device 100, allowing, using AR, a user to place
furniture, home appliances, and the like in an environment where the electronic device
100 is arranged, identifying the distance between the electronic device 100 and the
user using the electronic device 100, and adjusting the size of an image used in the
electronic device 100 considering the viewing angle of the user. In addition, in an
embodiment of the disclosure, using the depth map 600, the electronic device 100 may
perform operations such as three-dimensional (3D) modeling, vehicle navigation, object
recognition, and human gesture detection. However, the disclosure is not limited thereto.
In an embodiment of the disclosure, the electronic device 100 of the disclosure may
provide the generated depth map 600 to nearby electronic devices. The nearby electronic
devices may perform, based on the provided depth map 600, operations such as provision
of an AR experience, 3D modeling, vehicle navigation, object recognition, and human
gesture detection.
[0227] A program executable by the electronic device 100 described herein may be implemented
as a hardware component, a software component, and/or a combination of hardware components
and software components. The program is executable by any system capable of executing
computer-readable instructions.
[0228] The software may include a computer program, code, instructions, or a combination
of one or more thereof, and may configure the processor to operate as desired or may
independently or collectively instruct the processor.
[0229] The software may be implemented as a computer program that includes instructions
stored in computer-readable storage media. The computer-readable storage media may
include, for example, magnetic storage media (e.g., ROM, RAM, floppy disks, hard disks,
etc.) and optical storage media (e.g., a compact disc ROM (CD-ROM), a digital video
disc (DVD), etc.). The computer-readable recording medium may be distributed in computer
systems connected via a network and may store and execute computer-readable code in
a distributed manner. The recording medium may be computer-readable, may be stored
in a memory, and may be executed by a processor.
[0230] The computer-readable storage medium may be provided in the form of a non-transitory
storage medium. The term 'non-transitory storage medium' may refer to a tangible device
and does not include a signal (e.g., an electromagnetic wave), and the term 'non-transitory
storage medium' does not distinguish between a case where data is stored in a storage
medium semi-permanently and a case where data is stored temporarily. For example,
the non-transitory storage medium may include a buffer in which data is temporarily
stored.
[0231] In addition, a program according to an embodiment of the disclosure may be provided
in a computer program product. The computer program products may be traded as commodities
between sellers and buyers.
[0232] The computer program product may include a software program and a computer-readable
recording medium storing the software program. For example, the computer program product
may include a product (e.g., a downloadable application) in the form of a software
program electronically distributed through a manufacturer of the electronic device
100 or an electronic market (e.g., Samsung Galaxy Store). For electronic distribution,
at least part of the software program may be stored in a storage medium or temporarily
generated. In this case, the storage medium may be a storage medium of a server of
the manufacturer of the electronic device 100, a server of the electronic market,
or a relay server that temporarily stores the software program.
[0233] Although embodiments of the disclosure have been described with the limited embodiment
and the drawings, various modifications and changes may be made by those of skill
in the art from the above description. For example, suitable results may be obtained
even when the described techniques are performed in a different order, or when components
in a described electronic device, architecture, device, or circuit are coupled or
combined in a different manner, or replaced or supplemented by other components or
their equivalents.