TECHNICAL FIELD
[0001] The present disclosure relates to a camera module, a controller, a camera monitoring system, and a moveable body.
BACKGROUND
[0002] A monitoring system in which a plurality of cameras are attached to an automobile is known. If one of the cameras fails in such a system, a replacement camera to replace the failed camera can be moved to the position of the failed camera, thereby allowing the driver to continue monitoring the video. For example, see patent literature (PTL) 1.
[0003] DE 10 2010 023199 A1 discloses a method for operating an image capturing device for realizing outer mirror functions of a motor vehicle involving monitoring the momentary state of cameras in dependence of a result of comparison and/or merging of environmental characteristics.
[0004] JP 2014 039096 A discloses a multi-eye camera photographing system including a plurality of cameras which photograph respective imaging regions, convert them into image data, wherein partial parts of the imaging regions are overlapped in the adjacently arranged cameras, and a pan/tilt mechanism for panning and/or tilting the respective cameras.
[0005] US2015314736 A1 discloses an electronic rearview mirror system for a vehicle.
CITATION LIST
Patent Literature
SUMMARY
[0007] to [0010] The invention provides a camera module according to claim 1, a camera monitoring system according to claim 9, and a movable body according to claim 10. Preferred embodiments are described in the dependent claims.
BRIEF DESCRIPTION OF THE DRAWINGS
[0008] In the accompanying drawings:
FIG. 1A is a schematic side view of a moveable body, according to a first embodiment of the present disclosure, to which a camera module and a monitor are attached;
FIG. 1B is an enlarged view of a region R in FIG. 1A;
FIG. 2 is a functional block diagram of the camera module and the monitor of FIG. 1;
FIG. 3A illustrates a portion of an imaging range of a first camera as seen from above the moveable body;
FIG. 3B illustrates the imaging range of the first camera as seen from the side of the moveable body;
FIG. 3C illustrates an imaging range of a second camera as seen from the side of the moveable body;
FIG. 4A illustrates the appearance of the camera module at a first rotation position;
FIG. 4B illustrates the appearance of the camera module at a second rotation position;
FIG. 5A is a flowchart illustrating operations of a controller in the camera module of FIG. 1;
FIG. 5B is a sequence diagram illustrating operations of the controller and a selector in the camera module of FIG. 1;
FIG. 6 is a schematic top view of a moveable body, according to a second embodiment, to which a camera module is attached;
FIG. 7A is an enlarged view of the region R of FIG. 6, illustrating the appearance of the camera module at a first rotation position in the second embodiment;
FIG. 7B is an enlarged view of the region R of FIG. 6, illustrating the appearance of the camera module at a second rotation position in the second embodiment; and
FIG. 8 illustrates a portion of an imaging range of a first camera in a modification as seen from above the moveable body.
DETAILED DESCRIPTION
[0009] Embodiments of the present disclosure are described below with reference to the drawings.
First Embodiment
[0010] FIG. 1A is a schematic side view of a moveable body 9 according to a first embodiment of the present disclosure. The moveable body 9 includes a camera module 1 and a monitor 2. The camera module 1 and the monitor 2 are configured to be capable of transmitting signals over a dedicated line or a network such as a controller area network (CAN).
[0011] FIG. 1B is an enlarged view of a region R at the upper-front portion of the front door in FIG. 1A. The camera module 1 includes a first camera 11a and a second camera 11b and captures images in a predetermined range. Details are provided below.
[0012] The term "moveable body" in the present disclosure includes vehicles, ships, and aircraft. The term "vehicle" in the present disclosure includes, but is not limited to, automobiles, railway vehicles, industrial vehicles, and vehicles for daily life. For example, the term "vehicle" may include airplanes that travel down a runway. Examples of automobiles include, but are not limited to, passenger vehicles, trucks, buses, motorcycles, and trolley buses, and may include other vehicles that travel on the road. Railway vehicles include, but are not limited to, locomotives, freight cars, passenger cars, streetcars, guided railway vehicles, ropeways, cable cars, linear motor cars, and monorails, and may include other vehicles that travel along a track. Industrial vehicles include industrial vehicles for agriculture and for construction. Industrial vehicles include, but are not limited to, forklifts and golf carts. Industrial vehicles for agriculture include, but are not limited to, tractors, cultivators, transplanters, binders, combines, and lawnmowers. Industrial vehicles for construction include, but are not limited to, bulldozers, scrapers, backhoes, cranes, dump cars, and road rollers. Vehicles for daily life include, but are not limited to, bicycles, wheelchairs, baby carriages, wheelbarrows, and motorized, two-wheeled standing vehicles. Power engines for the vehicle include, but are not limited to, internal-combustion engines including diesel engines, gasoline engines, and hydrogen engines, and electrical engines including motors. The term "vehicle" includes human-powered vehicles. The vehicle is not limited to the above-listed types. For example, automobiles may include industrial vehicles that can drive on the road, and the same vehicle may be included in multiple categories.
[0013] FIG. 2 is a functional block diagram illustrating the schematic configuration of a camera monitoring system 8. The camera monitoring system 8 includes the camera module 1 and the monitor 2. The camera module 1 includes the first camera 11a, the second camera 11b, a rotary member 12a, a drive unit 12b, a controller 13, and a first selector 13a.
[0014] The first camera 11a and the second camera 11b each include an optical system and an image sensor. The optical system is configured by an optical element including at least one lens and has desired optical properties. The desired optical properties are, for example, the focal depth or the angle of view. The optical system forms a subject image on the image sensor. The image sensor is, for example, a charge-coupled device (CCD) image sensor or a complementary MOS (CMOS) image sensor. The image sensor generates an image signal corresponding to the subject imaged on an optical detection surface. A plurality of pixels are arrayed in two dimensions on the optical detection surface of the image sensor. Each pixel generates a pixel signal corresponding to the detected amount of light. An image signal of one frame is formed by the pixel signals generated by the pixels constituting the optical detection surface. The first camera 11a and the second camera 11b output their captured images to the controller 13.
[0015] FIG. 3A illustrates the relationship between the imaging range of the camera module 1 and the moveable body 9. FIG. 3B illustrates the relationship between an imaging direction D0 and a first imaging range R1. FIG. 3C illustrates the relationship between a first imaging direction D1 and a second imaging range R2. In an example, the first camera 11a is a camera for an electronic mirror. At the first rotation position, described below, the orientation of the first camera 11a is determined to allow imaging of the view in the imaging range established by law, for example. The imaging range is as follows in the "Proposal for Supplement 2 to the 04 series of amendments to Regulation No. 46 (Devices for indirect vision)" submitted by the Economic and Social Council of the United Nations, for example. Specifically, as illustrated in FIG. 3A, the imaging range has an outward width of 1 m from the side of the moveable body 9 at a position 4 m behind the driver on the ground. The imaging range is a range including an outward width of 4 m from the side of the moveable body 9 at a position 20 m or more behind the driver (see the shaded portion). The imaging range is set to increase monotonically, from the position 4 m behind the driver to the position 20 m behind the driver, in correspondence with the distance from the driver.
[0016] The first camera 11a faces the imaging direction D0 to be capable of capturing images in such an imaging range. Facing the imaging direction D0 refers to a state in which the optical axis of the first camera 11a is parallel to the imaging direction D0, as illustrated in FIG. 3B, and in which the first camera 11a can capture an image of a subject in the imaging direction D0. The imaging direction D0 is towards the rear of the moveable body 9 and somewhat downwards. In an example, the depression angle between the horizontal direction and the imaging direction D0 is approximately 15°. In an example, the angle between the horizontal component of the imaging direction D0 and the front-back direction of the moveable body 9 is approximately 20°. The first camera 11a captures an image in the first imaging range R1, which is a predetermined angle of view, such as 45°, centered on the optical axis direction.
[0017] In an example, the second camera 11b is a camera for monitoring to the side. In the first rotation position, described below, the optical axis of the second camera 11b is parallel to the first imaging direction D1 illustrated in FIG. 3C. In an example, the first imaging direction D1 faces downward in the vertical direction. The second camera 11b captures an image in the second imaging range R2, which is an angle of view of 170°, centered on the optical axis. The lens included in the optical system of the second camera 11b is a lens with a wider angle of view than the lens included in the optical system of the first camera 11a. This lens is, for example, a fish-eye lens.
[0018] The rotary member 12a is a rotatable mechanism. The rotary member 12a rotates to switch between the first rotation position and the second rotation position. At least the second camera 11b rotates as a result of switching of the rotation position of the rotary member 12a. The imaging direction of the second camera 11b changes by rotation of the rotary member 12a. The first camera 11a may rotate as a result of switching of the rotation position of the rotary member 12a. The imaging direction of the first camera 11a may change by rotation of the rotary member 12a.
[0019] The drive unit 12b includes at least one motor, for example. The drive unit 12b is controlled by the controller 13, for example, to drive the rotary member 12a. The rotary member 12a may be a mechanism allowing manual rotation. The drive unit 12b may be omitted.
[0020] The drive unit 12b may rotate the second camera 11b about its optical axis. The drive unit 12b may include a mechanism to rotate the second camera 11b directly. The drive unit 12b may rotate the second camera 11b about its optical axis when changing the position of the second camera 11b via the rotary member 12a. The drive unit 12b may include a plurality of motors.
[0021] FIG. 4A and FIG. 4B illustrate rotation of the camera module 1. FIG. 4A illustrates a first rotation position L1 of the camera module 1. At this time, the first camera 11a, the second camera 11b, and the rotary member 12a are at the first rotation position L1. FIG. 4B illustrates rotation of the camera module 1. The camera module 1 rotates in the direction indicated by the arrow A1 from the first rotation position L1 to the second rotation position L2. As a result of rotation, the first camera 11a, the second camera 11b, and the rotary member 12a are positioned at the second rotation position L2. As illustrated in FIG. 4A, the first camera 11a faces the imaging direction D0 at the first rotation position L1 and captures an image in the first imaging range R1. The second camera 11b faces the first imaging direction D1 at the first rotation position and captures an image in the second imaging range R2.
[0022] As illustrated in FIG. 4B, the second camera 11b faces the second imaging direction D2 at the second rotation position. At this time, a third imaging range R3 may include the imaging range illustrated in FIG. 3A. In an example, the second imaging direction D2 substantially matches the imaging direction D0. The second imaging direction D2 is not limited to matching the imaging direction D0. The second imaging direction D2 may have the same orientation as the imaging direction D0, or the second imaging direction D2 may be a different direction from the imaging direction D0.
[0023] References to the rotation position of the camera module 1 may be synonymous with references to the imaging direction of the second camera 11b. For example, stating that the camera module 1 is at the first rotation position may be synonymous with stating that the second camera 11b is facing the first imaging direction D1. Stating that the camera module 1 is at the second rotation position may be synonymous with stating that the second camera 11b is facing the second imaging direction D2.
[0024] As a result of rotation of the rotary member 12a, the overlapping range between the third imaging range R3 and the first imaging range R1 of the first camera 11a is greater than the overlapping range between the second imaging range R2 and the first imaging range R1. The overlapping range between the first imaging range R1 when the first camera 11a is facing the imaging direction D0 and the third imaging range R3 is wider than the overlapping range between the first imaging range R1 and the second imaging range R2.
[0025] The description now returns to FIG. 2. The controller 13 includes one or more processors. The controller 13 may include one or more memories that store programs for various processes and store information during calculations. The term "memory" encompasses volatile memory and non-volatile memory. The memory may be configured as a memory independent from the processor and/or a memory built into the processor. The term "processor" encompasses universal processors that execute particular functions by reading particular programs and dedicated processors that are specialized for particular processing. The controller 13 may be either a system-on-a-chip (SoC) or a system in a package (SiP) with one processor or a plurality of processors that work together. The controller 13 may drive the drive unit 12b at least to change the imaging direction of the second camera 11b.
[0026] The controller 13 acquires a first captured image from the first camera 11a. The controller 13 executes a normal process (first process) on the first captured image to generate a first processed image. The controller 13 outputs the generated first processed image. The normal process of the present embodiment includes, for example, white balance adjustment, γ correction, or the like on the first captured image. The white balance adjustment, γ correction, or the like may be termed pre-processing. The normal process of the present embodiment may further include color interpolation, contour enhancement, or the like on the first captured image. The color interpolation, contour enhancement, or the like may be performed on the first captured image subjected to pre-processing. The color interpolation, contour enhancement, or the like may be termed post-processing. In addition to the aforementioned processes, the normal process of the present embodiment may include at least one of a correction process to correct distortion of the first captured image, an extraction process to extract a partial region from the first captured image, an electronic zoom process to extract and expand a partial region from the first captured image, a sharpness process to sharpen the first captured image, and a super-resolution process to increase the resolution of the first captured image by digital processing. The extraction process may be omitted from the electronic zoom process for a first captured image already subjected to the extraction process.
[0027] The controller 13 acquires a second captured image from the second camera 11b. The controller 13 executes a normal process (second process) on the second captured image to generate a second processed image. The controller 13 outputs the generated second processed image. The normal process executed on the second captured image may be the same as the normal process executed on the first captured image, or a portion thereof may differ.
[0028] The controller 13 executes a conversion process (third process), described below, on the second captured image to generate a third processed image. The controller 13 outputs the generated third processed image.
[0029] The conversion process of the present embodiment differs from the normal process in the following way. Distortion correction within the conversion process is executed on the basis of different parameters from distortion correction within the normal process. In the distortion correction within the conversion process, a different substitution table than that of the distortion correction within the normal process is used. By the distortion correction in the conversion process, the third processed image is brought closer to the first processed image than the second processed image in terms of visual distortion after correction. The extraction process within the conversion process is executed on the basis of different parameters from the extraction process within the normal process. The extraction process within the conversion process extracts a narrower range than the range extracted by the extraction process within the normal process. In the extraction process within the conversion process, the aspect ratio differs from that of the range extracted by the extraction process within the normal process. The aspect ratio extracted by the extraction process within the conversion process is closer to the aspect ratio of the first processed image than the aspect ratio of the second processed image. The aspect ratio extracted by the extraction process within the conversion process may be the same as the aspect ratio of the first processed image. The electronic zoom process within the conversion process is executed so that the size of an object in the third processed image is brought closer to the size of an object in the first processed image.
[0030] For example, in the conversion process on the second captured image, the third processed image is generated by combining the processes so that the display conditions are brought closer to, or made to match, the first processed image. The imaging range (angle of view) and aspect ratio of a cut-out image and the imaging range (angle of view) and aspect ratio of the first processed image are brought closer, or made to match, by the controller 13. A predetermined region is established in advance on the basis of the difference in the angles of view of the lens included in the first camera 11a and the lens included in the second camera 11b. The controller 13 executes an electronic zoom process on a cut-out image so that the size of the image after the electronic zoom and the size of the first processed image are closer or match. The controller 13 may execute at least one of a sharpness process and a super-resolution process.
[0031] During the above-described conversion process, the processes included in the conversion process may be executed in any order. The above-described conversion process may be executed by coordinate transformation of the pixels, for example. The controller 13 may, for example, transform the coordinates of the pixels in the second captured image into coordinates of a different coordinate system determined in advance.
[0032] At least a portion of the processing differs between the conversion process and the normal process. At least a portion of the processing is considered to differ when one of the normal process and the conversion process includes a process not included in the other. At least one process by the controller 13 may be considered a portion of the conversion process and the normal process. The conversion process may include the same processes as the normal process. The controller 13 may execute a new process on the second processed image to generate the third processed image.
[0033] Periodically, or at a predetermined point in time, the controller 13 judges whether the first camera 11a has failed as follows. The controller 13 may judge that the first camera 11a has failed when a portion is missing in the first captured image acquired from the first camera 11a. The controller 13 may judge that the first camera 11a has failed when the first camera 11a does not operate. The controller 13 may judge that the first camera 11a has failed when it is detected that the lens has cracked or that the optical axes of the lens and the image sensor are misaligned.
[0034] The controller 13 maintains the first rotation position when judging that the first camera 11a has not failed, i.e. that the first camera 11a is normal.
[0035] When judging that the first camera 11a has failed, the controller 13 provides notification of the failure. The notification may, for example, include displaying an indication of failure on the first monitor 2a or the second monitor 2b described below. The notification may be provided by audio or by turning on a lamp.
[0036] When judging that the first camera 11a has failed, the controller 13 may control the drive unit 12b to set the rotary member 12a to the second rotation position. The rotary member 12a may be rotated automatically by the drive unit 12b when it is judged that the first camera 11a has failed. The rotary member 12a may be rotated by the drive unit 12b when a switching instruction is received from the driver or a passenger. The controller 13 may provide notification on the first monitor 2a or the second monitor 2b that the first camera 11a is being switched to the second camera 11b.
[0037] When rotation of the second camera 11b is complete, the controller 13 may, for example, provide notification on the first monitor 2a or the second monitor 2b that switching from the first camera 11a to the second camera 11b is complete. The controller 13 may provide notification on the first monitor 2a or the second monitor 2b that the range that was being imaged by the first camera 11a is being imaged by the second camera 11b.
[0038] The first selector 13a acquires the first processed image, the second processed image, and the third processed image from the controller 13. The first selector 13a receives the rotation position of the rotary member 12a (the first rotation position or the second rotation position) from the controller 13. When the rotary member 12a is at the first rotation position, the first selector 13a outputs at least one of the first processed image and the second processed image. When the rotary member 12a is at the second rotation position, the first selector 13a outputs the third processed image instead of the first processed image. The first selector 13a may judge whether the rotation position of the rotary member 12a is the first rotation position or the second rotation position.
[0039] The above-described monitor 2 is, for example, a liquid crystal display (LCD) or an organic or inorganic EL display. The monitor 2 is provided in the cabin of the moveable body 9 at a position visible to the driver. In the present embodiment, the monitor 2 includes the first monitor 2a, which is a dedicated monitor, and the second monitor 2b, which is a general-purpose monitor. The first monitor 2a is a screen continually displaying the captured image of the area behind the moveable body 9. The general-purpose monitor displays the captured image of the area to the side of the moveable body 9, a car navigation system screen, or the like in correspondence with the driving conditions. The first monitor 2a can display the first processed image and the third processed image. The first monitor 2a may display the first processed image when the second camera 11b is facing the first imaging direction D1. The first monitor 2a may display the third processed image when the second camera 11b is facing the second imaging direction D2. The second monitor 2b can display the second processed image. The second monitor 2b may display the second processed image when the second camera 11b is facing the first imaging direction D1. The second monitor 2b may suspend display of the second processed image when the second camera 11b is facing the second imaging direction D2.
[0040] The processing executed by the controller 13 in the camera module 1 periodically or at a predetermined point in time is described with reference to FIG. 5A.
[0041] FIG. 5A is a flowchart illustrating operations of the controller 13. The controller 13 executes the processing in FIG. 5A periodically or at a predetermined point in time. The controller 13 may, for example, execute the processing in FIG. 5A at predetermined intervals. The controller 13 may for example, execute the processing in FIG. 5A when acquiring the first captured image. In step S1, the controller 13 acquires at least the first captured image from the first camera 11a. In step S2, the controller 13 judges whether the first camera 11a has failed. When the judgment result in step S2 is Yes, the controller 13 judges in step S3 whether the rotation position is at the first rotation position. When the judgment result in step S3 is Yes, the controller 13 drives the drive unit 12b to set the rotation position to the second rotation position and change the imaging direction of the second camera 11b. When the judgment result in step S3 is No, the controller 13 maintains the rotation position at the second rotation position in step S5.
[0042] When the judgment result in step S2 is No, the controller 13 judges in step S6 whether the rotation position is at the second rotation position. When the judgment result in step S6 is Yes, the controller 13 sets the rotation position to the first rotation position to change the imaging direction of the second camera 11b in step S7. When the judgment result in step S6 is No, the controller 13 maintains the rotation position at the first rotation position in step S5.
[0043] The processing executed by the controller 13 and the first selector 13a in the camera module 1 each time the first captured image of one frame is acquired from the first camera 11a and the second captured image of one frame is acquired from the second camera 11b is described with reference to FIG. 5B.
[0044] In step S11, the controller 13 acquires the first captured image from the first camera 11a and the second captured image from the second camera 11b. In step S12, the controller 13 executes a normal process on the first captured image to generate a first processed image. In step S13, the controller 13 executes a normal process and a conversion process on the second captured image to generate a second processed image and a third processed image, respectively.
[0045] In step S14, the controller 13 outputs the first processed image, the second processed image, the third processed image, and information indicating the rotation position of the rotary member 12a to the first selector 13a.
[0046] In step S15, the first selector 13a judges whether the second camera 11b is facing the first imaging direction D1 or the second imaging direction D2 on the basis of the information indicating the rotation position of the rotary member 12a.
[0047] When the second camera 11b is facing the first imaging direction D1 in step S15, the first selector 13a outputs the first processed image and the second processed image in step S16.
[0048] When the second camera 11b is facing the second imaging direction D2 in step S15, the first selector 13a outputs the third processed image to the first monitor 2a in step S17. The first selector 13a does not output the second processed image.
[0049] The controller 13 in the present embodiment can generate the third processed image on which the third process differing from the second process is executed. The overlapping range between the first imaging range R1 and the third imaging range R3 is wider than the overlapping range between the first imaging range R1 and the second imaging range R2. The view and subjects in the first imaging range R1 of the first camera 11a can be captured over a wider range in the third imaging range R3 than in the second imaging range R2. Even when the first camera 11a has failed, the second camera 11b can cover the first imaging range R1, thereby improving convenience for the driver.
[0050] The controller 13 in the present embodiment can generate the third processed image yielded by execution of the third process, which differs from the second process, on the acquired second captured image. For example, the controller 13 can therefore acquire the second captured image with the second camera 11b and generate the third processed image by executing the conversion process on the acquired second captured image to bring the display conditions closer to those of the first processed image. Accordingly, even if the first camera 11a and the second camera 11b have different display conditions in accordance with use, the first processed image and the third processed image when the first camera 11a is functioning normally can be made to appear more similar. The risk of the driver perceiving images as unnatural is thereby reduced, further improving convenience for the driver.
[0051] The controller 13 of the present embodiment outputs the third processed image when it is judged that the second camera 11b is facing the second imaging direction D2. The controller 13 generates the second processed image when the second camera 11b is facing the first imaging direction D1. The necessary image can thus be generated in accordance with the imaging direction of the second camera 11b.
[0052] The first selector 13a of the present embodiment outputs the second processed image when the second camera 11b is facing the first imaging direction D1 and outputs the third processed image when the second camera 11b is facing the second imaging direction D2. Therefore, the appropriate image can be output in accordance with the imaging direction of the second camera 11b.
[0053] The controller 13 of the present embodiment executes at least one of a sharpness process and a super-resolution process to generate the third processed image. The occurrence of blurriness or decreased resolution, which may occur when bringing the display conditions of the first processed image closer to the display conditions of the third processed image, can therefore be reduced. This can improve the quality of the third processed image.
Second Embodiment
[0054] In the first embodiment, the second camera 11b is a camera for monitoring to the side of the moveable body 9. In the second embodiment, the second camera 11b is a camera for confirming the diagonally-left direction in front of the moveable body 9, for example.
[0055] Since the functional block diagram of the second embodiment is the same as the functional block diagram of the first embodiment, a description thereof is omitted. A description of other matter identical to the first embodiment is also omitted.
[0056] FIG. 6 is a top view of the moveable body 9. As illustrated in FIG. 6, the camera module 1 is attached to the left side of the moveable body 9. The first camera 11a can capture images of the area behind the moveable body 9, as in the first embodiment. The second camera 11b can capture images in the diagonally-left direction in front of the moveable body 9. A camera module including a camera capable of capturing images of the area behind the moveable body 9 and a camera capable of capturing images in the diagonally-right direction in front of the moveable body 9 may be attached to the right side of the moveable body 9 at a position corresponding to the camera module 1.
[0057] FIG. 7A illustrates a region R at the upper-front portion of the front door when the first camera 11a and the second camera 11b are at the first rotation position in FIG. 6. As in FIG. 4A, the first camera 11a attached to the moveable body 9 faces the imaging direction D0 at the first rotation position and captures an image in the first imaging range R1. In an example in the second embodiment, the imaging direction D0 is a substantially horizontal direction. The angle between the imaging direction D0 and the front-back direction of the moveable body 9 is approximately 30°. The first camera 11a captures an image in the first imaging range R1, which is a predetermined angle of view, such as 60°, centered on the optical axis direction. The second camera 11b faces the first imaging direction D1 at the first rotation position and captures an image in the second imaging range R2. In an example in the second embodiment, the first imaging direction D1 is a substantially horizontal direction. The angle between the first imaging direction D1 and the front-back direction of the moveable body 9 is approximately 45°. The second camera 11b captures an image in the second imaging range R2, which is a predetermined angle of view, such as 90°, centered on the optical axis direction. Unlike the first embodiment, no overlapping range exists between the first imaging range R1 and the second imaging range R2 in the second embodiment.
[0058] Like the first embodiment, the controller 13 in the second embodiment generates the first processed image, the second processed image, and the third processed image on the basis of the first captured image and the second captured image acquired from the first camera 11a and the second camera 11b. The controller 13 outputs the generated first processed image, second processed image, and third processed image to the first selector 13a.
[0059] The controller 13 judges whether the first camera 11a has failed on the basis of the first captured image. When judging that the first camera 11a has not failed, the controller 13 controls the first selector 13a to output the first processed image and the second processed image.
[0060] When judging that the first camera 11a has failed, the controller 13 controls the drive unit 12b to set the rotary member 12a to the second rotation position. FIG. 7B illustrates a region R at the upper-front portion of the front door when the first camera 11a and the second camera 11b are at the second rotation position in FIG. 6. As in the first embodiment, the overlapping range between the third imaging range R3 and the first imaging range R1 is wider at this time than the overlapping range between the second imaging range R2 and the first imaging range R1. The controller 13 at this time controls the first selector 13a to output the third processed image.
[0061] In the second embodiment, the overlapping range between the first imaging range R1 and the third imaging range R3 is wider than the overlapping range between the first imaging range R1 and the second imaging range R2. Therefore, the view and subjects in the first imaging range R1 of the first camera 11a can be captured over a wider range in the third imaging range R3 than in the second imaging range R2. Even when the first camera 11a has failed, the second camera 11b can cover the first imaging range R1, thereby improving convenience for the driver.
[0062] Although the present disclosure is based on embodiments and drawings, it is to be noted that various changes and modifications will be apparent to those skilled in the art based on the present disclosure. Therefore, such changes and modifications are to be understood as included within the scope of the present disclosure. For example, the functions and the like included in the various members, units, steps, and the like may be reordered in any logically consistent way. Furthermore, when embodying a method according to the present disclosure, units or steps may be combined into one or divided.
[0063] The camera module 1 includes two cameras in the first embodiment and the second embodiment but may include three or more cameras. When the first camera 11a fails in this case, the controller 13 may select the camera having the angle of view closer to that of the first camera 11a as a substitute for the first camera 11a.
[0064] In the first and second embodiments, the drive unit 12b rotates the rotary member 12a to change the imaging direction of the first camera 11a and the second camera 11b. The drive unit 12b may, however, change the imaging direction of the first camera 11a and the second camera 11b using any method for changing orientation other than a method for rotation.
[0065] In the first and second embodiments, the controller 13 changes the imaging direction of the second camera 11b so that the overlapping range between the third imaging range R3 and the first imaging range R1 is greater than the overlapping range between the second imaging range R2 and the first imaging range R1. The controller 13 may, however, change the imaging direction of the second camera 11b so that the second imaging direction D2 of the second camera 11b substantially matches the imaging direction D0 of the first camera 11a. Alternatively, the controller 13 may change the imaging direction of the second camera 11b so that the third imaging range R3 includes the first imaging range R1.
[0066] In the first and second embodiments, the controller 13 changes the imaging direction of the first camera 11a and the second camera 11b and furthermore executes a conversion process to convert the second captured image into the third processed image. The controller 13 may, however, simply change the imaging direction of the first camera 11a and the second camera 11b, and an electronic control unit (ECU) or a car navigation system of the moveable body 9 may execute the conversion process to yield the third processed image.
[0067] In the first and second embodiments, the first camera 11a and the second camera 11b are attached to the side of the moveable body 9 but may instead be attached at a different position inside the cabin, on top, at the back, or the like of the moveable body 9. For example, the case of the camera module 1 being attached inside the cabin of the moveable body 9 is described. At the first rotation position, the orientation of the first camera 11a is determined to allow imaging of the view in the imaging range established by law, for example. The above-described "Proposal for Supplement 2 to the 04 series of amendments to Regulation No. 46 (Devices for indirect vision)", for example, stipulates that the imaging range have a left-right width of 20 m from a center line parallel to the front-back direction of the moveable body 9 at any position 60 m or more behind the driver on the ground, as illustrated in FIG. 8. When judging that the first camera 11a has failed, the controller 13 may therefore change the imaging direction of the second camera 11b so that the third imaging range R3 of the second camera 11b includes the imaging range illustrated in FIG. 8.
[0068] In the first and second embodiments, the camera module 1 is attached only to one side of the moveable body 9 but may instead be attached to both sides of the moveable body 9.
[0069] The controller 13 may generate both the second processed image and the third processed image regardless of the imaging direction of the second camera 11b. The controller 13 may generate the third processed image when the second camera 11b is facing the second imaging direction D2. The controller 13 may suspend generation of the third processed image when the second camera 11b is facing the first imaging direction D1. The controller 13 may generate the second processed image regardless of the imaging direction of the second camera 11b and generate the third processed image only when the second camera 11b is facing the second imaging direction D2.
[0070] The controller 13 may generate the second processed image when the second camera 11b is facing the first imaging direction D1. The controller 13 may suspend generation of the second processed image when the second camera 11b is facing the second imaging direction D2. The controller 13 may generate the third processed image regardless of the imaging direction of the second camera 11b and generate the second processed image only when the second camera 11b is facing the first imaging direction D1.
[0071] By causing the controller 13 to generate only one of the third processed image and the second processed image, the camera module 1 can reduce the processing load of the controller 13 as compared to when both images are always generated.
[0072] The controller 13 may generate the first processed image regardless of the imaging direction of the second camera 11b. The controller 13 may generate the first processed image when the second camera 11b is facing the first imaging direction D1. The controller 13 may suspend generation of the first processed image when the second camera 11b is facing the second imaging direction D2.
[0073] In the first and second embodiments, the controller 13 and the first selector 13a are separate components. The first selector 13a may, however, be provided internal to the controller 13. Furthermore, the first selector 13a may be a separate component from the camera module 1.
[0074] In the first and second embodiments, the first selector 13a outputs the second processed image when the second camera 11b is facing the first imaging direction D1 and outputs the third processed image when the second camera 11b is facing the second imaging direction D2. The first selector 13a may be replaced by a second selector 13b that is a separate selector, as described below. The second selector 13b outputs the first processed image to the first monitor 2a when the second camera 11b is facing the first imaging direction D1 and outputs the third processed image to the first monitor 2a when the second camera 11b is facing the second imaging direction D2. The second selector 13b may output the second processed image in addition to the first processed image to the first monitor 2a when the second camera 11b is facing the first imaging direction D1.
[0075] Much of the subject matter of the present disclosure is described as a series of operations executed by a computer system and other hardware that can execute program instructions. It should be noted that in each embodiment, various operations are executed by a dedicated circuit (for example, individual logical gates interconnected in order to execute a particular function) implemented by program instructions (software), or by a logical block, program module, or the like executed by one or more processors. The one or more processors that execute a logical block, program module, or the like include, for example, one or more of a microprocessor, CPU, application specific integrated circuit (ASIC), digital signal processor (DSP), programmable logic device (PLD), field programmable gate array (FPGA), processor, controller, microcontroller, microprocessor, electronic device, other apparatus designed to be capable of executing the functions disclosed here, and/or a combination of any of the above. The disclosed embodiments are, for example, implemented by hardware, software, firmware, middleware, microcode, or a combination of any of these. The instructions may be program code or a code segment for executing the necessary tasks. The instructions may be stored on a machine-readable, non-transitory storage medium or other medium. The code segment may indicate a combination of any of the following: procedures, functions, subprograms, programs, routines, subroutines, modules, software packages, classes, instructions, data structures, or program statements. The code segment may transmit and/or receive information, data arguments, variables, or memory content to or from another code segment or hardware circuit in order for the code segment to connect to another code segment or hardware circuit.
[0076] A system including various modules and/or units that execute particular functions has been disclosed. It should be noted that these modules and units have been illustrated schematically to explain their functionality briefly and do not necessarily designate particular hardware and/or software. In this sense, it suffices for the modules, units, and other constituent elements to be hardware and/or software implemented so as to substantially execute the particular functions described here. The various functions or different constituent elements may be combined with or separated from hardware and/or software in any way, and each may be used individually or in some combination. An input/output (I/O) device or user interface including, but not limited to, a keyboard, display, touchscreen, or pointing device may be connected to the system directly or via an I/O controller. In this way, the various subject matter disclosed herein may be embodied in a variety of forms. All such embodiments are included in the scope of the present disclosure.
[0077] The references to "first", "second", and "third" in the present disclosure are identifiers for distinguishing between elements. As necessary, the identifiers may be switched for the members designated by the references to "first", "second", and "third" in the present disclosure. For example, the identifiers "first" and "second" of the first camera 11a and the second camera 11b may be switched. The identifiers may be removed as necessary. Elements from which the identifiers are removed are distinguished by their reference sign. The identifiers may be changed to any number as necessary. Any identifier may be added as necessary to a component mentioned in the present disclosure. For example, "first" may be added to refer to the controller 13 as the first controller 13. The identifiers such as "first", "second", and "third" in the present disclosure may not be used as an interpretation of the order of components.
REFERENCE SIGNS LIST
[0078]
- 1
- Camera module
- 11a
- First camera
- 11b
- Second camera
- 12a
- Rotary member
- 12b
- Drive unit
- 13
- Controller
- 13a
- First selector
- 2
- Monitor
- 2a
- First monitor
- 2b
- Second monitor
- 8
- Camera monitoring system
- 9
- Moveable body
1. A camera module (1) comprising:
a first camera (11a), the first camera (11a) configured to capture images in a first imaging range (R1) ; and
a second camera (11b) capable of switching between a first imaging direction (D1) and a second imaging direction (D2), the second camera (11b) configured to
capture images in a second imaging range (R2) when facing the first imaging direction (D1), and
capture images in a third imaging range (R3) when facing the second imaging direction (D2); and
wherein an overlapping range between the first imaging range (R1) and the third imaging range (R3) is greater than an overlapping range between the first imaging range (R1) and the second imaging range (R2)
the camera module (1) further comprising
a rotary member (12a) configured to rotate to switch between a first rotation position (L1), allowing the first camera (11a) to capture an image in the first imaging range (R1) and the second camera (11b) to capture an image in the second imaging range (R2), and a second rotation position (L2),allowing the second camera (11b) to capture an image in the third imaging range (R3),
a controller (13), the controller (13) configured to
judge whether the first camera (11a) has failed, maintaining the first rotation position (L1) when judging that the first camera (11a) has not failed and control a drive unit (12b) to set the rotary member (12a) to the second rotation position (L2) when judging that the first camera (11a) has failed;
acquire a first captured image captured by the first camera (11a);
execute a first process on the acquired first captured image to generate a first processed image;
acquire a second captured image captured by the second camera;
generate a second processed image yielded by execution of a second process on the acquired second captured image; and
generate a third processed image yielded by execution of a third process on the acquired second captured image, the third process being different from the second process; and
a first selector (13b), the first selector (13b) being configured to
output the first processed image to a first monitor (2a) when the second camera (11b) is facing the first imaging direction (D1); and
output the third processed image to the first monitor (2a) when the second camera (11b) is facing the second imaging direction (D2).
2. The camera module (1) of claim 1, wherein an overlapping range between the first processed image and the third processed image is greater than an overlapping range between the first processed image and the second processed image.
3. The camera module (1) of claim 1 or 2, wherein the controller (13) is configured to generate the third processed image when the second camera (11b) is facing the second imaging direction (D2).
4. The camera module (1) of any one of claims 1 to 3, wherein the controller (13) is configured to generate the second processed image when the second camera (11b) is facing the first imaging direction (D1).
5. The camera module (1) of claim 1 or 2, wherein the controller (13) is configured to generate the third processed image regardless of the imaging direction of the second camera (11b).
6. The camera module (1) of any one of claims 1 to 3 or 5, wherein the controller (13) is configured to generate the second processed image regardless of the imaging direction of the second camera (11b).
7. The camera module (1) of claim 5, wherein the controller (13) is configured to generate the second processed image when the second camera (11b) is facing the first imaging direction (D1).
8. The camera module (1) of any one of claims 1 to 7, further comprising a second selector (13a) configured to
output the second processed image when the second camera (11b) is facing the first imaging direction (D1); and
output the third processed image when the second camera (11b) is facing the second imaging direction (D2) .
9. A camera monitoring system (8) comprising:
the camera module (1) of any one of claims 1 to 8; and
a monitor (2), the monitor (2) configured to
display the first processed image when the second camera (11b) is facing the first imaging direction (D1); and
display the third processed image when the second camera (11b) is facing the second imaging direction (D2) .
10. A moveable body (9) comprising:
the camera module (1) of any one of claims 1 to 8; and
a monitor (2), the monitor (2) configured to
display the first processed image when the second camera (11b) is facing the first imaging direction (D1); and
display the third processed image when the second camera (11b) is facing the second imaging direction (D2) .
1. Kameramodul (1), aufweisend:
eine erste Kamera (11a), wobei die erste Kamera (11a) konfiguriert ist, um Bilder in einem ersten Abbildungsbereich (R1) zu erfassen, und
eine zweite Kamera (11b), die in der Lage ist, zwischen einer ersten Abbildungsrichtung (D1) und einer zweiten Abbildungsrichtung (D2) umzuschalten, wobei die zweite Kamera (11b) konfiguriert ist, um
Bilder in einem zweiten Abbildungsbereich (R2) zu erfassen, wenn sie der ersten Abbildungsrichtung (D1) zugewandt ist, und
Bilder in einem dritten Abbildungsbereich (R3) zu erfassen, wenn sie der zweiten Abbildungsrichtung (D2) zugewandt ist, und
wobei ein Überlappungsbereich zwischen dem ersten Abbildungsbereich (R1) und dem dritten Abbildungsbereich (R3) größer als ein Überlappungsbereich zwischen dem ersten Abbildungsbereich (R1) und dem zweiten Abbildungsbereich (R2) ist,
wobei das Kameramodul (1) ferner aufweist:
ein Drehelement (12a), das konfiguriert ist, um zu drehen, um zwischen einer ersten Drehposition (L1), die es der ersten Kamera (11a) ermöglicht, ein Bild in dem ersten Abbildungsbereich (R1) zu erfassen, und der zweiten Kamera (11b) ermöglicht, ein Bild in dem zweiten Abbildungsbereich (R2) zu erfassen, und einer zweiten Drehposition (L2) umzuschalten, die es der zweiten Kamera (11b) ermöglicht, ein Bild in dem dritten Abbildungsbereich (R3) zu erfassen,
einen Controller (13), wobei der Controller (13) konfiguriert ist zum:
Beurteilen, ob die erste Kamera (11a) ausgefallen ist, wobei die erste Drehposition (L1) beibehalten wird, wenn beurteilt wird, dass die erste Kamera (11a) nicht ausgefallen ist, und Steuern einer Antriebseinheit (12b), um das Drehelement (12a) in die zweite Drehposition (L2) zu bringen, wenn beurteilt wird, dass die erste Kamera (11a) ausgefallen ist,
Erlangen eines ersten erfassten Bildes, das von der ersten Kamera (11a) erfasst wurde,
Ausführen eines ersten Prozesses an dem erlangten ersten erfassten Bild, um ein erstes verarbeitetes Bild zu erzeugen,
Erlangen eines zweiten erfassten Bildes, das von der zweiten Kamera erfasst wurde,
Erzeugen eines zweiten verarbeiteten Bildes, das sich durch die Ausführung eines zweiten Prozesses an dem erlangten zweiten erfassten Bild ergibt, und
Erzeugen eines dritten verarbeiteten Bildes, das sich durch die Ausführung eines dritten Prozesses an dem erlangten zweiten erfassten Bild ergibt, wobei sich der dritte Prozess von dem zweiten Prozess unterscheidet, und
einen ersten Selektor (13b), wobei der erste Selektor (13b) konfiguriert ist zum:
Ausgeben des ersten verarbeiteten Bildes an einen ersten Monitor (2a), wenn die zweite Kamera (11b) der ersten Abbildungsrichtung (D1) zugewandt ist, und
Ausgeben des dritten verarbeiteten Bildes an den ersten Monitor (2a), wenn die zweite Kamera (11b) der zweiten Abbildungsrichtung (D2) zugewandt ist.
2. Kameramodul (1) gemäß Anspruch 1, wobei ein Überlappungsbereich zwischen dem ersten verarbeiteten Bild und dem dritten verarbeiteten Bild größer als ein Überlappungsbereich zwischen dem ersten verarbeiteten Bild und dem zweiten verarbeiteten Bild ist.
3. Kameramodul (1) gemäß Anspruch 1 oder 2, wobei der Controller (13) konfiguriert ist, um das dritte verarbeitete Bild zu erzeugen, wenn die zweite Kamera (11b) der zweiten Abbildungsrichtung (D2) zugewandt ist.
4. Kameramodul (1) gemäß irgendeinem der Ansprüche 1 bis 3, wobei der Controller (13) konfiguriert ist, um das zweite verarbeitete Bild zu erzeugen, wenn die zweite Kamera (11b) der ersten Abbildungsrichtung (D1) zugewandt ist.
5. Kameramodul (1) gemäß Anspruch 1 oder 2, wobei der Controller (13) konfiguriert ist, um das dritte verarbeitete Bild unabhängig von der Abbildungsrichtung der zweiten Kamera (11b) zu erzeugen.
6. Kameramodul (1) gemäß irgendeinem der Ansprüche 1 bis 3 oder 5, wobei der Controller (13) konfiguriert ist, um das zweite verarbeitete Bild unabhängig von der Abbildungsrichtung der zweiten Kamera (11b) zu erzeugen.
7. Kameramodul (1) gemäß Anspruch 5, wobei der Controller (13) konfiguriert ist, um das zweite verarbeitete Bild zu erzeugen, wenn die zweite Kamera (11b) der ersten Abbildungsrichtung (D1) zugewandt ist.
8. Kameramodul (1) gemäß irgendeinem der Ansprüche 1 bis 7, ferner aufweisend einen zweiten Selektor (13a), der konfiguriert ist zum:
Ausgeben des zweiten verarbeiteten Bildes, wenn die zweite Kamera (11b) der ersten Abbildungsrichtung (D1) zugewandt ist, und
Ausgeben des dritten verarbeiteten Bildes, wenn die zweite Kamera (11b) der zweiten Abbildungsrichtung (D2) zugewandt ist.
9. Kameraüberwachungssystem (8), aufweisend:
das Kameramodul (1) gemäß irgendeinem der Ansprüche 1 bis 8 und
einen Monitor (2), wobei der Monitor (2) konfiguriert ist zum:
Anzeigen des ersten verarbeiteten Bildes, wenn die zweite Kamera (11b) der ersten Abbildungsrichtung (D1) zugewandt ist, und
Anzeigen des dritten verarbeiteten Bildes, wenn die zweite Kamera (11b) der zweiten Abbildungsrichtung (D2) zugewandt ist.
10. Bewegbarer Körper (9), aufweisend:
das Kameramodul (1) gemäß irgendeinem der Ansprüche 1 bis 8 und
einen Monitor (2), wobei der Monitor (2) konfiguriert ist zum:
Anzeigen des ersten verarbeiteten Bildes, wenn die zweite Kamera (11b) der ersten Abbildungsrichtung (D1) zugewandt ist, und
Anzeigen des dritten verarbeiteten Bildes, wenn die zweite Kamera (11b) der zweiten Abbildungsrichtung (D2) zugewandt ist.
1. Module de caméra (1), comprenant :
une première caméra (11a), la première caméra (11a) configurée pour capturer des images dans une première plage d'imagerie (R1) ; et
une deuxième caméra (11b) capable de commuter entre une première direction d'imagerie (D1) et une deuxième direction d'imagerie (D2), la deuxième caméra (11b) étant configurée pour
capturer des images dans une deuxième plage d'imagerie (R2) lorsqu'elle fait face à la première direction d'imagerie (D1), et
capturer des images dans une troisième plage d'imagerie (R3) lorsqu'elle fait face à la deuxième direction d'imagerie (D2) ; et
dans lequel une plage de chevauchement entre la première plage d'imagerie (R1) et la troisième plage d'imagerie (R3) est supérieure à une plage de chevauchement entre la première plage d'imagerie (R1) et la deuxième plage d'imagerie (R2),
le module de caméra (1) comprenant en outre :
un élément rotatif (12a) configuré pour tourner afin de commuter entre une première position de rotation (L1), permettant à la première caméra (11a) de capturer une image dans la première plage d'imagerie (R1) et à la deuxième caméra (11b) de capturer une image dans la deuxième plage d'imagerie (R2), et une deuxième position de rotation (L2), permettant à la deuxième caméra (11b) de capturer une image dans la troisième plage d'imagerie (R3),
un contrôleur (13), le contrôleur (13) étant configuré pour
juger si la première caméra (11a) est tombée en panne, maintenant la première position de rotation (L1) lorsqu'il est jugé que la première caméra (11a) n'est pas tombée en panne et commander une unité d'entraînement (12b) pour placer l'élément rotatif (12a) dans la deuxième position de rotation (L2) lorsqu'il est jugé que la première caméra (11a) est tombée en panne ;
acquérir une première image capturée qui est capturée par la première caméra (11a) ;
exécuter un premier processus sur la première image capturée acquise pour générer une première image traitée ;
acquérir une deuxième image capturée qui est capturée par la deuxième caméra ;
générer une deuxième image traitée résultant de l'exécution d'un deuxième processus sur la deuxième image capturée acquise ; et
générer une troisième image traitée résultant de l'exécution d'un troisième processus sur la deuxième image capturée acquise, le troisième processus étant différent du deuxième processus ; et
un premier sélecteur (13b), le premier sélecteur (13b) étant configuré pour
émettre la première image traitée sur un premier moniteur (2a) lorsque la deuxième caméra (11b) fait face à la première direction d'imagerie (D1) ; et
émettre la troisième image traitée sur le premier moniteur (2a) lorsque la deuxième caméra (11b) fait face à la deuxième direction d'imagerie (D2).
2. Module de caméra (1) selon la revendication 1, dans lequel une plage de chevauchement entre la première image traitée et la troisième image traitée est supérieure à une plage de chevauchement entre la première image traitée et la deuxième image traitée.
3. Module de caméra (1) selon la revendication 1 ou 2, dans lequel le contrôleur (13) est configuré pour générer la troisième image traitée lorsque la deuxième caméra (11b) fait face à la deuxième direction d'imagerie (D2).
4. Module de caméra (1) selon l'une quelconque des revendications 1 à 3, dans lequel le contrôleur (13) est configuré pour générer la deuxième image traitée lorsque la deuxième caméra (11b) fait face à la première direction d'imagerie (D1).
5. Module de caméra (1) selon la revendication 1 ou 2, dans lequel le contrôleur (13) est configuré pour générer la troisième image traitée indépendamment de la direction d'imagerie de la deuxième caméra (11b).
6. Module de caméra (1) selon l'une quelconque des revendications 1 à 3 ou 5, dans lequel le contrôleur (13) est configuré pour générer la deuxième image traitée indépendamment de la direction d'imagerie de la deuxième caméra (11b).
7. Module de caméra (1) selon la revendication 5, dans lequel le contrôleur (13) est configuré pour générer la deuxième image traitée lorsque la deuxième caméra (11b) fait face à la première direction d'imagerie (D1).
8. Module de caméra (1) selon l'une quelconque des revendications 1 à 7, comprenant en outre un deuxième sélecteur (13a) configuré pour
émettre la deuxième image traitée lorsque la deuxième caméra (11b) fait face à la première direction d'imagerie (D1) ; et
émettre la troisième image traitée lorsque la deuxième caméra (11b) fait face à la deuxième direction d'imagerie (D2).
9. Système de surveillance de caméra (8), comprenant :
le module de caméra (1) selon l'une quelconque des revendications 1 à 8 ; et
un moniteur (2), le moniteur (2) étant configuré pour
afficher la première image traitée lorsque la deuxième caméra (11b) fait face à la première direction d'imagerie (D1) ; et
afficher la troisième image traitée lorsque la deuxième caméra (11b) fait face à la deuxième direction d'imagerie (D2).
10. Corps mobile (9), comprenant :
le module de caméra (1) selon l'une quelconque des revendications 1 à 8 ; et
un moniteur (2), le moniteur (2) étant configuré pour
afficher la première image traitée lorsque la deuxième caméra (11b) fait face à la première direction d'imagerie (D1) ; et
afficher la troisième image traitée lorsque la deuxième caméra (11b) fait face à la deuxième direction d'imagerie (D2).