(19)
(11)EP 3 726 429 A1

(12)EUROPEAN PATENT APPLICATION
published in accordance with Art. 153(4) EPC

(43)Date of publication:
21.10.2020 Bulletin 2020/43

(21)Application number: 18889848.0

(22)Date of filing:  14.12.2018
(51)International Patent Classification (IPC): 
G06K 9/62(2006.01)
(86)International application number:
PCT/CN2018/121008
(87)International publication number:
WO 2019/114807 (20.06.2019 Gazette  2019/25)
(84)Designated Contracting States:
AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR
Designated Extension States:
BA ME
Designated Validation States:
KH MA MD TN

(30)Priority: 15.12.2017 CN 201711346193

(71)Applicant: Nio Nextev Limited
Hong Kong (CN)

(72)Inventor:
  • PENG, Siwei
    Hong Kong (CN)

(74)Representative: Vossius & Partner Patentanwälte Rechtsanwälte mbB 
Siebertstrasse 3
81675 München
81675 München (DE)

  


(54)MULTI-SENSOR TARGET INFORMATION FUSION


(57) The invention relates to multi-sensor target information fusion. An optimization method for multi-sensor target information fusion comprises the steps of: obtaining, for each time, a fusion prediction result for all sensors concerning a target state at the current time; obtaining actual measurement results from each sensor concerning the target state at the current time; obtaining, for each set of actual measurement results, an optimal estimation result for a corresponding sensor concerning the target state at the current time based on the fusion prediction result and the set of actual measurement results; and fusing optimal estimation results for all the sensors to determine a corresponding weight of each of the optimal estimation results and thus to obtain an optimal fusion estimation result concerning the target state at the current time.




Description

Technical Field



[0001] The invention belongs to the field of automobiles, relates to multi-sensor target information fusion, and more particularly to an optimization method for multi-sensor target information fusion and synchronization of multi-sensor target information fusion and multi-sensor sensing.

Background Art



[0002] Currently, in the field of automobiles around the world, "intelligentization" is a clear development direction. The National Highway Traffic Safety Administration (NHTSA) has also given a standard for levels of automobile driving, from auxiliary driving to automated driving, but whether it is the auxiliary driving or the automated driving, it is necessary to fuse results from target sensing sensors, so as to reduce overlapped targets and compensate for deficiencies of the results from different sensors.

[0003] However, when it comes to the fusion of sensor results related to a target, this will involve the issue of weight assignment among the results from multiple sensors used for sensing the target. Such fusion requires processing of a large number of sensor results in real time and filtering of these results to provide a decision module with a stably and linearly varying target group, and thus usually involves the consumption of a large amount of processing power, and a conventional embedded processor chip has a relatively low processing performance, which result in the assignment of a fusion weight being artificially determined only according to characteristics of the sensors and the maximum error of the results when such a fusion process is carried out by means of the conventional embedded processor chip. The fusion weight obtained in this way is a weight defined in advance, and therefore, especially in the case of a sensor jump, there are problems that the weight assignment is irrational and is not universally applicable, and then an accurate target fusion estimation result cannot be obtained.

[0004] Moreover, in actual processing, due to the differences in algorithms of various sensor suppliers, the strength of hardware performance of sensors, etc., the update speeds of the sensors concerning a target will be affected to varying degrees, and different update speeds of different sensors will also affect the processing of target information fusion.

[0005] In this regard, a general processing method in the industry is as follows: when a target information fusion task is triggered, the previous sensing result from a sensor is usually taken as the latest input for target information fusion. However, when a recognition result from a sensor concerning a target is updated slowly or not updated, the problem that optimal estimation of a fusion result is not accurate enough is caused.

Summary of the Invention



[0006] In order to overcome one or more of the above-mentioned shortcomings or other shortcomings, the invention is implemented by using the technical solutions as follows.

[0007] According to a first aspect of the invention, an optimization method for multi-sensor target information fusion is provided, which comprises: step S1: obtaining, for each time, a fusion prediction result for all sensors concerning a target state at the current time; step S2: obtaining actual measurement results from each sensor concerning the target state at the current time; step S3: obtaining, for each set of actual measurement results, an optimal estimation result for a corresponding sensor concerning the target state at the current time based on the fusion prediction result and the set of actual measurement results; and step S4: fusing optimal estimation results for all the sensors to determine a corresponding weight of each of the optimal estimation results and thus to obtain an optimal fusion estimation result concerning the target state at the current time.

[0008] Further, in the first aspect of the invention, step S3 comprises: step S31: calculating, for each set of actual measurement results, a corresponding transformation matrix based on the fusion prediction result and the set of actual measurement results; step S32: calculating a corresponding covariance of each set of actual measurement results; step S33: calculating, for each set of actual measurement results, a corresponding Kalman gain based on the corresponding transformation matrix and the corresponding covariance; and step S34: calculating, for each set of actual measurement results, the corresponding optimal estimation result for the corresponding sensor concerning the target state at the current time based on the fusion prediction result, the corresponding Kalman gain, the set of actual measurement results, and the corresponding transformation matrix.

[0009] Further, in the first aspect of the invention, step S4 comprises: step S41: determining the corresponding weight of each of the optimal estimation results according to the corresponding covariance of each set of actual measurement results; and step S42: calculating the optimal fusion estimation result for all the sensors concerning the target state at the current time based on each of the optimal estimation results and the corresponding weight.

[0010] Further, in the first aspect of the invention, the method also comprises: step S5: correcting the covariance obtained in step S32 according to the transformation matrix obtained in step S31 and the Kalman gain obtained in step S33, so as to obtain a corrected covariance.

[0011] Further, in the first aspect of the invention, in step S32, the corresponding covariance of each set of actual measurement results at the current time is obtained with a corrected covariance obtained at a previous time.

[0012] Further, in the first aspect of the invention, in step S1, the fusion prediction result at the current time is obtained with an optimal fusion estimation result concerning the target state at the previous time.

[0013] According to an aspect of the invention, a method for synchronizing multi-sensor target information fusion with multi-sensor sensing is provided, which comprises: step P1: obtaining actual measurement results from each sensor concerning a target state and recording a time at which each set of actual measurement results is obtained, so as to record a first time stamp for each set of actual measurement results; step P2: recording, as a second time stamp, a time at which target information fusion processing starts to be performed; step P3: calculating a time difference between each first time represented by each first time stamp and a second time represented by the second time stamp respectively; step P4: updating corresponding actual measurement results obtained at the first time based on each calculated time difference to obtain corresponding estimated measurement results at the second time; step P5: obtaining a fusion prediction result for all sensors concerning the target state at the second time; and step P6: obtaining, for each set of estimated measurement results, an optimal estimation result for a corresponding sensor concerning the target state at the second time based on the fusion prediction result and the set of estimated measurement results.

[0014] Further, in the second aspect of the invention, step P6 comprises: step P61: calculating, for each set of estimated measurement results, a corresponding transformation matrix based on the fusion prediction result and the set of estimated measurement results; step P62: calculating a corresponding covariance of each set of estimated measurement results; step P63: calculating, for each set of estimated measurement results, a corresponding Kalman gain based on the corresponding transformation matrix and the corresponding covariance; and step P64: calculating, for each set of estimated measurement results, the corresponding optimal estimation result for the corresponding sensor concerning the target state at the second time based on the fusion prediction result, the corresponding Kalman gain, the set of estimated measurement results, and the corresponding transformation matrix.

[0015] Further, in the second aspect of the invention, the method also comprises: step P7: fusing optimal estimation results for all the sensors to determine a corresponding weight of each of the optimal estimation results and thus to obtain an optimal fusion estimation result concerning the target state at the second time.

[0016] Further, in the second aspect of the invention, the method also comprises: step P8: correcting the covariance obtained in step P62 according to the transformation matrix obtained in step P61 and the Kalman gain obtained in step P63, so as to obtain a corrected covariance.

[0017] Further, in the second aspect of the invention, in step P62, the corresponding covariance of each set of estimated measurement results at the second time is obtained with a corrected covariance at the first time.

[0018] Further, in the second aspect of the invention, in step P1, the fusion prediction result at the second time is obtained with an optimal fusion estimation result concerning the target state at the first time.

[0019] According to a third aspect of the invention, a computer device is provided, which comprises a memory, a processor, and a computer program stored on the memory and operable on the processor, wherein the program implements, when executed by the processor, the steps of the method according to the first aspect and/or the second aspect of the invention.

[0020] According to a fourth aspect of the invention, a recording medium having a computer program stored thereon is provided, wherein the program is executed by a computer to implement the steps of the method according to the first aspect and/or the second aspect of the invention.

[0021] According to a fifth aspect of the invention, an auxiliary driving method is provided, which comprises an optimization method for multi-sensor target information fusion according to the first aspect of the invention.

[0022] According to a sixth aspect of the invention, an auxiliary driving method is provided, which comprises an optimization method for multi-sensor target information fusion according to the second aspect of the invention.

[0023] According to a seventh aspect of the invention, an auxiliary driving system is provided, which comprises an optimization apparatus for multi-sensor target information fusion according to the invention.

[0024] The foregoing features and operations of the invention will become more obvious according to the following descriptions and the accompanying drawings.

Brief Description of the Drawings



[0025] The above-mentioned and other objectives and advantages of the invention will be more thorough and clearer from the following detailed description in conjunction with the accompanying drawings, wherein the same or similar elements are represented by the same numerals.

Fig. 1 is an example flowchart of an optimization method for multi-sensor target information fusion according to one implementation of the invention.

Fig. 2 is an example sub-flowchart of step S3 in Fig. 1 according to one embodiment of the invention.

Fig. 3 is an example sub-flowchart of step S4 in Fig. 1 according to one embodiment of the invention.

Fig. 4 is an example block diagram of an optimization apparatus for multi-sensor target information fusion according to one implementation of the invention.

Fig. 5 is an example block diagram of a computer device according to one implementation of the invention that is configured to perform an optimization method for multi-sensor target information fusion according to one implementation of the invention.

Fig. 6 is an example flowchart of a method for synchronizing multi-sensor target information fusion with multi-sensor sensing according to one implementation of the invention.

Fig. 7 is an example sub-flowchart of step P6 in Fig. 6 according to one embodiment of the invention.

Fig. 8 is an example block diagram of an apparatus for synchronizing multi-sensor target information fusion with multi-sensor sensing according to one implementation of the invention.

Fig. 9 is an example block diagram of a computer device according to one implementation of the invention that is configured to perform a method for synchronizing multi-sensor target information fusion with multi-sensor sensing according to one implementation of the invention.


Detailed Description of Embodiments



[0026] Methods, apparatuses, computer devices, and a recording medium according to the invention will be further described in detail below in conjunction with the accompanying drawings. It should be noted that the following detailed description of embodiments is exemplary rather than limiting, and is intended to provide a basic understanding of the invention, but not to confirm key or decisive elements of the invention or limit the scope of protection.

[0027] The invention is described below with reference to block diagram illustrations, block diagrams and/or flowcharts of the methods and apparatuses in the embodiments of the invention. It should be understood that each block of the flowchart illustrations and/or the block diagrams and combinations of the flowchart illustrations and/or the block diagrams may be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general-purpose computer, a special-purpose computer, or other programmable data processing devices to form a machine, such that the instructions, which are executed by the processor of the computer or the other programmable data processing devices, create components for implementing functions/operations specified in the flowcharts and/or blocks and/or one or more flow block diagrams.

[0028] These computer program instructions may be stored in a computer-readable memory and may instruct a computer or other programmable processors to implement functions in a specific manner, so that these instructions stored in the computer-readable memory constitute an article of manufacture, which contains instruction components for implementing the functions/operations specified in one or more blocks of the flowcharts and/or block diagrams.

[0029] These computer program instructions may be loaded onto a computer or other programmable data processors to enable a series of operational steps to be executed on the computer or the other programmable processors, so as to constitute a computer-implemented process, so that these instructions executed on the computer or the other programmable data processors provide steps for implementing the functions or operations specified in one or more blocks of the flowcharts and/or block diagrams. It should also be noted that in some alternative implementations, the functions/operations shown in the blocks may occur in an order different from that shown in the flowcharts. For example, two blocks shown in sequence may actually be executed substantially simultaneously or may sometimes be executed in a reverse order, which depends on the functions/operations involved.

[0030] The optimization method and apparatus for multi-sensor target information fusion according to the invention can be applied, for example, in a scenario where targets around a vehicle are sensed. In such a scenario, for example, a motion state of any one of targets around the vehicle can be characterized by a relative longitudinal position to the vehicle, a longitudinal speed, a longitudinal acceleration, a relative transverse position to the vehicle, a transverse speed, and a transverse acceleration, and each set of measurement results sensed by each of multiple sensors on the vehicle is related to the six items above or can be calculated to obtain values of the six items. Where there are differences in terms of the characteristic of each sensor and the measurement error for each item, the optimization method and apparatus for multi-sensor target information fusion according to the invention that are described in detail below can be used to obtain an optimal fusion result for any target and rationally determine a fusion weight.

[0031] Fig. 1 is an example flowchart of an optimization method for multi-sensor target information fusion according to one implementation of the invention. As shown in Fig. 1, method S100 comprises the step of: obtaining, for each time, a fusion prediction result for all sensors concerning a target state at the current time (step S1).

[0032] In one example, for a time t, the following equation (1) can be used to determine a fusion prediction result concerning the target state of a target:

where (t) is a fusion prediction result for all sensors concerning the target state at the time t, F is a system state transition matrix, X(t-1) is an optimal fusion estimation result concerning the target state at a time t-1 (described later), and W(t) is a system noise.

[0033] In one embodiment, as shown in Fig. 1, method S100 may further comprise the step of: obtaining actual measurement results from each sensor concerning the target state at the current time (step S2).

[0034] In one embodiment, as shown in Fig. 1, method S100 may further comprise the step of: obtaining, for each set of actual measurement results, an optimal estimation result for a corresponding sensor concerning the target state at the current time based on the fusion prediction result and the set of actual measurement results (step S3).

[0035] The details of step S3 will be described in detail in conjunction with Fig. 2.

[0036] Specifically, in one embodiment, as shown in Fig. 2, step S3 comprises the step of: calculating, for each set of actual measurement results, a corresponding transformation matrix based on the fusion prediction result and the set of actual measurement results (step S31).

[0037] In one example, for each set of actual measurement results obtained by each sensor, the following equation (2) can be used to determine the corresponding transformation matrix:

where Zik(t) is a kth set of actual measurement results from an ith sensor concerning the target state at the time t, Hik is a corresponding transformation matrix of the kth set of actual measurement results from the ith sensor, (t) is the fusion prediction result for all the sensors concerning the target state at the time t, and V(t) is a measurement noise.

[0038] In one embodiment, as shown in Fig. 2, step S3 may further comprise the step of: calculating a corresponding covariance of each set of actual measurement results (step S32).

[0039] In one example, for each set of actual measurement results obtained by each sensor, the following equation (3) can be used to determine the corresponding covariance:

where ik(t) is a corresponding covariance of the kth set of actual measurement results from the ith sensor concerning the target state at the time t, F is the system state transition matrix, FT is a transposed matrix of the system state transition matrix, Pik(t - 1) is a corresponding covariance of an optimal estimation result of the kth set of actual measurement results from the ith sensor at the time t-1 (described later), and Q is the covariance of a system process noise.

[0040] In one embodiment, as shown in Fig. 2, step S3 may further comprise the step of: calculating, for each set of actual measurement results, a corresponding Kalman gain based on the corresponding transformation matrix and the corresponding covariance (step S33).

[0041] In one example, for each set of actual measurement results obtained by each sensor, the following equation (4) can be used to calculate the corresponding Kalman gain:

where kgik(t) is a corresponding Kalman gain of the kth set of actual measurement results from the ith sensor concerning the target state at the time t, ik(t) is the corresponding covariance of the kth set of actual measurement results from the ith sensor concerning the target state at the time t, Hik is the corresponding transformation matrix of the kth set of actual measurement results from the ith sensor, HikT is a transposed matrix of the corresponding transformation matrix of the kth set of actual measurement results from the ith sensor, and R is the covariance of a measurement process noise.

[0042] In one embodiment, as shown in Fig. 2, step S3 may further comprise the step of: calculating, for each set of actual measurement results, the corresponding optimal estimation result for the corresponding sensor concerning the target state at the current time based on the fusion prediction result, the corresponding Kalman gain, the set of actual measurement results, and the corresponding transformation matrix (step S34).

[0043] In one example, for each set of actual measurement results obtained by each sensor, the following equation (5) can be used to calculate the corresponding optimal estimation result:

where Xik(t) is a corresponding optimal estimation result of the kth set of actual measurement results from the ith sensor at the time t, (t) is the fusion prediction result for all the sensors concerning the target state at the time t, kgik(t) is the corresponding Kalman gain of the kth set of actual measurement results from the ith sensor concerning the target state at the time t, Zik(t) is the kth set of actual measurement results from the ith sensor concerning the target state at the time t, and Hik is the corresponding transformation matrix of the kth set of actual measurement results from the ith sensor.

[0044] After the optimal estimation result for the corresponding sensor concerning the target state at the current time is calculated for each set of actual measurement results, the process returns to Fig. 1. In one embodiment, as shown in Fig. 1, method S100 may further comprise the step of: fusing optimal estimation results for all the sensors to determine a corresponding weight of each of the optimal estimation results and thus to obtain an optimal fusion estimation result concerning the target state at the current time (step S4).

[0045] In one example, the following equation (6) can be used to obtain the optimal fusion estimation result concerning the target state at the current time:

where X(t) is an optimal fusion estimation result for all the sensors concerning the target state at the time t, f is a fusion function, Xik(t) is the corresponding optimal estimation result of the kth set of actual measurement results from the ith sensor at the time t, and ik(t) is the corresponding covariance of the kth set of actual measurement results from the ith sensor concerning the target state at the time t.

[0046] The details of step S4 will be described in detail in conjunction with Fig. 3.

[0047] Specifically, in one embodiment, as shown in Fig. 3, step S4 comprises the step of: determining the corresponding weight of each of the optimal estimation results according to the corresponding covariance of each set of actual measurement results (step S41).

[0048] In one example, after the equation (3) is used to calculate corresponding covariances (i.e., 11(t),P12(t),...,ik(t) of all sets of actual measurement results from all the sensors, a weight (i.e., w11(t), w12(t),...,wik(t) is assigned to each corresponding optimal estimation result (i.e., X11(t), X12(t),..., Xik(t) of each set of actual measurement results according to the magnitude of the covariances.

[0049] In one embodiment, as shown in Fig. 3, step S4 may further comprise the step of: calculating the optimal fusion estimation result for all the sensors concerning the target state at the current time based on each of the optimal estimation results and the corresponding weight (step S42).

[0050] In one example, the assigned weights (i.e., w11(t), w12(t),..., wik(t) are used to perform a weighted operation on the corresponding optimal estimation results (i.e., X11(t), X12(t),..., Xik(t) so as to obtain the optimal fusion estimation result for all the sensors concerning the target state at the time t. In addition, as shown in the equation (1), X(t) at the time t can also be used to calculate (t+1) at a time t+1.

[0051] Optionally, in one embodiment, method S100 may further comprise the step of: correcting the covariance obtained in step S32 according to the transformation matrix obtained in step S31 and the Kalman gain obtained in step S33, so as to obtain a corrected covariance (step S5, not shown), wherein the corrected covariance can be used to calculate a corresponding covariance of corresponding actual measurement results at the next time of the current time (see equation (3) above).

[0052] In one example, the following equation (7) is used to obtain the corrected covariance at the current time:

where Pik(t) is a corrected covariance of the kth set of actual measurement results from the ith sensor at the time t, I is an identity matrix, kgik(t) is the corresponding Kalman gain of the kth set of actual measurement results from the ith sensor concerning the target state at the time t that is obtained in step S33, Hik is the corresponding transformation matrix of the kth set of actual measurement results from the ith sensor that is obtained in step S31, and ik(t) is the corresponding covariance of the kth set of actual measurement results from the ith sensor concerning the target state at the time t that is obtained in step S32, and in addition, Pik(t) at the time t can also be used to calculate ik(t+1) at the time t+1.

[0053] Weights can be adjusted in real time by means of the steps above, so as to obtain accurate target fusion estimation result.

[0054] In addition, it should be noted that although the sequence between steps is shown in Figs. 1, 2 and 3, those skilled in the art should understand that Figs. 1, 2 and 3 are merely examples, and the sequential relationship between the steps is not limited to the case shown in Figs. 1, 2 and 3. For example, step S1 in Fig. 1 may be performed after step S2, or the two steps may be performed simultaneously. For another example, step S31 in Fig. 2 may be performed after step S32, or the two steps may be performed simultaneously, etc.

[0055] Next, an optimization apparatus for multi-sensor target information fusion for performing the method shown in Fig. 1 is described with reference to Fig. 4.

[0056] As shown in Fig. 4, the apparatus 100 comprises unit 1 101, which is configured to obtain, for each time, a fusion prediction result for all sensors concerning a target state at the current time.

[0057] In one example, for a time t, the following equation (8) can be used to determine a fusion prediction result concerning the target state of a target:

where (t) is a fusion prediction result for all sensors concerning the target state at the time t, F is a system state transition matrix, X(t-1) is an optimal fusion estimation result concerning the target state at a time t-1 (described later), and W(t) is a system noise.

[0058] In one embodiment, as shown in Fig. 4, the apparatus 100 may further comprise unit 2 102, which is configured to obtain actual measurement results from each sensor concerning the target state at the current time.

[0059] In one embodiment, as shown in Fig. 4, the apparatus 100 may further comprise unit 3, which is configured to obtain, for each set of actual measurement results, an optimal estimation result for a corresponding sensor concerning the target state at the current time based on the fusion prediction result and the set of actual measurement results.

[0060] The internal structure of unit 3 103 will be described below in detail.

[0061] Specifically, in one embodiment, unit 3 103 comprises unit 3A (not shown), which is configured to calculate, for each set of actual measurement results, a corresponding transformation matrix based on the fusion prediction result and the set of actual measurement results.

[0062] In one example, for each set of actual measurement results obtained by each sensor, the following equation (9) can be used to determine the corresponding transformation matrix:

where Zik(t) is a kth set of actual measurement results from an ith sensor concerning the target state at the time t, Hik is a corresponding transformation matrix of the kth set of actual measurement results from the ith sensor, (t) is the fusion prediction result for all the sensors concerning the target state at the time t, and V(t) is a measurement noise.

[0063] In one embodiment, unit 3 103 may further comprise unit 3B (not shown), which is configured to calculate a corresponding covariance of each set of actual measurement results.

[0064] In one example, for each set of actual measurement results obtained by each sensor, the following equation (10) can be used to determine the corresponding covariance:

where ik(t) is a corresponding covariance of the kth set of actual measurement results from the ith sensor concerning the target state at the time t, F is the system state transition matrix, FT is a transposed matrix of the system state transition matrix, Pik(t-1) is a corresponding covariance of an optimal estimation result of the kth set of actual measurement results from the ith sensor at the time t-1 (described later), and Q is the covariance of a system process noise.

[0065] In one embodiment, unit 3 103 may further comprise unit 3C (not shown), which is configured to calculate, for each set of actual measurement results, a corresponding Kalman gain based on the corresponding transformation matrix and the corresponding covariance.

[0066] In one example, for each set of actual measurement results obtained by each sensor, the following equation (11) can be used to calculate the corresponding Kalman gain:

where kgik(t) is a corresponding Kalman gain of the kth set of actual measurement results from the ith sensor concerning the target state at the time t, ik(t) is the corresponding covariance of the kth set of actual measurement results from the ith sensor concerning the target state at the time t, Hik is the corresponding transformation matrix of the kth set of actual measurement results from the ith sensor, HikT is a transposed matrix of the corresponding transformation matrix of the kth set of actual measurement results from the ith sensor, and R is the covariance of a measurement process noise.

[0067] In one embodiment, unit 3 103 may further comprise unit 3D (not shown), which is configured to calculate, for each set of actual measurement results, the corresponding optimal estimation result for the corresponding sensor concerning the target state at the current time based on the fusion prediction result, the corresponding Kalman gain, the set of actual measurement results, and the corresponding transformation matrix.

[0068] In one example, for each set of actual measurement results obtained by each sensor, the following equation (12) can be used to calculate the corresponding optimal estimation result:

where Xik(t) is a corresponding optimal estimation result of the kth set of actual measurement results from the ith sensor at the time t, (t) is the fusion prediction result for all the sensors concerning the target state at the time t, kgik(t) is the corresponding Kalman gain of the kth set of actual measurement results from the ith sensor concerning the target state at the time t, Zik(t) is the kth set of actual measurement results from the ith sensor concerning the target state at the time t, and Hik is the corresponding transformation matrix of the kth set of actual measurement results from the ith sensor.

[0069] After the optimal estimation result for the corresponding sensor concerning the target state at the current time is calculated for each set of actual measurement results, the process returns to Fig. 4. In one embodiment, as shown in Fig. 4, the apparatus 100 may further comprise unit 4 104, which is configured to fuse optimal estimation results for all the sensors to determine a corresponding weight of each of the optimal estimation results and thus to obtain an optimal fusion estimation result concerning the target state at the current time.

[0070] In one example, the following equation (13) can be used to obtain the optimal fusion estimation result concerning the target state at the current time:

where X(t) is an optimal fusion estimation result for all the sensors concerning the target state at the time t, f is a fusion function, Xik(t) is the corresponding optimal estimation result of the kth set of actual measurement results from the ith sensor at the time t, and ik(t) is the corresponding covariance of the kth set of actual measurement results from the ith sensor concerning the target state at the time t.

[0071] The internal structure of unit 4 104 will be described below in detail.

[0072] Specifically, in one embodiment, unit 4 104 comprises unit 4A (not shown), which is configured to determine the corresponding weight of each of the optimal estimation results according to the corresponding covariance of each set of actual measurement results.

[0073] In one example, after the equation (10) is used to calculate corresponding covariances (i.e., 11(t), 12(t),...,ik(t) of all sets of actual measurement results from all the sensors, a weight (i.e., w11(t), w12(t),..., wik(t) is assigned to each corresponding optimal estimation result (i.e., X11(t), X12(t),..., Xik(t) of each set of actual measurement results according to the magnitude of the covariances.

[0074] In one embodiment, unit 4 104 may further comprise unit 4B (not shown), which is configured to calculate the optimal fusion estimation result for all the sensors concerning the target state at the current time based on each of the optimal estimation results and the corresponding weight.

[0075] In one example, the assigned weights (i.e., w11(t), w12(t),..., wik(t) are used to perform a weighted operation on the corresponding optimal estimation results (i.e., X11(t), X12(t),.., Xik(t) so as to obtain the optimal fusion estimation result for all the sensors concerning the target state at the time t. In addition, as shown in the equation (8), X(t) at the time t can also be used to calculate (t+1) at a time t+1.

[0076] Optionally, in one embodiment, the apparatus 100 may further comprise unit 5 (not shown), which is configured to correct the covariance obtained in unit 3B according to the transformation matrix obtained in unit 3A and the Kalman gain obtained in unit 3C, so as to obtain a corrected covariance, wherein the corrected covariance can be used to calculate a corresponding covariance of corresponding actual measurement results at the next time of the current time (see equation (10) above).

[0077] In one example, the following equation (14) is used to obtain the corrected covariance at the current time:

where Pik(t) is a corrected covariance of the kth set of actual measurement results from the ith sensor at the time t, I is an identity matrix, kgik(t) is the corresponding Kalman gain of the kth set of actual measurement results from the ith sensor concerning the target state at the time t that is obtained in step S33, Hik is the corresponding transformation matrix of the kth set of actual measurement results from the ith sensor that is obtained in step S31, and ik(t) is the corresponding covariance of the kth set of actual measurement results from the ith sensor concerning the target state at the time t that is obtained in step S32, and in addition, Pik(t) at the time t can also be used to calculate ik(t+1) at the time t+1.

[0078] Weights can be adjusted in real time by means of the units above, so as to obtain an accurate target fusion estimation result.

[0079] When applied to auxiliary driving, the optimization method and apparatus for multi-sensor target information fusion according to one implementation of the invention can enable an auxiliary driving system to employ more optimized data, thereby facilitating its decision-making and control, for example, making, based on the optimized data, a better decision for auxiliary driving functions or scenarios, such as adaptive cruise and emergency braking, with such functions or scenarios further including vehicle stability control and the like.

[0080] Although implementations of the optimization method and apparatus for multi-sensor target information fusion have been mainly described in the invention, the invention is not limited to these implementations, and the invention may be implemented in a way of: an auxiliary driving method comprising the method described above or an auxiliary driving system comprising the apparatus described above, or a computer device for performing the method described above or a computer program for performing the method described above, or a computer program for implementing functions of the apparatus described above or a computer-readable recording medium having the computer program recorded thereon.

[0081] Fig. 5 shows a computer device according to one implementation of the invention that is configured to perform an optimization method for multi-sensor target information fusion according to one implementation of the invention. As shown in Fig. 5, a computer device 200 comprises a memory 201 and a processor 202. Although not shown, the computer device 200 further comprises a computer program that is stored on the memory 201 and operable on the processor 202. The program implements, when executed by the processor, the steps, for example as shown in Figs. 1, 2 and 3, of an optimization method for multi-sensor target information fusion according to one implementation of the invention.

[0082] Compared with the prior art, the method S100 or the apparatus 100 according to one implementation of the invention can obtain one or more of the beneficial effects as follows:
  1. (1) a fusion weight can be calculated in real time to ensure that the result of weight assignment is appropriate in most scenarios; and
  2. (2) Kalman filtering is applied to each set of measurement results from each sensor to obtain a corresponding covariance and optimal estimation result for all the measurement results, and then an optimal fusion estimation result concerning a target state is calculated, so that each corresponding weight of the optimal estimation result can be determined in conjunction with the covariance of the measurement results from the sensor. Therefore, even if the sensor is affected by hardware performance and environment and a detection result does not conform to a physical motion principle, the weight can be adjusted in real time to realize the optimal fusion estimation of the target and to improve the performance of fusion estimation.


[0083] Fig. 6 is an example flowchart of a method for synchronizing multi-sensor target information fusion with multi-sensor sensing according to one implementation of the invention. As shown in Fig. 6, method P100 comprises the step of: obtaining actual measurement results from each sensor concerning a target state and recording a time at which each set of actual measurement results is obtained, so as to record a first time stamp for each set of actual measurement results (step P1).

[0084] In one example, after a set of actual measurement results is received, a corresponding receiving time stamp is recorded. For example, a receiving time stamp of a first set of actual measurement results from a first sensor is marked as t111, a receiving time stamp of a second set of actual measurement results from the first sensor is marked as t112, ..., and a receiving time stamp of a kth set of actual measurement results from an ith sensor is marked as tlik. These time stamps are collectively referred to as a first time stamp (i.e., a receiving time stamp) herein.

[0085] In one embodiment, as shown in Fig. 6, method P100 may further comprise the step of: recording, as a second time stamp, a time at which target information fusion processing starts to be performed (step P2).

[0086] In one example, a time at which fusion processing is performed on all sets of actual measurement results from all sensors is recorded as a second time stamp (i.e., a fusion time stamp).

[0087] In one embodiment, as shown in Fig. 6, method P100 may further comprise the step of: calculating a time difference between each first time represented by each first time stamp and a second time represented by the second time stamp respectively (step P3).

[0088] In one example, the following equation (15) can be used to calculate the time difference:

where Δtik is a corresponding time difference of the kth set of actual measurement results from the ith sensor, tlik is a first time represented by a corresponding first time stamp of the kth set of actual measurement results from the ith sensor, and t2 is the second time represented by the second time stamp.

[0089] In one embodiment, as shown in Fig. 6, method P100 may further comprise the step of: updating corresponding actual measurement results obtained at the first time based on each calculated time difference to obtain corresponding estimated measurement results at the second time (step P4).

[0090] In one example, assuming that the displacement of a vehicle changes within the time of Δtik, the following equations (16) to (18) can be used to update actual measurement results obtained at tlik:





where vcs is a longitudinal speed of the vehicle, vcs is a transverse speed of the vehicle, ΔXvcs_ik is the displacement of the vehicle in the longitudinal direction within Δt that corresponds to the kth set of actual measurement results from the ith sensor, ΔYvcs_ik is the displacement of the vehicle in the transverse direction within Δt that corresponds to the kth set of actual measurement results from the ith sensor, ik(t2|tlik) is estimated measurement results at the second time t2 that correspond to the kth set of actual measurement results from the ith sensor, ω is a deflection angle of the vehicle within Δt that corresponds to the kth set of actual measurement results from the ith sensor, Zik_x(tlik) is a longitudinal component of the kth set of actual measurement results from the ith sensor that are obtained at the first time, and Zik_y(tlik) is a transverse component of the kth set of actual measurement results from the ith sensor that are obtained at the first time.

[0091] In one embodiment, as shown in Fig. 6, method P100 may further comprise the step of: obtaining a fusion prediction result for all sensors concerning the target state at the second time (step P5).

[0092] In one example, for the second time t2, the following equation (19) can be used to determine a fusion prediction result concerning the target state of a target:

where X(t2) is the fusion prediction result for all the sensors concerning the target state at the second time t2, F is a system state transition matrix, X(t1) is an optimal fusion estimation result concerning the target state at the first time t1 (described later), and W(t2) is a system noise.

[0093] In one embodiment, as shown in Fig. 6, method P100 may further comprise the step of: obtaining, for each set of estimated measurement results, an optimal estimation result for a corresponding sensor concerning the target state at the second time based on the fusion prediction result and the set of estimated measurement results (step P6).

[0094] The details of step P6 will be described in detail in conjunction with Fig. 7.

[0095] Specifically, in one embodiment, as shown in Fig. 7, step P6 comprises the step of: calculating, for each set of estimated measurement results, a corresponding transformation matrix based on the fusion prediction result and the set of estimated measurement results (step P61).

[0096] In one example, for each set of estimated measurement results, the following equation (20) can be used to determine the corresponding transformation matrix:

where ik(t2|tlik) is estimated measurement results at the second time t2 that correspond to the kth set of actual measurement results from the ith sensor, Hik is a transformation matrix corresponding to the kth set of actual measurement results from the ith sensor, (t2) is the fusion prediction result for all the sensors concerning the target state at the second time t2, V(t2) is a measurement noise, ftik) is a function for calculating a weight of the measurement noise according to Δt, and the larger Δtik indicates the larger measurement noise.

[0097] In one embodiment, as shown in Fig. 7, step P6 may further comprise the step of: calculating a covariance corresponding to each set of estimated measurement results (step P62).

[0098] In one example, for each set of estimated measurement results, the following equation (21) can be used to determine the corresponding covariance:

where ik(t2) is a corresponding covariance of the kth set of estimated measurement results from the ith sensor concerning the target state at the second time t2, F is the system state transition matrix, FT is a transposed matrix of the system state transition matrix, Pik(tlik) is a corresponding covariance of an optimal estimation result of the kth set of estimated measurement results from the ith sensor at the first time t1 (described later), and Q is the covariance of a system process noise.

[0099] In one embodiment, as shown in Fig. 7, step P6 may further comprise the step of: calculating, for each set of estimated measurement results, a corresponding Kalman gain based on the corresponding transformation matrix and the corresponding covariance (step P63).

[0100] In one example, for each set of estimated measurement results, the following equation (22) can be used to calculate the corresponding Kalman gain:

where kgik(t2) is a corresponding Kalman gain of the kth set of estimated measurement results from the ith sensor concerning the target state at the second time t2, ik(t2) is the corresponding covariance of the kth set of estimated measurement results from the ith sensor concerning the target state at the second time t2, Hik is the corresponding transformation matrix of the kth set of actual measurement results from the ith sensor, HikT is a transposed matrix of the corresponding transformation matrix of the kth set of actual measurement results from the ith sensor, and R is the covariance of a measurement process noise.

[0101] In one embodiment, as shown in Fig. 7, step P6 may further comprise the step of: calculating, for each set of estimated measurement results, the corresponding optimal estimation result for the corresponding sensor concerning the target state at the second time based on the fusion prediction result, the corresponding Kalman gain, the set of estimated measurement results, and the corresponding transformation matrix (step P64).

[0102] In one example, for each set of estimated measurement results, the following equation (23) can be used to calculate the corresponding optimal estimation result:

where Xik(t2) is a corresponding optimal estimation result of the kth set of estimated measurement results from the ith sensor at the second time t2, (t2) is the fusion prediction result for all the sensors concerning the target state at the second time t2, kgik(t2) is the corresponding Kalman gain of the kth set of estimated measurement results from the ith sensor concerning the target state at the second time t2, ik(t2|tlik) is the estimated measurement results at the second time t2 that correspond to the kth set of actual measurement results from the ith sensor, and Hik is the corresponding transformation matrix of the kth set of actual measurement results from the ith sensor.

[0103] Through the above steps, an accurate optimal estimation result concerning the target state can be obtained for each set of actual measurement results.

[0104] Optionally, in one embodiment, after the optimal estimation result corresponding to each set of actual measurement results from each sensor is calculated, method P100 may further comprise the step of: fusing optimal estimation results for all the sensors to determine a corresponding weight of each of the optimal estimation results and thus to obtain an optimal fusion estimation result concerning the target state at the second time (step P7, not shown).

[0105] In one example, the following equation (24) can be used to obtain the optimal fusion estimation result concerning the target state at the second time t2:

where X(t2) is an optimal fusion estimation result for all the sensors concerning the target state at the second time t2, f is a fusion function, Xik(t2) is the corresponding optimal estimation result of the kth set of estimated measurement results from the ith sensor at the second time t2, and ik(t2) is the corresponding covariance of the kth set of estimated measurement results from the ith sensor concerning the target state at the second time t2. In addition, as shown in the equation (19), an optimal fusion estimation result X(t1) at the current time (for example, the first time t1) can also be used to calculate a fusion prediction result (t2) at the next time (for example, the second time t2).

[0106] Optionally, in one embodiment, method P100 may further comprise the step of: correcting the covariance obtained in step P62 according to the transformation matrix obtained in step P61 and the Kalman gain obtained in step P63, so as to obtain a corrected covariance (step P8, not shown), wherein the corrected covariance can be used to calculate a corresponding covariance of estimated measurement results at the next time (for example, a third time t3) of the current time (for example, the second time t2) (see equation (21) above).

[0107] In one example, the following equation (25) is used to obtain a corrected covariance at the current time (for example, the second time t2):

where Pik(t2) is a corrected covariance of the kth set of estimated measurement results from the ith sensor at the second time t2, I is an identity matrix, kgik(t2) is the corresponding Kalman gain of the kth set of estimated measurement results from the ith sensor concerning the target state at the second time t2 that is obtained in step P63, Hik is the corresponding transformation matrix of the kth set of actual measurement results from the ith sensor that is obtained in step P61, and ik(t2) is the corresponding covariance of the kth set of estimated measurement results from the ith sensor concerning the target state at the second time t2 that is obtained in step P62, and in addition, Pik(t2) at the current time (for example, the second time t2) can also be used to calculate ik(t3) at the next time (for example, the third time t3).

[0108] In addition, it should be noted that although the sequence between steps is shown in Figs. 6 and 7, those skilled in the art should understand that Figs. 6 and 7 are merely examples, and the sequential relationship between the steps is not limited to the case shown in Figs. 6 and 7. For example, step P5 in Fig. 6 may be performed before step P4, or the two steps may be performed simultaneously. For another example, step P61 in Fig. 7 may be performed after step P62, or the two steps may be performed simultaneously, etc.

[0109] Next, an apparatus for synchronizing multi-sensor target information fusion with multi-sensor sensing is described with reference to Fig. 8.

[0110] As shown in Fig. 8, the apparatus 800 comprises unit 6 801, which is configured to obtain actual measurement results from each sensor concerning a target state and record a time at which each set of actual measurement results is obtained, so as to record a first time stamp for each set of actual measurement results.

[0111] In one example, after a set of actual measurement results is received, a corresponding receiving time stamp is recorded. For example, a receiving time stamp of a first set of actual measurement results from a first sensor is marked as t111, a receiving time stamp of a second set of actual measurement results from the first sensor is marked as t112, ..., and a receiving time stamp of a kth set of actual measurement results from an ith sensor is marked as tlik. These time stamps are collectively referred to as a first time stamp (i.e., a receiving time stamp) herein.

[0112] In one embodiment, as shown in Fig. 8, the apparatus 800 may further comprise unit 7 802, which is configured to record, as a second time stamp, a time at which target information fusion processing starts to be performed.

[0113] In one example, a time at which fusion processing is performed on all sets of actual measurement results from all sensors is recorded as a second time stamp (i.e., a fusion time stamp).

[0114] In one embodiment, as shown in Fig. 8, the apparatus 800 may further comprise unit 8 803, which is configured to calculate a time difference between each first time represented by each first time stamp and a second time represented by the second time stamp respectively.

[0115] In one example, the following equation (26) can be used to calculate the time difference:

where Δtik, is a corresponding time difference of the kth set of actual measurement results from the ith sensor, tlik is a first time represented by a corresponding first time stamp of the kth set of actual measurement results from the ith sensor, and t2 is the second time represented by the second time stamp.

[0116] In one embodiment, as shown in Fig. 8, the apparatus 800 may further comprise unit 9 804, which is configured to update corresponding actual measurement results obtained at the first time based on each calculated time difference to obtain corresponding estimated measurement results at the second time.

[0117] In one example, assuming that the displacement of a vehicle changes within the time of Δtik, the following equations (27) to (29) can be used to update actual measurement results obtained at tlik:





where vcs is a longitudinal speed of the vehicle, vcs is a transverse speed of the vehicle, ΔXvcs_ik is the displacement of the vehicle in the longitudinal direction within Δt that corresponds to the kth set of actual measurement results from the ith sensor, ΔYvcs_ik is the displacement of the vehicle in the transverse direction within Δt that corresponds to the kth set of actual measurement results from the ith sensor, ik(t2|tlik) is estimated measurement results at the second time t2 that correspond to the kth set of actual measurement results from the ith sensor, ω is a deflection angle of the vehicle within Δt that corresponds to the kth set of actual measurement results from the ith sensor, Zik_x(tlik) is a longitudinal component of the kth set of actual measurement results from the ith sensor that are obtained at the first time, and Zik_y(tlik) is a transverse component of the kth set of actual measurement results from the ith sensor that are obtained at the first time.

[0118] In one embodiment, as shown in Fig. 8, the apparatus 800 may further comprise unit 10 805, which is configured to obtain a fusion prediction result for all sensors concerning the target state at the second time.

[0119] In one example, for the second time t2, the following equation (30) can be used to determine a fusion prediction result concerning the target state of a target:

where (t2) is the fusion prediction result for all the sensors concerning the target state at the second time t2, F is a system state transition matrix, X(t1) is an optimal fusion estimation result concerning the target state at the first time t1 (described later), and W(t2) is a system noise.

[0120] In one embodiment, as shown in Fig. 8, the apparatus 800 may further comprise unit 11 806, which is configured to obtain, for each set of estimated measurement results, an optimal estimation result for a corresponding sensor concerning the target state at the second time based on the fusion prediction result and the set of estimated measurement results.

[0121] The details of unit 11 806 will be described below in detail.

[0122] Specifically, in one embodiment, unit 11 806 comprises unit 6A (not shown), which is configured to calculate, for each set of estimated measurement results, a corresponding transformation matrix based on the fusion prediction result and the set of estimated measurement results.

[0123] In one example, for each set of estimated measurement results, the following equation (31) can be used to determine the corresponding transformation matrix:

where ik(t2|tlik) is estimated measurement results at the second time t2 that correspond to the kth set of actual measurement results from the ith sensor, Hik is a transformation matrix corresponding to the kth set of actual measurement results from the ith sensor, (t2) is the fusion prediction result for all the sensors concerning the target state at the second time t2, V(t2) is a measurement noise, ftik) is a function for calculating a weight of the measurement noise according to Δt, and the larger Δtik indicates the larger measurement noise.

[0124] In one embodiment, unit 11 806 may further comprise unit 6B (not shown), which is configured to calculate a covariance corresponding to each set of estimated measurement results.

[0125] In one example, for each set of estimated measurement results, the following equation (32) can be used to determine the corresponding covariance:

where ik(t2) is a corresponding covariance of the kth set of estimated measurement results from the ith sensor concerning the target state at the second time t2, F is the system state transition matrix, FT is a transposed matrix of the system state transition matrix, ik(tlik) is a corresponding covariance of an optimal estimation result of the kth set of estimated measurement results from the ith sensor at the first time t1 (described later), and Q is the covariance of a system process noise.

[0126] In one embodiment, unit 11 806 may further comprise unit 6C (not shown), which is configured to calculate, for each set of estimated measurement results, a corresponding Kalman gain based on the corresponding transformation matrix and the corresponding covariance.

[0127] In one example, for each set of estimated measurement results, the following equation (33) can be used to calculate the corresponding Kalman gain:

where kgik(t2) is a corresponding Kalman gain of the kth set of estimated measurement results from the ith sensor concerning the target state at the second time t2, ik(t2) is the corresponding covariance of the kth set of estimated measurement results from the ith sensor concerning the target state at the second time t2, Hik is the corresponding transformation matrix of the kth set of actual measurement results from the ith sensor, HikT is a transposed matrix of the corresponding transformation matrix of the kth set of actual measurement results from the ith sensor, and R is the covariance of a measurement process noise.

[0128] In one embodiment, unit 11 806 may further comprise unit 6D, which is configured to calculate, for each set of estimated measurement results, the corresponding optimal estimation result for the corresponding sensor concerning the target state at the second time based on the fusion prediction result, the corresponding Kalman gain, the set of estimated measurement results, and the corresponding transformation matrix.

[0129] In one example, for each set of estimated measurement results, the following equation (34) can be used to calculate the corresponding optimal estimation result:

where Xik(t2) is a corresponding optimal estimation result of the kth set of estimated measurement results from the ith sensor at the second time t2, (t2) is the fusion prediction result for all the sensors concerning the target state at the second time t2, kgik(t2) is the corresponding Kalman gain of the kth set of estimated measurement results from the ith sensor concerning the target state at the second time t2, ik(t2|tlik) is the estimated measurement results at the second time t2 that correspond to the kth set of actual measurement results from the ith sensor, and Hik is the corresponding transformation matrix of the kth set of actual measurement results from the ith sensor.

[0130] Through the above units, an accurate optimal estimation result concerning the target state can be obtained for each set of actual measurement results.

[0131] Optionally, in one embodiment, the apparatus 800 may further comprise unit 8 (not shown), which is configured to fuse optimal estimation results for all the sensors to determine a corresponding weight of each of the optimal estimation results and thus to obtain an optimal fusion estimation result concerning the target state at the second time.

[0132] In one example, the following equation (35) can be used to obtain the optimal fusion estimation result concerning the target state at the second time t2:

where X(t2) is an optimal fusion estimation result for all the sensors concerning the target state at the second time t2, f is a fusion function, Xik(t2) is the corresponding optimal estimation result of the kth set of estimated measurement results from the ith sensor at the second time t2, and ik(t2) is the corresponding covariance of the kth set of estimated measurement results from the ith sensor concerning the target state at the second time t2. In addition, as shown in the equation (30), an optimal fusion estimation result X(t1) at the current time (for example, the first time t1) can also be used to calculate a fusion prediction result (t2) at the next time (for example, the second time t2).

[0133] Optionally, in one embodiment, the apparatus 800 may further comprise unit 9 (not shown), which is configured to correct the covariance obtained in unit 6B according to the transformation matrix obtained in unit 6A and the Kalman gain obtained in unit 6C, so as to obtain a corrected covariance, wherein the corrected covariance can be used to calculate a corresponding covariance of estimated measurement results at the next time (for example, a third time t3) of the current time (for example, the second time t2) (see equation (32) above).

[0134] In one example, the following equation (36) is used to obtain a corrected covariance at the current time (for example, the second time t2):

where Pik(t2) is a corrected covariance of the kth set of estimated measurement results from the ith sensor at the second time t2, I is an identity matrix, kgik(t2) is the corresponding Kalman gain of the kth set of estimated measurement results from the ith sensor concerning the target state at the second time t2 that is obtained in step P63, Hik is the corresponding transformation matrix of the kth set of actual measurement results from the ith sensor that is obtained in step P61, and ik(t2) is the corresponding covariance of the kth set of estimated measurement results from the ith sensor concerning the target state at the second time t2 that is obtained in step P62, and in addition, Pik(t2) at the current time (for example, the second time t2) can also be used to calculate ik(t3) at the next time (for example, the third time t3).

[0135] When applied to auxiliary driving, the method and apparatus for synchronizing multi-sensor target information fusion with multi-sensor sensing according to one implementation of the invention can enable an auxiliary driving system to employ more optimized data, thereby facilitating its decision-making and control, for example, making, based on the optimized data, a better decision for auxiliary driving functions or scenarios, such as adaptive cruise and emergency braking, with such functions or scenarios further including vehicle stability control and the like.

[0136] Although implementations of the method and apparatus for synchronizing multi-sensor target information fusion with multi-sensor sensing have been mainly described in the invention, the invention is not limited to these implementations, and the invention may be implemented in a way of: an auxiliary driving method comprising the method described above or an auxiliary driving system comprising the apparatus described above, or a computer device for performing the method described above or a computer program for performing the method described above, or a computer program for implementing functions of the apparatus described above or a computer-readable recording medium having the computer program recorded thereon.

[0137] Fig. 9 shows a computer device according to one implementation of the invention that is configured to perform a method for synchronizing multi-sensor target information fusion with multi-sensor sensing according to one implementation of the invention. As shown in Fig. 9, a computer device 900 comprises a memory 901 and a processor 902. Although not shown, the computer device 900 further comprises a computer program that is stored on the memory 901 and operable on the processor 902. The program implements, when executed by the processor, the steps, for example as shown in Figs. 6 and 7, of a method for synchronizing multi-sensor target information fusion with multi-sensor sensing according to one implementation of the invention.

[0138] Compared with the prior art, the method P100 or the apparatus 800 according to one implementation of the invention can obtain one or more of the beneficial effects as follows:
  1. (1) it can be ensured that a sufficiently accurate fusion result can be obtained even if sensing results of sensors are not updated or have a relatively long update cycle; and
  2. (2) in addition to a scenario in which the sensing results of the sensors are not updated or have a relatively long update cycle, it can also be ensured that a sufficiently accurate fusion result can be obtained even if sensors have different cycles for updating sensing results thereof.


[0139] In addition, as described above, the invention may also be implemented as a recording medium, which stores a program for enabling a computer to perform an optimization method for multi-sensor target information fusion according to one implementation of the invention.

[0140] The invention may also be implemented as a recording medium, which stores a program for enabling a computer to perform a method for synchronizing multi-sensor target information fusion with multi-sensor sensing according to another implementation of the invention.

[0141] Here, various types of recording media, such as disks (e.g., a magnetic disk, an optical disk, etc.), cards (e.g., a memory card, an optical card, etc.), semiconductor memories (e.g., a ROM, a non-volatile memory, etc.), and tapes (e.g., a magnetic tape, a cassette tape, etc.), can be used as the recording medium.

[0142] By recording, in these recording media, a computer program that enables a computer to perform the optimization method for multi-sensor target information fusion in the implementations above or a computer program that enables a computer to implement functions of the optimization apparatus for multi-sensor target information fusion in the implementations above and circulating the computer program, costs are reduced, and portability and versatility are improved.

[0143] Furthermore, the recording medium is loaded on a computer, the computer reads the computer program recorded in the recording medium and stores same in a memory, and a processor (central processing unit (CPU) and a micro processing unit (MPU)) provided on the computer reads out and executes the computer program from the memory, whereby the optimization method for multi-sensor target information fusion in the implementations above can be performed, and functions of the optimization apparatus for multi-sensor target information fusion in the implementations above can be implemented.

[0144] Those of ordinary skill in the art should understand that the invention is not limited to the implementations above, and the invention can be implemented in many other forms without departing from the essence and scope thereof. Therefore, the presented examples and implementations are regarded to be schematic rather than restrictive, and without departing from the spirit and scope of the invention that are defined by the appended claims, the invention may cover various changes and replacements.


Claims

1. An optimization method for multi-sensor target information fusion, characterized by comprising:

step S1: obtaining, for each time, a fusion prediction result for all sensors concerning a target state at the current time;

step S2: obtaining actual measurement results from each sensor concerning the target state at the current time;

step S3: obtaining, for each set of actual measurement results, an optimal estimation result for a corresponding sensor concerning the target state at the current time based on the fusion prediction result and the set of actual measurement results; and

step S4: fusing optimal estimation results for all the sensors to determine a corresponding weight of each of the optimal estimation results and thus to obtain an optimal fusion estimation result concerning the target state at the current time.


 
2. The optimization method according to claim 1, characterized in that step S3 comprises:

step S31: calculating, for each set of actual measurement results, a corresponding transformation matrix based on the fusion prediction result and the set of actual measurement results;

step S32: calculating a corresponding covariance of each set of actual measurement results;

step S33: calculating, for each set of actual measurement results, a corresponding Kalman gain based on the corresponding transformation matrix and the corresponding covariance; and

step S34: calculating, for each set of actual measurement results, the corresponding optimal estimation result for the corresponding sensor concerning the target state at the current time based on the fusion prediction result, the corresponding Kalman gain, the set of actual measurement results, and the corresponding transformation matrix.


 
3. The optimization method according to claim 2, characterized in that step S4 comprises:

step S41: determining the corresponding weight of each of the optimal estimation results according to the corresponding covariance of each set of actual measurement results; and

step S42: calculating the optimal fusion estimation result for all the sensors concerning the target state at the current time based on each of the optimal estimation results and the corresponding weight.


 
4. The optimization method according to claim 2, characterized by further comprising:
step S5: correcting the covariance obtained in step S32 according to the transformation matrix obtained in step S31 and the Kalman gain obtained in step S33, so as to obtain a corrected covariance.
 
5. The optimization method according to claim 4, characterized in that in step S32, the corresponding covariance of each set of actual measurement results at the current time is obtained with a corrected covariance obtained at a previous time.
 
6. The method according to any of claims 1 to 5, characterized in that in step S1, the fusion prediction result at the current time is obtained with an optimal fusion estimation result concerning the target state at the previous time.
 
7. A method for synchronizing multi-sensor target information fusion with multi-sensor sensing, characterized by comprising:

step P1: obtaining actual measurement results from each sensor concerning a target state and recording a time at which each set of actual measurement results is obtained, so as to record a first time stamp for each set of actual measurement results;

step P2: recording, as a second time stamp, a time at which target information fusion processing starts to be performed;

step P3: calculating a time difference between each first time represented by each first time stamp and a second time represented by the second time stamp respectively;

step P4: updating corresponding actual measurement results obtained at the first time based on each calculated time difference to obtain corresponding estimated measurement results at the second time;

step P5: obtaining a fusion prediction result for all sensors concerning the target state at the second time; and

step P6: obtaining, for each set of estimated measurement results, an optimal estimation result for a corresponding sensor concerning the target state at the second time based on the fusion prediction result and the set of estimated measurement results.


 
8. The method according to claim 7, characterized in that step P6 comprises:

step P61: calculating, for each set of estimated measurement results, a corresponding transformation matrix based on the fusion prediction result and the set of estimated measurement results;

step P62: calculating a corresponding covariance of each set of estimated measurement results;

step P63: calculating, for each set of estimated measurement results, a corresponding Kalman gain based on the corresponding transformation matrix and the corresponding covariance; and

step P64: calculating, for each set of estimated measurement results, the corresponding optimal estimation result for the corresponding sensor concerning the target state at the second time based on the fusion prediction result, the corresponding Kalman gain, the set of estimated measurement results, and the corresponding transformation matrix.


 
9. The method according to claim 8, characterized by further comprising:
step P7: fusing optimal estimation results for all the sensors to determine a corresponding weight of each of the optimal estimation results and thus to obtain an optimal fusion estimation result concerning the target state at the second time.
 
10. The method according to claim 8, characterized by further comprising:
step P8: correcting the covariance obtained in step P62 according to the transformation matrix obtained in step P61 and the Kalman gain obtained in step P63, so as to obtain a corrected covariance.
 
11. The method according to claim 10, characterized in that in step P62, the corresponding covariance of each set of estimated measurement results at the second time is obtained with a corrected covariance at the first time.
 
12. The method according to claim 9, characterized in that in step P1, the fusion prediction result at the second time is obtained with an optimal fusion estimation result concerning the target state at the first time.
 
13. A computer device, comprising a memory, a processor, and a computer program stored on the memory and operable on the processor, characterized in that the program implements, when executed by the processor, the steps of the method according to any of claims 1 to 12.
 
14. A recording medium having a computer program stored thereon, characterized in that the program can be executed by a computer to implement the steps of the method according to any of claims 1 to 12.
 
15. An auxiliary driving method, characterized by comprising:
an optimization method for multi-sensor target information fusion according to any of claims 1 to 6; and/or
a method for synchronizing multi-sensor target information fusion with multi-sensor sensing according to any of claims 7 to 12.
 




Drawing



















Search report