(19)
(11) EP 4 036 882 A1

(12) EUROPEAN PATENT APPLICATION

(43) Date of publication:
03.08.2022 Bulletin 2022/31

(21) Application number: 21153893.9

(22) Date of filing: 28.01.2021
(51) International Patent Classification (IPC): 
G08B 13/193(2006.01)
(52) Cooperative Patent Classification (CPC):
G08B 13/193
(84) Designated Contracting States:
AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR
Designated Extension States:
BA ME
Designated Validation States:
KH MA MD TN

(71) Applicant: Hochschule für angewandte Wissenschaften Landshut
84036 Landshut (DE)

(72) Inventor:
  • IVANOV, Artem
    84032 Landshut (DE)

(74) Representative: Koplin, Moritz 
Anne-Conway-Straße 1
28359 Bremen
28359 Bremen (DE)

   


(54) DETECTION OF A MOVEMENT OF A BODY IN AN AREA WHICH IS TO BE MONITORED


(57) Provided is a device for detecting a movement of a body in an area which is to be monitored by the device. The device comprises a first sensor which is configured to output a first signal that is indicative of a first intensity of electromagnetic radiation incident upon a first surface region of said first sensor and a second sensor which is configured to output a second signal that is indicative of a second intensity of electromagnetic radiation incident upon a second surface region of said second sensor. The device further comprises an arrangement which images discrete first portions of the area onto the first surface region and discrete second portions of the area onto the second surface region, thereby segmenting the area along a first curve traversing the area from a first side of the area to a second side of the area, into a sequence comprising alternating, overlapping or non-overlapping, discrete first and discrete second portions.




Description

FIELD



[0001] The present disclosure relates to the detection of a movement of a body in an area which is to be monitored. In particular, the present disclosure relates to the detection of a movement of a body based on pyroelectric sensors.

BACKGROUND



[0002] The use of pyroelectric sensors for detecting a movement of a body in an area which is to be monitored is known, for example, from DE 42 18 151 C2. However, conventional pyroelectric sensors may suffer from blind spots or weak detection accuracy in parts of the area.

SUMMARY



[0003] The present invention is directed at a device for and a method of detecting a movement of a body in an area.

[0004] The device comprises a first sensor which is configured to output a first signal that is indicative of a first intensity of electromagnetic radiation incident upon a first surface region of said first sensor, a second sensor which is configured to output a second signal that is indicative of a second intensity of electromagnetic radiation incident upon a second surface region of said second sensor, and an arrangement which images discrete first portions of the area onto the first surface region and discrete second portions of the area onto the second surface region, thereby segmenting the area along a first curve traversing the area from a first side of the area to a second side of the area, into a sequence comprising alternating, overlapping or non-overlapping, discrete first and discrete second portions.

[0005] In this regard, the term "electromagnetic radiation", as used throughout the description and the claims, particularly refers to electromagnetic waves having a wavelength within the infrared spectrum (e.g., 700 nm to 1 mm). Moreover, the term "body", as used throughout the description and the claims, particularly refers, but is not limited, to a body of a human being. I.e., the term "body", as used throughout the description and the claims, shall encompass any object of whatever shape or form that is moved (or moves) through the area.

[0006] Furthermore, the wording "discrete first portions", "discrete second portions", "discrete third portions", and "discrete fourth portions", as used throughout the description and the claims, particularly refers to, first portions which neither overlap with nor abut on any other first portion, second portions which neither overlap with nor abut on any other second portion, third portions which neither overlap with nor abut on any other third portion and fourth portions which neither overlap with nor abut on any other fourth portion, respectively. In this regard, the term "area", as used throughout the description and the claims, refers to the three-dimensional space (volume) in which a moving body may be detected.

[0007] Moreover, the words "first", "second", "third", etc., are used in a nominal sense rather than in an ordinal sense throughout the description and the claims, unless the context clearly provides otherwise. So, for example, the wording "a second surface region of said second sensor" does not necessarily mean that the second sensor also has a first surface region, although that is entirely possible. Instead, this wording is used to avoid confusion which might otherwise occur when referring to the "first surface region". In fact, each sensor may have exactly one (active) surface region so that a signal which is output by a sensor may only be indicative of electromagnetic radiation incident upon said (single) surface region.

[0008] In addition, the term "sequence", as used throughout the description and the claims, particularly refers to a succession of individual portions, wherein the succession contains each individual portion only once. In this regard, the formulation "alternating first and second portions", as used throughout the description and the claims, particularly refers to a sequence of the type "ABABA ..." or "BABAB ..." where "A" stands for a first portion and "B" stands for a second portion. Analogously, the formulation "alternating discrete third and fourth portions", as used throughout the description and the claims, particularly refers to a sequence of the type "CDCDC ..." or "DCDCD ..." where "C" stands for a discrete third portion and "D" stands for a discrete fourth portion. Finally, the formulation that the arrangement "images discrete portions of the area onto a surface region", as used throughout the description and the claims, particularly refers to a projection where each discrete portion projected onto the surface region covers the whole surface region. In other words, the surface region may "see" through the arrangement all of, but not more than, the discrete portions imaged onto it.

[0009] The first curve may lie in a first plane. For example, the first curve may be a horizontal curve.

[0010] The first curve may be a (straight) line, a curve comprising one or more straight segments, a curve consisting of straight segments or a curve that has no straight segment.

[0011] The first sensor may comprise a first pyroelectric element and the second sensor may comprise a second pyroelectric element, wherein the first surface region is a surface region of the first pyroelectric element and the second surface region is a surface region of the second pyroelectric element.

[0012] For example, the first signal may be a first voltage generated by the first pyroelectric element and the second signal may be a second voltage generated by the second pyroelectric element.

[0013] Alternatively, or in addition, the first sensor may comprise a first thermopile element and the second sensor may comprise a second thermopile element, wherein the first surface region is a surface region of the first thermopile element and the second surface region is a surface region of the second thermopile element.

[0014] Alternatively, or in addition, the first sensor may be a first bolometer and the second sensor may be a second bolometer, wherein the first surface region is a surface region of an absorptive element of the first bolometer and the second surface region is a surface region of an absorptive element of the second bolometer.

[0015] Hence, a body may only be detectable if its surface temperature differs from the temperature of a background which it temporarily occludes. Accordingly, the formulation "detecting a movement of a body in an area" is to be understood to mean that detecting a movement of a body in an area is desired but may not be achieved (with sufficient accuracy) under each and any circumstances.

[0016] The arrangement may comprise a plurality of optical elements. The optical elements may be mirrors and/or lenses. Moreover, the plurality of optical elements may be distinct sections of a single lens or mirror (wherein different sections differ with regard to their imaging properties).

[0017] For example, the arrangement may comprise a first lens and a second lens, wherein a first fraction of the electromagnetic radiation incident upon the first surface region of the first sensor travels through the first lens, a second fraction of the electromagnetic radiation incident upon the first surface region of the first sensor travels through the second lens, and a first fraction of the electromagnetic radiation incident upon the second surface region of the second sensor travels through the first lens, and a second fraction of the electromagnetic radiation incident upon the second surface region of the second sensor travels through the second lens.

[0018] A first optical axis of the first lens and a second optical axis of the second lens may be parallel and/or the first surface region and the second surface region may be coplanar.

[0019] Accordingly, keeping the images focused on the surface regions may be facilitated. For example, a first sensor-side focal point of the first lens and a second sensor-side focal point of the second lens may lie in a plane defined by the first surface region and the second surface region.

[0020] There may be a horizontal gap and/or a vertical gap between the first surface region and the second surface region.

[0021] The horizontal gap between the first surface region and the second surface region may be substantially zero or a multiple of a horizontal distance between the first optical axis and the second optical axis.

[0022] The horizontal distance between the first optical axis and the second optical axis may be equal to or larger than a combined width of the first surface region and the second surface region.

[0023] This may avoid (horizontal) gaps between the discrete first and discrete second portions. For example, if the horizontal distance between the first optical axis and the second optical axis is equal to the combined width of the first surface region and the second surface region, a (horizontal) overlap between the discrete first and discrete second portions may be avoided. Accordingly, there may be no space (between the portions) that is not covered by the device, yet the width of the portions may be adjusted to increase the device's sensitivity.

[0024] The width of the first surface region may be equal to the width of the second surface region.

[0025] Accordingly, the segmentation may be mirror-symmetric. For example, the segmentation may be mirror-symmetric to a horizontal axis.

[0026] The vertical gap between the first surface region and the second surface region may be substantially zero or a multiple of a vertical distance between the first optical axis and the second optical axis.

[0027] The vertical distance between the first optical axis and the second optical axis may be equal to or larger than a combined height of the first surface region and the second surface region.

[0028] This may avoid (vertical) gaps between the discrete first and discrete second portions. For example, if the vertical distance between the first optical axis and the second optical axis is equal to the combined height of the first surface region and the second surface region, a (vertical) overlap between the discrete first and discrete second portions may be avoided. Accordingly, there may be no space (between the portions) that is not covered by the device, yet the height of the portions may be adjusted to increase the device's sensitivity.

[0029] The height of the first surface region may be equal to the height of the second surface region.

[0030] Accordingly, the segmentation may be mirror-symmetric. For example, the segmentation may be mirror-symmetric to a vertical axis.

[0031] The device may further comprise a first output for outputting the first signal and a second output for outputting the second signal, or circuitry which is configured to combine the first signal and the second signal, wherein said circuitry may be further configured to ensure that an effect of a decreasing first intensity on the combined signal is not compensated by an effect of an increasing second intensity on the combined signal, and vice versa.

[0032] Hence, the device may be able to detect a body that moves (e.g., in the horizontal direction) out of a discrete first portion and (directly) into an (abutting) discrete second portion, or out of a discrete second portion and (directly) into an (abutting) discrete first portion.

[0033] The device may further comprise a third sensor which is configured to output a third signal that is indicative of a third intensity of electromagnetic radiation incident upon a third surface region of said third sensor, wherein the arrangement images discrete third portions of the area onto the third surface region, thereby segmenting the area along a second curve, the first curve and the second curve being lines that are perpendicular to each other, into a sequence comprising alternating, overlapping or non-overlapping, discrete third and discrete fourth portions.

[0034] The second curve may lie in a second plane, wherein the second plane is preferably not aligned coplanar/parallel to the first plane. For example, the second curve may be a vertical curve.

[0035] Thus, a two-dimensional segmentation may be achieved.

[0036] For example, a horizontal segmentation and a vertical segmentation of the area may be combined, thereby facilitating the accurate detection of both, horizontal and vertical movements of a body within said area.

[0037] The arrangement may image the discrete fourth portions of the area onto a fourth surface region of a fourth sensor of the device, which may be configured to output a fourth signal that is indicative of a fourth intensity of electromagnetic radiation incident upon the fourth surface region. Moreover, instead of adding a fourth sensor and imaging the discrete fourth portions of the area onto the fourth surface region of the fourth sensor, the discrete fourth portions may also be imaged onto the first surface region of the first sensor.

[0038] The arrangement may comprise a third lens and a fourth lens, wherein a first fraction of the electromagnetic radiation incident upon the third surface region of the third sensor travels through the third lens, and a second fraction of the electromagnetic radiation incident upon the third surface region of the third sensor travels through the fourth lens. Moreover, a third fraction of the electromagnetic radiation incident upon the third surface region of the third sensor may travel through the first lens. For example, in a scenario where only two lenses (or two lens rows) are vertically stacked, the electromagnetic radiation incident upon the third surface region of the third sensor may also only travel through the first lens and the third lens.

[0039] If, however, the arrangement images the discrete fourth portions of the area onto the fourth surface region of the fourth sensor, a first fraction of the electromagnetic radiation incident upon the fourth surface region of the fourth sensor may travel through the third lens, and a second fraction of the electromagnetic radiation incident upon the fourth surface region of the fourth sensor may travel through the fourth lens. Again, if, for instance, only two lenses (or two lens rows) are vertically stacked, the electromagnetic radiation incident upon the third surface region of the third sensor may also only travel through the first lens and the third lens.

[0040] Yet, if the arrangement images the discrete fourth portions of the area onto the first surface region of the first sensor, a third fraction of the electromagnetic radiation incident upon the first surface region of the first sensor may travel through the third lens, and a fourth fraction of the electromagnetic radiation incident upon the first surface region of the first sensor may travel through the fourth lens. Again, if only two lenses (or two lens rows) are vertically stacked, the electromagnetic radiation incident upon the first surface region of the first sensor may only travel through the first lens and the third lens.

[0041] A third optical axis of the third lens and a fourth optical axis of the fourth lens may be parallel.

[0042] A vertical distance between the third optical axis and the fourth optical axis may be equal to or larger than two times a height of the third surface region.

[0043] The method comprises segmenting the area into a pattern of non-overlapping portions, imaging a first subset of said portions onto a first surface region of a first sensor which is configured to output a first signal and imaging a second subset of said portions onto a second surface region of a second sensor which is configured to output a second signal, wherein any pair of portions of said first subset is separated by at least one portion of said second subset and, wherein any pair of portions of said second subset is separated by at least one portion of said first subset, and evaluating the first signal and evaluating the second signal, or evaluating a combined signal, wherein the combined signal is indicative of each change of either the first signal, the second signal, or both.

[0044] In this regard, the formulation that "the combined signal is indicative of each change of either the first signal, the second signal, or both", as used throughout the description and the claims, particularly refers to signals which are not formed by merely summing the first signal and the second signal. For example, the first sensor and the second sensor may not be merely serially (or anti-serially) connected.

[0045] The portions of the first and second subsets may be imaged onto the first surface region and the second surface region, respectively, by a first lens and a second lens, wherein a first optical axis of the first lens and a second optical axis of the second lens are parallel, and the first surface region and the second surface region are coplanar.

[0046] There may be a horizontal gap and/or a vertical gap between the first surface region and the second surface region.

[0047] The horizontal gap between the first surface region and the second surface region may be substantially zero or a multiple of a horizontal distance between the first optical axis and the second optical axis, and said horizontal distance between the first optical axis and the second optical axis may be equal to or larger than a combined width of the first surface region and the second surface region.

[0048] The vertical gap between the first surface region and the second surface region may be substantially zero or a multiple of a vertical distance between the first optical axis and the second optical axis, and said vertical distance between the first optical axis and the second optical axis may be equal to or larger than a combined height of the first surface region and the second surface region.

[0049] Notably, all other features of the device may also be features of the method and vice versa.

BRIEF DESCRIPTION OF THE DRAWINGS



[0050] The foregoing aspects and many of the attendant advantages will become more readily appreciated as the same become better understood by reference to the following description of embodiments, when taken in conjunction with the accompanying drawings, wherein like reference numerals refer to like parts throughout the various views, unless otherwise specified.

Fig. 1a schematically illustrates a top view and Fig. 1b schematically illustrates a side view of a device for detecting a movement of a body in an area which is to be monitored by the device.

Fig. 2a, Fig. 2b and Fig. 2c schematically illustrate the sensor signals caused by a body moving through the area.

Fig. 3a and Fig. 3b schematically illustrate a possible modification of the device which results in overlapping first and second portions.

Fig. 4 schematically illustrates an option for the horizontal placement of the sensors in the device.

Fig. 5 schematically illustrates another option for the horizontal placement of the sensors in the device.

Fig. 6 schematically illustrates a possible modification of the size of the sensors shown in Fig. 4.

Fig. 7 schematically illustrates a possible modification of the size of the sensors shown in Fig. 5.

Fig. 8a and Fig. 8b schematically illustrate another possible modification of the device which results in a vertical segmentation of the area.

Fig. 8c and Fig. 8d schematically illustrate further sensor signals caused by a body moving through the area.

Fig. 9 schematically illustrates a modification where two sensors are vertically stacked.

Fig. 10 schematically illustrates another option for the vertical placement of the sensors shown in Fig. 9.

Fig. 11 schematically illustrates a modification where three lenses are vertically stacked.

Fig. 12 schematically illustrates another option for the vertical placement of the sensors shown in Fig. 11.

Fig. 13 shows a flow chart of the method.



[0051] Notably, the drawings are not drawn to scale and unless otherwise indicated, they are merely intended to conceptually illustrate the structures and procedures described herein.

DESCRIPTION OF EMBODIMENTS



[0052] Fig. 1a schematically illustrates a top view and Fig. 1b schematically illustrates a side view of device 10 which is configured to detect a movement of body 12 in area 14. Area 14 is segmented along a horizontal curve H into a sequence of alternating, non-overlapping discrete first portions 16 and discrete second portions 18 by arrangement 20 which images discrete first portions 16 onto first surface region 22a (of first pyroelectric sensor 22) and discrete second portions 18 onto second surface region 24a (of second pyroelectric sensor 24). Hence, discrete first portions 16 may be monitored by first pyroelectric sensor 22 and discrete second portions 18 may be monitored by second pyroelectric sensor 24. As stated above, first pyroelectric sensor 22 and second pyroelectric sensor 24 may be replaced by thermopile sensors or bolometer sensors.

[0053] As indicated by the arrow in Fig. 1a and the circles in Fig. 2a, body 12 may move through area 14 from one side to the other. If the surface temperature of body 12 differs from the surface temperature of the structures which it occludes from view, signal I1 (which depends on the intensity of electromagnetic radiation incident upon first surface region 22a) may undergo a change whenever body 12 enters into one of discrete first portions 16 and signal I2 (which depends on the intensity of electromagnetic radiation incident upon second surface region 24a) may undergo a change whenever body 12 enters into one of discrete second portions 18, as schematically indicated in Fig. 2b and Fig. 2c. Likewise, signal I1 may undergo a change whenever body 12 leaves one of discrete first portions 16 and signal I2 may undergo a change whenever body 12 leaves one of discrete second portions 18, as schematically indicated in Fig. 2b and Fig. 2c.

[0054] To detect the movement of body 12, signal I1 output by first sensor 22 and signal I2 output by second sensor 24 may be analyzed individually. For example, if either I1 or 12 undergoes a change (within a specific time) that is larger than a threshold, device 10 may output a signal indicating that body 12 (is present in area 14 and/or) moves. This signal may be used to activate another device (not shown). Furthermore, signal I1 output by first sensor 22 and signal 12 output by second sensor 24 may be combined (by circuitry in device 10). The combined signal may ensure that a change having an impact on signal I1 output by first sensor 22, that could otherwise be observed in the combined signal, is not (easily) compensated by a change having an impact on signal 12 output by second sensor 24. For example, a combined signal may be formed by summing ±I12 and ±I22. Signal I1 and I2 may also be output separately by device 10 and may be processed by an apparatus connected to device 10 (e.g., by electric wires).

[0055] As schematically illustrated in Fig. 3a and Fig. 3b, area 14 may also be segmented along H into a sequence comprising alternating overlapping discrete first and second portions 16 and 18. For example, a width of first surface region 22a and a width of second surface region 24a may be increased to provide for a buffer to prevent gaps between neighboring ones of discrete first and second portions 16 and 18. Alternatively, a width of first surface region 22a and a width of second surface region 24a may be decreased to provide for gaps between neighboring ones of discrete first and second portions 16 and 18 (not shown).

[0056] Fig. 4 shows a top view of arrangement 20 which comprises a plurality of lenses 26 (of substantially the same size and shape). Lenses 26 are positioned along a horizontal (straight) line and have equidistantly spaced parallel optical axes A1, A2, and A3. First surface region 22a and second surface region 24a are coplanar and have a combined width which equals a distance D of the optical axes of neighboring ones of lenses 26. Alternatively, or in addition, arrangement 20 may comprise a plurality of mirrors (not shown). Moreover, as illustrated in Fig. 5, there may be a (horizontal) gap/distance between first sensor 22 and second sensor 24 which may be a multiple of D (i.e., the width of the gap may be n·D with n being a non-negative integer).

[0057] Although Fig. 5 and Fig. 6 show first and second surface regions 22a and 24a having a same width (i.e., a width of 0.5·D each), a width of first surface region 22a and a width of second surface region 24a may differ, as illustrated in Fig. 6 and Fig. 7. This may cause discrete first portions 16 to shrink in width and discrete second portions 18 to gain in width.

[0058] As illustrated in Fig. 8a and Fig. 8b, area 14 may also be segmented (by arrangement 20) along a vertical curve V into a sequence of alternating non-overlapping discrete third and fourth portions 28 and 30. Discrete third portions 28 may be imaged onto third surface region 32a of third sensor 32 which may be configured to output signal I3 that is indicative of an intensity of electromagnetic radiation incident upon third surface region 32a. Discrete fourth portions 30 may be imaged onto first surface region 22a of first sensor 22 (or onto a surface region of an additional sensor, e.g., a fourth sensor). As indicated by the circles in Fig. 8b, body 12 may move through area 14 from one side to the other.

[0059] If the surface temperature of body 12 differs from the surface temperature of the structures which it occludes from view, signal I3 (which depends on the intensity of electromagnetic radiation incident upon third surface region 32a) may undergo a change whenever body 12 enters into one of discrete third portions 28 and signal I1 (which depends on the intensity of electromagnetic radiation incident upon first surface region 22a) may undergo a change whenever body 12 enters into one of discrete fourth portions 30, as schematically indicated in Fig. 8c and Fig. 8d. Likewise, signal I3 may undergo a change whenever body 12 leaves one of discrete third portions 28 and signal I1 may undergo a change whenever body 12 leaves one of discrete fourth portions 30, as schematically indicated in Fig. 8c and Fig. 8d.

[0060] As illustrated in Fig. 9 which depicts a side view of arrangement 20, arrangement 20 may comprise two lenses 26 which are equidistantly spaced apart along a vertical (straight) line and have parallel optical axes A1 and A4. Alternatively, or in addition, arrangement 20 may comprise a plurality of mirrors (not shown). Moreover, first surface region 22a and third surface region 32a may be coplanar and have a combined height which equals a distance D of the optical axes A1 and A4. Furthermore, as illustrated in Fig. 10, there may be a (vertical) gap between first sensor 22 and third sensor 32 which may be a multiple of D (i.e., the height of the gap may be n·D with n being a non-negative integer).

[0061] Although Fig. 9 and Fig. 10 show first and third surface regions 22a and 32a having a same height (i.e., a height of 0.5·D each), a height of first surface region 22a and a height of third surface region 32a may differ, which may cause discrete fourth portions 30 to shrink in height and discrete third portions 28 to gain in height and vice versa. Moreover, area 14 may also be segmented along V into a sequence comprising alternating overlapping discrete third and fourth portions 28 and 30 (not shown). For example, a height of first surface region 22a and a height of third surface region 32a may be increased to provide for a buffer to prevent gaps between neighboring ones of discrete third and fourth portions 28 and 30. Alternatively, a height of first surface region 22a and a height of third surface region 32a may be decreased to provide for gaps between neighboring ones of discrete third and fourth portions 28 and 30.

[0062] As illustrated in Fig. 11 and Fig. 12 which depict side views of arrangement 20, one or more further lenses may be added to arrangement 20. As a result, arrangement 20 comprises three lenses 26 which are equidistantly spaced apart along a vertical (straight) line and have parallel optical axes A1, A4, and A5. As an additional lens 26 has been added, additional discrete third and fourth portions 28 and 30 emerge. Furthermore, the electromagnetic radiation incident upon first surface region 22a and third surface region 32a travels through the vertically stacked lenses 26.

[0063] Fig. 13 shows a flow chart of the method. The method starts at step 34 with segmenting area 14 into a pattern (horizontal, vertical, diagonal, or a combination thereof) of non-overlapping portions 16, 18, 28, and 30. At step 36, the method is continued with imaging a first subset of portions 16, 18, 28, and 30 onto one of surface regions 22a, 24a, and 32a, and imaging a second subset of portions 16, 18, 28, and 30 onto another one of surface regions 22a, 24a, and 32a. When body 12 moves in area 14, the imaging causes different sensors to detect body 12, depending on which portions 16, 18, 28, and 30 body 12 enters or leaves. The method may thus be continued at step 38 with evaluating signal I1 and evaluating signal 12, or at step 40 with evaluating a combined signal, wherein the combined signal may be indicative of each change of either the first signal Ii, the second signal I2, or both.

REFERENCE SIGNS LIST



[0064] 
10
device
12
body
14
area
16
portion
18
portion
20
arrangement
22
sensor
22a
surface region
24
sensor
24a
surface region
26
lens
28
portion
30
portion
32
sensor
32a
surface region
34
step
36
step
38
step
40
step
A1
optical axis
A2
optical axis
A3
optical axis
A4
optical axis
A5
optical axis
D
gap
H
horizontal curve
I1
signal
I2
signal
I3
signal
V
vertical curve



Claims

1. A device (10) for detecting a movement of a body (12) in an area (14) which is to be monitored by the device (10), the device (10) comprising:

a first sensor (22) which is configured to output a first signal (Ii) that is indicative of a first intensity of electromagnetic radiation incident upon a first surface region (22a) of said first sensor (22);

a second sensor (24) which is configured to output a second signal (I2) that is indicative of a second intensity of electromagnetic radiation incident upon a second surface region (24a) of said second sensor (24); and

an arrangement (20) which images discrete first portions (16) of the area (14) onto the first surface region (22a) and discrete second portions (18) of the area (14) onto the second surface region (24a), thereby segmenting the area (14) along a first curve (H, V) traversing the area (14) from a first side of the area (14) to a second side of the area (14), into a sequence comprising alternating, overlapping or non-overlapping, discrete first and discrete second portions (16, 18).


 
2. The device (10) of claim 1, wherein the first sensor (22) comprises a first pyroelectric element and the second sensor (24) comprises a second pyroelectric element, the first surface region (22a) being a surface region of the first pyroelectric element and the second surface region (24a) being a surface region of the second pyroelectric element.
 
3. The device (10) of claim 2, wherein the arrangement (20) comprises a plurality of optical elements.
 
4. The device (10) of claim 3, wherein the arrangement (20) comprises a first lens (26) and a second lens (26) and wherein:

a first fraction of the electromagnetic radiation incident upon the first surface region (22a) of the first sensor (22) travels through the first lens (26) and a second fraction of the electromagnetic radiation incident upon the first surface region (22a) of the first sensor (22) travels through the second lens (26); and

a first fraction of the electromagnetic radiation incident upon the second surface region (24a) of the second sensor (24) travels through the first lens (26) and a second fraction of the electromagnetic radiation incident upon the second surface region (24a) of the second sensor (24) travels through the second lens (26).


 
5. The device (10) of claim 4, wherein a first optical axis (A1) of the first lens (26) and a second optical axis (A2) of the second lens (26) are parallel and/or the first surface region (22a) and the second surface region (24a) are coplanar.
 
6. The device (10) of claim 5, wherein a horizontal gap between the first surface region (22a) and the second surface region (24a) is substantially zero or a multiple of a horizontal distance (D) between the first optical axis (A1) and the second optical axis (A2).
 
7. The device (10) of claim 6, wherein the horizontal distance (D) between the first optical axis (A1) and the second optical axis (A2) is equal to or larger than a combined width of the first surface region (22a) and the second surface region (24a).
 
8. The device (10) of claim 7, wherein the width of the first surface region (22a) is equal to the width of the second surface region (24a).
 
9. The device of any one of claims 1 to 8, further comprising:

a first output for outputting the first signal (Ii) and a second output for outputting the second signal (I2); or

circuitry which is configured to combine the first signal (Ii) and the second signal (I2), wherein, preferably, said circuitry is further configured to ensure that an effect of a decreasing first intensity on the combined signal is not compensated by an effect of an increasing second intensity on the combined signal, and vice versa.


 
10. The device (10) of any one of claims 1 to 9, further comprising:

a third sensor (32) which is configured to output a third signal (I3) that is indicative of a third intensity of electromagnetic radiation incident upon a third surface region (32a) of said third sensor (32);

wherein the arrangement (20) images discrete third portions (28) of the area (14) onto the third surface region (32a), thereby segmenting the area (14) along a second curve (H, V), the first curve (H, V) and the second curve (H, V) being lines that are perpendicular to each other, into a sequence comprising alternating, overlapping or non-overlapping, discrete third and discrete fourth portions (28, 30).


 
11. The device (10) of claim 10, wherein the arrangement (20) comprises a third lens (26) and a fourth lens (26) and wherein:
a first fraction of the electromagnetic radiation incident upon the third surface region (32a) of the third sensor (32) travels through the third lens (26) and a second fraction of the electromagnetic radiation incident upon the third surface region (32a) of the third sensor (32) travels through the fourth lens (26).
 
12. The device (10) of claim 11, wherein a third optical axis (A4) of the third lens (26) and a fourth optical axis (A5) of the fourth lens (26) are parallel.
 
13. The device (10) of claim 12, wherein the vertical distance between the third optical axis (A4) and the fourth optical axis (A5) is equal to or larger than two times a height of the third surface region.
 
14. A method of detecting a movement of a body (12) in an area (14), comprising:

segmenting the area (14) into a pattern of non-overlapping portions (16, 18, 28, 30);

imaging a first subset of said portions (16, 18, 28, 30) onto a first surface region (22a) of a first sensor (22) which is configured to output a first signal (Ii) and imaging a second subset of said discrete portions (16, 18, 28, 30) onto a second surface region (24a) of a second sensor (24) which is configured to output a second signal (I2), wherein any pair of portions (16, 18, 28, 30) of said first subset is separated by at least one portion (16, 18, 28, 30) of said second subset and any pair of portions (16, 18, 28, 30) of said second subset is separated by at least one portion (16, 18, 28, 30) of said first subset; and

evaluating the first signal (Ii) and evaluating the second signal (I2); or

evaluating a combined signal, wherein the combined signal is indicative of each change of either the first signal (Ii), the second signal (I2), or both.


 
15. The method of claim 14, wherein:

the portions (16, 18, 28, 30) of the first and second subsets are imaged onto the first and second surface region (22a, 24a), respectively, by a first lens (26) and a second lens (26);

a first optical axis (A1) of the first lens (26) and a second optical axis (A2) of the second lens (26) are parallel;

the first surface region (22a) and the second surface region (24a) are coplanar;

a horizontal gap between the first surface region (22a) and the second surface region (24a) is substantially zero or a multiple of a horizontal distance (D) between the first optical axis (A1) and the second optical axis (A2); and

said horizontal distance (D) between the first optical axis (A1) and the second optical axis (A2) is equal to or larger than a combined width of the first surface region (22a) and the second surface region (24a).


 




Drawing































Search report


















Search report




Cited references

REFERENCES CITED IN THE DESCRIPTION



This list of references cited by the applicant is for the reader's convenience only. It does not form part of the European patent document. Even though great care has been taken in compiling the references, errors or omissions cannot be excluded and the EPO disclaims all liability in this regard.

Patent documents cited in the description