BACKGROUND OF THE INVENTION
1. Field of the Invention
[0001] The present invention relates to the system employed and circuitry used with an ensemble
of clocks to obtain an ensemble time. More particularly, the present invention relates
to an improved algorithm defining ensemble time that can be, for example, implemented
with Kalman filters for obtaining an improved estimate of time from an ensemble of
clocks.
2. Description of the Related Art
[0002] For a number of years, groups of precision clocks used in combination have provided
the "time" in situations in which high precision timekeeping is required. For example,
an "official" time for the United States is provided by the atomic time scale at the
National Bureau of Standards, the UTC(NBS), which depends upon an ensemble of continuously
operating cesium clocks. The time interval known as the "second" has been defined
in terms of the cesium atom by the General Conference of Weights and Measures to be
the duration of 9,192,631,770 periods of the radiation corresponding to the transition
between the two hyperfine levels of the ground state of the cesium-133 atom. Other
clocks may be calibrated according to this definition. Thus, while each clock in a
group or ensemble of clocks is typically some type of atomic clock, each clock need
not be a cesium clock.
[0003] Even though one such atomic clock alone is theoretically quite accurate, in many
applications demanding high accuracy it is preferred that an ensemble of atomic clocks
be used to keep time for a number of reasons. Typically, no two identical clocks will
keep precisely the identical time. This is due to a number of factors, including differing
frequencies, noise, frequency aging, etc. Further, such clocks are not 100% reliable;
that is, they are subject to failure. Accordingly, by using an ensemble of clocks
in combination, a more precise estimate of the time can be maintained.
[0004] When an ensemble of clocks is utilized to provide an estimate of time, various techniques
may be employed for processing the signals output by the clocks to obtain the "time".
Typically, interclock time comparisons are made to determine the relative time and
frequency of each clock. The noise spectrum of each clock is represented by a mathematical
model, with noise parameters determined by the behavior of the individual clock. Clock
readings are combined based on these comparisons and models to produce the time scale.
[0005] A problem with known systems for providing an estimate of time or a time scale based
upon an ensemble of clocks is that the individual states of the clocks are not observable
and only clock differences can be measured. As a result, there are an infinite number
of solutions or estimates of time possible. Stated another way, the systems are underdetermined.
For example, in systems that implement a Kalman filter approach to estimating time,
the lack of observability and resultant underdetermined character of the system manifest
themselves in covariance matrix elements that grow on each cycle of the Kalman recursion.
Since the computation is implemented on a computer system with finite accuracy, this
growth eventually causes computational problems. In addition to reducing the consequences
of lack of observabililty, it is also desirable to obtain a substantially unbiased
estimate of clock performance that can be used to achieve a ensemble time system that
is adaptive. Moreover, there is a need to achieve this adaptive quality while also
maintaining or improving the robustness of the resulting ensemble time.
[0006] There is yet a further need for an ensemble time system that permits estimates of
all of the spectral densities of the clocks in the system, detects steps in state
estimates, and provides improved performance over a wider range of averaging times.
[0007] Based on the foregoing, there is a need for a system for estimating time based on
a clock ensemble that addresses the aforementioned needs.
SUMMARY OF THE INVENTION
[0008] Accordingly, one object of the present invention is to address the consequences of
the lack of observability of individual clock states in an ensemble clock system and,
in particular, the plurality of solutions to the ensemble time definition, or, stated
differently, the underdetermined character of such systems.
[0009] To achieve this objective, an ensemble time system is provided that includes a plurality
or ensemble of clocks, a mechanism for measuring differences in time related parameters
between pairs of clocks, and a device for providing an ensemble time that uses the
differences and an ensemble time definition with a constrained number of solutions
for the ensemble time. By providing a mechanism to constrain the number of solutions,
an improved ensemble time is realized. Further, in systems that implement the ensemble
time definition with Kalman filters, the constraining mechanism reduces computational
problems associated with the covariance matrix.
[0010] An ensemble time system is also provided that provides a substantially unbiased estimate
of the state of each clock in the system thereby permiting the system to be adaptive.
In one embodiment, the estimates of the state of each clock is weighted in a fashion
that is proportional to the rms noise of the state of the clock. In another embodiment,
the ensemble time (in contrast to the estimates of the states) is also weighted but
with a different set of weights than the state estimates, to achieve robustness and
other desirable characteristics in the system.
[0011] In yet a further embodiment, a first ensemble time with a first sampling interval
that provides substantially optimal performance is combined with a second ensemble
time with a second sampling interval that has substantially optimal performance over
a second range to provide a third ensemble time that has optimal performance over
a greater range than either the first or second ensemble time. In this way, estimates
of the spectral densities of all the clocks can be obtained, steps in the states being
estimated can be detected, and improved performance over a wider range of averaging
times obtained. In one embodiment, a phase-locked loop is used to combine the first
and second ensemble times to obtain the third ensemble time.
[0012] Other objects and advantages of the present invention will be set forth in part in
the description and drawing figure which follow, and will further be apparent to those
skilled in the art.
BRIEF DESCRIPTION OF THE DRAWING
[0013] Figure 1 is a circuit diagram of an implementation of the present invention.
DETAILED DESCRIPTION OF THE INVENTION
[0014] As discussed previously, one method for processing the output of a plurality of clocks
(i.e., oscillators) included in an ensemble is referred to as the Kalman approach.
In this Kalman approach, one of the clocks in the ensemble is temporarily designated
as the reference clock, with the remaining clocks "aiding" the time provided by the
reference clock. Kalman filters provide state estimation and forecasting functions.
Generally, Kalman filters are used to model the performance of quartz oscillators
and atomic clocks. Kalman filters act as minimum square error state estimators and
are applicable to dynamic systems, that is, systems whose state evolves in time. Kalman
filters are recursive and therefore have modest data storage requirements. When employed
to provide time from an ensemble of clocks, Kalman filters can, of course, only provide
estimates that reflect the algorithms which they embody.
[0015] The novel clock model utilized in the present invention takes into account the time,
the frequency, and the frequency aging. The general form of the clock model consists
of a series of integrations. The frequency aging is the integral of white noise, and
therefore exhibits a random walk. The frequency is the integral of the frequency aging
and an added white noise term, allowing for the existence of random walk frequency
noise. The time is the integral of the frequency and an added white noise term which
produces random walk phase noise, usually called white frequency noise. An unintegrated
additive white noise on the phase state produces additive white phase noise.
[0016] When two clocks are compared, the relative states are the differences between the
state vectors of the individual clocks. Hereinbelow, the state vector of a clock i
will be referred to as

. Only the differences between clocks can be measured. In terms of the state vectors,
the differences between a clock j and a clock k at time t is denoted by

The same approach will be used below to denote the time of a clock with respect to
an ensemble. The ensemble is designated by the subscript e. Since ensemble time is
a computed quantity, the ensemble is only realizable in terms of its difference from
a physical clock.
[0017] In the present invention, the individual clock state vector is four-dimensional.
In prior approaches, the comparable state vector has more typically been a two-dimensional
state vector, taking into account only a phase component and a frequency component.
In contrast, the present invention utilizes a system model which incorporates the
time, the time without white phase noise, the frequency, and the frequency aging into
a four-dimensional state vector, such that a four-dimensional state vector

(t) is as follows:

where u(t) is the time of the system (in this case a clock pair) at sample (t), x(t)
is the time of the system without white phase noise at sample (t), y(t) is the frequency
of the system at sample (t), and w(t) is the frequency aging of the system at sample
(t). The state vector evolves from time t to time t+δ according to

where Φ(δ) is a 4 x 4 dimensional state transition matrix, Γ

is the plant noise and Γ

(t+δ|t) is a four-dimensional vector containing the noise inputs to the system during
the time interval from t to t+δ, and

(t) is a four-dimensional vector containing the control inputs made at time t.
[0018] The 4 x 4 dimensional state transition matrix Φ(δ) embodies the system model described
above. The state transition matrix is assumed to depend on the length of the interval,
but not on the origin, such that

[0019] The four-dimensional vector Γ(δ)

(t+δ|t) contains the noise input to the system during the interval from t to t+δ,
where


and where β'
jk(t+δ) is the white time noise input between clocks j and k at time (t+δ), ε'
jk(t+δ|t) is the white frequency noise input at time t+δ, η'
jk(t+δ|t) is the random walk frequency noise input at time t+δ, and α'
jk(t+δ|t) is the random walk frequency aging noise input at time t+δ. Each element of

(t+δ|t) is normally distributed with zero mean and is uncorrelated in time. The four-dimensional
vector

(t) contains the control input made at time t.
[0020] Equation 2 generates a random walk in the elements of the state vector.
[0021] A single observation z(t) can be described by a measurement equation. Such an equation
relative to clocks j and k can take the following form:

where H(t) is a 1 x 4 dimensional measurement matrix and v(t) is the scalar white
noise. An observation made at time t is linear-related to the four elements of the
state vector (Equation 1) by the 1 x 4 dimensional measurement matrix H(t) and the
scalar white noise v(t).
[0022] The noise covariance matrix of the measurement noise, R(t), is defined as follows:

where E[] is an expectation operator and

(t)
T is the transpose of the noise vector.
[0023] Phase measurements of the clock relative to the reference are described by

where

is the variance of the phase measurement process.
[0024] Similarly, the frequency measurements are described by

[0025] Q
jk(t+δ|t) is the covariance matrix of the system (or plant) noise generated during an
interval from t to t+δ, and is defined by

[0026] The system covariance matrix can be expressed in terms of the spectral densities
of the noises such that

where f
h is an infinitely sharp high-frequency cutoff. Without this bandwidth limitation,
the variance of the white phase additive noise would be infinite. The clock pair spectral
densities are the sum of the individual contributions from each of the clocks,

where S
j and S
k are the spectral densities of clocks j and k, respectively.
[0028] It is this form of the plant covariance (i.e., Equations 13 - 22) which will be used
to calculate the plant covariance of the reference clock versus the ensemble.
[0029] As discussed briefly above, one of the clocks in the ensemble is used as a reference
and is designated as clock j. The choice of clock j as the reference clock is arbitrary
and may be changed computationally. The role of the reference clock j is to provide
initial estimates and to be the physical clock whose differences from the ensemble
are calculated. Given that the ensemble consists of N clocks, each of the other N-1
clocks is used as an aiding source. That is, each of the remaining clocks provides
an independent representation of the states of clock j with respect to the ensemble.
As indicated, these states are time, frequency, and frequency aging. The present invention
defines the states of each clock with respect to the ensemble to be the weighted average
of these representations, and the present invention provides a user with full control
over the weighting scheme. Given

(t₂|t₁) denotes a forecast of

at time t₂ based on the true state through time t₁, time, frequency, and frequency
aging of a multiple weight ensemble can be defined as follows:



Each new time of a clock j with respect to the ensemble depends only on the prior
states of all the clocks with respect to the ensemble and the current clock difference
states. The ensemble definition uses the forecasts of the true states from time t
to t+δ, that is,

where

(t+δ|t) is the forecasted state vector at time (t+δ) based on the true state through
time t. No unsupported estimated quantities are involved in the definition.
[0030] Prior approaches have frequently used relations superficially similar to that found
in Equation 23 to define ensemble time. However, as the present inventor has found,
Equation 23 alone does not provide a complete definition of the ensemble time. Since
the prior art does not provide a complete definition of the ensemble time, the filters
employed in the prior art do not yield the best estimate of ensemble time. The present
invention provides a more complete definition of ensemble time based not only on the
time equation (Equation 23), but also on the frequency and frequency aging relations
(Equations 24 and 25).
[0031] The time-scale is used to correct the time of its individual members. Thus the time
of the time-scale is obtained from the relation

Since we can only estimate the time of a clock with respect to the time-scale, the
computed correction is

Equation 26-2 means that we use the correction û
je (t|t) to the time of clock j in order to perform an action according to the time-scale.
Thus if clock j is estimated to be 2 seconds fast compared to the time-scale, all
actions that are synchronized to the time-scale are performed when clock j reads 2
seconds after the designated time. The error in the correction u
ecorr(t) relative to the true time-scale, u
e(t), has two components - the error in the estimate of the ensemble and the error
in the estimate of the clock j.
[0032] Equations 23, 24, and 25 are still not a unique definition because the individual
clock states are not observable and only clock differences can be measured. Thus,
an infinite number of solutions satisfy Equations 23, 24, and 25. For example, if
u₁, u₂,..., u
n, u
e is a solution, then u₁+p, u₂+p,..., u
n+p, u
e+p is also a solution. What the clocks actually did is unknown and difference measurements
cannot detect what truly happened. The correction term discussed above is not affected
by the ambiguity of the individual clock and time-scale estimates since it involves
their differences. However, the ambiguity produces practical problems. For example,
when a Kalman filter is used to estimate the individual states, the covariance matrix
elements grow on each cycle of the Kalman recursion since these states are not observable.
The increasing size of the covariance matrix elements will eventually cause computational
problems due to the finite accuracy of double precision math in a computer. The mixture
of the two components of the covariance matrix elements may also cause problems with
algorithms that use the covariance matrix elements to estimate the uncertainty of
the state estimates. For example, the covariance matrix elements are used in parameter
estimation and also to detect outlier observations before they are used in the state
estimation process.
[0033] One way to eliminate the ambiguity that arises from the impossibility of making absolute
time measurements is to introduce an independent relationship among the individual
clock states that constrains the solution. To see how this might be done, we rewrite
the ensemble definition in a different form:



We interpret these equations as follows. The difference between the true time-scale
state at time t+δ and the forecast based on the true time-scale state at time t is
the weighted mean of the differences between the true clock states at time t+δ and
their forecasts based on the true clock states at time t. That is, the time-scale
is the centroid of the individual clocks.
[0034] When the ensemble is composed of high quality atomic clocks, we believe
a priori that the deviations of the clock times from a uniform time scale are random and uncorrelated.
It is possible to force the estimates of the individual clock states to reflect this
assumption. This is accomplished by introducing three constraint equations that select
the solution with the desired properties:



The parameters a'
i, b'
i and c'
i may be chosen arbitrarily. However, the constraints have the undesirable effect of
introducing correlation between the clocks. For example, in the case of two clocks,
if one has a positive deviation the other must have a negative deviation. The resulting
variances of the noise inputs are not equal to the true variances for arbitrary choices
of the coefficients. To resolve this limitation, the coefficients a'
i, b'
i, and c'
i are selected so that the noises on the estimated states of each clock are nearly
equal to the true noises of that clock for all possible combinations of clocks with
different levels of noise. We find that the coefficients are proportional to the
rms noise on the clock states:



When these coefficients are used to scale the noise inputs to the clocks, the scaled
noises all have unit variance. The constraint equations say that the sum of the unit
variance noise inputs from all the ensemble members is zero. Heuristically, this seems
very reasonable.
[0035] When the constraint equations are used one obtains estimates for the true states
of each clock and an independent estimate of the time scale. The former are important
in obtaining unbiased estimates of the spectral densities of the clocks. The latter
permits the generation of system time using arbitrarily chosen weights. Separating
the estimation of the truth from generation of the time scale increases flexibility
and reduces biases.
[0036] The parameters
ai,
bi, and
ci of equations 26-3 through 26-5 may be chosen arbitrarily. However, we note that when
a'
i = a, b'
i = b, and c'
i = c, then the ensemble states are identically zero and the correction term for realizing
the time-scale via a clock reduces to the time state of that clock. The constraining
equations result in solutions for the clock states that are centered about zero and
is a particular case of the time-scale definition. Using the constraint equations
has a beneficial effect on the Kalman filter solution to the estimation problem. They
relieve the nonobservability of the problem, and the covariance matrix approaches
a finite asymptotic value. Such constraining equations are applicable to ensembles
of other types of clocks than high quality atomic clocks. Further, comparable constraining
equations can be applied to the ensemble time definition taught in the article by
Richard H. Jones and Peter V. Tryon entitled "Estimating Time From Atomic Clocks",
Journal of Research of the National Bureau of Standards, Vol. 88, No. 1, pp.17-24,
January-February 1983, incorporated herein by reference.
[0037] As noted above, a
i(t), b
i(t), and c
i(t) represent weights to be chosen for each of the three relations described in Equations
23 through 25 for each of the N clocks in the ensemble. The weights may be chosen
in any way subject to the restrictions that all of the weights are positive or 0 and
the sum of the weights is 1. That is,

The weights may be chosen to optimize the performance (e.g., by heavily weighting
a higher quality clock relative to the others) and/or to minimize the risk of disturbance
due to any single clock failure.
[0038] In contrast to the known prior approaches, the present invention provides a time
scale algorithm that utilizes more than one weighting factor for each clock. Accordingly,
the present invention is actually able to enhance performance at both short and long
times even when the ensemble members have wildly different characteristics, such as
cesium standards, active hydrogen masers and mercury ion frequency standards.
[0039] The use of multiple weights per clock improves the performance of the time scale
compared to the use of a single weight per clock. However, it is possible to achieve
additional performance improvements by combining independent time scales that are
computed using different sampling rates. Let us consider heuristically why this would
be so.
[0040] Suppose time scale A has a sampling interval of δ that is in the region where all
the clocks are dominated by short-term noise such as white frequency noise. Further
suppose that the clocks have widely varying levels of long-term noise such as random
walk frequency. The apportionment of noise and the estimation of frequencies for time
scale A is done based on time predictions over the short interval δ where the effects
of the long-term noises are visible with a signal-to-noise ratio that is significantly
less than 1.
[0041] Now suppose that time scale B has a sampling interval of
kδ, where k is large compared to 1, that is in the region where all of the clocks are
dominated by the long-term random walk frequency noise. The apportionment of noise
and the estimation of frequencies for time scale B is done based on the time predictions
over the long interval
kδ where the effects of the long-term noises are visible with a signal-to-noise ratio
that is significantly greater than 1. This time scale will have more accurate apportionment
of the long-term noises between the clocks because there will be no errors made in
the apportionment of the short-term noise (an incorrect fraction of 0 is still 0).
[0042] Time scale A may be phase locked to time scale B to produce a single time scale with
more nearly optimal performance over a wider range of sample times than the individual
time scales. This method may be used to combine as many time scales as necessary.
Each time scale is phase locked to the time scale with the next largest sampling interval.
The time scale with the longest sampling interval is allowed to free run. A second
order phase lock loop should be used to combine two time scales. This type of loop
results in zero average closed loop time error between two time scales no matter how
large the open loop frequency differences between the two scales.
[0044] This version of the ensemble definition is in the form required for the application
Kalman filter techniques. As discussed above, the advantage of the Kalman approach
is the inclusion of the system dynamics, which makes it possible to include a high
degree of robustness and automation in the algorithm.
[0045] Next we describe two methods of using Kalman filters to estimate the times of the
clocks with respect to the ensemble. The first method uses N separate Kalman filters
to estimate the states of each of the N clocks.
[0046] Kalman Method 1: In order to apply Kalman filters to the problem of estimating the
states of a clock obeying the state equations provided above, it is necessary to describe
the observations in the form of Equation 6. This is accomplished by a transformation
of coordinates on the raw clock time difference measurements or clock frequency difference
measurements. Since z may denote either a time or a frequency observation, a pseudomeasurement
may be defined such that

This operation translates the actual measurements by a calculable amount that depends
on the past ensemble state estimates and the control inputs.
[0047] An additional requirement for the use of the usual form of Kalman filters is that
the measurement noise, v
je, is uncorrelated with the plant noise, Γ

However, this is not true for the measurement model of Equation 33. Through algebraic
manipulations, it has been found that the noise perturbing the pseudomeasurements
can be characterized as

This pseudonoise depends on the true state at time t₂ and is therefore correlated
with the plant noise which entered into the evolution of the true state from time
t₁ to time t₂. The correlation of these noises is represented by a matrix C defined
by

For the case of a single time measurement,

is a scalar and C is a 4 x 1 matrix where,

For a single frequency measurement,

One method of resolving this difficulty is to extend the Kalman filter equations to
allow correlated measurement and plant noise.
[0048] In this regard, it is possible to have a Kalman recursion with correlated measurement
and plant noise. The error in the estimate of the state vector after the measurement
at time t₁ is

(t₁|t₁)-

(t₁) and the error covariance matrix is defined to be

[0049] The diagonal elements of this n x n matrix are the variances of the estimates of
the components of

(t₁) after the measurement at time t₁. The error covariance matrix just prior to the
measurement at time t₂ is defined as

The error covariance matrix evolves according to the system model, such that

The new estimate of the state vector depends on the previous estimate and the current
measurement,

where the gain matrix, K(t₂), determines how heavily the new measurements are weighted.
The desired or Kalman gain, K
opt, is determined by minimizing the square of the length of the error vector, that is,
the sum of the diagonal elements (i.e., the trace) of the error covariance matrix,
such that

Finally, the updated error covariance matrix is given by

where I is the identity matrix.
[0050] Equations 40 - 43 define the Kalman filter. As defined, the Kalman filter is an optimal
estimator in the minimum squared error sense. Each application of the Kalman recursion
yields an estimate of the state of the system, which is a function of the elapsed
time since the last filter update. Updates may occur at any time. In the absence of
observations, the updates are called forecasts. The interval between updates, δ =
t₂ - t₁, is arbitrary and is specifically not assumed to be constant. It is possible
to process simultaneous measurements either all at once or sequentially. In the present
invention, simultaneous measurements are processed sequentially, since sequential
processing avoids the need for matrix inversions and is compatible with outlier rejection.
[0051] As will be appreciated by those skilled in the art, implementation of the relationships
defined in Equations 40 - 43 as a Kalman filter is a matter of carrying out known
techniques.
[0052] For the estimation of the reference clock versus the ensemble, the first step is
the selection of a reference clock for this purpose. The reference clock referred
to herein is distinguished from a hardware reference clock, which is normally used
as the initial calculation reference. However, this "software" reference clock normally
changes each time the ensemble is calculated for accuracy.
[0053] As discussed above, the ensemble consists of N clocks and therefore N estimates of
the ensemble time exist. Thus, the first estimate of the ensemble time cannot be rejected
and must be robust. To obtain this robust initial estimate, the median of the pseudomeasurements
is computed. The clock which yields the median pseudomeasurement is selected as the
calculation reference clock, and is designated clock r. In this regard

Of the N pseudomeasurements, one pseudomeasurement is a forecast and the remainder
of the pseudomeasurements add new information. New pseudomeasurements must be calculated
if the reference for the calculation has changed. To change reference clocks from
one clock to another, i.e., from clock j to clock r, it is necessary only to form
the difference, such that

This procedure works even if the initial reference clock (clock r) has been corrupted
by some large error.
[0055] The initial state estimate at time t₂ is a forecast via the reference clock r. The
initial covariance matrix is the covariance before measurement. The data from all
the remaining clocks are used to provide N-1 updates. The pseudomeasurements are processed
in order of increasing difference from the current estimate of the time of the reference
clock r with respect to the ensemble. Pseudomeasurement I(k) is the "k"th pseudomeasurement
processed and I(1) is the reference clock forecast. Outliers (i.e., data outside an
anticipated data range) are "de-weighted" when processing pseudomeasurments 2 through
N using the statistic

where the

(t₂) is the innovation or difference between the pseudomeasurment and the forecast,
such that

This equation can be rearranged in the form

After squaring and taking the expectation value, the result is

[0056] To preserve the robustness of the state estimation process, deweighting of the outlier
data is used rather than rejection. This preserves the continuity of the state estimates.
A nonoptimum Kalman gain is calculated from

where

is the Hampel's Ψ function.
[0057] When this calculation is concluded, the estimates of the states of the reference
clock r with respect to the ensemble have been provided. The corresponding estimates
for the remaining clocks are obtained by their values with respect to the reference
clock r. This procedure is used rather than estimating the clock parameters directly
with respect to the ensemble because the innovations of this process are used in parameter
estimation.
[0058] The estimates of the clocks relative to reference clock r are obtained from N-1 independent
Kalman filters of the type described above. The four dimensional state vectors are
for the clock states relative to the reference clock r

Every clock pair has the same state transition matrix and Γ matrix, which are provided
for above in Equations 3 and 5. The system covariance matrices are Q
ir(t+ δ|t). The white phase noise is given by the measurement model

where each measurement is described by the same 4 x 1 row matrix

[0059] The updated difference dates are provided in Equation 41, which is one of the equations
which define the Kalman filter. No attempt is made to independently detect outliers.
Instead, the deweighting factors determined in the reference clock versus ensemble
calculation are applied to the Kalman gains in the clock difference filters. The state
estimates for the clocks with respect to the ensemble are calculated from the previously
estimated states of the reference clock r with respect to the ensemble and the clock
difference states, such that

[0060] This essentially completes the calculation of ensemble time. The remaining task is
to update all of the parameters used in the computation. The parameter estimation
problem is discussed more completely below. Briefly, the parameter estimates are obtained
from prediction errors of all possible clock pairs. Accordingly, rather than computing
Kalman filters for N-1 clock pairs, the calculations are performed for N(N-1)/2 pairs,
ij, for i=1 to N-1 and j=i+1 to N. Certainly, in a large ensemble, this may entail
significant computation. But little information is added by comparison of noisy clocks
with one another. For each noise type, a list of the five clocks having the lowest
noise can be formed. If the index i is restricted to this more limited range, then
only 5N-15 filters are required for each parameter estimated.
[0061] Next we describe an alternate method of estimating states of the clocks by combining
these states into a single state vector and using one Kalman filter.
[0062] Kalman Filter Method 2: The ensemble definition can be implemented using the Kalman
filter method by combining the states of all the clocks into a single state vector
which also explicitly contains the states of the time-scale. The clock model is written
in matrix form as follows:

The state vector evolves from time t to time t+δ according to

where the 4(N + 1) x 4(N + 1) dimensional state transition matrix Φ(δ) embodies the
system model described above. The state transition matrix, Φ, depends on the length
of the interval, but not on the origin. It consists of N + 1 identical 4 + 4 submatrices,
φ, arranged on the diagonal of a 4(N + 1) dimensional matrix that has zeros in all
other positions. φ is state transition matrix for an individual clock (or clock pair)
given in Equation 3.
[0063] The 4(N + 1) dimensional vector

(t + δ|t) contains the noise inputs to the system during the interval from t to t
+ δ. Thus,

Each element of

(t + δ|t) is normally distributed with zero mean and is uncorrelated in time. Finally,
the 4(N + 1) dimensional vector

(t) contains the control inputs made at time t. Equation 64-2 generates a random
walk in the elements of the state vector.
[0064] A set of r observations,

(t) is described by the measurement equation

which means that the observations made at time t are related linearly to the 4(N +
1) elements of the state vector by the r x 4(N + 1) dimensional measurement matrix
H(t) and the r dimensional white noise vector

(t). The noise covariance is defined as:

For example, the phase measurements of clocks 2 and 3 relative to clock 1 in a three
clock ensemble are described by:

If the measurements are made simultaneously by a dual mixer measurement system the
measurement covariance matrix is:

where σ²
vx is the variance of the phase measurement process. Similarly, the frequency measurements
are described by

[0065] The system (or plant) covariance matrix is defined as follows

and can be expressed in terms of the spectral densities of the noises of the individual
clocks using Equations 13 - 22.
[0066] The Kalman recursion described here is the discrete Kalman filter documented in "Applied
Optimal Estimation", A. Gelb, ed., MIT Press, Cambridge, 1974. His derivation assumes
that the perturbing noises are white and are normally distributed with mean zero.
He further assumes that the measurement noise and the process noise are uncorrelated.
[0067] The error in the estimate of the state vector after the measurement at time t₁ is

(t₁|t₁)-

(t₁) and the error covariance matrix is defined to be

The diagonal elements of this n x n matrix are the variances of the error in the estimates
of the components of

(t₁) after the measurement at time t₁. Next, the error covariance matrix just prior
to the measurement at time t₂ is defined as

The error covariance matrix evolves according to the system model.

The new estimate of the state vector depends on the previous estimate and the current
measurement

where the gain matrix,
K(t₂), determines how heavily the new measurements are weighted. The optimum or Kalman
gain,
Kopt, is determined by minimizing the "square of the length of the error vector,",
i.e., the sum of the diagonal elements (the trace) of the error covariance matrix

Finally, the update error covariance matrix is

where
I is the identity matrix. Equations 64-12 through 64-15 define the Kalman filter, and
so defined it as an optimal estimator in the minimum squared-error sense. Each application
of the Kalman recursion yields an estimate of the state of the system, which is a
function of the elapsed time since the last filter update. Updates may occur at any
time. In the absence of observations, the updates are equal to the forecasts. The
interval between updates, δ = t₂ - t₁, is arbitrary and is not assumed to be constant
over time. It is possible to process simultaneous measurements either all at once
or sequentially. Simultaneous processing is used here since it is more compatible
than sequential processing with outlier deweighting.
[0068] The ensemble definition is used to generate those plant covariance matrix elements
that involve ensemble states. It should also be used to connect the individual clock
states to the ensemble states. This may be accomplished in an approximate manner by
implementing three pseudomeasurements. We rearrange the ensemble definition and replace
the forecasts based on the true states by the forecasts based on the prior estimates
to obtain



The measurements on the left side of these equations are described in the
H by adding three additional rows.

The constraint equations may also be implemented using pseudomeasurements. This is
particularly simple when the coefficients of the constraint equations are the weights.
In that case the pseudomeasurements are:



The measurements on the left side of these equations are also described in the
H matrix by adding three more rows.

[0069] Outliers are detected using the method developed by Jones et al., "Estimating Time
From Atomic Clocks", NBS Journal of Research, Vol. 80, pp. 17-24, January-February
1983, and are deweighted following the recommendation of Percival, "Use of Robust
Statistical Techniques in Time Scale Formation", 2nd Symposium on Atomic Time Scale
Algorithms, June, 1982.
[0070] The innovation,

(t₂), is the difference between the observations and the forecasts,

It has a covariance matrix given by:

Outliers are not detected by dividing each element of the innovation vector by the
corresponding element of the covariance matrix since clock differences are being measured
and the elements of the innovation vector are correlated. When there is an error in
the reference clock for the measurements, a constant bias will occur in all of the
measurements. When there is an error in any other clock, the error will occur in only
those measurements involving that clock (usually one). If clock j changes time by
an amount f
j, then the innovation may be written

where the error vector,

, has covariance matrix
C. The column vector,

has all ones when clock j is the reference clock. When clock j is not the reference
clock,

has zeros for measurements that don't involve that clock and minus one for any measurement
that does. Following Jones, the minimum variance least squares estimate of f
j, for a given

is

and has standard error
If the test statistic,


is large in absolute value, then clock j is assumed to have an error.
[0071] Deweighting of erroneous measurements, rather than rejection, of the outlier data
is used to enhance the robustness of the state estimation process. Deweighting preserves
the continuity of the state estimates. On the other hand, rejecting outliers results
in discontinuous state estimates that can destabilize the time-scale computation by
causing the rejection of all members of the time-scale. A non-optimum Kalman gain
is calculated from

where

is Hampel's Ψ function.
[0072] Since the reference clock appears in every measurement, it can't be deweighted. Thus,
the first step in the outlier deweighting process is to find a good reference clock.
Once this is accomplished, all the measurements are processed and deweighting factors
are computed for the remaining clocks as necessary.
[0073] The outlier detection algorithm of the ensemble calculation identifies the measurements
which are unlikely to have originated from one of the processes included in the model.
These measurements are candidate time steps. The immediate response to a detected
outlier in the primary ensemble Kalman filter is to reduce the Kalman gain toward
zero so that the measurement does not unduly influence the state estimates. However,
the occurrence of M₁ successive outliers is interpreted to be a time step. The time
state of the clock that experienced the time step is reset to agree with the last
measurement and all other processing continues unmodified. If time steps continue
until M₂ successive outliers have occurred, as might happen after an extremely large
frequency step, then the clock should be reinitialized. The procedure for frequency
steps should be used to reinitialize the clock.
[0074] Most frequency steps are too small to produce outliers in the primary ensemble Kalman
filter. This is because the small frequency steps do not result in the accumulation
of large time errors during a single sample interval. Thus, all but the largest frequency
steps are detected in secondary ensemble Kalman filters that are computed solely for
this purpose. A set of filters with a range of sample intervals will result in the
early detection of frequency steps and also produce near optimum sensitivity for a
variety of clocks and performances. Recommended sample intervals are one hours, twelve
hours, one day, two days, four days, eight days and sixteen days. Since time steps
have already been detected (and rejected) using the primary ensemble filter, outliers
detected by the secondary filters are considered to have resulted from frequency steps.
[0075] When a frequency step is detected in one of the clocks, for example, clock k, it
is desirable to reduce the time constant for learning the new frequency. Therefore,
a new value is calculated for the spectral density of the random walk frequency noise.
First, the estimate of

is increased sufficiently so that the detected outlier would have been considered
normal. Then, the weights of the clock k are decreased to small values or zero to
protect the ensemble. The clock k is then reinitialized using a clock addition procedure.
[0076] As discussed previously, the clock weights are positive, semidefinite, and sum to
one, without any other restriction. It is possible to calculate a set of weights which
minimizes the total noise variance of the ensemble. First, the variance of the noise
in the ensemble states is calculated. This is represented by the following equations:




The weights which minimize the noise on the states u
e, y
e, and w
e are obtained by minimizing the appropriate diagonal elements of

such that



Alternatively, the weights can be chosen to have equal weighting for each member of
the ensemble. In this case, a
k=b
k=c
k=1/N .
[0077] Minimizing the noise of the time-scale is not necessarily the best approach to clock
weighting. The time-scale is realized by following Equation 26-2 and subtracting the
estimate of a clock with respect to the time-scale from the actual time of the clock.
Thus, the physical realization of the time scale is both perturbed by the time-scale
noise and also is in error due to the need to estimate the difference between the
time of the clock and the time-scale. When the error in this estimation process is
dominant, using clock weights that minimize the theorectical noise of the time-scale
is not appropriate. It has been found empirically that when the ensemble members differ
dramatically in performance, the weights that minimize the time-scale noise do not
optimize the overall performance of the physically realized time-scale using Equation
26-2. The situation can be improved by changing the weights associated with the time
state and keeping the frequency and frequency aging weights. Thus the weights

are used with the optimum b and c weights. This weighting scheme produces optimum
long-term performance and slightly suboptimum short-term performance in some circumstances,
but is much superior to the time-scale noise minimizing weights when the clocks have
very dissimilar performances. The weights of Equation 71-1 are strictly optimum when
the clocks have equal performance.
[0078] Whatever the method used, the clock weights are chosen in advance of the calculation.
However, if there is one or more outliers, the selected weights are modified by the
outlier rejection process. The actual weights used can be calculated from

where K'
I(1) is defined as 1 and the indexing scheme is as previously described. To preserve the
reliability of the ensemble, one usually limits the weights of each of the clocks
to some maximum value a
max. Thus, it may be necessary to readjust the initial weight assignments to achieve
the limitation or other requirements. If too few clocks are available, it may not
be possible to satisfy operational requirements. Under these conditions, it may be
possible to choose not to compute the ensemble time until the requirements can be
met. However, if the time must be used, it is always better to compute the ensemble
than to use a single member clock.
[0079] Another problem to be considered in the Kalman approach is the estimation of the
parameters required by a Kalman filter. The techniques that are normally applied are
Allan variance analysis and maximum likelihood analysis. However, in using the Allan
variance, there is a problem in that the Allan variance is defined for equally spaced
data. In an operational scenario, where there are occasional missing data, the gaps
may be bridged. But when data are irregularly spaced, a more powerful approach is
required.
[0080] The maximum likelihood approach determines the parameter set most likely to have
resulted in the observations. Equally spaced data are not required, but the data are
batch processed. Furthermore, each step of the search for the maximum requires a complete
recomputation of the Kalman filter, which results in an extremely time consuming procedure.
Both the memory needs and computation time are incompatible with real time or embedded
applications.
[0081] A variance analysis technique compatible with irregular observations has been developed.
The variance of the innovation sequence of the Kalman filter is analyzed to provide
estimates of the parameters of the filter. Like the Allan variance analysis, which
is performed on the unprocessed measurements, the innovation analysis requires only
a limited memory of past data. However, the forecast produced by the Kalman filter
allows the computation to be performed at arbitrary intervals once the algebraic form
of the innovation variance has been calculated.
[0082] The innovation sequence has been used to provide real time parameter estimates for
Kalman filters with equal sampling intervals. The conditions for estimating all the
parameters of the filter include (1) the system must be observable, (2) the system
must be invariant, (3) number of unknown parameters in Q (the system covariance) must
be less than the product of the dimension of the state vector and the dimension of
the measurement vector, and (4) the filter must be in steady state. This approach
was developed for discrete Kalman filters with equal sampling intervals, and without
modification, cannot be used for mixed mode filters because of the irregular sampling
which prevents the system from ever reaching steady state. However, it is possible
to proceed in a similar fashion by calculating the variance of the innovations in
terms of the true values of the parameters and the approximate gain and actual covariance
of the suboptimal Kalman filter that produces the innovation sequence. We describe
two methods of parameter estimation. The first method uses clock time difference estimates,
while the second method uses the estimates of the clocks with respect to the ensemble.
[0083] Parameter Estimation Method 1: The innovation vector is the difference between the
observation and the prediction, as follows:

By substituting Equation 73 in the measurement model (Equation 6)

since the measurement noise is uncorrolated with system noise for the clock difference
filters.
[0084] Adaptive modeling begins with an approximate Kalman filter gain K. As the state estimates
are computed, the variance of the innovations on the left side of Equation 74 is also
computed. The right side of this equation is written in terms of the actual filter
element values (covariance matrix elements) and the theoretical parameters. Finally,
the equations are inverted to produce improved estimates for the parameters. The method
of solving the parameters for discrete Kalman filters with equal sampling intervals
is inappropriate here because the autocovariance function is highly correlated from
one lag to the next and the efficiency of data utilization is therefore small. Instead,
only the autocovariance of the innovations for zero lags, i.e., the covariance of
the innovations, is used. The variances are given by

for the case of a time measurement, and

for the case of a frequency measurement. It is assumed the oscillator model contains
no hidden noise processes. This means that each noise in the model is dominant over
some region of the Fourier frequency space. The principal of parsimony encourages
this approach to modeling. Inspection of Equation 75 leads to the conclusion that
each of the parameters dominates the variance of the innovations in a unique region
of prediction time interval, δ, making it possible to obtain high quality estimates
for each of the parameters through a bootstrapping process. It should be noted that
the white phase measurement noise can be separated from the clock noise only by making
an independent assessment of the measurement system noise floor.
[0085] For each parameter to be estimated, a Kalman filter is computed using a subset of
the data chosen to maximize the number of predictions in the interval for which that
parameter makes the dominant contribution to the innovations. The filters are designated
0 through 4, starting with zero for the main state estimation filter, which runs as
often as possible. Each innovation is used to compute a single-point estimate of the
variance of the innovations for the corresponding δ. Substituting the estimated values
of the remaining parameters, Equation 75 is solved for the dominant parameter, and
the estimate of that parameter is updated in an exponential filter of the appropriate
length, for example,

[0086] If the minimum sampling interval is too long, it may not be possible to estimate
one or more of the parameters. However, there is no deleterious consequence of the
situation, since a parameter that cannot be estimated is not contributing appreciably
to the prediction errors. Simulation testing has shown that the previously described
method combines good data efficiency and high accuracy.
[0087] Each time a clock pair filter runs, a single estimate is obtained for one of the
noise spectral densities or variances of the clock, represented by F
ij. A Kalman filter can be used to obtain an optimum estimate for all F
i, given all possible measurements F
ij. The F
i for a given noise type are formed into an N dimensional vector

The state transition matrix is just the N dimensional identity matrix. The noise vector
is chosen to be nonzero in order to allow the estimates to change slowly with time.
This does not mean that the clock noises actually experience random walk behavior,
only that this is the simplest model that does not permanently lock in fixed values
for the noises. The variances of the noises perturbing the clock parameters can be
chosen based on the desired time constant of the Kalman filter. Assuming that the
noise is small, the Kalman gain is approximately σ
F/σ
meas. The parameter value will refresh every M measurements when its variance is set to
1/M² of the variance of the single measurement estimate of the parameter. A reasonable
value for the variance of a single measurement is

being approximately equal to 2F
i. The measurement matrix for the "ij"th measurement is a 1 x N row vector whose "i"th
and "j"th elements are unity and whose remaining elements are zero, such that H
ij =[0...010...010...0]. All the individual clock parameters are updated each cycle
of the Kalman recursion, even though the measurement involves only two clocks, because
the prior state estimates depend on the separation of variances involving all of the
clocks.
[0088] The storage requirements for this approach are minimal. There are five N element
state vectors, one for each of the possible noise types (white phase noise, white
frequency noise, white frequency measurement noise, random walk frequency noise, and
random walk frequency noise aging). There are also five N x N covariance matrixes.
A total of 5N(N-1)/2 cycles of the Kalman recursion are currently believed necessary
for the parameter update.
[0089] Parameter Estimation Method 2: A variance analysis technique compatible with irregular
observations has been developed. The variance of the innovation sequence of the time-scale
calculation is analyzed to provide estimates of the parameters of the filter. Like
the Allan variance analysis, which is performed on unprocessed measurements, the innovation
analysis requires only limited memory of past data. However, the forecasts allow the
computation to be performed at arbitrary intervals once the algebraic form of the
innovation variance has been calculated.
[0090] The innovation,

(t₂), is calculated from Equation 64-24. This equation can be rearranged in the form

After squaring and taking the expectation value one finds

[0091] Consider a Kalman filter to compute the estimates of

. Suppose the measurements for this filter are the û
j that were computed during the time-scale computation. Such a filter has
R = 0. Examination of Eq. 20 leads to the conclusion that

Evaluation of the terms on the right hand side leads to the equation

[0092] Each cycle of the filter is used to compute a primitive estimate of the variance
of the innovations, where the innovation is a scalar equal to û
j(t + δ|t + δ) - û
j(t + δ|t).

where

The spectral densities are calculated based on the assumption that the oscillator
model contains no hidden noise processes. This means that each noise in the model
is dominant over some region of Fourier frequency space. The principle of parsimony
encourages this approach to modeling. Inspection of Equation 64-37 leads to the conclusion
that each of the parameters dominates the variance of the innovations in a unique
region of prediction time interval, δ, making it possible to obtain high-quality estimates
for each of the parameters through a bootstrapping process. Note that the white phase
measurement noise can be separated from the clock noise only by making an independent
assessment of the measurement system noise floor.
[0093] For each parameter to be estimated, a Kalman filter is computed using a subset of
the data chosen to maximize the number of predictions in the interval for which that
parameter makes the dominant contribution to the innovations. Five filters should
be sufficient to estimate parameters in a time-scale of standard and high performance
cesium clocks, active hydrogen masers, and mercury ion frequency standards. The filter
are designated 1 through 5, starting with 1 for the main state estimation filter.
Each innovation is used to compute a single-point estimate of the variance of the
innovations for the corresponding δ. Substituting the estimated values of the remaining
parameters, Eq. 24 is solved for the dominant parameter, and the estimate of that
parameter is updated in an exponential filter of the appropriate length. If the minimum
sampling interval is too long, it may not be possible to estimate one or more of the
parameters. However, there is no deleterious consequence of this situation, since
a parameter that cannot be estimated is not contributing appreciably to the prediction
errors. Simulation testing in the case of a single clock has shown that the previously
described method combines good data efficiency and high accuracy. The following is
an example of an appropriate procedure. The measurement noise is assumed to lie between
σ
y(1s) = 1 x 10⁻¹² and 1 x 10⁻¹¹.
FILTER 1 Runs at a nominal sampling rate of 300 seconds. Such fast data acquisition
is important for real-time control. It is also fast enough to permit the estimation
of the white phase noise for active hydrogen masers. The source of this noise is the
measurement system. The same level of white phase noise should be assigned to all
time measurement channels. No other parameters should be estimated using this filter.
To determine whether the white phase noise may be measured for a given clock, verify
that Ŝ
β'h > Ŝ
ξδ.
FILTER 2 Runs at a nominal sampling rate of 3 hours. This sampling rate is appropriate
for estimating the white frequency noise of all the clocks including the mercury ion
devices. To determine whether the white frequency noise may be estimated for a given
clock, verify that Ŝ
β'h < Ŝ
ξδ and Ŝ
ξδ > Ŝ
µδ³/3. Frequency steps may be detected.
FILTER 3 Runs at a nominal sampling rate of 1 day. This sampling rate is appropriate
for estimating the random walk frequency noise for active hydrogen masers. To determine
whether the random walk frequency noise may be estimated for a given clock, verify
that Ŝ
ξδ < Ŝ
µδ³/3 and Ŝ
µδ³/3 > Ŝ
ζδ⁵/20. Frequency steps may be detected.
FILTER 4 Runs at a nominal sampling rate of 4 days. To determine whether the random
walk frequency noise may be estimated for a given clock, verify that Ŝ
ξδ < Ŝ
µδ³/3 and Ŝ
µδ³/3 > Ŝ
ζδ⁵/20. Frequency steps may be detected.
FILTER 5 Runs at a nominal sampling rate of 16 days. To determine whether the random
walk frequency noise may be estimated for a given clock, verify that Ŝ
ξδ < Ŝ
µδ³ and Ŝ
µδ³/3 > Ŝ
ζδ⁵/20. Frequency steps may be detected.
[0094] An implementation of the present invention will now be described with respect to
Figure 1. Figure 1 illustrates a circuit for obtaining a computation of ensemble time
from an ensemble of clocks 10. The ensemble 10 includes N clocks 12. The clocks 12
can be any combination of clocks suitable for use with precision time measurement
systems. Such clocks may include, but are not limited to cesium clocks, rubidium clocks,
hydrogen maser clocks and quartz crystal oscillators. Additionally, there is no limit
on the number of clocks.
[0095] Each of the N clocks 12 produces a respective signal µ₁, µ₂, µ₃,..., µ
N which is representative of its respective frequency output. The respective frequency
signals are passed through a passive power divider circuit 14 to make them available
for use by a time measurement system 16, which obtains the time differences between
designated ones of the clocks 12. As discussed above, the desired time differences
are the differences between the one of the clocks 12 designated as a hardware reference
clock and the remaining clocks 12. The clock 12 which acts as the reference clock
can be advantageously changed as desired by an operator. For example, if clock 12
designated "clock 1" is chosen to be the reference clock, the time measurement system
16 determines the differences between the reference clock and the remaining clocks,
which are represented by z₁₂, z₁₃, z₁₄,... z
1N. These data are input to a computer 18 for processing in accordance with the features
of the present invention as described above, namely, the complete ensemble definition
as provided above. When the ensemble definition as provided by Equations 23 - 25 is
provided for in Kalman filters, and since the Kalman filters are software-implemented,
the Kalman filters can be stored in memory 20. The computer 18 accesses the memory
20 for the necessary filters as required by the system programming in order to carry
out the time scale computation. The weights and other required outside data are input
by operator through a terminal 22. Upon completion of the processing of the clock
data via the Kalman filters according to the present invention, an estimate of the
ensemble time is output from the computer 20 to be manipulated in accordance with
the requirements of the user.
[0096] As discussed above, Kalman filters have been previously used in connection with ensembles
to obtain ensemble time estimates. These Kalman filters embodied the previous incomplete
ensemble definitions in Kalman form for the appropriate processing. Accordingly, it
will be appreciated by those skilled in the art that the actual implementation of
the Kalman equations into a time measurement system as described above and the appropriate
programming for the system are procedures known in the art. As also should be appreciated,
by providing a complete definition of the ensemble, the present system generally provides
a superior calculation of the ensemble time with respect to prior art.
[0097] While one embodiment of the invention has been discussed, it will be appreciated
by those skilled in the art that various modifications and variations are possible
without departing from the spirit and scope of the invention.