Home | About | Journals | Submit | Contact Us | Français |

**|**Sensors (Basel)**|**v.16(6); 2016 June**|**PMC4934273

Formats

Article sections

- Abstract
- 1. Introduction
- 2. Problem Formulation and Model Description
- 3. Distributed Fusion Linear Filter
- 4. Centralized LS Fusion Linear Filter
- 5. Numerical Simulation Example
- 6. Conclusions
- References

Authors

Related links

Sensors (Basel). 2016 June; 16(6): 847.

Published online 2016 June 8. doi: 10.3390/s16060847

PMCID: PMC4934273

Xue-Bo Jin, Academic Editor

Received 2016 May 5; Accepted 2016 June 3.

Copyright © 2016 by the authors; licensee MDPI, Basel, Switzerland.

This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC-BY) license (http://creativecommons.org/licenses/by/4.0/).

This article has been cited by other articles in PMC.

This paper is concerned with the distributed and centralized fusion filtering problems in sensor networked systems with random one-step delays in transmissions. The delays are described by Bernoulli variables correlated at consecutive sampling times, with different characteristics at each sensor. The measured outputs are subject to uncertainties modeled by random parameter matrices, thus providing a unified framework to describe a wide variety of network-induced phenomena; moreover, the additive noises are assumed to be one-step autocorrelated and cross-correlated. Under these conditions, without requiring the knowledge of the signal evolution model, but using only the first and second order moments of the processes involved in the observation model, recursive algorithms for the optimal linear distributed and centralized filters under the least-squares criterion are derived by an innovation approach. Firstly, local estimators based on the measurements received from each sensor are obtained and, after that, the distributed fusion filter is generated as the least-squares matrix-weighted linear combination of the local estimators. Also, a recursive algorithm for the optimal linear centralized filter is proposed. In order to compare the estimators performance, recursive formulas for the error covariance matrices are derived in all the algorithms. The effects of the delays in the filters accuracy are analyzed in a numerical example which also illustrates how some usual network-induced uncertainties can be dealt with using the current observation model described by random matrices.

Estimation problems in networked stochastic systems have been widely studied, especially in the past decade, due to the wide range of potential applications in many areas, such as target tracking, air traffic control, fault diagnosis, computer vision, and so on, and important advances in the design of multisensor fusion techniques have been achieved [1]. The basic question in the fusion estimation problems is how to merge the measurement data of different sensors; usually, the centralized and distributed fusion methods are used. Centralized fusion estimators are obtained by processing in the fusion center the measurements received from all sensors; in the distributed fusion method, first, local estimators, based on the measurements received from each sensor, are obtained and, then, these local estimators are combined according to a certain information fusion criterion. Several centralized and/or distributed fusion estimation algorithms have been proposed for conventional systems (see, e.g., [2,3,4,5], and the references therein), where the sensor measured outputs are affected only by additive noises, and assuming that the data are sent directly to the fusion center without any kind of transmission error; that is, assuming perfect connections.

However, although the use of sensor networks offers several advantages, due to restrictions of the physical equipment and uncertainties in the external environment, new problems associated with network-induced phenomena inevitably arise in both the output and the transmission of the sensor measurements [6]. Multiplicative noise uncertainties, sensor gain degradations and missing measurements are some of the random phenomena that usually arise in the sensor measured outputs, which cannot only be described by the usual additive disturbances. The existence of these uncertainties can severely degrade the performance of the conventional estimators, and has encouraged the need of designing new fusion estimation algorithms for such systems (see, e.g., [7,8,9,10,11], and references therein). Clearly, these situations with uncertainties in the measurements of the network sensors can be modeled by systems with random parameter measurement matrices, which have important practical significance due to the large number of realistic situations and application areas, such as digital control of chemical processes, radar control, navigation systems, or economic systems, in which this kind of systems with stochastic parameters are found (see, e.g., [12,13], among others). Therefore, it is not surprising that, in the past few years, much attention has been focused on the design of new fusion estimation algorithms for systems with random parameter matrices (see, e.g., [14,15,16,17], and references therein).

In addition to the aforementioned uncertainties, when the data packets are sent from the sensors to the processing center via a communication network, some further network-induced phenomena, such as random delays or measurement loss, inevitably arise during this transmission process, due to the unreliable network characteristics, imperfect communication channels and failures in the transmission. These additional transmission uncertainties can spoil the fusion estimators performance and motivate the need of designing fusion estimation algorithms that take their effects into consideration. In recent years, in response to the popularity of networked stochastic systems, the fusion estimation problem from observations with random delays and packet dropouts, which may happen during the data transmission, has been one of the mainstream research topics (see, e.g., [18,19,20,21,22,23,24,25,26,27,28,29,30], and references therein). All the above papers on signal estimation with random transmission delays assume independent random delays in each sensor and mutually independent delays between the different network sensors; in [31] this restriction was weakened and random delays featuring correlation at consecutive sampling times were considered, thus allowing to deal with common practical situations (e.g., those in which two consecutive observations cannot be delayed).

It should also be noted that in many real-world problems the measurement noises are usually correlated; for example, when all the sensors operate in the same noisy environment or when the sensor noises are state dependent. For this reason, the fairly conservative assumption that the measurement noises are uncorrelated is commonly weakened in most of the aforementioned research on signal estimation, including systems with both deterministic and random parameter matrices. Namely, the optimal Kalman filtering fusion problem in systems with cross-correlated noises at consecutive sampling times is addressed, for example, in [25]; also, under different types of noise correlation, centralized and distributed fusion algorithms are obtained in [11,26] for systems with multiplicative noise, and in [7] for systems where the measurements might have partial information about the signal. Autocorrelated and cross-correlated noises have been also considered in systems with random parameter matrices and transmission uncertainties; some results on the fusion estimation problems in these systems can be found in [22,24,27].

In this paper, covariance information is used to address the distributed and centralized fusion estimation problems for a class of linear networked stochastic systems with measured outputs perturbed by random parameter matrices, and subject to one-step random transmission delays. It is assumed that the sensor measurement noises are one-step autocorrelated and cross-correlated, and that the Bernoulli variables describing the measurement delays in the different sensors are correlated at the same and consecutive sampling times. The proposed observation model can describe the case of one-step delay, packet loss, or re-received measurements. To the best of the authors’ knowledge, fusion estimation problems in this framework with random measurement matrices and cross-correlated sensor noises in the measured outputs, together with correlated random delays in transmission, has not been investigated; so, encouraged by the above considerations, we reckon that it constitutes an interesting research challenge.

The main contributions of the current research are highlighted as follows: (i) The treatment used to address the estimation problem, based on covariance information, does not require the evolution model generating the signal process; nonetheless, the proposed fusion algorithms are also applicable to the conventional state-space model formulation; (ii) Random parameter matrices are considered in the measured outputs, which provide a fairly comprehensive and unified framework to describe some network-induced phenomena, such as multiplicative noise uncertainties or missing measurements, and correlation between the different sensor measurement noises is simultaneously considered; (iii) As in [31], random correlated delays in the transmission, with different delay characteristics at each sensor, are considered; however, the observation model in this paper is more general than that in [31], as the latter does not take random measurement matrices and correlated noises into account; (iv) Unlike [22,24,31], where only centralized fusion estimators are obtained, in this paper, both centralized and distributed estimation problems are addressed; the estimators are obtained under the innovation approach and recursive algorithms, very simple computationally and suitable for online applications, are proposed; (v) Finally, it must be noted that we do not use the augmented approach to deal with the delayed measurements, thus reducing the computational cost compared with the augmentation method.

The rest of the paper is organized as follows. Both, the measurement outputs with random parameter matrices, and the one-step random delay observation models are presented in Section 2, including the model hypotheses under which the distributed and centralized estimation problems are addressed. The distributed fusion method is applied in Section 3; specifically, the local least-squares linear filtering algorithms are derived in Section 3.1 using an innovation approach and, in Section 3.2, the proposed distributed fusion filter is obtained by a matrix-weighted linear combination of the local filtering estimators, using the mean squared error as optimality criterion. In Section 4, a recursive algorithm for the centralized least-squares linear filtering estimator is proposed. Section 5 is devoted to analyze the effectiveness of the proposed estimation algorithms by a simulation example, in which the effects of the sensor random delays on the estimators are compared. Some conclusions are drawn in Section 6.

The notations throughout the paper are standard. The notation $a\wedge b$ indicates the minimum value of two real numbers $a,b$. ${\mathbb{R}}^{n}$ and ${\mathbb{R}}^{m\times n}$ denote the *n*-dimensional Euclidean space and the set of all $m\times n$ real matrices, respectively. For a matrix *A*, the symbols ${A}^{T}$ and ${A}^{-1}$ denote its transpose and inverse, respectively; the notation $A\otimes B$ represents the Kronecker product of the matrices $A,B$. If the dimensions of vectors or matrices are not explicitly stated, they are assumed to be compatible with algebraic operations. In particular, *I* and 0 denote the identity matrix and the zero matrix of appropriate dimensions, respectively. For any function ${G}_{k,s}$, depending on the time instants *k* and *s*, we will write ${G}_{k}={G}_{k,k}$ for simplicity; analogously, ${F}^{\left(i\right)}={F}^{\left(ii\right)}$ will be written for any function ${F}^{\left(ij\right)}$, depending on the sensors *i* and *j*. Finally, ${\delta}_{k,s}$ denotes the Kronecker delta function.

This paper deals with the distributed and centralized fusion filtering problems from randomly delayed observations coming from networked sensors; the signal measurements at the different sensors are noisy linear functions with random parameter matrices, and the sensor noises are assumed to be correlated and cross-correlated at the same and consecutive sampling times. The estimation is performed in a processing center, which is connected to all sensors, where the measurements are transmitted through unreliable communication channels which may lead to one-step random delays, due to network congestion or other causes. In the centralized filtering problem, the estimators are obtained by fusion of all the network observations at each sampling time, whereas in the distributed filtering, local estimators, based only on the observations of each individual sensor, are first obtained and then, a fusion estimator based on the local ones is calculated.

Our aim is to design recursive algorithms for the optimal linear distributed and centralized filters under the least-squares (LS) criterion, requiring only the first and second-order moments of the processes involved in the model describing the observations from the different sensors. Next, we present the model and formulate the hypotheses necessary to address the estimation problem.

*Description of the observation model.* Let us consider a discrete-time second-order ${n}_{x}$-dimensional signal process, $\{{x}_{k};\phantom{\rule{4pt}{0ex}}k\ge 1\}$, which is measured in *m* sensor nodes of the network, $i=1,\dots ,m$, that generate the outputs ${z}_{k}^{\left(i\right)}\in {\mathbb{R}}^{{n}_{z}},\phantom{\rule{4pt}{0ex}}k\ge 1$, described by

$${z}_{k}^{\left(i\right)}={H}_{k}^{\left(i\right)}{x}_{k}+{v}_{k}^{\left(i\right)},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 1;\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}i=1,\dots ,m$$

(1)

where $\{{H}_{k}^{\left(i\right)};\phantom{\rule{4pt}{0ex}}k\ge 1\}$ are independent random parameter matrices of appropriate dimensions and $\{{v}_{k}^{\left(i\right)};\phantom{\rule{4pt}{0ex}}k\ge 1\}$ is the process describing the measurement noise in sensor *i*, which is assumed to be one-step correlated. We also assume that all the sensors operate in the same noisy environment and there exists correlation between different sensor noises at the same and consecutive sampling times.

In order to estimate the signal process, $\{{x}_{k};\phantom{\rule{4pt}{0ex}}k\ge 1\}$, the measurement outputs are transmitted to a processing center via unreliable channels, causing one-step random delays in such transmissions. When a measurement, ${z}_{k}^{\left(i\right)}$, suffers delay and, hence, it is unavailable at time *k*, the processor is assumed to use the previous one, ${z}_{k-1}^{\left(i\right)}$; this way of dealing with delays requires that the first measurement is always available and so, considering zero-one random variables ${\gamma}_{k}^{\left(i\right)},\phantom{\rule{4pt}{0ex}}k\ge 2$, the observations used in the estimation are described by

$${y}_{k}^{\left(i\right)}=(1-{\gamma}_{k}^{\left(i\right)}){z}_{k}^{\left(i\right)}+{\gamma}_{k}^{\left(i\right)}{z}_{k-1}^{\left(i\right)}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 2;\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}{y}_{1}^{\left(i\right)}={z}_{1}^{\left(i\right)};\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}i=1,\dots ,m$$

(2)

In this paper, the variables modelling the delays are assumed to be one-step correlated, thus covering many practical situations; for example those in which two consecutive observations through the same channel cannot be delayed, and situations where there is some sort of link between the different communications channels.

*Model hypotheses.* The LS estimation problem will be addressed under the following hypotheses about the processes involved in Equations (1) and (2), which formally specify the above assumptions:

*(H1)*The signal process, $\left\{{x}_{k};\phantom{\rule{4pt}{0ex}}k\ge 1\right\}$, has zero mean and its covariance function can be expressed in a separable form; namely, $E\left[{x}_{k}{x}_{s}^{T}\right]={A}_{k}{B}_{s}^{T},\phantom{\rule{4pt}{0ex}}s\le k$, where ${A}_{k},{B}_{k},\phantom{\rule{4pt}{0ex}}k\ge 1,$ are known matrices of dimensions ${n}_{x}\times N$.*(H2)*$\left\{{H}_{k}^{\left(i\right)};\phantom{\rule{4pt}{0ex}}k\ge 1\right\}$, $i=1,\dots ,m$, are independent sequences of independent random parameter matrices whose entries have known means and known second-order moments; we will denote ${\overline{H}}_{k}^{\left(i\right)}\equiv E[{H}_{k}^{\left(i\right)}],\phantom{\rule{4pt}{0ex}}k\ge 1$.*(H3)*The sensor measurement noises $\left\{{v}_{k}^{\left(i\right)};\phantom{\rule{4pt}{0ex}}k\ge 1\right\},\phantom{\rule{4pt}{0ex}}i=1,\dots ,m$, are zero-mean sequences with known second-order moments, defined by:$$E\left[{v}_{k}^{\left(i\right)}{v}_{s}^{\left(j\right)T}\right]={R}_{k}^{\left(ij\right)}{\delta}_{k,s}+{R}_{k,k-1}^{\left(ij\right)}{\delta}_{k-1,s}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}s\le k;\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}i,j=1,\dots ,m$$*(H4)*$\left\{{\gamma}_{k}^{\left(i\right)};\phantom{\rule{4pt}{0ex}}k\ge 2\right\},\phantom{\rule{4pt}{0ex}}i=1,\dots ,m,$ are sequences of Bernoulli random variables with known means, ${\overline{\gamma}}_{k}^{\left(i\right)}\equiv E\left[{\gamma}_{k}^{\left(i\right)}\right],\phantom{\rule{4pt}{0ex}}k\ge 2$. It is assumed that ${\gamma}_{k}^{\left(i\right)}$ and ${\gamma}_{s}^{\left(j\right)}$ are independent for $|k-s|\ge 2$, and the second-order moments, ${\overline{\gamma}}_{k,s}^{\left(ij\right)}\equiv E\left[{\gamma}_{k}^{\left(i\right)}{\gamma}_{s}^{\left(j\right)}\right],\phantom{\rule{4pt}{0ex}}s=k-1,k,$ and $i,j=1,\dots ,m,$ are also known.*(H5)*For $i=1,\dots ,m$, the processes $\{{x}_{k};\phantom{\rule{4pt}{0ex}}k\ge 1\},\phantom{\rule{4pt}{0ex}}\{{H}_{k}^{\left(i\right)};\phantom{\rule{4pt}{0ex}}k\ge 1\},\phantom{\rule{4pt}{0ex}}\{{v}_{k}^{\left(i\right)};\phantom{\rule{4pt}{0ex}}k\ge 1\}$ and $\left\{{\gamma}_{k}^{\left(i\right)};\phantom{\rule{4pt}{0ex}}k\ge 2\right\}$ are mutually independent.

Hypotheses *(H1)–(H5)* guarantee that the observation processes in the different sensors have zero mean, and that the matrices ${\Sigma}_{k,s}^{{y}^{\left(ij\right)}}\equiv E\left[{y}_{k}^{\left(i\right)}{y}_{s}^{\left(j\right)T}\right],\phantom{\rule{4pt}{0ex}}s=k-1,k$, can be obtained from ${\Sigma}_{k,s}^{{z}^{\left(ij\right)}}\equiv E\left[{z}_{k}^{\left(i\right)}{z}_{s}^{\left(j\right)T}\right],\phantom{\rule{4pt}{0ex}}s=k-2,k-1,k$, and their transposes, by the following expressions:

$$\begin{array}{c}{\Sigma}_{k,s}^{{y}^{\left(ij\right)}}=\left(1-{\overline{\gamma}}_{k}^{\left(i\right)}-{\overline{\gamma}}_{s}^{\left(j\right)}+{\overline{\gamma}}_{k,s}^{\left(ij\right)}\right){\Sigma}_{k,s}^{{z}^{\left(ij\right)}}+\left({\overline{\gamma}}_{s}^{\left(j\right)}-{\overline{\gamma}}_{k,s}^{\left(ij\right)}\right){\Sigma}_{k,s-1}^{{z}^{\left(ij\right)}}+\left({\overline{\gamma}}_{k}^{\left(i\right)}-{\overline{\gamma}}_{k,s}^{\left(ij\right)}\right){\Sigma}_{k-1,s}^{{z}^{\left(ij\right)}}+\hfill \\ \phantom{{\Sigma}_{k,s}^{{y}^{\left(ij\right)}}=}+{\overline{\gamma}}_{k,s}^{\left(ij\right)}{\Sigma}_{k-1,s-1}^{{z}^{\left(ij\right)}},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k,s\ge 2\hfill \\ {\Sigma}_{2,1}^{{y}^{\left(ij\right)}}=\left(1-{\overline{\gamma}}_{2}^{\left(i\right)}\right){\Sigma}_{2,1}^{{z}^{\left(ij\right)}}+{\overline{\gamma}}_{2}^{\left(i\right)}{\Sigma}_{1}^{{z}^{\left(ij\right)}};\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}{\Sigma}_{1}^{{y}^{\left(ij\right)}}={\Sigma}_{1}^{{z}^{\left(ij\right)}}\hfill \\ {\Sigma}_{k,s}^{{z}^{\left(ij\right)}}=E\left[{H}_{k}^{\left(i\right)}{A}_{k}{B}_{s}^{T}{H}_{s}^{\left(j\right)T}\right]+{R}_{k}^{\left(ij\right)}{\delta}_{k,s}+{R}_{k,k-1}^{\left(ij\right)}{\delta}_{k-1,s}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}s\le k\hfill \end{array}$$

(3)

*The term $E\left[{H}_{k}^{\left(i\right)}{A}_{k}{B}_{s}^{T}{H}_{s}^{\left(j\right)T}\right]$ in ${\Sigma}_{k,s}^{{z}^{\left(ij\right)}}$ becomes from $E\left[{H}_{k}^{\left(i\right)}{x}_{k}{x}_{s}^{T}{H}_{s}^{\left(j\right)T}\right]$ which, from the independence between $({x}_{k},{x}_{s})$ and $({H}_{k}^{\left(i\right)},{H}_{s}^{\left(j\right)})$, is equal to $E\left[{H}_{k}^{\left(i\right)}E\left[{x}_{k}{x}_{s}^{T}\right]{H}_{s}^{\left(j\right)T}\right]$. From* (H2)*, $E\left[{H}_{k}^{\left(i\right)}{A}_{k}{B}_{s}^{T}{H}_{s}^{\left(j\right)T}\right]={\overline{H}}_{k}^{\left(i\right)}{A}_{k}{B}_{s}^{T}{\overline{H}}_{s}^{\left(j\right)T}$ for $j\ne i$ or $s\ne k$, and $E\left[{H}_{k}^{\left(i\right)}{A}_{k}{B}_{k}^{T}{H}_{k}^{\left(i\right)T}\right]$ is computed from its entries, according the following formula:*

$${\left(E\left[H{A}_{k}{B}_{k}^{T}H\right]\right)}_{\phantom{\rule{-0.166667em}{0ex}}pq}=\sum _{a=1}^{{n}_{x}}\sum _{b=1}^{{n}_{x}}E\left[{h}_{pa}{h}_{qb}\right]{\left({A}_{k}{B}_{k}^{T}\right)}_{ab}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}p,q=1,\dots ,{n}_{z}\phantom{\rule{4pt}{0ex}}$$

*where $H\in {\mathbb{R}}^{{n}_{z}\times {n}_{x}}$ is any random matrix and ${h}_{pq}$ denote its entries.*

The aim of this section is to address the distributed fusion linear filtering problem of the signal ${x}_{k}$, from the randomly delayed observations defined by Equations (1) and (2), under the LS criterion. The proposed distributed filter is designed as the LS matrix-weighted linear combination of the local LS linear filters and therefore, in a first step, such local filters need being derived.

With the purpose of obtaining the signal LS linear filters based on the available observations from each sensor, we will use an innovation approach, which provides recursive algorithms for the local estimators, that will be denoted by ${\widehat{x}}_{k/k}^{\left(i\right)}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}i=1,\dots ,m$.

For each sensor $i=1,\dots ,m$, the innovation at time *k*, which represents the new information provided by the *k*-th observation, is defined by ${\mu}_{k}^{\left(i\right)}={y}_{k}^{\left(i\right)}-{\widehat{y}}_{k/k-1}^{\left(i\right)},\phantom{\rule{4pt}{0ex}}k\ge 1$, where ${\widehat{y}}_{k/k-1}^{\left(i\right)}$ is the LS linear estimator of ${y}_{k}$ based on ${y}_{s}^{\left(i\right)},\phantom{\rule{4pt}{0ex}}s\le k-1$, with ${\widehat{y}}_{1/0}^{\left(i\right)}=E\left[{y}_{1}^{\left(i\right)}\right]=0$. As it is known [32], the innovations, $\{{\mu}_{k}^{\left(i\right)};\phantom{\rule{4pt}{0ex}}k\ge 1\}$, constitute a zero-mean white process, and the LS linear estimator of any random vector ${\xi}_{k}$ based on the observations ${y}_{1}^{\left(i\right)},\dots ,{y}_{L}^{\left(i\right)}$, denoted by ${\widehat{\xi}}_{k/L}^{\left(i\right)}$, can be calculated as a linear combination of the corresponding innovations, ${\mu}_{1}^{\left(i\right)},\dots {\mu}_{L}^{\left(i\right)}$; namely,

$${\widehat{\xi}}_{k/L}^{\left(i\right)}=\sum _{h=1}^{L}E\left[{\xi}_{k}{\mu}_{h}^{\left(i\right)T}\right]{\Pi}_{h}^{\left(i\right)-1}{\mu}_{h}^{\left(i\right)}$$

(4)

where ${\Pi}_{k}^{\left(i\right)}\equiv E\left[{\mu}_{k}^{\left(i\right)}{\mu}_{k}^{\left(i\right)T}\right]$ denotes the covariance matrix of ${\mu}_{k}^{\left(i\right)}$. This general expression is derived from the *Orthogonal Projection Lemma* (OPL), which establishes that the estimation error is uncorrelated with all the observations or, equivalently, uncorrelated with all the innovations.

Using the following alternative expression for the observations ${y}_{k}^{\left(i\right)}$ given by Equation (2),

$$\begin{array}{c}{y}_{k}^{\left(i\right)}=\left(1-{\overline{\gamma}}_{k}^{\left(i\right)}\right){H}_{k}^{\left(i\right)}{x}_{k}+{\overline{\gamma}}_{k}^{\left(i\right)}{\overline{H}}_{k-1}^{\left(i\right)}{x}_{k-1}+{w}_{k}^{\left(i\right)},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 2\hfill \\ {w}_{k}^{\left(i\right)}={\overline{\gamma}}_{k}^{\left(i\right)}\left({H}_{k-1}^{\left(i\right)}-{\overline{H}}_{k-1}^{\left(i\right)}\right){x}_{k-1}+\left(1-{\overline{\gamma}}_{k}^{\left(i\right)}\right){v}_{k}^{\left(i\right)}+{\overline{\gamma}}_{k}^{\left(i\right)}{v}_{k-1}^{\left(i\right)}-\left({\gamma}_{k}^{\left(i\right)}-{\overline{\gamma}}_{k}^{\left(i\right)}\right)\left({z}_{k}^{\left(i\right)}-{z}_{k-1}^{\left(i\right)}\right),\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 2\hfill \end{array}$$

(5)

and taking into account the independence hypotheses stated on the model, it is easy to see, from Equation (4), that

$${\widehat{y}}_{k/k-1}^{\left(i\right)}=\left(1-{\overline{\gamma}}_{k}^{\left(i\right)}\right){\overline{H}}_{k}^{\left(i\right)}{\widehat{x}}_{k/k-1}^{\left(i\right)}+{\overline{\gamma}}_{k}^{\left(i\right)}{\overline{H}}_{k-1}^{\left(i\right)}{\widehat{x}}_{k-1/k-1}^{\left(i\right)}+\sum _{h=1}^{(k-1)\wedge 2}{\mathcal{W}}_{k,k-h}^{\left(i\right)}{\Pi}_{k-h}^{\left(i\right)-1}{\mu}_{k-h}^{\left(i\right)}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 2$$

(6)

where ${\mathcal{W}}_{k,k-h}^{\left(i\right)}\equiv E\left[{w}_{k}^{\left(i\right)}{\mu}_{k-h}^{\left(i\right)T}\right],\phantom{\rule{4pt}{0ex}}h=1,2.$

The general Expression (4) for the LS linear estimators as linear combination of the innovations, together with Expression (6) for the one-stage observation predictor, are the starting point to derive the local recursive filtering algorithms presented below in Theorem 1; these algorithms provide also the filtering error covariance matrices, ${P}_{k/k}^{\left(i\right)}\equiv E\left[({x}_{k}-{\widehat{x}}_{k/k}^{\left(i\right)}){({x}_{k}-{\widehat{x}}_{k/k}^{\left(i\right)})}^{T}\right]$, which measure the accuracy of the estimators ${\widehat{x}}_{k/k}^{\left(i\right)}$ when the LS optimality criterion is used.

Hereafter, for the matrices ${A}_{k}$ and ${B}_{k}$ involved in the signal covariance factorization *(H1)*, the following operator will be used:

$${\overline{\mathcal{H}}}_{{C}_{k}}^{\left(i\right)}=\left(1-{\overline{\gamma}}_{k}^{\left(i\right)}\right){\overline{H}}_{k}^{\left(i\right)}{C}_{k}+{\overline{\gamma}}_{k}^{\left(i\right)}{\overline{H}}_{k-1}^{\left(i\right)}{C}_{k-1},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 2;\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}{\overline{\mathcal{H}}}_{{C}_{1}}^{\left(i\right)}={\overline{H}}_{1}^{\left(i\right)}{C}_{1},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}C=A,B$$

(7)

*For each $i=1,\dots ,m$, the local LS linear filters, ${\widehat{x}}_{k/k}^{\left(i\right)},$ and the corresponding error covariance matrices, ${P}_{k/k}^{\left(i\right)}$, are given by*

$$\begin{array}{c}{\widehat{x}}_{k/k}^{\left(i\right)}={A}_{k}{O}_{k}^{\left(i\right)},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 1\hfill \end{array}$$

(8)

*and*

$${P}_{k/k}^{\left(i\right)}={A}_{k}{\left({B}_{k}-{A}_{k}{r}_{k}^{\left(i\right)}\right)}^{T},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 1$$

(9)

*where the vectors ${O}_{k}^{\left(i\right)}$ and the matrices ${r}_{k}^{\left(i\right)}=E\left[{O}_{k}^{\left(i\right)}{O}_{k}^{\left(i\right)T}\right]$ are recursively obtained from*

$${O}_{k}^{\left(i\right)}={O}_{k-1}^{\left(i\right)}+{J}_{k}^{\left(i\right)}{\Pi}_{k}^{\left(i\right)-1}{\mu}_{k}^{\left(i\right)},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 1;\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}{O}_{0}^{\left(i\right)}=0$$

(10)

$${r}_{k}^{\left(i\right)}={r}_{k-1}^{\left(i\right)}+{J}_{k}^{\left(i\right)}{\Pi}_{k}^{\left(i\right)-1}{J}_{k}^{\left(i\right)T},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 1;\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}{r}_{0}^{\left(i\right)}=0$$

(11)

*and the matrices ${J}_{k}^{\left(i\right)}=E\left[{O}_{k}^{\left(i\right)}{\mu}_{k}^{\left(i\right)T}\right]$ satisfy*

$${J}_{k}^{\left(i\right)}={\overline{\mathcal{H}}}_{{B}_{k}}^{\left(i\right)T}-{r}_{k-1}^{\left(i\right)}{\overline{\mathcal{H}}}_{{A}_{k}}^{\left(i\right)T}-\sum _{h=1}^{(k-1)\wedge 2}{J}_{k-h}^{\left(i\right)}{\Pi}_{k-h}^{\left(i\right)-1}{\mathcal{W}}_{k,k-h}^{\left(i\right)T}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 2;\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}{J}_{1}^{\left(i\right)}={\overline{\mathcal{H}}}_{{B}_{1}}^{\left(i\right)T}$$

(12)

*The innovations ${\mu}_{k}^{\left(i\right)}$, and their covariance matrices, ${\Pi}_{k}^{\left(i\right)}$, are given by*

$${\mu}_{k}^{\left(i\right)}={y}_{k}^{\left(i\right)}-{\overline{\mathcal{H}}}_{{A}_{k}}^{\left(i\right)}{O}_{k-1}^{\left(i\right)}-\sum _{h=1}^{(k-1)\wedge 2}{\mathcal{W}}_{k,k-h}^{\left(i\right)}{\Pi}_{k-h}^{\left(i\right)-1}{\mu}_{k-h}^{\left(i\right)}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 2;\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}{\mu}_{1}^{\left(i\right)}={y}_{1}^{\left(i\right)}$$

(13)

*and*

$$\begin{array}{c}{\displaystyle {\Pi}_{k}^{\left(i\right)}={\Sigma}_{k}^{{y}^{\left(i\right)}}-{\overline{\mathcal{H}}}_{{A}_{k}}^{\left(i\right)}\left({\overline{\mathcal{H}}}_{{B}_{k}}^{\left(i\right)T}-{J}_{k}^{\left(i\right)}\right)-\sum _{h=1}^{(k-1)\wedge 2}{\mathcal{W}}_{k,k-h}^{\left(i\right)}{\Pi}_{k-h}^{\left(i\right)-1}{\left({\overline{\mathcal{H}}}_{{A}_{k}}^{\left(i\right)}{J}_{k-h}^{\left(i\right)}+{\mathcal{W}}_{k,k-h}^{\left(i\right)}\right)}^{T},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 2}\hfill \\ {\Pi}_{1}^{\left(i\right)}={\Sigma}_{1}^{{y}^{\left(i\right)}}\hfill \end{array}$$

(14)

*The coefficients ${\mathcal{W}}_{k,k-h}^{\left(i\right)}=E\left[{w}_{k}^{\left(i\right)}{\mu}_{k-h}^{\left(i\right)T}\right]$, $h=1,2$, are calculated as*

$$\begin{array}{c}{\mathcal{W}}_{k,k-1}^{\left(i\right)}={\Sigma}_{k,k-1}^{{y}^{\left(i\right)}}-{\overline{\mathcal{H}}}_{{A}_{k}}^{\left(i\right)}{\overline{\mathcal{H}}}_{{B}_{k-1}}^{\left(i\right)T}-{\mathcal{W}}_{k,k-2}^{\left(i\right)}{\Pi}_{k-2}^{\left(i\right)-1}{\left({\overline{\mathcal{H}}}_{{A}_{k-1}}^{\left(i\right)}{J}_{k-2}^{\left(i\right)}+{\mathcal{W}}_{k-1,k-2}^{\left(i\right)}\right)}^{T},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 3\hfill \\ {\mathcal{W}}_{2,1}^{\left(i\right)}={\Sigma}_{2,1}^{{y}^{\left(i\right)}}-{\overline{\mathcal{H}}}_{{A}_{2}}^{\left(i\right)}{\overline{\mathcal{H}}}_{{B}_{1}}^{\left(i\right)T}\hfill \end{array}$$

(15)

$$\phantom{\rule{-91.04872pt}{0ex}}{\mathcal{W}}_{k,k-2}^{\left(i\right)}={\overline{\gamma}}_{k}^{\left(i\right)}(1-{\overline{\gamma}}_{k-2}^{\left(i\right)}){R}_{k-1,k-2}^{\left(i\right)}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 4;\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}{\mathcal{W}}_{3,1}^{\left(i\right)}={\overline{\gamma}}_{3}^{\left(i\right)}{R}_{2,1}^{\left(i\right)}$$

(16)

Finally, the matrices ${\Sigma}_{k,s}^{{y}^{\left(i\right)}}$ and ${\overline{\mathcal{H}}}_{{A}_{s}}^{\left(i\right)},\phantom{\rule{4pt}{0ex}}{\overline{\mathcal{H}}}_{{B}_{s}}^{\left(i\right)},\phantom{\rule{4pt}{0ex}}s=k-1,k$, are given in Equations (3) and (7), respectively.

The local filter ${\widehat{x}}_{k/k}^{\left(i\right)}$ will be obtained from the general Expression (4), starting from the computation of the coefficients ${\mathcal{X}}_{k,h}^{\left(i\right)}=E\left[{x}_{k}{\mu}_{h}^{\left(i\right)T}\right]=E\left[{x}_{k}{y}_{h}^{\left(i\right)T}\right]-E\left[{x}_{k}{\widehat{y}}_{h/h-1}^{\left(i\right)T}\right],\phantom{\rule{4pt}{0ex}}1\le h\le k$.

The independence hypotheses and the separable structure of the signal covariance *(H1)* lead to $E\left[{x}_{k}{y}_{h}^{\left(i\right)T}\right]={A}_{k}{\overline{\mathcal{H}}}_{{B}_{h}}^{\left(i\right)T}$, with ${\overline{\mathcal{H}}}_{{B}_{h}}^{\left(i\right)}$ given by Equation (7). From Expression (6) for ${\widehat{y}}_{h/h-1}^{\left(i\right)},\phantom{\rule{4pt}{0ex}}h\ge 2,$ we have:

$$E\left[{x}_{k}{\widehat{y}}_{h/h-1}^{\left(i\right)T}\right]=\left(1-{\overline{\gamma}}_{h}^{\left(i\right)}\right)E\left[{x}_{k}{\widehat{x}}_{h/h-1}^{\left(i\right)T}\right]{\overline{H}}_{h}^{\left(i\right)T}+{\overline{\gamma}}_{h}^{\left(i\right)}E\left[{x}_{k}{\widehat{x}}_{h-1/h-1}^{\left(i\right)T}\right]{\overline{H}}_{h-1}^{\left(i\right)T}+\sum _{j=1}^{(h-1)\wedge 2}\phantom{\rule{-2.84544pt}{0ex}}{\mathcal{X}}_{k,h-j}^{\left(i\right)}{\Pi}_{h-j}^{\left(i\right)-1}{\mathcal{W}}_{h,h-j}^{\left(i\right)T}$$

Hence, using now Equation (4) for ${\widehat{x}}_{h/h-1}^{\left(i\right)}$ and ${\widehat{x}}_{h-1/h-1}^{\left(i\right)}$, the filter coefficients are expressed as

$$\begin{array}{c}{\displaystyle {\mathcal{X}}_{k,h}^{\left(i\right)}={A}_{k}{\overline{\mathcal{H}}}_{{B}_{h}}^{\left(i\right)T}-\sum _{j=1}^{h-1}{\mathcal{X}}_{k,j}^{\left(i\right)}{\Pi}_{j}^{\left(i\right)-1}\left(\left(1-{\overline{\gamma}}_{h}^{\left(i\right)}\right){\mathcal{X}}_{h,j}^{\left(i\right)T}{\overline{H}}_{h}^{\left(i\right)T}+{\overline{\gamma}}_{h}^{\left(i\right)}{\mathcal{X}}_{h-1,j}^{\left(i\right)T}{\overline{H}}_{h-1}^{\left(i\right)T}\right)}\hfill \\ {\displaystyle \phantom{{\mathcal{X}}_{k,h}^{\left(i\right)}=}-\sum _{j=1}^{(h-1)\wedge 2}{\mathcal{X}}_{k,h-j}^{\left(i\right)}{\Pi}_{h-j}^{\left(i\right)-1}{\mathcal{W}}_{h,h-j}^{\left(i\right)T}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}2\le h\le k;}\hfill \\ {\mathcal{X}}_{k,1}^{\left(i\right)}={A}_{k}{\overline{\mathcal{H}}}_{{B}_{1}}^{\left(i\right)T}\hfill \end{array}$$

which guarantees that ${\mathcal{X}}_{k,h}^{\left(i\right)}={A}_{k}{J}_{h}^{\left(i\right)},\phantom{\rule{4pt}{0ex}}1\le h\le k$, with ${J}_{h}^{\left(i\right)}$ given by

$$J}_{h}^{\left(i\right)}={\overline{\mathcal{H}}}_{{B}_{h}}^{\left(i\right)T}-\sum _{j=1}^{h-1}{J}_{j}^{\left(i\right)}{\Pi}_{j}^{\left(i\right)-1}{J}_{j}^{\left(i\right)T}{\overline{\mathcal{H}}}_{{A}_{h}}^{\left(i\right)T}-\phantom{\rule{-0.166667em}{0ex}}\phantom{\rule{-0.166667em}{0ex}}\sum _{j=1}^{(h-1)\wedge 2}{J}_{h-j}^{\left(i\right)}{\Pi}_{h-j}^{\left(i\right)-1}{\mathcal{W}}_{h,h-j}^{\left(i\right)T}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}h\ge 2;\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}{J}_{1}^{\left(i\right)}={\overline{\mathcal{H}}}_{{B}_{1}}^{\left(i\right)T$$

(17)

Therefore, by defining $O}_{k}^{\left(i\right)}=\sum _{h=1}^{k}{J}_{h}^{\left(i\right)}{\Pi}_{h}^{\left(i\right)-1}{\mu}_{h}^{\left(i\right)$ and ${r}_{k}^{\left(i\right)}=E\left[{O}_{k}^{\left(i\right)}{O}_{k}^{\left(i\right)T}\right]$, Expression (8) for the filter follows immediately from Equation (4), and Equation (9) is obtained by using the OPL to express ${P}_{k/k}^{\left(i\right)}=E\left[{x}_{k}{x}_{k}^{T}\right]-E\left[{\widehat{x}}_{k/k}^{\left(i\right)}{\widehat{x}}_{k/k}^{\left(i\right)T}\right]$, and applying *(H1)* and Equation (8).

The recursive Expressions (10) and (11) are directly obtained from the corresponding definitions, taking into account that $r}_{k}^{\left(i\right)}=\sum _{h=1}^{k}{J}_{h}^{\left(i\right)}{\Pi}_{h}^{\left(i\right)-1}{J}_{h}^{\left(i\right)T$ which, in turn, from Equation (17), leads to Equation (12) for ${J}_{k}^{\left(i\right)}$.

From now on, using that ${\widehat{x}}_{k/k-1}^{\left(i\right)}={A}_{k}{O}_{k-1}^{\left(i\right)},\phantom{\rule{4pt}{0ex}}{\widehat{x}}_{k-1/k-1}^{\left(i\right)}={A}_{k-1}{O}_{k-1}^{\left(i\right)}$ and Equation (7), Expression (6) for the observation predictor will be equivalently written as follows:

$${\widehat{y}}_{k/k-1}^{\left(i\right)}={\overline{\mathcal{H}}}_{{A}_{k}}^{\left(i\right)}{O}_{k-1}^{\left(i\right)}+\sum _{h=1}^{(k-1)\wedge 2}{\mathcal{W}}_{k,k-h}^{\left(i\right)}{\Pi}_{k-h}^{\left(i\right)-1}{\mu}_{k-h}^{\left(i\right)}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 2$$

(18)

From Equation (18), Expression (13) for the innovation is directly obtained and, applying the OPL to express its covariance matrix as ${\Pi}_{k}^{\left(i\right)}=E\left[{y}_{k}^{\left(i\right)}{y}_{k}^{\left(i\right)T}\right]-E\left[{\widehat{y}}_{k/k-1}^{\left(i\right)}{\widehat{y}}_{k/k-1}^{\left(i\right)T}\right]$, the following identity holds:

$${\Pi}_{k}^{\left(i\right)}={\Sigma}_{k}^{{y}^{\left(i\right)}}-{\overline{\mathcal{H}}}_{{A}_{k}}^{\left(i\right)}E\left[{O}_{k-1}^{\left(i\right)}{\widehat{y}}_{k/k-1}^{\left(i\right)T}\right]-\sum _{h=1}^{(k-1)\wedge 2}{\mathcal{W}}_{k,k-h}^{\left(i\right)}{\Pi}_{k-h}^{\left(i\right)-1}E\left[{\mu}_{k-h}^{\left(i\right)}{\widehat{y}}_{k/k-1}^{\left(i\right)T}\right],\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 2;\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}{\Pi}_{1}^{\left(i\right)}={\Sigma}_{1}^{{y}^{\left(i\right)}}$$

Now, using again Equation (18), it is deduced from Expression (12) that $E\left[{O}_{k-1}^{\left(i\right)}{\widehat{y}}_{k/k-1}^{\left(i\right)T}\right]={\overline{\mathcal{H}}}_{{B}_{k}}^{\left(i\right)T}-{J}_{k}^{\left(i\right)}$ and, since $E\left[{\widehat{y}}_{k/k-1}^{\left(i\right)}{\mu}_{k-h}^{\left(i\right)T}\right]={\overline{\mathcal{H}}}_{{A}_{k}}^{\left(i\right)}E\left[{O}_{k-1}^{\left(i\right)}{\mu}_{k-h}^{\left(i\right)T}\right]+{\mathcal{W}}_{k,k-h}^{\left(i\right)}$ and $E\left[{O}_{k-1}^{\left(i\right)}{\mu}_{k-h}^{\left(i\right)T}\right]={J}_{k-h}^{\left(i\right)}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}h=1,2$, Expression (14) for ${\Pi}_{k}^{\left(i\right)}$ is obtained.

To complete the proof, the expressions for ${\mathcal{W}}_{k,k-h}^{\left(i\right)}=E\left[{w}_{k}^{\left(i\right)}{\mu}_{k-h}^{\left(i\right)T}\right],\phantom{\rule{4pt}{0ex}}h=1,2$, with ${w}_{k}^{\left(i\right)}$ given in Equation (5), are derived taking into account that ${w}_{k}^{\left(i\right)}$ is uncorrelated with ${y}_{h}^{\left(i\right)},\phantom{\rule{4pt}{0ex}}h\le k-3$. Consequently, ${\mathcal{W}}_{k,k-2}^{\left(i\right)}=E\left[{w}_{k}^{\left(i\right)}{y}_{k-2}^{\left(i\right)T}\right]$, and Expression (16) is directly obtained from Equations (1), (2) and (5), using the hypotheses stated on the model. Next, using Equation (4) for ${\widehat{y}}_{k-1/k-2}^{\left(i\right)}$ in ${\mathcal{W}}_{k,k-1}^{\left(i\right)}=E\left[{w}_{k}^{\left(i\right)}{y}_{k-1}^{\left(i\right)T}\right]-E\left[{w}_{k}^{\left(i\right)}{\widehat{y}}_{k-1/k-2}^{\left(i\right)T}\right]$, we have:

$${\mathcal{W}}_{k,k-1}^{\left(i\right)}=E\left[{w}_{k}^{\left(i\right)}{y}_{k-1}^{\left(i\right)T}\right]-{\mathcal{W}}_{k,k-2}^{\left(i\right)}{\Pi}_{k-2}^{\left(i\right)-1}{\left(E\left[{y}_{k-1}^{\left(i\right)}{\mu}_{k-2}^{\left(i\right)T}\right]\right)}^{T}$$

(19)

To compute the first expectation involved in this formula, we express ${w}_{k}^{\left(i\right)}={y}_{k}^{\left(i\right)}-(1-{\overline{\gamma}}_{k}^{\left(i\right)}){H}_{k}^{\left(i\right)}{x}_{k}-{\overline{\gamma}}_{k}^{\left(i\right)}{H}_{k-1}^{\left(i\right)}{x}_{k-1}$ and we apply the OPL to rewrite $E\left[{x}_{s}{y}_{k-1}^{\left(i\right)T}\right]=E\left[{\widehat{x}}_{s/k-1}^{\left(i\right)}{y}_{k-1}^{\left(i\right)T}\right],\phantom{\rule{4pt}{0ex}}s=k,k-1$, thus obtaining that $E\left[{w}_{k}^{\left(i\right)}{y}_{k-1}^{\left(i\right)T}\right]={\Sigma}_{k,k-1}^{{y}^{\left(i\right)}}-{\overline{\mathcal{H}}}_{{A}_{k}}^{\left(i\right)}E\left[{O}_{k-1}^{\left(i\right)}{y}_{k-1}^{\left(i\right)T}\right]$; then, by expressing $E\left[{O}_{k-1}^{\left(i\right)}{y}_{k-1}^{\left(i\right)T}\right]=E\left[{O}_{k-1}^{\left(i\right)}{\mu}_{k-1}^{\left(i\right)T}\right]+E\left[{O}_{k-1}^{\left(i\right)}{\widehat{y}}_{k/k-1}^{\left(i\right)T}\right]$ and using Equations (12) and (18), it follows that $E\left[{w}_{k}^{\left(i\right)}{y}_{k-1}^{\left(i\right)T}\right]={\Sigma}_{k,k-1}^{{y}^{\left(i\right)}}-{\overline{\mathcal{H}}}_{{A}_{k}}^{\left(i\right)}{\overline{\mathcal{H}}}_{{B}_{k-1}}^{\left(i\right)T}.$ The second expectation in Equation (19) is easily computed taking into account that, from the OPL, it is equal to $E\left[{\widehat{y}}_{k-1/k-2}^{\left(i\right)}{\mu}_{k-2}^{\left(i\right)T}\right]$ and using Equation (18). So the proof is completed.

As it has been mentioned previously, a linear matrix-weighted fusion filter is now generated from the local filters by applying the LS optimality criterion. The distributed fusion filter at any time *k* is hence designed as a product, ${\mathcal{F}}_{k}{\widehat{X}}_{k/k}$, where ${\widehat{X}}_{k/k}={\left({\widehat{x}}_{k/k}^{\left(1\right)T},\dots ,{\widehat{x}}_{k/k}^{\left(m\right)T}\right)}^{T}$ is the vector constituted by the local filters, and ${\mathcal{F}}_{k}\in {\mathbb{R}}^{{n}_{x}\times m{n}_{x}}$ is a matrix such that the mean squared error, $E\left[{\left({x}_{k}-{\mathcal{F}}_{k}{\widehat{X}}_{k/k}\right)}^{T}\left({x}_{k}-{\mathcal{F}}_{k}{\widehat{X}}_{k/k}\right)\right]$, is minimized.

As it is known, the solution of this problem is given by ${\mathcal{F}}_{k}^{opt}=E\left[{x}_{k}{\widehat{X}}_{k/k}^{T}\right]{\left(E\left[{\widehat{X}}_{k/k}{\widehat{X}}_{k/k}^{T}\right]\right)}^{-1}$ and, consequently, the proposed distributed filter is expressed as:

$${\widehat{x}}_{k/k}^{\left(D\right)}=E\left[{x}_{k}{\widehat{X}}_{k/k}^{T}\right]{\widehat{\Sigma}}_{k/k}^{-1}{\widehat{X}}_{k/k}$$

(20)

with ${\widehat{\Sigma}}_{k/k}\equiv E\left[{\widehat{X}}_{k/k}{\widehat{X}}_{k/k}^{T}\right]={\left(E\left[{\widehat{x}}_{k/k}^{\left(i\right)}{\widehat{x}}_{k/k}^{\left(j\right)T}\right]\right)}_{i,j=1,\dots ,m}$.

In view of Equation (20), and since the OPL guarantees that $E\left[{x}_{k}{\widehat{X}}_{k/k}^{T}\right]=\left(E\left[{\widehat{x}}_{k/k}^{\left(1\right)}{\widehat{x}}_{k/k}^{\left(1\right)T}\right],\dots ,E\left[{\widehat{x}}_{k/k}^{\left(m\right)}{\widehat{x}}_{k/k}^{\left(m\right)T}\right]\right)$, the derivation of ${\widehat{x}}_{k/k}^{\left(D\right)}$ only requires the knowledge of the matrices ${\widehat{\Sigma}}_{k/k}^{\left(ij\right)}\equiv E\left[{\widehat{x}}_{k/k}^{\left(i\right)}{\widehat{x}}_{k/k}^{\left(j\right)T}\right],\phantom{\rule{4pt}{0ex}}i,j=1,\dots ,m$.

The following theorem provides a recursive algorithm to compute the matrices ${\widehat{\Sigma}}_{k/k}^{\left(ij\right)}$ which not only determine the proposed distributed fusion filter, but also the filtering error covariance matrix, ${P}_{k/k}^{\left(D\right)}=E\left[\left({x}_{k}-{\widehat{x}}_{k/k}^{\left(D\right)}\right){\left({x}_{k}-{\widehat{x}}_{k/k}^{\left(D\right)}\right)}^{T}\right]$.

*Let ${\widehat{X}}_{k/k}={\left({\widehat{x}}_{k/k}^{\left(1\right)T},\dots ,{\widehat{x}}_{k/k}^{\left(m\right)T}\right)}^{T}$ denote the vector constituted by the local LS filters given in Theorem 1, and ${\widehat{\Sigma}}_{k/k}={\left({\widehat{\Sigma}}_{k/k}^{\left(ij\right)}\right)}_{i,j=1,\dots ,m}$ , with ${\widehat{\Sigma}}_{k/k}^{\left(ij\right)}=E\left[{\widehat{x}}_{k/k}^{\left(i\right)}{\widehat{x}}_{k/k}^{\left(j\right)T}\right]$. Then, the distributed filtering estimator, ${\widehat{x}}_{k/k}^{\left(D\right)}$, and the error covariance matrix, ${P}_{k/k}^{\left(D\right)}$, are given by*

$${\widehat{x}}_{k/k}^{\left(D\right)}=\left({\widehat{\Sigma}}_{k/k}^{\left(1\right)}\phantom{\rule{4pt}{0ex}},\dots ,{\widehat{\Sigma}}_{k/k}^{\left(m\right)}\right){\widehat{\Sigma}}_{k/k}^{-1}{\widehat{X}}_{k/k}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 1$$

(21)

*and*

$${P}_{k/k}^{\left(D\right)}={A}_{k}{B}_{k}^{T}-\left({\widehat{\Sigma}}_{k/k}^{\left(1\right)}\phantom{\rule{4pt}{0ex}},\dots ,{\widehat{\Sigma}}_{k/k}^{\left(m\right)}\right){\widehat{\Sigma}}_{k/k}^{-1}{\left({\widehat{\Sigma}}_{k/k}^{\left(1\right)}\phantom{\rule{4pt}{0ex}},\dots ,{\widehat{\Sigma}}_{k/k}^{\left(m\right)}\right)}^{T},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 1$$

(22)

*The matrices ${\widehat{\Sigma}}_{k/k}^{\left(ij\right)}$, $i,j=1,\dots ,m$, are computed by*

$${\widehat{\Sigma}}_{k/k}^{\left(ij\right)}={A}_{k}{r}_{k}^{\left(ij\right)}{A}_{k}^{T},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 1$$

(23)

*with ${r}_{k}^{\left(ij\right)}=E\left[{O}_{k}^{\left(i\right)}{O}_{k}^{\left(j\right)T}\right]$ satisfying*

$${r}_{k}^{\left(ij\right)}={r}_{k-1}^{\left(ij\right)}+{J}_{k-1,k}^{\left(ij\right)}{\Pi}_{k}^{\left(j\right)-1}{J}_{k}^{\left(j\right)T}+{J}_{k}^{\left(i\right)}{\Pi}_{k}^{\left(i\right)-1}{J}_{k}^{\left(ji\right)T},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 1;\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}{r}_{0}^{\left(ij\right)}=0$$

(24)

*where ${J}_{k-1,k}^{\left(ij\right)}=E\left[{O}_{k-1}^{\left(i\right)}{\mu}_{k}^{\left(j\right)T}\right]$ are given by*

$$\begin{array}{c}{\displaystyle {J}_{k-1,k}^{\left(ij\right)}=\left({r}_{k-1}^{\left(i\right)}-{r}_{k-1}^{\left(ij\right)}\right){\overline{\mathcal{H}}}_{{A}_{k}}^{\left(j\right)T}+\sum _{h=1}^{(k-1)\wedge 2}{J}_{k-h}^{\left(i\right)}{\Pi}_{k-h}^{\left(i\right)-1}{\mathcal{W}}_{k,k-h}^{\left(ji\right)T}-\sum _{h=1}^{(k-1)\wedge 2}{J}_{k-1,k-h}^{\left(ij\right)}{\Pi}_{k-h}^{\left(j\right)-1}{\mathcal{W}}_{k,k-h}^{\left(j\right)T}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 2}\hfill \\ {J}_{0,1}^{\left(ij\right)}=0\hfill \end{array}$$

(25)

*and ${J}_{k,s}^{\left(ij\right)}=E\left[{O}_{k}^{\left(i\right)}{\mu}_{s}^{\left(j\right)T}\right],$ for $s=k-1,k$, satisfy*

$${J}_{k,s}^{\left(ij\right)}={J}_{k-1,s}^{\left(ij\right)}+{J}_{k}^{\left(i\right)}{\Pi}_{k}^{\left(i\right)-1}{\Pi}_{k,s}^{\left(ij\right)},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 2;\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}{J}_{1}^{\left(ij\right)}={J}_{1}^{\left(i\right)}{\Pi}_{1}^{\left(i\right)-1}{\Pi}_{1}^{\left(ij\right)}$$

(26)

*The innovation cross-covariance matrices ${\Pi}_{k}^{\left(ij\right)}=E\left[{\mu}_{k}^{\left(i\right)}{\mu}_{k}^{\left(j\right)T}\right]$ are obtained from*

$$\begin{array}{c}{\Pi}_{k}^{\left(ij\right)}={\Sigma}_{k}^{{y}^{\left(ij\right)}}-{\overline{\mathcal{H}}}_{{A}_{k}}^{\left(i\right)}\left({\overline{\mathcal{H}}}_{{B}_{k}}^{\left(j\right)T}-{J}_{k}^{\left(j\right)}-{J}_{k-1,k}^{\left(ij\right)}\right)\hfill \\ {\displaystyle \phantom{{\Pi}_{k}^{\left(ij\right)}=}-\sum _{h=1}^{(k-1)\wedge 2}{\mathcal{W}}_{k,k-h}^{\left(ij\right)}{\Pi}_{k-h}^{\left(j\right)-1}{\left({\overline{\mathcal{H}}}_{{A}_{k}}^{\left(j\right)}{J}_{k-h}^{\left(j\right)}+{\mathcal{W}}_{k,k-h}^{\left(j\right)}\right)}^{T}-\sum _{h=1}^{(k-1)\wedge 2}{\mathcal{W}}_{k,k-h}^{\left(i\right)}{\Pi}_{k-h}^{\left(i\right)-1}{\Pi}_{k-h,k}^{\left(ij\right)}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 2}\hfill \\ {\Pi}_{1}^{\left(ij\right)}={\Sigma}_{1}^{{y}^{\left(ij\right)}}\hfill \end{array}$$

(27)

*where ${\Pi}_{k,s}^{\left(ij\right)}=E\left[{\mu}_{k}^{\left(i\right)}{\mu}_{s}^{\left(j\right)T}\right],\phantom{\rule{4pt}{0ex}}s=k-2,k-1$, are given by*

$${\Pi}_{k,s}^{\left(ij\right)}={\overline{\mathcal{H}}}_{{A}_{k}}^{\left(i\right)}\left({J}_{s}^{\left(j\right)}-{J}_{k-1,s}^{\left(ij\right)}\right)+{\mathcal{W}}_{k,s}^{\left(ij\right)}-\sum _{h=1}^{(k-1)\wedge 2}\phantom{\rule{-2.84544pt}{0ex}}{\mathcal{W}}_{k,k-h}^{\left(i\right)}{\Pi}_{k-h}^{\left(i\right)-1}{\Pi}_{k-h,s}^{\left(ij\right)}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 2$$

(28)

*The coefficients ${\mathcal{W}}_{k,k-h}^{\left(ij\right)}=E\left[{w}_{k}^{\left(i\right)}{\mu}_{k-h}^{\left(j\right)T}\right],\phantom{\rule{4pt}{0ex}}h=1,2$, involved in the above expressions, are computed by*

$$\begin{array}{c}{\mathcal{W}}_{k,k-1}^{\left(ij\right)}={\Sigma}_{k,k-1}^{{y}^{\left(ij\right)}}-{\overline{\mathcal{H}}}_{{A}_{k}}^{\left(i\right)}{\overline{\mathcal{H}}}_{{B}_{k-1}}^{\left(j\right)T}-{\mathcal{W}}_{k,k-2}^{\left(ij\right)}{\Pi}_{k-2}^{\left(j\right)-1}{\left({\overline{\mathcal{H}}}_{{A}_{k-1}}^{\left(j\right)}{J}_{k-2}^{\left(j\right)}+{\mathcal{W}}_{k-1,k-2}^{\left(j\right)}\right)}^{T},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 3\hfill \\ {\mathcal{W}}_{3,2}^{\left(ij\right)}={\Sigma}_{2,1}^{{y}^{\left(ij\right)}}-{\overline{\mathcal{H}}}_{{A}_{2}}^{\left(i\right)}{\overline{\mathcal{H}}}_{{B}_{1}}^{\left(j\right)T}\hfill \end{array}$$

(29)

$$\phantom{\rule{-85.35826pt}{0ex}}{\mathcal{W}}_{k,k-2}^{\left(ij\right)}={\overline{\gamma}}_{k}^{\left(i\right)}\left(1-{\overline{\gamma}}_{k-2}^{\left(j\right)}\right){R}_{k-1,k-2}^{\left(ij\right)}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 4;\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}{\mathcal{W}}_{3,1}^{\left(ij\right)}={\overline{\gamma}}_{3}^{\left(i\right)}{R}_{2,1}^{\left(ij\right)}$$

(30)

Finally, the matrices ${\Sigma}_{k,s}^{{y}^{\left(ij\right)}}$, and ${\overline{\mathcal{H}}}_{{A}_{s}}^{\left(l\right)}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}{\overline{\mathcal{H}}}_{{B}_{s}}^{\left(l\right)},\phantom{\rule{4pt}{0ex}}s=k-1,k,\phantom{\rule{4pt}{0ex}}l=i,j$, are given in Equations (3) and (7), respectively.

As it has been discussed previously, Expression (21) is immediately derived from Equation (20), while Equation (22) is obtained from ${P}_{k/k}^{\left(D\right)}=E\left[{x}_{k}{x}_{k}^{T}\right]-E\left[{\widehat{x}}_{k/k}^{\left(D\right)}{\widehat{x}}_{k/k}^{\left(D\right)T}\right]$, using *(H1)* and (21). Moreover, Equation (23) for ${\widehat{\Sigma}}_{k/k}^{\left(ij\right)}$ is directly obtained using Equation (8) for the local filters and defining ${r}_{k}^{\left(ij\right)}=E\left[{O}_{k}^{\left(i\right)}{O}_{k}^{\left(j\right)T}\right]$.

Next, we derive the recursive formulas to obtain the matrices ${r}_{k}^{\left(ij\right)}$, which clearly satisfy Equation (24) by simply using Equation (10) and defining ${J}_{s,k}^{\left(ij\right)}=E\left[{O}_{s}^{\left(i\right)}{\mu}_{k}^{\left(j\right)T}\right],\phantom{\rule{4pt}{0ex}}s=k-1,k$.

For subsequent derivations, the following expression of the one-stage predictor of ${y}_{k}^{\left(j\right)}$ based on the observations of sensor *i* will be used; this expression is obtained from Equation (5), taking into account, as proven in Theorem 1, that ${\widehat{x}}_{k/s}^{\left(i\right)}={A}_{k}{O}_{s}^{\left(i\right)},\phantom{\rule{4pt}{0ex}}s=k-1,k$, and defining ${\mathcal{W}}_{k,k-h}^{\left(ji\right)}=E\left[{w}_{k}^{\left(j\right)}{\mu}_{k-h}^{\left(i\right)T}\right],\phantom{\rule{4pt}{0ex}}h=1,2$:

$${\widehat{y}}_{k/k-1}^{(j/i)}={\overline{\mathcal{H}}}_{{A}_{k}}^{\left(j\right)}{O}_{k-1}^{\left(i\right)}+\sum _{h=1}^{(k-1)\wedge 2}{\mathcal{W}}_{k,k-h}^{\left(ji\right)}{\Pi}_{k-h}^{\left(i\right)-1}{\mu}_{k-h}^{\left(i\right)}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 2$$

(31)

As Expression (31) is a generalization of Equation (18), hereafter we will also refer to it for the local predictors ${\widehat{y}}_{k/k-1}^{\left(i\right)},\phantom{\rule{4pt}{0ex}}k\ge 2$.

By applying the OPL, is clear that $E\left[{O}_{k-1}^{\left(i\right)}{y}_{k}^{\left(j\right)T}\right]=E\left[{O}_{k-1}^{\left(i\right)}{\widehat{y}}_{k/k-1}^{(j/i)T}\right]$ and, consequently, we can rewrite ${J}_{k-1,k}^{\left(ij\right)}=E\left[{O}_{k-1}^{\left(i\right)}{\left({\widehat{y}}_{k/k-1}^{(j/i)}-{\widehat{y}}_{k/k-1}^{\left(j\right)}\right)}^{T}\right]$; then, using Equation (31) for both predictors, Expression (25) is easily obtained. Also, Equation (26) for ${J}_{k,s}^{\left(ij\right)},\phantom{\rule{4pt}{0ex}}s=k-1,k$, is immediate from Equation (10), by simply defining ${\Pi}_{k,s}^{\left(ij\right)}=E\left[{\mu}_{k}^{\left(i\right)}{\mu}_{s}^{\left(j\right)T}\right]$.

To obtain Equation (27), firstly we apply the OPL to express ${\Pi}_{k}^{\left(ij\right)}={\Sigma}_{k}^{{y}^{\left(ij\right)}}-E\left[{\widehat{y}}_{k/k-1}^{(i/j)}{\widehat{y}}_{k/k-1}^{\left(j\right)T}\right]-E\left[{\widehat{y}}_{k/k-1}^{\left(i\right)}{\mu}_{k}^{\left(j\right)T}\right]$. Then, using Equation (31) for ${\widehat{y}}_{k/k-1}^{(i/j)}$ and ${\widehat{y}}_{k/k-1}^{\left(i\right)}\phantom{\rule{4pt}{0ex}}$, and definitions of ${J}_{k-1,k}^{\left(ij\right)}$ and ${\Pi}_{k-h,k}^{\left(ij\right)}\phantom{\rule{4pt}{0ex}}$, we have

$$\begin{array}{c}{\displaystyle E\left[{\widehat{y}}_{k/k-1}^{(i/j)}{\widehat{y}}_{k/k-1}^{\left(j\right)T}\right]={\overline{\mathcal{H}}}_{{A}_{k}}^{\left(i\right)}E\left[{O}_{k-1}^{\left(j\right)}{\widehat{y}}_{k/k-1}^{\left(j\right)T}\right]+\sum _{h=1}^{(k-1)\wedge 2}{\mathcal{W}}_{k,k-h}^{\left(ij\right)}{\Pi}_{k-h}^{\left(j\right)-1}E\left[{\mu}_{k-h}^{\left(j\right)}{\widehat{y}}_{k/k-1}^{\left(j\right)T}\right],}\hfill \\ {\displaystyle E\left[{\widehat{y}}_{k/k-1}^{\left(i\right)}{\mu}_{k}^{\left(j\right)T}\right]={\overline{\mathcal{H}}}_{{A}_{k}}^{\left(i\right)}{J}_{k-1,k}^{\left(ij\right)}+\sum _{h=1}^{(k-1)\wedge 2}{\mathcal{W}}_{k,k-h}^{\left(i\right)}{\Pi}_{k-h}^{\left(i\right)-1}{\Pi}_{k-h,k}^{\left(ij\right)}\phantom{\rule{4pt}{0ex}}}\hfill \end{array}$$

and Equation (27) is obtained taking into account that $E\left[{O}_{k-1}^{\left(j\right)}{\widehat{y}}_{k/k-1}^{\left(j\right)T}\right]={\overline{\mathcal{H}}}_{{B}_{k}}^{\left(j\right)T}-{J}_{k}^{\left(j\right)}$ and $E\left[{\widehat{y}}_{k/k-1}^{\left(j\right)}{\mu}_{k-h}^{\left(j\right)T}\right]={\overline{\mathcal{H}}}_{{A}_{k}}^{\left(j\right)}{J}_{k-h}^{\left(j\right)}+{\mathcal{W}}_{k,k-h}^{\left(j\right)}\phantom{\rule{4pt}{0ex}}$, as it has been derived in the proof of Theorem 1.

Next, Expression (28) for ${\Pi}_{k,s}^{\left(ij\right)}=E\left[{y}_{k}^{\left(i\right)}{\mu}_{s}^{\left(j\right)T}\right]-E\left[{\widehat{y}}_{k/k-1}^{\left(i\right)}{\mu}_{s}^{\left(j\right)T}\right],$ with $s=k-2,k-1$, is obtained from $E\left[{y}_{k}^{\left(i\right)}{\mu}_{s}^{\left(j\right)T}\right]={\mathcal{H}}_{{A}_{k}}^{\left(i\right)}{J}_{s}^{\left(j\right)}+{\mathcal{W}}_{k,s}^{\left(ij\right)}$, and using Equation (31) in $E\left[{\widehat{y}}_{k/k-1}^{\left(i\right)}{\mu}_{s}^{\left(j\right)T}\right]$.

Finally, the reasoning for obtaining the coefficients ${\mathcal{W}}_{k,k-h}^{\left(ij\right)}=E\left[{w}_{k}^{\left(i\right)}{\mu}_{k-h}^{\left(j\right)T}\right],\phantom{\rule{4pt}{0ex}}h=1,2,$ is also similar to that of ${\mathcal{W}}_{k,k-h}^{\left(i\right)}$ in Theorem 1, so it is omitted. Then the proof of Theorem 2 is completed.

In the centralized fusion filtering, the observations of the different sensors are jointly processed at each sampling time to yield the optimal filter of the signal ${x}_{k}$, which will be denoted by ${\widehat{x}}_{k/k}^{\left(C\right)}$. To carry out this process, at each time $k\ge 1$ we will work with the vector constituted by the observations of all sensors, ${y}_{k}={\left({y}_{k}^{\left(1\right)T},\dots {y}_{k}^{\left(m\right)T}\right)}^{T}$, which, from Equation (2), can be expressed by

$${y}_{k}=\left(I-{\Gamma}_{k}\right){z}_{k}+{\Gamma}_{k}{z}_{k-1},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 2;\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}{y}_{1}={z}_{1}$$

(32)

where ${z}_{k}={\left({z}_{k}^{\left(1\right)T},\dots {z}_{k}^{\left(m\right)T}\right)}^{T}$ is the vector constituted by the sensor measured outputs given in Equation (1), and ${\Gamma}_{k}=Diag\left({\gamma}_{k}^{\left(1\right)},\dots ,{\gamma}_{k}^{\left(m\right)}\right)\otimes {I}_{{n}_{z}}$. Let us note that, analogously to the sensor measured outputs ${z}_{k}^{\left(i\right)}$, the stacked vector ${z}_{k}$ is a noisy linear function of the signal, with random parameter matrices defined by ${H}_{k}={\left({H}_{k}^{\left(1\right)T},\dots {H}_{k}^{\left(m\right)T}\right)}^{T}$, and noise ${v}_{k}={\left({v}_{k}^{\left(1\right)T},\dots {v}_{k}^{\left(m\right)T}\right)}^{T}$:

$${z}_{k}={H}_{k}{x}_{k}+{v}_{k}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 1$$

(33)

The processes involved in Equations (32) and (33) satisfy the following properties, which are immediately derived from the model hypotheses *(H1)–(H5)*:

*(P1)*$\left\{{H}_{k};\phantom{\rule{4pt}{0ex}}k\ge 1\right\}$ is a sequence of independent random parameter matrices whose entries have known means and second-order moments.*(P2)*The noise $\left\{{v}_{k};\phantom{\rule{4pt}{0ex}}k\ge 1\right\}$ is a zero-mean sequence with known second-order moments defined by the matrices ${R}_{k,s}\equiv {\left({R}_{k,s}^{\left(ij\right)}\right)}_{i,j=1,\dots ,m}$.*(P3)*The matrices $\left\{{\Gamma}_{k};\phantom{\rule{4pt}{0ex}}k\ge 2\right\}$ have known means, ${\overline{\Gamma}}_{k}\equiv E\left[{\Gamma}_{k}\right]=Diag\left({\overline{\gamma}}_{k}^{\left(1\right)},\dots ,{\overline{\gamma}}_{k}^{\left(m\right)}\right)\otimes {I}_{{n}_{z}}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}k\ge 2$, and ${\Gamma}_{k}$ and ${\Gamma}_{s}$ are independent for $|k-s|\ge 2$.*(P4)*The processes $\{{x}_{k};\phantom{\rule{4pt}{0ex}}k\ge 1\},\phantom{\rule{4pt}{0ex}}\{{H}_{k};\phantom{\rule{4pt}{0ex}}k\ge 1\},\phantom{\rule{4pt}{0ex}}\{{v}_{k};\phantom{\rule{4pt}{0ex}}k\ge 1\}$ and $\left\{{\Gamma}_{k};\phantom{\rule{4pt}{0ex}}k\ge 2\right\}$ are mutually independent.

In view of Equations (32) and (33) and the above properties, the study of the LS linear filtering problem based on the stacked observations, $\{{y}_{k};\phantom{\rule{4pt}{0ex}}k\ge 1\}$, is completely similar to that of the local filtering problem carried out in Section 3. Therefore, the centralized filtering algorithm described in the following theorem is derived by an analogous reasoning to that used in Theorem 1 and, hence, its proof is omitted.

*The centralized LS linear filter, ${\widehat{x}}_{k/k}^{\left(C\right)},$ and the corresponding error covariance matrix, ${P}_{k/k}^{\left(C\right)}$, are given by*

$$\begin{array}{c}{\widehat{x}}_{k/k}^{\left(C\right)}={A}_{k}{O}_{k}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 1\hfill \end{array}$$

*and*

$${P}_{k/k}^{\left(C\right)}={A}_{k}{\left({B}_{k}-{A}_{k}{r}_{k}\right)}^{T},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 1$$

*where the vectors ${O}_{k}$ and the matrices ${r}_{k}=E\left[{O}_{k}{O}_{k}^{T}\right]$ are recursively obtained from*

$${O}_{k}={O}_{k-1}+{J}_{k}{\Pi}_{k}^{-1}{\mu}_{k}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 1;\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}{O}_{0}=0$$

$${r}_{k}={r}_{k-1}+{J}_{k}{\Pi}_{k}^{-1}{J}_{k}^{T},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 1;\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}{r}_{0}=0$$

*The matrices ${J}_{k}=E\left[{O}_{k}{\mu}_{k}^{T}\right]$ satisfy*

$$J}_{k}={\overline{\mathcal{H}}}_{{B}_{k}}^{T}-{r}_{k-1}{\overline{\mathcal{H}}}_{{A}_{k}}^{T}-\sum _{h=1}^{(k-1)\wedge 2}{J}_{k-h}{\Pi}_{k-h}^{-1}{\mathcal{W}}_{k,k-h}^{T}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 2;\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}{J}_{1}={\overline{\mathcal{H}}}_{{B}_{1}}^{T$$

*The innovations, ${\mu}_{k}$, and their covariance matrices, ${\Pi}_{k}$, are given by*

$$\mu}_{k}={y}_{k}-{\overline{\mathcal{H}}}_{{A}_{k}}{O}_{k-1}-\sum _{h=1}^{(k-1)\wedge 2}{\mathcal{W}}_{k,k-h}{\Pi}_{k-h}^{-1}{\mu}_{k-h}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 2;\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}{\mu}_{1}={y}_{1$$

*and*

$$\begin{array}{c}{\displaystyle {\Pi}_{k}={\Sigma}_{k}^{y}-{\overline{\mathcal{H}}}_{{A}_{k}}\left({\overline{\mathcal{H}}}_{{B}_{k}}^{T}-{J}_{k}\right)-\sum _{h=1}^{(k-1)\wedge 2}{\mathcal{W}}_{k,k-h}{\Pi}_{k-h}^{-1}{\left({\overline{\mathcal{H}}}_{{A}_{k}}{J}_{k-h}+{\mathcal{W}}_{k,k-h}\right)}^{T},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 2;\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}{\Pi}_{1}={\Sigma}_{1}^{y}}\hfill \end{array}$$

*and the coefficients ${\mathcal{W}}_{k,k-h}=E\left[{w}_{k}{\mu}_{k-h}^{T}\right],$$h=1,2$, verify*

$$\begin{array}{c}{\mathcal{W}}_{k,k-1}={\Sigma}_{k,k-1}^{y}-{\overline{\mathcal{H}}}_{{A}_{k}}{\overline{\mathcal{H}}}_{{B}_{k-1}}^{T}-{\mathcal{W}}_{k,k-2}{\Pi}_{k-2}^{-1}{\left({\overline{\mathcal{H}}}_{{A}_{k-1}}{J}_{k-2}+{\mathcal{W}}_{k-1,k-2}\right)}^{T},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 3\hfill \\ {\mathcal{W}}_{2,1}={\Sigma}_{2,1}^{y}-{\overline{\mathcal{H}}}_{{A}_{2}}{\overline{\mathcal{H}}}_{{B}_{1}}^{T}\hfill \end{array}$$

$$\phantom{\rule{-96.73918pt}{0ex}}{\mathcal{W}}_{k,k-2}={\overline{\Gamma}}_{k}{R}_{k-1,k-2}(I-{\overline{\Gamma}}_{k-2}),\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 4;\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}{\mathcal{W}}_{3,1}={\overline{\Gamma}}_{3}{R}_{2,1}$$

In the above formulas, the matrices ${\Sigma}_{k}^{y}$ and ${\Sigma}_{k,k-1}^{y}$ are computed by ${\Sigma}_{k,s}^{y}={\left({\Sigma}_{k,s}^{{y}^{\left(ij\right)}}\right)}_{i,j=1,\dots ,m}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}s=k,k-1$, with ${\Sigma}_{k,s}^{{y}^{\left(ij\right)}}$ given in Equation (3), and ${\overline{\mathcal{H}}}_{{C}_{k}}={\left({\overline{\mathcal{H}}}_{{C}_{k}}^{\left(1\right)T},\dots ,{\overline{\mathcal{H}}}_{{C}_{k}}^{\left(m\right)T}\right)}^{T},\phantom{\rule{4pt}{0ex}}C=A,B$, with ${\overline{\mathcal{H}}}_{{C}_{k}}^{\left(i\right)}$ defined in Equation (7).

This section is devoted to analyze the effectiveness of the proposed distributed and centralized filtering algorithms by a simulation example. Let us consider a zero-mean scalar signal process, $\{{x}_{k};\phantom{\rule{0.166667em}{0ex}}k\ge 1\}$, with autocovariance function $E\left[{x}_{k}{x}_{s}\right]=1.025641\times 0.{95}^{k-s},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}s\le k,$ which is factorizable according to *(H1)* just taking, for example, ${A}_{k}=1.025641\times 0.{95}^{k}$ and ${B}_{k}=0.{95}^{-k}.$

The measured outputs of this signal, which are provided by four different sensors, are described by Equation (1):

$${z}_{k}^{\left(i\right)}={H}_{k}^{\left(i\right)}{x}_{k}+{v}_{k}^{\left(i\right)},\phantom{\rule{1.em}{0ex}}k\ge 1,\phantom{\rule{1.em}{0ex}}i=1,2,3,4$$

where the processes $\{{H}_{k}^{\left(i\right)};\phantom{\rule{0.166667em}{0ex}}k\ge 1\}$ and $\{{v}_{k}^{\left(i\right)};\phantom{\rule{0.166667em}{0ex}}k\ge 1\}$, $i=1,2,3,4,$ are defined as follows:

- ${H}_{k}^{\left(1\right)}=0.8{\lambda}_{k}^{\left(1\right)}$, ${H}_{k}^{\left(2\right)}={\lambda}_{k}^{\left(2\right)}\left(0.75+0.95{\epsilon}_{k}\right)$, ${H}_{k}^{\left(3\right)}=0.8{\lambda}_{k}^{\left(3\right)}$ and ${H}_{k}^{\left(4\right)}=0.75{\lambda}_{k}^{\left(4\right)}$, where $\{{\epsilon}_{k};\phantom{\rule{4pt}{0ex}}k\ge 1\}$ is a zero-mean Gaussian white process with unit variance, and $\{{\lambda}_{k}^{\left(i\right)};\phantom{\rule{4pt}{0ex}}k\ge 1\}$, $i=1,2,3,4$, are white processes with the following time-invariant probability distributions:
- – For $i=1,2$, ${\lambda}_{k}^{\left(i\right)}$ are Bernoulli random variables with $P[{\lambda}_{k}^{\left(i\right)}=1]=0.6$.
- – ${\lambda}_{k}^{\left(3\right)}$ is uniformly distributed over $[0.1,0.9]$.
- – $P[{\lambda}_{k}^{\left(4\right)}=0]=0.3,\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}P[{\lambda}_{k}^{\left(4\right)}=0.5]=0.3,\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}P[{\lambda}_{k}^{\left(4\right)}=1]=0.4.$

It is also assumed that the sequences $\{{\epsilon}_{k};\phantom{\rule{4pt}{0ex}}k\ge 1\}$ and $\{{\lambda}_{k}^{\left(i\right)};\phantom{\rule{4pt}{0ex}}k\ge 1\},\phantom{\rule{4pt}{0ex}}i=1,2,3,4,$ are mutually independent. - The additive noises are defined as ${v}_{k}^{\left(i\right)}={c}_{i}({\eta}_{k}+{\eta}_{k+1}),\phantom{\rule{4pt}{0ex}}i=1,2,3,4$, where ${c}_{1}={c}_{4}=0.75$, ${c}_{2}=1$, ${c}_{3}=0.5$, and $\{{\eta}_{k};\phantom{\rule{4pt}{0ex}}k\ge 1\}$ is a zero-mean Gaussian white process with unit variance.

Note that the sequences of random variables $\{{H}_{k}^{\left(i\right)};\phantom{\rule{0.166667em}{0ex}}k\ge 1\}$, $i=1,2,3,4,$ model different types of uncertainty in the measured outputs: *missing measurements* in sensor 1; both *missing measurements and multiplicative noise* in sensor 2; and *continuous and discrete gain degradation* in sensors 3 and 4, respectively. Moreover, it is clear that the additive noises $\{{v}_{k}^{\left(i\right)};\phantom{\rule{0.166667em}{0ex}}k\ge 1\}$, $i=1,2,3,4,$ are only correlated at the same and consecutive time instants, with ${R}_{k}^{\left(ij\right)}=2{c}_{i}{c}_{j}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}{R}_{k,k-1}^{\left(ij\right)}={c}_{i}{c}_{j}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}i,j=1,2,3,4.$

Next, according to our theoretical observation model, it is supposed that, at any sampling time $k\ge 2$, the data transmissions are subject to random one-step delays with different rates and such delays are correlated at consecutive sampling times. More precisely, let us assume that the available measurements ${y}_{k}^{\left(i\right)}$ are given by Equation (2):

$${y}_{k}^{\left(i\right)}=(1-{\gamma}_{k}^{\left(i\right)}){z}_{k}^{\left(i\right)}+{\gamma}_{k}^{\left(i\right)}{z}_{k-1}^{\left(i\right)}\phantom{\rule{4pt}{0ex}},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}k\ge 2,\phantom{\rule{1.em}{0ex}}i=1,2,3,4$$

where the variables ${\gamma}_{k}^{\left(i\right)}$ modeling this type of correlated random delays are defined using three independent sequences of independent Bernoulli random variables, $\{{\theta}_{k}^{\left(i\right)};\phantom{\rule{4pt}{0ex}}k\ge 1\}$, $i=1,2,3$, with constant probabilities, $P[{\theta}_{k}^{\left(i\right)}=1]={\overline{\theta}}^{\left(i\right)}$, for all $k\ge 1$; specifically, for $i=1,2,3,$ ${\gamma}_{k}^{\left(i\right)}={\theta}_{k+1}^{\left(i\right)}(1-{\theta}_{k}^{\left(i\right)}),$ and ${\gamma}_{k}^{\left(4\right)}={\theta}_{k}^{\left(1\right)}(1-{\theta}_{k+1}^{\left(1\right)}).$

It is clear that the sensor delay probabilities are time-invariant: ${\overline{\gamma}}^{\left(i\right)}={\overline{\theta}}^{\left(i\right)}(1-{\overline{\theta}}^{\left(i\right)})$, for $i=1,2,3$, and ${\overline{\gamma}}^{\left(4\right)}={\overline{\gamma}}^{\left(1\right)}$. Moreover, the independence of the sequences $\{{\theta}_{k}^{\left(i\right)}\phantom{\rule{0.166667em}{0ex}};\phantom{\rule{0.166667em}{0ex}}k\ge 1\}$, $i=1,2,3$, together with the independence of the variables in each sequence, guarantee that the random variables ${\gamma}_{k}^{\left(i\right)}$ and ${\gamma}_{s}^{\left(j\right)}$ are independent if $|k-s|\ge 2$, for any $i,j=1,2,3,4$. Also, it is clear that, at each sensor, the variables $\{{\gamma}_{k}^{\left(i\right)}\phantom{\rule{0.166667em}{0ex}};\phantom{\rule{4pt}{0ex}}k\ge 2\}$ are correlated at consecutive sampling times and ${\overline{\gamma}}_{k,s}^{\left(ii\right)}=0$, for $i=1,2,3,4$ and $|k-s|=1$. Finally, we have that $\{{\gamma}_{k}^{\left(4\right)};\phantom{\rule{4pt}{0ex}}k\ge 2\}$ is independent of $\{{\gamma}_{k}^{\left(i\right)}\phantom{\rule{0.166667em}{0ex}};\phantom{\rule{4pt}{0ex}}k\ge 2\}$, $i=2,3$, but correlated with $\{{\gamma}_{k}^{\left(1\right)}\phantom{\rule{0.166667em}{0ex}};\phantom{\rule{4pt}{0ex}}k\ge 2\}$ at consecutive sampling times, with ${\overline{\gamma}}_{k,k-1}^{\left(14\right)}={\overline{\gamma}}^{\left(1\right)}{\overline{\theta}}^{\left(1\right)}$ and ${\overline{\gamma}}_{k,k-1}^{\left(41\right)}={\overline{\gamma}}^{\left(1\right)}(1-{\overline{\theta}}^{\left(1\right)})$.

Let us observe that, for each sensor $i=1,2,3,4$, if ${\gamma}_{k}^{\left(i\right)}=1$, then ${\gamma}_{k+1}^{\left(i\right)}=0$; this fact guarantees that, when the measurement at time *k* is delayed, the available measurement at time $k+1$ is well-timed. Therefore, this correlation model avoids the possibility of two consecutive delayed observations at the same sensor. Table 1 shows an example of data transmission in sensor 1 when ${\overline{\theta}}^{\left(1\right)}=0.5$.

Note that if ${\gamma}_{k}^{\left(1\right)}=0$, there is no delay at time *k*; *i.e.*, the output ${z}_{k}^{\left(1\right)}$ is received on time by the fusion center and ${y}_{k}^{\left(1\right)}={z}_{k}^{\left(1\right)}$. If ${\gamma}_{k}^{\left(1\right)}=1$, the packet received at time *k* is the output at $k-1$; *i.e.*, there is one-step delay and the processed observation is ${y}_{k}^{\left(1\right)}={z}_{k-1}^{\left(1\right)}$. From Table 1, we see that ${z}_{4}^{\left(1\right)},\phantom{\rule{4pt}{0ex}}{z}_{7}^{\left(1\right)}$ and ${z}_{11}^{\left(1\right)}$ are are lost, and ${z}_{3}^{\left(1\right)},\phantom{\rule{4pt}{0ex}}{z}_{6}^{\left(1\right)}$ and ${z}_{10}^{\left(1\right)}$ are re-received.

To illustrate the feasibility and analyze the effectiveness of the proposed estimators, the algorithms were implemented in MATLAB, and a hundred iterations were run. In order to measure the estimation accuracy, the error variances of both distributed and centralized fusion estimators were calculated for several values of the delay probabilities at the different sensors, obtained from several values of ${\overline{\theta}}^{\left(i\right)}$. Let us observe that the delay probabilities, ${\overline{\gamma}}^{\left(i\right)}={\overline{\theta}}^{\left(i\right)}(1-{\overline{\theta}}^{\left(i\right)})$, for $i=1,2,3$, are the same if $1-{\overline{\theta}}^{\left(i\right)}$ is used instead of ${\overline{\theta}}^{\left(i\right)}$; for this reason, only the case ${\overline{\theta}}^{\left(i\right)}\le 0.5$ will be considered here.

First, the error variances of the local, distributed and centralized filters will be compared considering the same delay probabilities for the four sensors. Figure 1 displays these error variances when ${\overline{\theta}}^{\left(i\right)}=0.3$, $i=1,2,3$, which means that the delay probability at every sensor is ${\overline{\gamma}}^{\left(i\right)}=0.21$. This figure shows that the error variances of the distributed fusion filtering estimator are lower than those of every local estimators, but slightly greater than those of the centralized one. However, this slight difference is compensated by the fact that the distributed fusion structure reduces the computational cost and has better robustness and fault tolerance. Analogous results are obtained for other values of ${\overline{\theta}}^{\left(i\right)}$. Actually, Figure 2 displays the distributed and centralized filtering error variances for ${\overline{\theta}}^{\left(i\right)}=0.1,\phantom{\rule{0.166667em}{0ex}}0.2,\phantom{\rule{0.166667em}{0ex}}0.3,\phantom{\rule{0.166667em}{0ex}}0.4,\phantom{\rule{0.166667em}{0ex}}0.5$, which lead to the delay probabilities ${\overline{\gamma}}^{\left(i\right)}=0.09,\phantom{\rule{0.166667em}{0ex}}0.16,\phantom{\rule{0.166667em}{0ex}}0.21,\phantom{\rule{0.166667em}{0ex}}0.24,\phantom{\rule{0.166667em}{0ex}}0.25$, respectively. In this figure, both graphs (corresponding to the distributed and centralized fusion filters, respectively) show that the performance of the filters is poorer as ${\overline{\theta}}^{\left(i\right)}$ increases. This fact was expected as the increase of ${\overline{\theta}}^{\left(i\right)}$ yields a rise in the delay probabilities. This figure also confirms that both methods, distributed and centralized, have approximately the same accuracy, thus corroborating the previous results.

Local, distributed and centralized filtering error variances for ${\overline{\theta}}^{\left(i\right)}=0.3$, $i=1,2,3$.

(**a**) Distributed filtering error variances and (**b**) centralized filtering error variances for different values of ${\overline{\theta}}^{\left(i\right)}$.

In order to carry out a further discussion on the effects of the sensor random delays, Figure 3 shows a comparison of the filtering error variances in the following cases:

- Case I: error variances
*versus*${\overline{\theta}}^{\left(1\right)}$, when ${\overline{\theta}}^{\left(2\right)}={\overline{\theta}}^{\left(3\right)}=0.5$. In this case, as mentioned above, the values ${\overline{\theta}}^{\left(1\right)}=0.1,\phantom{\rule{0.166667em}{0ex}}0.2,\phantom{\rule{0.166667em}{0ex}}0.3,\phantom{\rule{0.166667em}{0ex}}0.4,\phantom{\rule{0.166667em}{0ex}}0.5$, lead to the values ${\overline{\gamma}}^{\left(1\right)}={\overline{\gamma}}^{\left(4\right)}=0.09,\phantom{\rule{0.166667em}{0ex}}0.16,\phantom{\rule{0.166667em}{0ex}}0.21,\phantom{\rule{0.166667em}{0ex}}0.24,\phantom{\rule{0.166667em}{0ex}}0.25$, respectively, for the delay probabilities of sensors 1 and 4, whereas the delay probabilities of sensors 2 and 3 are constant and equal to 0.25. - Case II: error variances
*versus*${\overline{\theta}}^{\left(1\right)}$, when ${\overline{\theta}}^{\left(2\right)}={\overline{\theta}}^{\left(1\right)}$ and ${\overline{\theta}}^{\left(3\right)}=0.5$. Varying ${\overline{\theta}}^{\left(1\right)}$ as in Case I, the delay probabilities of sensors 1, 2 and 4 are equal and take the aforementioned values, whereas the delay probability of sensor 3 is constant and equal to 0.25. - Case III: error variances
*versus*${\overline{\theta}}^{\left(1\right)}$, when ${\overline{\theta}}^{\left(2\right)}={\overline{\theta}}^{\left(3\right)}={\overline{\theta}}^{\left(1\right)}$. Now, as in Figure 2, the delay probabilities of the four sensors are equal, and they all take the aforementioned values.

(**a**) Distributed filtering error variances and (**b**) centralized filtering error variances at $k=100$, *versus*
${\overline{\theta}}^{\left(1\right)}$.

Since the behavior of the error variances is analogous in all the iterations, only the results of a specific iteration ($k=100$) are displayed in Figure 3, which shows that the performance of the distributed and centralized estimators is indeed influenced by the probability ${\overline{\theta}}^{\left(i\right)}$ and, as expected, better estimations are obtained as ${\overline{\theta}}^{\left(i\right)}$ becomes smaller, due to the fact that the delay probabilities, ${\overline{\gamma}}^{\left(i\right)}$, decrease with ${\overline{\theta}}^{\left(i\right)}$. Moreover, this figure shows that the error variances in Case III are less than those of Case II which, in turn, are lower than those of Case I. This is due to the fact that, while the delay probabilities of the four sensors are varied in Case III, only two and three sensors vary their delay probabilities in Cases I and II, respectively. Since the constant delay probabilities of the other sensors are assumed to take their greatest possible value, this figure confirms that the estimation accuracy improves as the delay probabilities decrease.

In this paper, distributed and centralized fusion filtering algorithms have been designed in multi-sensor systems from measured outputs with random parameter matrices and correlated noises, assuming correlated random delays in transmissions. The main outcomes and results can be summarized as follows:

*Information on the signal process:*our approach, based on covariance information, does not require the evolution model generating the signal process to design the proposed distributed and centralized filtering algorithms; nonetheless, they are also applicable to the conventional formulation using the state-space model.*Signal uncertain measured outputs:*random measurement matrices and cross-correlation between the different sensor noises are considered in the measured outputs, thus providing a unified framework to address different network-induced phenomena, such as missing measurements or sensor gain degradation, along with correlated measurement noises.*Random one-step transmission delays:*the fusion estimation problems are addressed assuming random one-step delays in the outputs transmission to the fusion center through the network communication channels; the delays have different characteristics at each sensor and they are assumed to be correlated and cross-correlated at consecutive sampling times. This correlation assumption covers many situations where the common assumption of independent delays is not realistic; for example, networked systems with stand-by sensors for the immediate replacement of a failed unit, thus avoiding the possibility of two successive delayed observations.*Fusion filtering algorithms:*firstly, recursive algorithms for the local LS linear signal filters based on the measured output data coming from each sensor have been designed by an innovation approach; the computational procedure of the local algorithms is very simple and suitable for online applications. After that, the matrix-weighted sum that minimizes the mean-squared estimation error is proposed as distributed fusion estimator. Also, using covariance information, a recursive centralized LS linear filtering algorithm, with analogous structure to that of the local algorithms, is proposed. The accuracy of the proposed fusion estimators, obtained under the LS optimality criterion, is measured by the error covariance matrices, which can be calculated offline as they do not depend on the current observed data set.*Simulations:*a numerical simulation example has illustrated the usefulness of the proposed algorithms for the estimation of a scalar signal. Error variance comparisons have shown that both distributed and centralized fusion filters outperform the local ones, as well as a slight superiority of the centralized fusion estimators over the distributed ones. The effects of the delays on the estimators performance have been also analyzed by the error variances. This example has also highlighted the applicability of the proposed algorithms to different multi-sensor systems with stochastic uncertainties, which can be dealt with using the observation model with random measurement matrices considered in this paper.

This research is supported by *Ministerio de Economía y Competitividad* and *Fondo Europeo de Desarrollo Regional FEDER* (grant no. MTM2014-52291-P).

The following abbreviations are used in this manuscript:

LS | Least-Squares |

OPL | Orthogonal Projection Lemma |

Author Contributions

All the authors contributed equally to this work. Raquel Caballero-Águila, Aurora Hermoso-Carazo and Josefa Linares-Pérez provided original ideas for the proposed model and collaborated in the derivation of the estimation algorithms; they participated equally in the design and analysis of the simulation results; and the paper was also written and reviewed cooperatively.

Conflicts of Interest

The authors declare no conflict of interest.

1. Li W., Wang Z., Wei G., Ma L., Hu J., Ding D. A Survey on multisensor fusion and consensus filtering for sensor networks. Discret. Dyn. Nat. Soc. 2015;2015:683701.

2. Ran C., Deng Z. Self-tuning weighted measurement fusion Kalman filtering algorithm. Comput. Stat. Data Anal. 2012;56:2112–2128. doi: 10.1016/j.csda.2012.01.001. [Cross Ref]

3. Feng J., Zeng M. Optimal distributed Kalman filtering fusion for a linear dynamic system with cross-correlated noises. Int. J. Syst. Sci. 2012;43:385–398. doi: 10.1080/00207721.2010.502601. [Cross Ref]

4. Yan L., Li X.R., Xia Y., Fu M. Optimal sequential and distributed fusion for state estimation in cross-correlated noise. Automatica. 2013;49:3607–3612. doi: 10.1016/j.automatica.2013.09.013. [Cross Ref]

5. Feng J., Wang Z., Zeng M. Distributed weighted robust Kalman filter fusion for uncertain systems with autocorrelated and cross-correlated noises. Inform. Fusion. 2013;14:78–86. doi: 10.1016/j.inffus.2011.09.004. [Cross Ref]

6. Hu J., Wang Z., Chen D., Alsaadi F.E. Estimation, filtering and fusion for networked systems with network-induced phenomena: New progress and prospects. Inform. Fusion. 2016;31:65–75. doi: 10.1016/j.inffus.2016.01.001. [Cross Ref]

7. Liu Y., He X., Wang Z., Zhou D. Optimal filtering for networked systems with stochastic sensor gain degradation. Automatica. 2014;50:1521–1525. doi: 10.1016/j.automatica.2014.03.002. [Cross Ref]

8. Caballero-Águila R., García-Garrido I., Linares-Pérez J. Information fusion algorithms for state estimation in multi-sensor systems with correlated missing measurements. Appl. Math. Comput. 2014;226:548–563. doi: 10.1016/j.amc.2013.10.068. [Cross Ref]

9. Fangfang P., Sun S. Distributed fusion estimation for multisensor multirate systems with stochastic observation multiplicative noises. Math. Probl. Eng. 2014;2014:373270. doi: 10.1155/2014/373270. [Cross Ref]

10. Pang C., Sun S. Fusion predictors for multi-sensor stochastic uncertain systems with missing measurements and unknown measurement disturbances. IEEE Sens. J. 2015;15:4346–4354. doi: 10.1109/JSEN.2015.2416511. [Cross Ref]

11. Tian T., Sun S., Li N. Multi-sensor information fusion estimators for stochastic uncertain systems with correlated noises. Inform. Fusion. 2016;27:126–137. doi: 10.1016/j.inffus.2015.06.001. [Cross Ref]

12. De Koning W.L. Optimal estimation of linear discrete-time systems with stochastic parameters. Automatica. 1984;20:113–115. doi: 10.1016/0005-1098(84)90071-2. [Cross Ref]

13. Luo Y., Zhu Y., Luo D., Zhou J., Song E., Wang D. Globally optimal multisensor distributed random parameter matrices Kalman filtering fusion with applications. Sensors. 2008;8:8086–8103. doi: 10.3390/s8128086. [Cross Ref]

14. Shen X.J., Luo Y.T., Zhu Y.M., Song E.B. Globally optimal distributed Kalman filtering fusion. Sci. China Inf. Sci. 2012;55:512–529. doi: 10.1007/s11432-011-4538-7. [Cross Ref]

15. Hu J., Wang Z., Gao H. Recursive filtering with random parameter matrices, multiple fading measurements and correlated noises. Automatica. 2013;49:3440–3448. doi: 10.1016/j.automatica.2013.08.021. [Cross Ref]

16. Linares-Pérez J., Caballero-Águila R., García-Garrido I. Optimal linear filter design for systems with correlation in the measurement matrices and noises: Recursive algorithm and applications. Int. J. Syst. Sci. 2014;45:1548–1562. doi: 10.1080/00207721.2014.909093. [Cross Ref]

17. Caballero-Águila R., García-Garrido I., Linares-Pérez J. Distributed fusion filtering in networked systems with random measurement matrices and correlated noises. Discret. Dyn. Nat. Soc. 2015;2015:398605. doi: 10.1155/2015/398605. [Cross Ref]

18. Shi Y., Fang H. Kalman filter based identification for systems with randomly missing measurements in a network environment. Int. J. Control. 2010;83:538–551. doi: 10.1080/00207170903273987. [Cross Ref]

19. Li H., Shi Y. Robust *H*_{∞} filtering for nonlinear stochastic systems with uncertainties and random delays modeled by Markov chains. Automatica. 2012;48:159–166. doi: 10.1016/j.automatica.2011.09.045. [Cross Ref]

20. Li N., Sun S., Ma J. Multi-sensor distributed fusion filtering for networked systems with different delay and loss rates. Digit. Signal Process. 2014;34:29–38. doi: 10.1016/j.dsp.2014.07.016. [Cross Ref]

21. Sun S., Ma J. Linear estimation for networked control systems with random transmission delays and packet dropouts. Inf. Sci. 2014;269:349–365. doi: 10.1016/j.ins.2013.12.055. [Cross Ref]

22. Caballero-Águila R., Hermoso-Carazo A., Linares-Pérez J. Covariance-based estimation from multisensor delayed measurements with random parameter matrices and correlated noises. Math. Probl. Eng. 2014;2014:958474. doi: 10.1155/2014/958474. [Cross Ref]

23. Chen B., Zhang W., Yu L. Networked fusion Kalman filtering with multiple uncertainties. IEEE Trans. Aerosp. Electron. Syst. 2015;51:2332–2349. doi: 10.1109/TAES.2015.130803. [Cross Ref]

24. Caballero-Águila R., Hermoso-Carazo A., Linares-Pérez J. Optimal state estimation for networked systems with random parameter matrices, correlated noises and delayed measurements. Int. J. Gen. Syst. 2015;44:142–154. doi: 10.1080/03081079.2014.973728. [Cross Ref]

25. Wang S., Fang H., Tian X. Recursive estimation for nonlinear stochastic systems with multi-step transmission delays, multiple packet dropouts and correlated noises. Signal Process. 2015;115:164–175. doi: 10.1016/j.sigpro.2015.03.022. [Cross Ref]

26. Chen D., Yu Y., Xu L., Liu X. Kalman filtering for discrete stochastic systems with multiplicative noises and random two-step sensor delays. Discret. Dyn. Nat. Soc. 2015;2015:809734. doi: 10.1155/2015/809734. [Cross Ref]

27. Caballero-Águila R., Hermoso-Carazo A., Linares-Pérez J. Fusion estimation using measured outputs with random parameter matrices subject to random delays and packet dropouts. Signal Process. 2016;127:12–23. doi: 10.1016/j.sigpro.2016.02.014. [Cross Ref]

28. Chen D., Xu L., Du J. Optimal filtering for systems with finite-step autocorrelated process noises, random one-step sensor delay and missing measurements. Commun. Nonlinear Sci. Numer. Simul. 2016;32:211–224. doi: 10.1016/j.cnsns.2015.08.015. [Cross Ref]

29. Wang S., Fang H., Tian X. Minimum variance estimation for linear uncertain systems with one-step correlated noises and incomplete measurements. Digit. Signal Process. 2016;49:126–136. doi: 10.1016/j.dsp.2015.10.007. [Cross Ref]

30. Gao S., Chen P., Huang D., Niu Q. Stability analysis of multi-sensor Kalman filtering over lossy networks. Sensors. 2016;16:566. doi: 10.3390/s16040566. [PMC free article] [PubMed] [Cross Ref]

31. Caballero-Águila R., Hermoso-Carazo A., Linares-Pérez J. Linear estimation based on covariances for networked systems featuring sensor correlated random delays. Int. J. Syst. Sci. 2013;44:1233–1244. doi: 10.1080/00207721.2012.659709. [Cross Ref]

32. Kailath T., Sayed A.H., Hassibi B. Linear Estimation. Prentice Hall; Upper Saddle River, NJ, USA: 2000.

Articles from Sensors (Basel, Switzerland) are provided here courtesy of **Multidisciplinary Digital Publishing Institute (MDPI)**

PubMed Central Canada is a service of the Canadian Institutes of Health Research (CIHR) working in partnership with the National Research Council's national science library in cooperation with the National Center for Biotechnology Information at the U.S. National Library of Medicine(NCBI/NLM). It includes content provided to the PubMed Central International archive by participating publishers. |