7

Decentralized Data Fusion

Formulation and Algorithms

Paul Thompson, Eric Nettleton, and Hugh Durrant-Whyte

CONTENTS

7.1    Decentralized Data Fusion Introduction

7.2    Information Form Introduction

7.3    Decentralized Fusion and Communication

7.3.1    Tree Network Topology, Channel Cache

7.3.2    Related Channel Filter Approaches

7.3.3    Summary

7.4    Dynamic Systems

7.4.1    State Dynamics

7.4.1.1     State Dynamic Model

7.4.1.2    Trajectory Information Approach

7.4.1.3    Equivalence to the Conventional Approach

7.4.1.4    Multiple Trajectory States

7.4.2    Dynamics in Decentralized Data Fusion

7.4.2.1    Common Process Noise Problem

7.4.2.2    Delayed and Asequent Observations

7.4.2.3    Burst Communications

7.4.2.4    Solution Using Trajectory States

7.4.2.5    Filtering the Trajectory State System

7.4.2.6    Filtering with Stored Filter Estimates

7.4.2.7    Operation of Channel Caches with Trajectory States

7.4.3    Summary

7.5    k-Tree Topologies for Redundant and Dynamic Networks

7.5.1    Decentralized Data Fusion on k-Trees

7.5.2    Data-Tagging Sets

7.5.3    Separator and Neighborhood Properties

7.5.3.1    Separator Property

7.5.3.2    Local Neighborhood Property

7.5.4    k-Tree Communications Algorithm

7.5.4.1    Data-Tag Set Elimination

7.5.5    Link and Node Failure Robustness

7.5.6    Summary

7.6    Conclusion

7.A   Appendix

7.A.1    Marginalization in the Information Form

7.A.2    Trajectory Information Form Equivalence

Acknowledgments

References

7.1    DECENTRALIZED DATA FUSION INTRODUCTION

A decentralized data fusion (DDF) system consists of a network of sensing and computing nodes that aim to cooperatively estimate a common state [12]. Fusion occurs on each node using locally obtained observations and communications from neighboring nodes, without relying on a centralized decision or fusion system. This chapter summarizes and builds on previous research in DDF, including [18,22].

DDF systems have been characterized by three constraints [9,12]:

1.  There should be no single central fusion center; no single node should be central to the successful operation of the network.

2.  There is no common communication facility; nodes cannot broadcast results and communication must be kept on a strictly node-to-node basis.

3.  Sensor nodes do not have any global knowledge of the network topology; nodes should only know about connections in their own neighborhood.

The resulting estimates in the decentralized system can be compared to an equivalent centralized estimator operating with the same observations and modeling assumptions. The focus of this chapter is on exact solutions for DDF, which are equivalent to centralized data fusion, in the following sense:

•  The use of consistent fusion with information terms which are conditionally independent given the state, as opposed to methods which double count or miss information terms or use conservative fusion methods.

•  The use of direct solution methods as opposed to iterative or convergent methods.

This chapter is organized as follows. Section 7.2 introduces the information form, which is used in this chapter as the expression for fusion operations. Section 7.3 discusses the fusion update and communication aspects of DDF and discusses the operation of DDF on tree topology networks. Section 7.4 introduces the trajectory state formulation for dynamics and uses this to operate decentralized networks for dynamic systems, including handling delayed, asequent and burst communications issues. Section 7.5 extends the tree topology to k-tree topologies for redundant and dynamic decentralized topologies.

7.2    INFORMATION FORM INTRODUCTION

For the decentralized algorithms presented in this chapter, it is convenient to express the fusion operations in terms of information by reformulating multiplication of probability as summation of log-probability.

The main properties that motivate the use of the information form are as follows:

1.  Additivity of fusion and observation updates

2.  Sparsity of the information matrix

Consider a random variable x with prior probability density function (PDF) described by a Gaussian PDF, together with a linear observation, described by a Gaussian likelihood:

p(x)=1b exp (12(xˆx)TP1(xˆx))

(7.1)

p(z|x)=1c exp (12(Hxz)TR1(Hxz))

(7.2)

Where the observation is modeled as

z=Hx+w E[w]=0 E[wwT]=R

(7.3)

Under Bayes’ rule, p(x | z) = p(z | x)p(x)/p(z), the posterior PDF given the prior and observation is

p(x|z)=1d exp (12(xˆx)TP1(xˆx)12(Hxz)TR1(Hxz))

(7.4)

=1d exp (12(xˆx+)TP1+(xˆx+))

(7.5)

The two expressions for the posterior must equate

(xˆx+)TP1+(xˆx+)=(xˆx+)TP1(xˆx)+(Hxz)TR1(Hxz)

(7.6)

By matching first and second derivatives of each side with respect to x, this results in

P1+=P1+HΤR1H

(7.7)

P1+ˆx+=P1ˆx++HΤR1z

(7.8)

The information form is defined by these terms P−1 and P1ˆx:

Y 2 log {p(x)}x2 =P1ˆy log {p(x)}x|@x=0=P1ˆx

(7.9)

Consequently, given Y and y^, the estimate is recovered by the solution of the linear system

Yx^=y^

(7.10)

The estimate x^ will be identical to that obtained by a covariance-based Gaussian estimator such as a Kalman filter operating under identical assumptions.

So, given a prior PDF described by information matrix Y and information vector y, the posterior following the observation is

Y+=Y+HTR1H

(7.11)

y+=y+HTR1z

(7.12)

It is convenient to label the observation as contributing observation information in the form

I=HTR1H

(7.13)

i=HTR1z

(7.14)

In general, the fusion of multiple, statistically independent terms is a straightforward addition:

Y+=iYiy+=iyi

(7.15)

7.3    DECENTRALIZED FUSION AND COMMUNICATION

This section discusses DDF with a focus on the fusion and update steps resulting from communication and observations. These aspects are highlighted by considering DDF on a static state variable. Section 7.4 extends this discussion by considering system dynamics and temporal aspects.

The static case highlights the basic properties of the fusion component of the decentralized system, since the present formulation of DDF is built on the additive properties of the information fusion operations.

The system then consists of a common static state x, which is to be estimated on the multiple platforms. The decentralized system is required to obtain an estimate in exact agreement with an equivalent centralized estimator using the same observations and modeling assumptions. In a static system, the posterior information is identically the sum of the individual independent observation information terms

Y=kHjTRk1Hky^=kHkTRk1zk

(7.16)

x^=Y/y

(7.17)

In essence, the DDF nodes communicate to obtain the Y and y^ sums in Equation 7.16. The required globally agreeing estimate is then obtained through the solution of Equation 7.17 separately at each node.

Different algorithms on different topologies operate different methods for obtaining the sums in Equation 7.16:

•  In a centralized estimator, each HkTRk1Hk and HkTRk1zk is communicated to the central estimator, which performs the sum in Equation 7.16.

•  In a fully connected decentralized topology, each node transmits HkTRk1Hk and HkTRk1zk to each other node. Each node is then able to separately perform the sum in Equation 7.16.

•  In a tree-connected decentralized topology, nodes accumulate partial sums of Equation 7.16 and communicate in a tree to obtain the global sums. This topology is discussed further later.

7.3.1    TREE NETWORK TOPOLOGY, CHANNEL CACHE

This section considers the singly connected or tree decentralized topology. Under this topology, the graph properties of the tree and the distributivity of the addition are exploited in order to perform the required summation in Equation 7.16.

A tree topology has no cycles. This means that for each node any communications to a neighbor cannot affect any other neighbor. Also, any communications from a neighbor cannot be affected by any other neighbor. This occurs because at any node, a, the neighbors of the neighbors of a, excluding a, are disjoint.

For every node a in a tree: { N(N(a)) }a are disjoint

(7.18)

where

N(a) is the neighbor of node a

Sa is the set S excluding a

This means that the sum in Equation 7.16 can be written as a hierarchy of partial sums over disjoint subsets:

Ya=Ia+iN(a){ Ii+i{ N(i)a }{ Ij+k{ N(j)i }{ Ik+ } } }

(7.19)

The tree topology guarantees that the terms inside each summation are disjoint (independent from each other) and therefore prevents double counting of observation information.

The algorithm developed in this section will be referred to as the channel cache algorithm. The channel cache algorithm is a variant of the well-known channel filter algorithm [6,9,12,18, 19, 20,22]. The channel cache algorithm is also inspired by junction tree algorithm for inference in graphical models [21] and [16].

The operation of a tree topology decentralized network is illustrated in Figure 7.1. Figure 7.1 shows a branch in a tree network.

Each node stores its own observation information (dark gray) in the form I = HTR−1H and i = HTR−1z. These correspond to the I terms in Equation 7.19. These observation information terms are required to be statistically independent information unique to one node.

The communicated term from a node i to a is an information matrix Cia (and its information vector counterpart). Cia consists of the transmit node’s own independent observation information plus the sum of all communicated terms received from the “upstream” part of the tree network:

Image

FIGURE 7.1 DDF with channel caches. Four nodes are arranged in the topology shown in the lower right. Each node stores its own fused observations (dark gray). Each node caches the received communication term from each of its neighbors (light gray). The total fused information at each node is the sum of each stack, since each layer consists of independent information. The transmitted communication term is the sum of the stack excluding the destination’s cache term.

Cia=Ii+j{ N(i)a }Cji

(7.20)

=Ii+j{ N(i)a }{ Ij+k{ N(j)i }{ Ckj } }

(7.21)

Each node locally caches the received communication term from each of its neigh-bors (light gray) in a so-called channel cache. All of the channel cache, C, and the observation, I, information terms are statistically independent.

Transmission of a communication term has no effect at the transmitting node, so transmissions can be lost without breaking the consistency of the estimates. On reception of a communication term, the received term is simply stored in the channel cache. This means that duplicate transmissions and/or duplicate receptions are acceptable.

Each node can obtain the total network sum, Y, by summing its own observation information together with all the locally cached communication terms, e.g., at node a:

Ya=Ia+iN(a)Cia

(7.22)

The net result is that the network computes a series of partial sums, with each node obtaining the sum as in Equation 7.19. For each node, the evaluation of Equation 7.19 operates as a series of messages propagating inward on the tree toward that node.

Nodes initialize their observation and communication cache information terms to zero, I = 0, C = 0, such that nodes can produce estimates even before the network has finished propagating terms across the network span.

7.3.2    RELATED CHANNEL FILTER APPROACHES

The previous section presented the channel cache algorithm for tree topology networks. The channel cache is closely related to the channel filter algorithm, which has been discussed in various papers [6,9,12,18, 19, 20,22].

The approach used in a channel filter is to maintain the total information estimate at each node and maintain the common information between pairs of nodes on a tree network. The channel filter’s use of the common information is motivated by the following equation for the fusion of a local YA and a received communication YB from a remote node:

YAB=YA+YBYAB

(7.23)

Each primary operation of the channel filter is described next (referring to operations at a node i, transmitting to a node a and receiving from a node j). Table 7.1 summarizes the channel algorithms. The channel filter consists of the following operations:

•  Observation: Yi + = I. Observation information I adds simply into the total information Yi without affecting any channel common information.

•  Transmit: The node’s current total information is transmitted, Cia = Yi. It is assumed that the destination node a will successfully receive the communication; therefore, the common information is set Yia = Yi.

•  Receive: The received information is Yji, so given the existing common information Cji, the total information is updated: Yi + = CjiYji. The nodes now have common information from the communication, so the node sets Yji = Cji.

•  Result: The total information is maintained in Yi.

TABLE 7.1
Summary of the Primary Operations for the Channel Cache and Channel Filter Algorithms

Obs Update

Channel Cache Ii+ = I

Channel Filter Yi+ = I

Transmit

Cia=Ii+j{ N(i)a }Cji

Cia = Yi Yia = Yi

Receive

Store Cji

Yi + = CjiYji Yji = Cji

Result

Yi=Ii+jN(i)Cji

Use Yi

Both the channel filter and the channel cache algorithms are designed to exploit a tree topology network. At each node, both the channel filter and channel cache algorithms store an information matrix and vector for each neighbor, intended to ensure correct consistent interaction with that neighbor. The basic difference between the channel filter and the channel cache algorithms is as follows:

•  The channel filter algorithm maintains the common information between the local total information and each neighbor’s total information.

•  The channel cache algorithm maintains the contributed information from each neighbor.

Using the common information requires both nodes to maintain identical copies of the common information, which is vulnerable to failure if the two copies differ (cases in which this can happen are discussed later). The common information maintained at both nodes on a channel is required to be identical, since the common information, Yij, is symmetrical between two nodes, i.e., Yij = Yji [12].

By contrast, the channel cache algorithm maintains a local record of the contributed information from the neighbor. This decouples the communication between nodes such that the changes only occur locally when information is received, not when it is transmitted. Table 7.2 shows a decentralized communication transaction between two nodes, showing both the contributed information and the common information, in the case of an ideal communication. The communication must update the common information at both nodes, which requires an assumption of successful communication for the sender. On the other hand, the communication only needs to update the contributed information record once (at the destination node) and only upon an actual successful communication.

Image

Table 7.2 also shows that the two contributed information terms sum to the common information:

YAcontributed toB+YBcontributed toA=YAB 

(7.24)

The channel filter algorithm can fall into cases where the two common-information records can differ due to miscommunication:

•  Asynchronous operation. If nodes send messages which “cross over,” then their common-information records can become misaligned. An example is shown in Table 7.3. Consider a pair of nodes i,j which transmit almost simultaneously at times ti, tj and receive at times ri, rj. This asynchronous case arises if rj > tj or ri > ti.

•  Lost transmissions. The channel filter algorithm can also become misaligned in the case of transmissions which are lost. This occurs if a node completes the “transmit” update to its channel filter but the destination fails to receive the message. An example is shown in Table 7.4.

7.3.3    SUMMARY

This section described DDF with a focus on the observation update and the decentralized communication, particularly in tree topology networks.

The fusion of independent observation and/or communicated information is additive when performed in the log-likelihood or information form. Therefore, the problem of forming a decentralized estimate which is identical to a centralized equivalent reduces down to a decentralized algorithm for forming a correct sum of the observation information terms.

When applied to tree topology networks, it suffices to maintain a local node information matrix and vector and one information matrix and vector for each neighbor in the network.

Image

This section presented the channel cache algorithm for handling the local node information and decentralized communications operations. The channel cache algorithm handles imperfect communications such as asynchronous transmissions, lost transmissions in a simple manner. The channel cache algorithm operates on records of contributed information from each neighbor. This is in addition to the capabilities previously possible with channel filters on tree networks, especially the avoidance of double counting, avoidance of conservative fusion, while achieving global agreement among nodes in a decentralized network.

The aforementioned discussion has focused on the observation update and the decentralized communication. The following section extends the discussion of DDF into dynamic systems.

7.4    DYNAMIC SYSTEMS

The observation and communication updates, as described in the previous section, were discussed with respect to a static system, i.e., a single-state vector x. This section extends the discussion into dynamic systems and reviews the smoothing or trajectory state formulation of dynamic systems to formulate DDF for estimation of dynamic systems. This trajectory state formulation of dynamics is then applied to address the issues of delayed and asequent observations and burst communications in DDF.

Image

When the decentralized system has observation and/or communication interruptions and delays, it becomes important to decide when and where the dynamic propagation of the estimate is to be applied. Furthermore, at each decentralized node, there are stored communication terms relating to other nodes, and so it is also necessary to consider the dynamic propagation of these.

This section presents the trajectory state approach to representing system dynamics. The trajectory state approach expands the state for a dynamic system into a joint state consisting of a sequence (trajectory) of states. The tools for manipulating joint probabilities in several dimensions and tools for manipulating probabilities and decentralization of static states then become applicable to the dynamic system.

The trajectory state approach relates to smoothing methods used in Kalman smoothing [17]. It is also known as delayed states and has been used to account for delayed decision making in estimation such as delayed associations [15]. The use of delayed states in the information form, with the resulting sparse structure, has been applied in localization and mapping [8,10]. Delayed states have more recently been applied to DDF as a tool for delayed measurements [1] and for delayed and asequent measurements and communications [4].

This section focuses on correct approaches to dealing with delayed, asequent, and burst communications with dynamic models that are known and can be applied by each node. The trajectory state approach can be extended to allow the decentralized system to distribute dynamic models which originate from one node (known as model distribution). The issue of dynamic communication topologies in DDF is discussed separately in Section 7.5.

7.4.1    STATE DYNAMICS

This section explains the state dynamics and trajectory state form, in general. Section 7.4.2 applies these to DDF specifically.

7.4.1.1    State Dynamic Model

We consider a basic, linear discrete time state dynamic model in the form

xk+1=Fxk+Buk+Gvk

(7.25)

where

xk is the state vector at instant k

F is the state transition matrix

vk is unknown, zero mean, white noise, E[ vk ]=0,E[ vkvkT ]=Q

uk is a known control signal, if available

The conventional treatment is to form a prediction by using a dynamic transformation of the estimate [2]

x^k+1=Fx^k+Buk

(7.26)

Pk+1=FPkFT+GQGT

(7.27)

where this is considered to be a transformation of the estimate, replacing the estimate for time k by that for time k + 1, as a discrete operation.

7.4.1.2    Trajectory Information Approach

Equation 7.27 can actually be considered (see later) to consist of an augmentation of the estimate into the latter timestep k + 1, followed immediately by a marginalization to remove timestep k. In this way, the prediction operation in Equation 7.27 moves the estimate forward in time but removes the state components for the past timestep. This removal of the past timestep makes it impossible (or difficult) to fuse late observations or communicated information. This transformative prediction approach thus requires observations to be fused in at the appropriate timestep.

This section describes an alternative approach known as delayed state or trajectory state approach, which is designed to address the aforementioned issues. In the trajectory state approach, instead of considering prediction equations to explicitly transform the estimate from time k to k + 1, we instead consider the trajectory described by a joint state X = [xkxk+1]

The key reason why this is useful is as follows: If we operate with a joint trajectory state vector X = [xk xk+1xk+n], then observations and communications in any time (k to k + n), and the dynamic model all act additively on the joint state X. Thus the methods of Section 7.3 remain applicable, since they are designed to exploit additive operations over decentralized networks.

Therefore, we rearrange Equation 7.25 to focus on the joint trajectory state X = [xkxk+1]:

Bu=Fxk+Ixk+1Gv

(7.28)

Bu=[ FI ][ xkxk+1 ]TGv

(7.29)

where I denotes the identity matrix. We can then consider Equation 7.29 in the form of an observation, as in Equation 7.3:

z=Hx+wE[ wwT ]=RBu=[ FI ][ xkxk+1 ]TGvE[ GvvTGT ]=GQGT

(7.30)

Considering the dynamic model in the form of an observation requires the following replacements:

H[ FI ]RGQGTzBu

(7.31)

By analogy with Equation 7.13, the dynamic model can then be represented as an information matrix and vector in the joint trajectory state X:

I=HTR1Hi=HTR1z

(7.32)

I=[ FTQ1FFTQ1Q1FQ1 ]i=(FTQ1BuQ1Bu)

(7.33)

where Q ≜ GQGT.

7.4.1.3    Equivalence to the Conventional Approach

We next show the equivalence of the trajectory state approach to existing prediction equations in the information and covariance forms. To show the equivalence, we setup the same initial conditions and steps as for a prediction:

1.  Define some prior information in the earlier timestep xk: Yk and yk satisfying Ykx^k=yk.

2.  Allow no prior information in the latter xk+1.

3.  Apply the dynamic model to the joint xk and xk+1.

4.  Evaluate the marginal information in the latter timestep xk+1.

The posterior information, Y and y, after step 3 (i.e., given the prior and the dynamic model) is

Y=(Yk000)+HTR1Hy=(yk0)+HTR1z

(7.34)

Y=(Yk+FTQ1FFTQ1Q1FQ1)y=(ykFTQ1BuQ1Bu)

(7.35)

Equation 7.35 is equivalent to the conventional prediction in Equation 7.27 (proofs are provided in the appendix):

•  The joint X^ satisfies YX^=y, with X^=(x^x^k+1)=(x^kFx^k+Bu).

•  The xk marginal of Y remains as the given Yk and yk.

•  The xk+1 marginal of Y yields known expressions [2,19] for the prediction in covariance and information forms:

Yk+1|k={ FPk|kFT+Q }1

(7.36)

=MMGΣ1GTM

(7.37)

yk+1|k=[ IMGΣ1GT ]FTyk|k+Yk+1|kBu

(7.38)

(Σ=GTMG+Q1)

(7.39)

(M=FTYk|kF1)

(7.40)

7.4.1.4    Multiple Trajectory States

Earlier we discussed the formation of a pair of joint successive dynamic states, xk and xk+1. Now consider a longer sequence of trajectory states. The original discrete time dynamic model in Equation 7.25 holds for each pair of successive dynamic states; therefore, each successive pair has the dynamic model information added, as in Equation 7.33. The information matrix and vector for the dynamic model between any successive states k and k + 1 is

Ikdyn=(FTQ1FFTQ1Q1FQ1)ikdyn=(FTQ1BuQ1Bu)

(7.41)

We later re-write Ikdyn=(ADDTC) and ikdyn=(ac)T to save space.

Over a sequence of trajectory states, these pairwise Idyn blocks add up to form a sparse banded matrix:

k[1,7]Ikdyn=(AD000000DTC+AD000000DTC+AD000000DTC+AD000000DTC+AD000000DTC+AD000000DTC+AD000000DTC)

(7.42)

The benefit of the trajectory state formulation is that observations of states within the trajectory appear additively in the information matrix. For example, the total information for the trajectory system from times 1 to 8, including a prior Y1 | 1 at time k = 1, an observation I5 at time k = 5, and the dynamic model information between each time is given by

Y1:8=Y1|1+I5+k[ 1,7 ]Ikdyn=(Y1|1+AD000000DTC+AD000000DTC+AD000000DTC+AD000000DTC+A+I5D000000DTC+AD000000DTC+AD000000DTC)y1:8=(y1|1+ac+ac+ac+ac+a+i5c+ac+ac)T

(7.43)

where the observation information I5 appears as an addition on the diagonal of Y corresponding to the state at the observed time.

To propagate the trajectory state system forward in time (maintaining a fixed duration trajectory), there are two steps:

1.  Augmenting the system with the additional timestep. This requires expanding the state vector for the new timestep (k + 1) and adding the dynamic model information Ikdyn and ikdyn.

2.  Marginalizing away the earliest timestep.

The system following propagation by one timestep is now given by

Y2:9=(A+Y211D000000DTC+AD000000DTC+AD000000DTC+A+I5D000000DTC+AD000000DTC+AD000000DTC+AD000000DTC)y2:9=(a+y2|1c+ac+ac+a+i5c+ac+ac+ac)T

(7.44)

where Y2|1 and y2|1 are

Y2|1=CDT(Y2|1+A)1D

(7.45)

=Q1Q1F(Y1|1+FTQ1F)1FTQ1

(7.46)

y2|1=cQT(Y1+A)1(y1+a)

(7.47)

=Q1Bu+Q1F(Y1|1+FTQ1F)1(y1|1FTQ1Bu)

(7.48)

Y2|1 is actually the same expression as for the predicted Yk+1|k in Equation 7.35. This is proven in the appendix. The earlier prior information, Y1|1, clearly resides in a nonadditive form, in the expression DT(Y1|1 + A)−1D.

In summary:

•  The augmentation process, which extends the system to further timesteps, continues the same sparse banded pattern in the information matrix.

•  The fusion of observations within the duration of the trajectory states is a straightforward addition in the information matrix and vector.

•  Marginalization of the earliest timestep in a succession of trajectory states follows the same pattern as for information filtering prediction, leaving any observations or prior information in the removed timestep k in a nonadditive form.

7.4.2    DYNAMICS IN DECENTRALIZED DATA FUSION

The previous section discussed the state dynamics generally, resulting in the formation of a trajectory state system. The key advantage of using a sequence of trajectory states is that for observations of states within the trajectory states, the observation information is additive, just as for observations of a static state. This additivity of observation information applies regardless of the timing or sequence of observations, as long as the state at the observed time exists in the current set of trajectory states.

This section describes the application of the trajectory state approach for handling timing issues in DDF. In particular, we consider the following problem cases:

•  Delayed and asequent data fusion, in which an observation from an earlier time becomes available after a prediction step, has been performed (delayed) or after other data have been fused for later times (asequent). Delayed and asequent observations usually refer to local sensor node observations.

•  Burst communication, which occurs when decentralized communications is resumed after a period of interruption. The communications that occurs after the interruption is referred to as burst communication, since it aims to deliver a large amount of information in a short time (or single message) to re-establish agreement between the nodes. Burst communications can also be thought of as delayed/asequent fusion across multiple decentralized nodes.

The key issue behind the aforementioned difficulties is additivity of observation and communicated information, and the fact that for states that have been replaced by predictions cannot be updated additively by other predictions. This is explained in further detail next.

7.4.2.1    Common Process Noise Problem

The underlying issue of concern relates to the common process noise problem. They are so called because separately predicted terms ignore their common use of the same process noise. This can also be expressed as the problem that the fusion of predicted information is unequal to the prediction of fused information:

Predict (Fuse(A,B))Fuse(Predict(A), Predict(B))

(7.49)

Consider a case where a fused estimate is predicted forward. The fused estimate at time k is obtained as the sum of two independent information terms, e.g., Yk,a and Yk,b.

Yk=Yk,a+Yk,b

(7.50)

The correct expression for the predicted information for time k + 1 requires the prediction of the sum

Yk+1exact=Predict(Yk)

(7.51)

={ FYk1FT+Q }1

(7.52)

If, however, the term Yk,a has already been predicted forward, a common approximation to Yk+1 is to take

Yk+1approx=Predict(Yk,a)+Predict(Yk,b)

(7.53)

={ FYk,a1FT+Q }1+{ FYk,b1FT+Q }1

(7.54)

The approximate form is not generally equal to the exact form Yk+1approxYk+1exact. The approximate form ignores the fact that there is only one underlying process; hence, the two prediction instances share common process noise, v (of which E[vvT] = Q). To consider the approximation further, consider a simpler worst case where Ya = Yb:

Yk+1approxYk+1exact={ FYa1FT+Q }1+{ FYb1FT+Q }1{ F(Ya+Yb)1FT+Q }1

(7.55)

={ I+2a }{ I+a }

(7.56)

where

a=FTYaF1Q=(FYa1FT)1Q

(7.57)

for which it can be seen that

lima0Yk+1approxYk+1exact=IlimaYk+1approxYk+1exact=2I

(7.58)

Yk+1exactYk+1approx2Yk+1exact

(7.59)

This shows that Yk+1approx is always slightly overconfident, but is close to Yk+1exact for small Q. However, for large Q, the Yk+1approx is overconfident, being up to 2Yk+1exact in the worst case. Predicting the Yk,a and Yk,b independently is equivalent to claiming that there are two independent process models available.

7.4.2.2    Delayed and Asequent Observations

A delayed observation occurs when an observation from an earlier time becomes available after a prediction step has been performed [19]. The problem of delayed observations can occur in any form of estimator, not only decentralized estimators. Delayed observations can occur as a result of processing and/or communication delays before observations are available at the estimator.

An asequent observation occurs when an observation from an earlier time becomes available after other data have been fused for later times [19]. Asequent observations may occur if multiple sensors are used locally on a single node, and these sensors have differing observation delays. The case of asequent observations occurring on distinct decentralized nodes is similar, but since it involves the communication aspect it is more similar to the burst communications case discussed later.

The problem with delayed or asequent data fusion is that once the estimator has predicted the local state forward to time k + 1, the (late) incoming information for time k needs to be considered. If the late arriving information is predicted forward separately, the common process noise problem applies (as discussed in Section 7.4.2) and the result will be approximate and over-confident.

The problem with delayed and asequent data fusion is basically caused by the filter architecture destructively predicting estimates forward. That is, applying the prediction equations in a way that replaces a local estimate.

The proposed solution instead applies the trajectory state approach to avoid destructively predicting estimates until after a window of time has passed, while still obtaining correct current-time filter estimates given all available past observations.

The trajectory information matrix is constructed as in Section 7.4.1:

Yk:k+4=(A+Yk|k1+IkD000DTC+A+Ik+1D000DΤC+A+Ik+2D000DTC+A+Ik+3D000DTC+Ik+4)

(7.60)

where Yk:k+4 is written with observation information on each timestep, indicating how current, delayed, and/or asequent observations can be fused additively in the trajectory information matrix and vector at their appropriate timestep, as long as that timestep is available within the trajectory state system.

Given the trajectory state system, the estimate solution for the current (latest) timestep will be equivalent to a filtered solution, correctly accounting for the late and asequent observations. Methods for obtaining the solution are discussed in Section 7.4.2.

The trajectory state approach with N timesteps of trajectory states defers the destructive prediction of the earliest state by N timesteps, allowing delayed and asequent observations in that duration. However, very late observations beyond N timesteps will still be subject to the same common process noise problem preventing their use. Very late observations beyond N timesteps are expected to occur less frequently and be less informative to the present estimate and should be discarded (which is conservative). Note that the intention of the trajectory state method is to use N timesteps such that the system can still benefit from the observations with small delay which are very likely to occur and very beneficial to the present estimate.

7.4.2.3    Burst Communications

Burst communication occurs when decentralized communications are resumed after a period of interruption. The communications that occurs after the interruption is referred to as burst communication since it aims to deliver a large amount of information in a short time (or single message) to re-establish agreement between the nodes.

The problem with burst communications occurs when the estimator predicts the local state forward during a period of interrupted communications. The problem is that other decentralized nodes will also perform the same prediction on their local estimates. When the nodes re-connect and communicate, the common process noise problem arises, since the information from each node will have been separately predicted.

The problem with delayed and asequent data fusion is again caused by the filter architecture destructively predicting estimates forward. That is, applying the prediction equations in a way that replaces a local estimate.

The proposed solution, as for asequent observations, involves using trajectory states in order to maintain a window of some duration in which communications can be late, but still fuse additively into states in the trajectory window. Estimates for the current time can still be obtained from the system, conditioned on all the available past observations.

Referring to Equation 7.60, the decentralized system can communicate the diagonal matrix consisting of the Ik blocks:

Ik:k+4=(Ik00000Ik+100000Ik+200000Ik+300000Ik+4)

(7.61)

This becomes equivalent to a sequence of static decentralized problems, one for each timestep. The band structure corresponding to the dynamic model can be applied locally at each node.

For normal operation, with frequent communications, the nodes transmit their current Ik block. But if the communications is blocked for an interval of time and then later resumed, the resulting burst communications will contain the diagonal blocks Ij:k for the fused observations for the blocked interval.

The methods for obtaining the solution are discussed in Section 7.4.2.

7.4.2.4    Solution Using Trajectory States

The cases of delayed and asequent observations and burst communications presented earlier can be addressed using a set of trajectory states. This section focuses on how to solve the resulting trajectory state system:

Yk4:k=(A+Ykn|kn1+IknD000DTC+A+Ikn+1D000DTC+A+ID000DTC+A+Ik1D000DTC+Ik)

(7.62)

yk:k+4=(a+ykn|kn1+ikna+c+ikn+1a+c+ia+c+ik1c+ik)

(7.63)

The system in Equation 7.62 is a block tridiagonal sparse linear system. Such a system can be solved very efficiently in O(n) time, for n trajectory states [11]. It is also possible to obtain smoothing estimates for the duration of the trajectory states by solving the joint system fully. This basically corresponds to solving for the latest estimate as described later, together with back-substitution for the smoothed estimates of the earlier states.

7.4.2.5    Filtering the Trajectory State System

The solution process for the filtered estimate of a trajectory state system is very similar to an online filtering process. In that case, the dynamic model in the trajectory state system need only be defined implicitly, leaving only the diagonal blocks (observations and prior) to be explicitly stored. So the current filtering estimate can be obtained basically by running the information filtering prediction cycles, starting from the prior information in the start of the trajectory and using the stored fused observation information at each time. Note that we only need to run this when requiring an estimate for the present state. Multiple observations can be added into the trajectory system without requiring this solve process. This approach is less general than the next, which will be described in greater detail.

7.4.2.6    Filtering with Stored Filter Estimates

In most cases, it is likely that observations and decentralized communications will arrive with only a small delay, and thus only affect the latter part of the trajectory state system. In that case, it is inefficient to process the entire trajectory state system for its whole duration. Also, allowing re-processing only the affected portion of the trajectory may allow a longer trajectory system to be used. When an estimate of the present state is required, it is only necessary to process forward from timestep kn to the present, where timestep kn is the earliest changed state in the trajectory. (Changes include local observations and decentralized communications.) In this way, the cost in computation to re-process following delayed or asequent observations or newly arrived burst communications depends on how far back the observation occurs, such that the normal case of short delays can proceed forward with little overhead.

This method is similar to that described for asequent data fusion in Ref. [19]. The difference is that we store the observations Ik in the trajectory state approach, not only the filtered estimates. This allows the handling of burst communications and also simplifies the case for asequent observations.

This method corresponds to filtering, but stores in memory the filtered estimates for a few key timesteps in the trajectory duration. The system needs to store Ij for each timestep, and Yj|j−1 and yj|j−1 for a few key timesteps.

The Ij terms for each j and the filtered Ykn|kn−1 for the earliest trajectory state timestep, kn are statistically independent of each other. These are regarded as “source” information. The stored filter Yj|j−1 estimates for the other timesteps are not independent of each other, and not independent of the Ik or Ykn|kn−1. These are to be regarded as a computational aid, storing partial results. These filter Yj|j1 could instead be re-processed from the initial Ykn|kn−1 and the observation information terms Ij.

This occurs as follows:

1.  The forward filtering can start from time kn where timestep kn is the earliest changed state in the trajectory, or wherever a starting or stored prior information exists. Starting from time kn, define a current information matrix and vector of the size of the state at a single time:

Yc=Ykn/kn1yc=yknkn1

(7.64)

2.  For each time j = kn: k

a.  Fuse observations for time j:

Yc=Yj|j=Yj|j1+Ijyc=yj|j=yj|j1+ij

(7.65)

b.  Predict the current Yj|j and yj|j to time j + 1 using Equation 7.37 (except at the last time k):

Yc=Yj+1|j=MMGΣ1GTM

(7.66)

yc=yj+1|j=[ IMGΣ1GT ]FTyj|j+YcBu

(7.67)

(Σ=GTMG+Q1)

(7.68)

(M=FTYj|jF1)

(7.69)

c.  The resulting Yc = Yj+1|j and yc = yj+1|j can be stored if desired, so processing can resume from time j + 1 later.

3.  The resulting Yc = Yk|k and yc = yk|k is the filtered information for the state at time k. X^k is obtained from

x^k=Yk|k1yk|k

(7.70)

Image

FIGURE 7.2 Illustration of combined trajectory state and channel cache–based DDF at a single node. The lower portion shows the whole system propagated forward by one timestep.

7.4.2.7    Operation of Channel Caches with Trajectory States

Figure 7.2 shows the combined trajectory state and channel cache–based DDF node. The node stores the observations Ij and ij for each timestep kn to k in the trajectory state window for each channel. These are the channel cache contributed information terms from the neighbors. The node similarly stores its own observations for each timestep. The node also stores the prior Ykn|kn and ykn|kn. The complete trajectory state system is formed by summing all the entries for each timestep, including the dynamic model information.

To shift the combined system forward by one timestep, at the back of the trajectory state window (earliest timestep), the prior Ykn|kn is propagated forward as in Equation 7.43. The observations Ikn+1 are added into Ykn+1|kn, leaving the next prior Ykn+1|kn+1. The observations for time kn + 1 are then popped out of the trajectory system, so that the guaranteed conditional independence of all information terms is maintained. The front end (latest and current timesteps) of the trajectory state window is extended into timestep k + 1, ready for new observations or channel terms.

7.4.3    SUMMARY

This section presented the smoothing or trajectory state formulation of dynamic systems to formulate DDF for estimation of dynamic systems. This trajectory state formulation of dynamics was then applied to address the issues of delayed and asequent observations and burst communications in DDF.

The solution method for the trajectory state formulation requires nodes to store the fused observation information for each timestep for a finite duration, allowing for observation and communication delays. For efficient solving it is also useful to store the results of filter estimates.

7.5    K-TREE TOPOLOGIES FOR REDUNDANT AND DYNAMIC NETWORKS

The algorithms presented in Section 7.3.1 relate to tree topology networks. Exact decentralized estimation has, in the past, largely been restricted to singly connected tree networks [7,9,18,20]. The key point relating to tree networks is that the DDF problem can be reduced down to a problem of finding a global sum of information terms, which is performed in an efficient local manner on a tree network. In tree networks, there is only one path between any two nodes. This is used in the decentralized algorithm to ensure exact fusion, especially avoiding cases of double counting or rumor propagation in the network. However, the single path property of tree networks also means that tree networks are vulnerable to the failure of nodes and links, since the failure of any nonleaf node or link would leave the network in multiple disconnected pieces. Tree networks include both “star” and chain topologies as well as branching trees.

This section presents an extension beyond tree communications network topologies into so-called k-tree network topologies. The k-tree topologies are more general than tree topologies but are more specialized for scalability than arbitrary topologies. The presentation of this section is based on Ref. [24]. The k-tree is an extension beyond tree topologies, which keeps an overall strict tree-like pattern on a large scale (N nodes ⨠ k), as shown in Figure 7.3f, but allows redundant, looped, dynamic topologies or other subsets of full connection within groups of nodes smaller or equal to k + 1. The costs in storage and communication grow with k but not with N, the total number of nodes in the network. The k-tree topologies are intended to be used with as small k as possible.

The motivation behind using k-tree topologies is to improve redundancy and dynamism while maintaining scalability and correctness. For redundancy and fault robustness, it is desirable to allow the network to include multiple redundant paths such that some links or nodes can fail without disconnecting the network topology. It is also desirable to improve dynamism so that some topology changes are able to allow for link failures and re-connections, especially for mobile decentralized networks. The dynamic topology capability is closely related to the link redundancy capability, because once the algorithm is capable of handling multiple paths redundantly, then the network can pick and choose among them dynamically. These capabilities are obtained while ensuring the scalability and correctness of the DDF network.

The k-tree topology is used to define an allowable topology; the allowable set of links for the decentralized network. Once this k-tree allowable topology is established, it defines which nodes can communicate on which links and establishes what each node needs to store and communicate in order to ensure correct and exact DDF, as will be described later in this section. This use of a defined restricted topology is similar to how spanning-tree algorithms can be used to define an allowable tree of links in an otherwise unstructured network for DDF [16,18]. A spanning-tree can then be used with tree-topology decentralized networks (as in Section 7.3.1), but these do not offer a simple, exact method for handling the data fusion aspects of changing topology or dealing with link failures.

Image

FIGURE 7.3 Example k-tree topologies. Each line is an allowable decentralized communication link, each vertex is a decentralized node. (a) A complete two-tree topology, (b) a mixed one-two-tree topology, (c) a one-tree topology over the same nodes, (d) a ring topology (black lines) is a subset of a two-tree (gray dashed lines), (e) a complete three-tree network, (f) a larger two-tree example showing the broad scale tree topology for Nk.

In a k-tree allowable topology (and using the k-tree algorithms presented here), nodes can communicate dynamically on all/any links within the k-tree topology, even if there are multiple redundant paths or loops and links fail and reconnect unpredictably.

A k-tree allowable topology becomes an arbitrary unrestricted topology for kN, the number of nodes. This would allow completely arbitrary dynamic and redundant decentralized communications, but would however result in expensive storage and communication.

The treewidth of a graph is well known for its role in limiting the complexity of algorithms in graph theory [3,13,14], graphical models [5], and sparse linear algebra [21]. Given the strong effect of the treewidth on the complexity of the algorithms and network, we considered generalizations of one-tree topologies into k-tree topologies, focusing in particular on the next-highest k; k = 2, since it is the simplest topology that demonstrates the novel properties of the k-tree approach. It is notable that arbitrarily large ring networks can be expressed as a two-tree network. Example k-tree topologies are shown in Figure 7.3.

A complete k-tree graph is made up of cliques of k + 1 nodes [13,14]. Each adjacent pair of cliques overlaps at k nodes (a junction or separator). The overall graph of connections between the cliques is a tree. A k-tree graph has treewidth of k, so called because the separators are made of k nodes.

Table 7.5 shows the number of links in various k-tree topologies compared with those of a fully connected topology. This shows that the number of k-tree links grows at O(n2) up until n = k + 1 (when the first k + 1 clique is formed), after which each additional node only adds an extra k links. Hence, the number of k-tree links grows as O(n) ultimately. By contrast, the fully connected topology always grows as O(n2).

Image

7.5.1    DECENTRALIZED DATA FUSION ON K-TREES

The decentralized algorithm defines what each node needs to store and communicate such that each node can obtain the global fused information. The algorithm actively limits the data sizes communicated and stored, leading to the scalable performance of the system. The goal of the topology and message passing is to produce a set of terms pi(x), such that the fusion of these is a consistent estimate for the state x:

p(x)=1cipi(x)

(7.71)

These pi(x) are probabilities which are conditionally independent of each other given x, or equivalently, that they have independent errors.

As shown earlier, it is convenient to express this as a sum of information terms:

Y=iYi

(7.72)

y=iyi

(7.73)

7.5.2    DATA-TAGGING SETS

The approach used here guarantees against double counting of information by using explicit “data-tagging” sets. A data-tagging set is a set of separate information terms, Yi, each with a unique identifier. Each data-tagging set stores only conditionally independent terms, so Equation 7.72 can be used on all items in a data-tagging set to recover a consistent fused estimate. Fusion of two or more sets is performed as a set union followed by Bayesian fusion (Equation 7.72). The set union step identifies any terms with matching labels and ensures that these are counted only once in the Bayesian fusion. Thus data-tagging avoids double counting of information.

The approach used here ensures scalability by summarizing every stored or communicated data-tagging set into a minimal size. This summarization process exploits the global k-tree property and uses the local topology around the sending and receiving nodes. The necessary local topology properties are guaranteed by designing the global network topology as a k-tree.

The proposed approach uses an efficient, minimal form of data-tagging. This is in contrast to the inefficient full data-tagging approach. In the full data-tagging method, each node maintains a set of independent information terms (conditionally independent of each other, given the true state), including its own sensor observations. In communicating out to any neighbor, the full set of information terms is sent. In receiving communication from a neighbor, the received set is merged (unioned) into the local set. The full data-tagging approach guarantees avoidance of double counting in arbitrary network topology and allows arbitrary dynamism, but is expensive for large-scale networks. Eventually every node’s storage and every communicated set have the full list of the conditionally independent information terms arising from every other node. In the full data-tagging approach, the node storage and communication size is O(n) for n nodes in the whole network. This increasing storage and communication size limits the scalability of the network for large n. The full data-tagging approach is equivalent to a k-tree operating with kn for n nodes in the network.

The proposed k-tree approach is obtained by reducing the data-tagging sets to exploit the tree nature of the communications network. The communications and storage scheme proposed achieves correct operation in k-tree networks without using full data-tagging, thus obtaining a decentralized algorithm which is scalable in the number of nodes.

The “stack” of channel cache terms, in Section 7.3.1, Figure 7.1 is actually a minimal data-tag set for the tree network. Each node has n + 1 entries corresponding to the n neighbors and a single entry for itself.

7.5.3    SEPARATOR AND NEIGHBORHOOD PROPERTIES

Before explaining the k-tree decentralized algorithm, it is necessary to discuss some properties of the k-tree.

7.5.3.1    Separator Property

An important k-tree property is the existence of tree separators, as shown in Figure 7.4. In a k-tree any k-clique is a separator. Each separator divides the network into distinct parts. Within each part, the effect of all other parts can be summarized into the separator. Separators enable efficient summarization of entire branches of the k-tree network. Separators use the k-tree separator property: in a k-tree, if any path between any two nodes i,k passes through the separator, then all paths between nodes i,k pass through the separator.

These separators are used at the borders of the local neighborhood L to summarize the fused total of the rest of the network beyond the local neighborhood. For example in Figure 7.4, the total information in each half can be expressed as

Image

FIGURE 7.4 Illustration of the separator property. In a k-tree any k-clique is a separator. Each separator divides the network into two parts. In this figure, bd is the separator. The two parts and the intersection are shown. Within each part, the effect of the other part can be summarized into the separator.

Yrhs={ Yinterior }+[ Yseparator ]

(7.74)

={ Yb+Ye+Yh+Yd+Yf }+[ Ybd ]

(7.75)

Ylhs={ Ya+Yb+Yc+Yd+Yg }+[ Ybd ]

(7.76)

where Ybd represents information in the separator bd.

The identifiers in the data-tagging sets are used to identify which node or branch of the tree network the information originates from. This means that the identifier should be a set of node labels, to allow reference to one neighbor, or k neighbors on a k-tree branch separator.

7.5.3.2    Local Neighborhood Property

A consequence of the separator property is that the local neighborhood around a node becomes a sufficient representation for that node’s interaction with the whole rest of the network. The k-tree networks allow an efficient decentralized and local neighborhood representation to serve as the only required topology awareness at the nodes. This is important for scalability, allowing the representation of a global network with only small local neighborhood representations. The local neighborhood is therefore an important data structure used in the algorithm proposed in this chapter.

At any node, Vi, the local neighborhood subgraph consists of Vi, the neighbors of Vi and the links and cliques between them, as shown in Figure 7.5.

The local neighborhood representation is motivated by the k-tree “junction path covering property”: in a k-tree, if any path between any two nodes i,k passes through the local neighborhood of a node j, then all paths between nodes i,k pass through the local neighborhood of j.

This junction path covering property means that the local neighborhood around a node j has control over how any messages can pass from one side to the other. The local neighborhood encodes which neighbors to communicate with, which information terms must be maintained separately in data-tag sets (for correctness), and which terms can be fused into others (for scalability).

Image

FIGURE 7.5 Illustration of the local neighborhood representation, L. In k-tree networks, the local neighborhood L is an efficient local summary of the relevant parts of the global topology: (a) global network topology, (b) local neighborhood representations, L, at a, b, e respectively.

For one-tree topologies used in prior works, the local neighborhood representation is simply the list of neighboring vertices and list of the corresponding edges to those neighbors.

7.5.4    K-TREE COMMUNICATIONS ALGORITHM

This section explains the decentralized communications algorithm for k-trees. We explain the algorithm in the case that the complete k-tree is present. Note, however, that the full set of links is not required.

The algorithm will be described by referring to the sending node, Vt (“transmitting vertex”) and the receiving node Vd (“destination vertex”). The transmitting node knows the topology of the allowable links within its own neighborhood of the allowable k-tree topology, denoted as L. The sending node has an existing data-tag set. The objective of the algorithm is to calculate a reduced data-tag set to send to the destination, Vd.

The communications algorithm is simply stated as follows:

•  The data-tag set is reduced into the intersection of the local and destination neighborhoods.

The communications algorithm is given in algorithm 1 and illustrated in Figure 7.6. In step 1, the algorithm initially copies the local data-tag set to the output data-tag set. This corresponds to the full data-tagging solution. The subsequent steps erase and/or summarize some of the entries, thus ensuring scalability. For step 2, data-tag terms involving the destination vertex are redundant and can be explicitly deleted. Step 3 eliminates any data-tag terms which are not neighbors of the destination vertex. This is explained in the following section.

Image

FIGURE 7.6 Summary of the communications algorithm. The local neighborhood at the source node is summarized into the neighborhood separator for the destination. This summarized separator set is sent to the destination and merged into the local set. (a) The full network, highlighting neighbor-hoods of Vh and Vm and their intersection. (b) At Vh the network beyond the immediate neighbors is already summarized within the neighborhood. (c) To prepare a communication set, the source Vh can summarize its local neighborhood set into the intersection with the destination neighbor Vm (resulting in the left hand set). This set is communicated to Vm. Vh keeps the union of the received set (left) with its local set (right).

Image

7.5.4.1    Data-Tag Set Elimination

This section explains the marginalization process which summarizes nonlocal information, in Algorithm 1.

Elimination proceeds at each step by eliminating a so-called leaf vertex, which reduces the size of the data-tag set. A leaf vertex in an ordinary tree would be any vertex with exactly one edge. More generally, however, there are k-tree leaves which are defined as follows: A vertex which is part of exactly one clique of k + 1 is a k-tree leaf.

Eliminating a k-tree leaf results in its former k + 1 clique being reduced to a clique of k. To eliminate a vertex v in a k + 1 clique

•  The result data-tag term r is that with identifier containing the k node labels of the neighbors of v

•  For each data-tag term, t, whose identifier contains v: add t into r

Examples of leaf vertex elimination for k ≤ 3 are shown in Table 7.6.

7.5.5    LINK AND NODE FAILURE ROBUSTNESS

The key properties of the proposed approach are correct fusion, scalability and robustness against node and link loss. Achieving these properties simultaneously is achieved by using the bounded treewidth network topology.

ALGORITHM 1: K-TREE DECENTRALIZED COMMUNICATIONS

Compute the communication output data-tag set to send

Input: L: a copy of the local neighborhood graph

Input: Vt: this transmitting node in L

Input: Vd: the destination neighbor in L

Input: localTags: the local data-tag set

Result: destTags: the output data-tag set to send

1. Starting case: No summarization:

Copy destTags ← localTags

2. Delete terms involving Vd:

Erase term Vd from destTags

Erase any terms for Vd separators from destTags

3. Summarize away parts not local to Vd:

Determine the region to summarize out, S:

S is all vertices in L except Vd and its neighbors

while S is not empty do

Find a leaf vertex V1 of L in S

Eliminate V1, updating destTags

Erase V1 from S

The proposed method is robust against link and node failures simply because it can send information terms on multiple paths. This still yields correct and consistent fusion since the method uses data-tagging to avoid double counting and/or the need for conservative fusion. Furthermore, the method still yields a scalable solution for large networks since the multiple-path and data-tagging is only performed within the nodes and separators of the k + 1 cliques.

Figure 7.7 shows the pattern of communication of individual information terms in the data-tag sets. In various cases, there are multiple sources redundantly communicating the same term. The receiving node always stores the incoming information terms into a given data-tag set entry. The receive process has no effect other than storing the information, so it is acceptable to receive the same term multiple times from different paths.

7.5.6    SUMMARY

This section presented an algorithm for scalable DDF based on k-tree topologies. The k-tree topologies are more densely connected than 1-trees, but still have an overall sparse (k-)tree topology which gives scalability for large networks. The k-tree topologies have some redundancy in the topology, which makes them more robust to node or link failures than 1-tree topologies. The k-tree topologies allow dynamic changes to the communications topology within subsets of links in the k-tree. Finally, k-tree topologies transition into the fully connected topology and fully data-tagged decentralized algorithm as k increases to N, the number of nodes in the network. Thus k-trees allow some trade-off via k between the tree-based approaches (kN) and the unstructured approaches (kN).

Image

FIGURE 7.7 Diagrams showing the individual data-tag terms which would be stored and communicated in the given topology. At each node the diagram shows the stored terms at that node (cluster of labels), including its own independent information (circled labels). Each arrow indicates the communication of an individual data-tag term. Terms which originate from each node are shown in different shades. Communications which result from the fusion of multiple terms are shown in dashed lines. Communications is strictly with nearest neighbors only, but the sum of all data-tag terms at each node equals the global sum of independent information.

7.6    CONCLUSION

This chapter presented and reviewed methods for DDF. This chapter focused on the channel cache algorithm for DDF in tree topologies and robustness to imperfect communications. In the second part, this chapter reviewed the trajectory state formulation of dynamic systems to formulate DDF for the estimation of dynamic systems. This trajectory state formulation of DDF was applied to address the issues of delayed and asequent observations and burst communications in DDF. In the final part, this chapter extended the operation of DDF on tree topology networks into so-called k-tree topologies. The k-tree topologies are tree-like on the broad scale, which gives good scalability for large networks of nodes. The k-tree topologies allow loops, dense connections, and hence redundancy and dynamic changes among groups of up to k + 1 nodes.

Taken together, these algorithms contribute significantly toward achieving DDF that is robust to communications latencies and failures, but still yield centralized equivalent estimator performance and are scalable for larger networks.

7.A APPENDIX

7.A.1 MARGINALIZATION IN THE INFORMATION FORM

This appendix states the expressions required for marginalization in the information form. Consider an information matrix partitioned into state variables xa and xc:

Y=(ABBTC)x^=(x^ax^c)y^=(ac)

(7.77)

where these satisfy YX^=y^

Then the marginal information matrix Ya and marginal information vector ya which satisfy YaX^a=y^a, and similarly for Yc are

Ya=ABC1BTya=aBC1cYax^a=y^a

(7.78)

Yc=CBTA1Byc=cBTA1aYcx^c=y^c

(7.79)

7.A.2 Trajectory Information Form Equivalence

As stated earlier, the following are both equivalent:

Y=(Yk+FTQ1FFTQ1Q1FQ1)y=(ykFTQ1BuQ1Bu)Pk+1=FPkFT+GQGTx^k+1=Fx^k+Buk

This can be shown in a few ways:

•  The joint X^ satisfies YX^=y, with X^=(x^kx^k+1)=(x^kFx^k+Bu):

YX^=(Yk+FTQ1FFTQ1Q1FQ1)(x^kFx^k+Bu)

(7.80)

=(Ykx^+FTQ1Fx^FTQ1{ Fx^k+Bu }Q1Fx^+Q1{ Fx^k+Bu })

(7.81)

=(Yk+FTQ1BuQ1Bu)

(7.82)

=y

(7.83)

•  The xk marginal of Y is equal to the prior Yk and yk. This means that augmenting a predicted state xk+1 onto a given xk system (including the addition of the dynamic model information) does not alter the marginal PDF for xk. This is shown as follows.

•  We write the xk marginal of Y as Ykmarg, leaving Yk to mean the prior information matrix of timestep k

Ykmarg={ Yk+FTQ1F }FTQ1{ Q1 }1Q1F

(7.84)

=Yk

(7.85)

•  The xk+1 marginal of Y yields known expressions [2,19] for the prediction in covariance and information forms:

Yk+1={ FPkFT+Q }1

(7.86)

=MMG(GTMG+Q1)1GTM

(7.87)

(M=FTYkF1)

(7.88)

The xk+1 marginal of Y, using Equation 7.79, is

Yk+1=Q1Q1F[ Yk+FTQ1F ]1FTQ1

(7.89)

Using the matrix inversion lemma:

[ BCD+A ]1=A1A1B[ C1+DA1B ]1DA1

(7.90)

With AQBFCYk1DFT

Yk+1=[ FYk1FT+Q ]1

(7.91)

=[ FPkFT+GQGT ]1

(7.92)

which is the covariance form prediction equation.

The information form prediction equation is obtained by a different use of the matrix inversion lemma, using

AFYk1FTBGCQDGT

Yk+1=[ FYk1FT+GQGT ]1

(7.93)

=MMG(Q1+GTMG)1GTM

(7.94)

where

M=[ FYk1FT ]1=FTYkF1

Equations 7.89, 7.92, and 7.94 can also be found systematically from the following augmented system [23]:

(Yk0F000I0FI0G00GTQ1)(xkxk+1vvk)=(yk0Buk0)

(7.95)

where ν is a vector of Lagrange multipliers [23] and vk is the (unknown) process noise, as in Equation 7.25.

•  Marginalizing (7.95) in the ordering (v, ν, then xk) results in the predicted information marginal:

Yk+1=Q1Q1F(Y+FTQ1F)1FTQ1

(7.96)

•  Marginalizing (7.95) in the ordering (xk, ν then v) results in the same predicted information marginal, in the form conventionally used in information filtering:

Yk+1=MMG(Q1+GTMG)1GTM

(7.97)

M=FTYF1

(7.98)

•  Marginalizing (7.95) in the ordering (xk, v then ν) results in the inverse of the expression used in the covariance form Kalman filtering:

Yk+1=(FPFT+GQGT)1

(7.99)

ACKNOWLEDGMENTS

The work in this chapter was supported by the Australian Centre for Field Robotics. The authors would also like to thank Tim Bailey and Chris Lloyd for related technical discussions.

REFERENCES

1.  T. Bailey and H. Durrant-Whyte. Decentralised data fusion with delayed states for consistent inference in mobile ad hoc networks. Technical report, 2007, http://www.personal.acfr.usyd.edu.au/tbailey/papers/delayedstated.pdf

2.  Y. Bar-Shalom, X. Rong Li, and T. Kirubarajan. Estimation with Applications to Tracking and Navigation. Wiley, Hoboken, NJ, 2001.

3.  B. Bollobás. Modern Graph Theory. Springer, New York, 1998.

4.  J. Capitan, L. Merino, F. Caballero, and A. Ollero. Decentralized delayed-state information filter (DDSIF): A new approach for cooperative decentralized tracking. Robotics and Autonomous Systems, 59(6):376–388, 2011.

5.  V. Chandrasekaran, N. Srebro, and P. Harsha. Complexity of inference in graphical models. In 24th Conference on Uncertainty in Artificial Intelligence and Statistics, Helsinki, Finland, pp. 70–78, 2008.

6.   K.C. Chang, C. Chong, and S. Mori. On scalable distributed sensor fusion. International Fusion, 2008 11th International conference Cologne, Germany, pp. 1–8, 2008.

7.  K. Chang, C.-Y. Chong, and S. Mori. Analytical and computational evaluation of scalable distributed fusion algorithms. IEEE Transactions on Aerospace and Electronic Systems, 46(4):2022–2034, October 2010.

8.  F. Dellaert and M. Kaess. Square root SAM: Simultaneous localization and mapping via square root information smoothing. International Journal of Robotics Research, 25(12):1181–1203, 2006.

9.  H. Durrant-Whyte, M. Stevens, and E. Nettleton. Data fusion in decentralised sensing networks. In 4th International Conference on Information Fusion, Montreal, Quebec, Canada, pp. 302–307, 2001.

10.  R. Eustice, H. Singh, and J. Leonard. Exactly sparse delayed-state filters. In Proceedings of the 2005 IEEE International Conference on Robotics and Automation, Barcelona, Spain, pp. 2417–2424, 2005.

11.  G.H. Golub and C.F. Van Loan. Matrix Computations, 3rd edn. The Johns Hopkins University Press, Baltimore, MD, 1996.

12.  S. Grime and H.F. Durrant-Whyte. Data fusion in decentralized sensor networks. Control Engineering Practice, 2(5):849–863, 1994.

13.  T. Kloks. Treewidth, computations and approximations. Lecture Notes in Computer Science. 1994.

14.  E. Korach and N. Solel. Tree-width, path-width and cut-width. Discrete Applied Mathematics, 43(1):97–101, 1993.

15.  J.J. Leonard and R J. Rikoski. Incorporation of delayed decision making into stochastic mapping. In International Symposium on Experimental Robotics, Montreal, Quebec, Canada, pp. 533–542, 2000.

16.  A. Makarenko, A. Brooks, T. Kaupp, H. Durrant-Whyte, and F. Dellaert. Decentralised data fusion: A graphical model approach. In Proceedings of the 12th International Conference on Information Fusion, Seattle, WA, pp. 545–554, July 2009.

17.  P.S. Maybeck. Stochastic models estimation and control. Mathematics in Science and Engineering, 1:423, 1979.

18.  E. Nettleton. Decentralised architectures for tracking and navigation with multiple flight vehicles. PhD thesis, Australian Centre for Field Robotics, Department of Aerospace, Mechanical and Mechatronic Engineering, The University of Sydney, Sydney, Australia, 2003.

19.  E.W. Nettleton and H.F. Durrant-Whyte. Delayed and asequent data in decentralised sensing networks. Proceedings of SPIE—The International Society for Optical Engineering, 4571:1–9, 2001.

20.  D. Nicholson, C.M. Lloyd, S.J. Julier, and J.K. Uhlmann. Scalable distributed data fusion. In Information Fusion, 2002. Proceedings of the Fifth International Conference, Annapolis, MD, Vol. 1, pp. 630–635, 2002.

21.  M.A. Paskin and G.D. Lawrence. Junction tree algorithms for solving sparse linear systems. Technical Report UCB/CSD-03-1271, University of California, Berkeley, CA, 2003.

22.  S. Sukkarieh, E. Nettleton, J.-H. Kim, M. Ridley, A. Goktogan, and H. Durrant-Whyte. The ANSER project: Data fusion across multiple uninhabited air vehicles. International Journal of Robotics Research, 22(7–8):505–539, 2003.

23.  P. Thompson. A novel augmented graph approach for estimation in localisation and mapping. PhD thesis, The University of Sydney, Sydney, Australia, March 2009.

24.  P. Thompson and H. Durrant-Whyte. Decentralised data fusion in 2-tree sensor networks. In Proceedings of the 13th International Conference on Information Fusion, Edinburgh, U.K., pp. 1–8, July 2010.

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset