Next Article in Journal
A Quantitative Comparison between Shannon and Tsallis–Havrda–Charvat Entropies Applied to Cancer Outcome Prediction
Next Article in Special Issue
Information Bottleneck Signal Processing and Learning to Maximize Relevant Information for Communication Receivers
Previous Article in Journal
Review and Update on Some Connections between a Spring-Block SOC Model and Actual Seismicity in the Case of Subduction Zones
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Distributed Quantization for Partially Cooperating Sensors Using the Information Bottleneck Method

by
Steffen Steiner
1,*,
Abdulrahman Dayo Aminu
2 and
Volker Kuehn
1
1
Institute of Communications Engineering, University of Rostock, 18119 Rostock, Germany
2
Pydro GmbH, 18119 Rostock, Germany
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(4), 438; https://doi.org/10.3390/e24040438
Submission received: 24 February 2022 / Revised: 16 March 2022 / Accepted: 18 March 2022 / Published: 22 March 2022
(This article belongs to the Special Issue Theory and Application of the Information Bottleneck Method)

Abstract

:
This paper addresses the optimization of distributed compression in a sensor network with partial cooperation among sensors. The widely known Chief Executive Officer (CEO) problem, where each sensor has to compress its measurements locally in order to forward them over capacity limited links to a common receiver is extended by allowing sensors to mutually communicate. This extension comes along with modified statistical dependencies among involved random variables compared to the original CEO problem, such that well-known outer and inner bounds do not hold anymore. Three different inter-sensor communication protocols are investigated. The successive broadcast approach allows each sensor to exploit instantaneous side-information of all previously transmitting sensors. As this leads to dimensionality problems for larger networks, a sequential point-to-point communication scheme is considered forwarding instantaneous side-information to only one successor. Thirdly, a two-phase transmission protocol separates the information exchange between sensors and the communication with the common receiver. Inspired by algorithmic solutions for the original CEO problem, the sensors are optimized in a greedy manner. It turns out that partial communication among sensors improves the performance significantly. In particular, the two-phase transmission can reach the performance of a fully cooperative CEO scenario, where each sensor has access to all measurements and the knowledge about all channel conditions. Moreover, exchanging instantaneous side-information increases the robustness against bad Wyner–Ziv coding strategies, which can lead to significant performance losses in the original CEO problem.

1. Introduction

This contribution considers a special case of the distributed source coding problem where each sensor observes the same source signal. In order to forward their measurements over capacity limited links to a common receiver, the sensors have to compress their measurements. In the case where a direct communication among operating sensors is not possible, this problem is termed as the CEO problem. Here, the compression at each sensor is optimized according to the Wyner–Ziv coding principle exploiting only statistical side-information. Within this paper, the term CEO problem always stands for this non-cooperative CEO problem, which means that sensors cannot communicate with each other during runtime.
We extend this scenario and allow sensors to cooperate with each other by exchanging instantaneous side-information. The fully cooperative Chief Executive Officer (fcCEO) problem is obtained if sensors can forward their uncompressed observations over inter-sensor links to all other sensors. The partially cooperative Chief Executive Officer (pcCEO) problem represents a scenario where instantaneous side-information is compressed before it is forwarded to other sensors.

1.1. The CEO Problem

The CEO problem has been investigated for various system assumptions which mainly differ in the distribution of the variable of interest and the applied distortion measure. The quadratic Gaussian CEO problem considers jointly Gaussian signals and the mean squared error (MSE) distortion measure [1,2,3,4,5,6]. Using an infinite number of encoders, Oohama analytically derived an asymptotic version of the sum-rate distortion in [1]. In [2], the authors investigated the influence of cooperating and non-cooperating encoders on the distortion measure. It turns out that the distortion decreases asymptotically with the reciprocal sum-rate R for non-cooperating encoder. In scenarios with cooperating sensors, the MSE distortion decays exponentially with 2 2 R . The non-asymptotic case was first investigated in [3] where the authors derived an upper bound on the sum-rate distortion function. Moreover, they showed that this bound is tight when each encoder has the same measurement SNR. Prabhakaran et al. [4], Oohama [5] and Wagner et al. [6] characterized the complete rate-region for this quadratic Gaussian CEO problem. Results also exist for multivariate Gaussian relevant processes [7,8]. The CEO problem for arbitrary discrete source distributions and the logarithmic loss distortion measure has been analyzed in [9,10,11]. For this scenario, Courtade and Weissman completely characterized the CEO rate-region in [9]. In [10,11], asymptotic analyses for an infinite number of sensors have been performed. Using the Hamming distance as a distortion measure, in [10] an inevitable loss in error rate performance due to non-cooperating sensors has been discovered. For arbitrary distortion measures, a scaling law on the sum-rate distortion function has been derived in [11]. In [12,13], the authors developed a variational bound on the optimal trade-off between relevance and complexity of the CEO setup and used neural networks for encoder and decoder to compute this bound.
There exist several algorithmic approaches to solve the CEO problem [8,14,15,16,17]. In our previous work [14,15], a Greedy Distributed Information Bottleneck (GDIB) algorithm based on the inner bound of the CEO problem with log-loss distortion measure defined in [9] has been introduced to determine extreme points of the contra-polymatroid solution space. Equivalently to Wyner–Ziv coding, the algorithm optimizes the quantizer of a specific sensor using the mappings of previously designed quantizers as statistical side-information. We showed that the GDIB algorithm outperforms an individual scalar information bottleneck optimization of sensors without Wyner–Ziv coding, especially for larger networks. For asymmetric scenarios, we demonstrated the dependency of the performance on the optimization order, i.e., the Wyner–Ziv coding strategy. Since the memory complexity of this optimization algorithm depends exponentially on the network size, we introduced a way to compress the Wyner–Ziv side-information by means of the information bottleneck principle. However, there still remains a large performance gap between non-cooperative and fully cooperative distributed compression.

1.2. Partially Cooperating Sensors

In order to close the gap between non-cooperative and fully cooperative distributed compression, we consider an extension of the CEO system model allowing partial cooperation among sensors. While a rich literature can be found on the classical CEO problem, less results are known for partially cooperating sensors. Most work has been done for jointly Gaussian signals because they allow an analytical treatment at least in parts. In [18], it is shown that cooperation among sensors can reduce the compression sum-rate except for the quadratic Gaussian CEO problem. In [19], the authors consider estimation problems under communication constraints and propose coding strategies for tree-structured sensor networks. Exploiting the Wyner–Ziv coding principle, the authors developed solutions for general trees and provide particular results for serial and parallel networks. A two sensor system with a Gaussian source was investigated in [20] for two transmission scenarios: orthogonal but rate-limited links between sensors and the common receiver as well as the Gaussian multiple access channel. It was shown that cooperation between the sensors over rate-limited inter-sensor links leads in both cases to substantial gains in terms of the compression sum-rate. Finally, a simple three node network consisting of encoder, helper and decoder was analyzed in [21]. The authors showed that side-information provided by the helper to the encoder need not exceed that given to the decoder.

1.3. Structure and Notation

This paper is structured as follows: Section 2 gives a short introduction to the information bottleneck principle. Section 3 and Section 4 introduce the non-cooperative distributed sensing scenario with the Greedy Distributed Information Bottleneck as an algorithmic solution defined in [14,15] and the fully cooperative distributed sensing scenario being equivalent to a centralized quantization approach, respectively. The main contribution of this paper can be found in Section 5, which introduces the partially cooperative distributed sensing scenario containing three different inter-sensor communication protocols, i.e., successive broadcasting, successive point to point transmission and two-phase transmission. Section 6 concludes this paper.
Throughout this paper, the following notation is used. Calligraphic letters 𝒳 , 𝒴 , Ƶ denote random variables with realizations x , y , z , which are elements of the sets 𝕏 , 𝕐 , with cardinalities | 𝕏 | , | 𝕐 | and | | , respectively. Bold letters y = [ y 1 y M ] T denote vectors while boldface calligraphic letters 𝓨 , Ƶ denote multivariate random variables. Note that Ƶ < m covers only the processes Ƶ 1 to Ƶ m 1 . I ( 𝒳 ; 𝒴 ) represents the mutual information between the random variables 𝒳 and 𝒴. Conditional and joint probability mass functions (pmfs) are termed p ( y | x ) and p ( x , y ) , respectively. The Kullback-Leibler (KL) divergence is given as D K L [ · · ] . Finally, the expectation of a function f ( x ) with respect to the random variable 𝒳 is denoted as 𝔼 𝒳 [ f ( 𝒳 ) ] .

2. The Information Bottleneck Principle

The information bottleneck (IB) principle was first introduced by Tishby et al. in [22,23] and defines a clustering framework based on information theoretic measures. An overview about algorithmic solutions for this basic optimization problem is given in [23,24]. The IB principle finds application in various fields in communications [25,26,27,28,29,30].
The general IB setup is depicted in Figure 1. It contains the relevant process 𝒳, a noisy observation 𝒴 of 𝒳 and a compressed version Ƶ of 𝒴. The IB approach aims to optimize the mapping p ( z | y ) in order to preserve as much information about the relevant process 𝒳 in Ƶ as possible. More precisely, it tries to maximize the relevant mutual information I ( 𝒳 ; Ƶ ) while fulfilling a rate constraint I ( 𝒴 ; Ƶ ) C . This general goal is summarized in Figure 2a. The optimization can be formulated as a maximization of the Lagrangian function
L IB = I ( 𝒳 ; Ƶ ) β I ( 𝒴 ; Ƶ ) .
It turns out to be a non-convex optimization problem, since I ( 𝒳 ; Ƶ ) and I ( 𝒴 ; Ƶ ) are both convex functions of the mapping p ( z | y ) . The parameter β is a trade-off parameter steering the focus between the preservation of relevant information and the compression of the observation. In the case of β = 0 , the focus only lies on preservation of relevant information. By increasing β the compression becomes more and more important up to the case of β . Here, the functional in (1) becomes maximal if I ( 𝒴 ; Ƶ ) = 0 , which means that all information is compressed to a single cluster. Therefore, the parameter β can be used to adjust the compression rate I ( 𝒴 ; Ƶ ) in order to fulfill a desired rate constraint I ( 𝒴 ; Ƶ ) C . Since the compression-rate curve is a monotonic increasing function in 1 β , a simple bi-section search can be applied. The optimization problem in (1) can be solved by taking the derivative with respect to the mapping p ( z | y ) and equating it to zero. It results in the implicit update equation
p ( z | y ) = e d β ( y , z ) z e d β ( y , z )
with
d β ( y , z ) = 1 β D K L [ p ( x | y ) p ( x | z ) ] log p ( z ) = 1 β 𝔼 𝒳 | y [ log p ( x | y ) p ( x | z ) ] log p ( z ) .
In (3), D K L [ p ( x | y ) p ( x | z ) ] denotes the Kullback–Leibler divergence. This implicit solution can be solved by an iterative Blahut–Arimoto-like algorithm.
In the case of focusing solely on preservation of relevant information with β = 0 , the optimization algorithm yields a deterministic clustering p ( z | y ) { 0 , 1 } . For β > 0 , the clustering p ( z | y ) [ 0 , 1 ] is generally stochastic. The IB method can easily be extended to multiple input values. A graphical tool for visualization are IB graphs [31]. Figure 2b illustrates an example where the observations y 1 , y 2 are compressed into the cluster index z. The trapezoid represents the IB compression with respect to the relevant variable written inside the trapezoid.

3. Non-Cooperative Distributed Sensing System

Figure 3 illustrates the CEO system model without communication among operating sensors. Here, M sensors observe noisy versions y m of the same relevant signal x. The measurement processes can be modeled as statistically independent memoryless channels (MCs).
Exemplarily, a measurement y m = x + w m represents the relevant signal x corrupted by zero mean white Gaussian measurement noise w m with measurement signal-to-noise-ratio (SNR) γ m = σ x 2 σ w m 2 , where σ x 2 , σ w m 2 denote signal and noise variances, respectively. In order to be able to forward the measurements over capacity limited links with capacities C 1 , , C M , each sensor has to compress its observations using a specific encoding process. More precisely, each sensor compresses its observations y m to a cluster index z m using the mapping p ( z m | y m ) leading to the Markov property:
p ( z , y , x ) = m = 1 M p ( z m | y m ) p ( y m | x ) p ( x ) .
The encoding process contains a second lossy compression step if the mapping p ( z m | y m ) is stochastic and lossless entropy coding if the mapping p ( z m | y m ) is deterministic. Therefore, a compressed version of the index z m is transmitted without any further loss to the common receiver. The optimization of p ( z m | y m ) for each sensor is done offline.
The mathematical analysis of the CEO problem and the structure of its rate-region for discrete input alphabets and the log-loss distortion measure was presented in [9] and exploits (4). It was proved that the extreme points of the contra-polymatroid solution space can be determined by greedy algorithms as the one described next. Since the communication among sensors during run-time is not possible in this approach, the solution represents a lower bound on the performance of cooperative distributed compression in this paper.
An algorithmic solution to solve the CEO problem has previously been proposed in [14,15] as the so called Greedy Distributed Information Bottleneck (GDIB) algorithm. It is based on the inner bound of the CEO rate-region for the logarithmic loss distortion measure [9] and optimizes the quantization at the sensors successively. Replacing the logarithmic loss function H ( 𝒳 | Ƶ ) by the relevant mutual information I ( 𝒳 ; Ƶ ) delivers the optimization problem
max P I ( 𝒳 ; Ƶ ) s . t . I ( 𝓨 𝕊 ; Ƶ 𝕊 | Ƶ 𝕊 ¯ ) m 𝕊 C m 𝕊 { 1 , 2 , , M } .
The set P =   p ( z 1 | y 1 ) p ( z M | y M ) defines the set of all mappings. According to [9], the compression rates I ( 𝓨 𝕊 ; Ƶ 𝕊 | Ƶ 𝕊 ¯ ) are supermodular set functions with respect to the sets 𝕊 [32], while the relevant information I ( 𝒳 ; Ƶ ) does not depend on 𝕊. Therefore, the greedy optimization structure of the GDIB algorithm is optimal and finds the extreme points of the solution space. It has to be emphasized that since the GDIB algorithm is based on the inner bound of the rate-region, it does not find the complete rate-region of the CEO problem. Following this approach, M IB related Lagrangian optimization problems are obtained, one for each sensor.
L GDIB ( 1 ) = I ( 𝒳 ; Ƶ 1 ) β 1 I ( 𝒴 1 ; Ƶ 1 )
L GDIB ( M ) = I ( 𝒳 ; Ƶ M | Ƶ < M ) β M I ( 𝒴 M ; Ƶ M | Ƶ < M )
Obviously, the optimization problem of the first sensor resembles the optimization problem for the scalar IB problem given in (1) since there is no predecessor. Subsequent sensors exploit the mappings of previously designed quantizers as statistical side-information leading to the well-known Wyner–Ziv coding strategy. Naturally, each Lagrange multiplier β m has to be chosen such that the corresponding compression rate fulfills the individual rate constraint I ( 𝒴 m ; Ƶ m | Ƶ < m ) C m . The objectives in (6) and (7) can be solved by equating the derivative with respect to the mapping p ( z m | y m ) to zero delivering the update rule
p ( z m | y m ) = e d β m ( y m , z m ) z m e d β m ( y m , z m )
with the exponent
d β m ( y m , z m ) E Ƶ < m | y m 1 β m D K L [ p ( x | y m , z < m ) p ( x | z m ) ] log p ( z m | z < m ) .
Similar to the scalar IB optimization, the implicit expression in (8) can be solved using a Blahut–Arimoto like algorithm, providing local optimal solutions. It has to be mentioned that for asymmetric scenarios, this optimization has to be performed for all M ! possible permutations of the optimization order to find the best solution.
A detailed derivation and performance analysis of this algorithm can be found in [14,15]. If the capacity is equally distributed over all sensors in the network, e.g., sensors share the same channel in an orthogonal way and a round robin fashion, numerical results demonstrate that the GDIB algorithm outperforms an individual scalar IB optimization at each sensor. However, there is still a large gap to the performance of a fcCEO scenario, which is defined in Section 4. Moreover, in asymmetric scenarios, the performance highly depends on the optimization order. Although no clear conclusion about the optimal Wyner–Ziv coding strategy can be drawn, a good solution can be expected when starting the optimization with the best forward channel conditions, i.e., the lowest compression (highest compression rate).

4. Fully Cooperative Distributed Sensing—A Centralized Quantization Approach

This section introduces the fcCEO scenario, which considers distributed sensors being able to forward their uncompressed observations to all other sensors in the network over ideal noiseless inter-sensor links. In this case, sensors can perfectly exchange their measurements y m before they jointly compress the received signals taking into account the rate constraints of all individual forward channels. Naturally, the exchange has to be done by a two-phase transmission protocol, consisting of a cooperation phase and a transmission phase. During the cooperation phase sensors exchange information until every sensor knows measurements y = [ y 1 y M ] T of all M sensors. The actual forwarding of the compressed observations to the common receiver is performed during the transmission phase. This full cooperation is equivalent to a single central quantizer having access to all measurements y as depicted in Figure 4. Applying the IB principle, this central quantizer can be designed in order to compress the vector y onto a cluster index z using the mapping p ( z | y ) , which motivates the name centralized IB (CIB) for the algorithmic solution in a fcCEO scenario. The optimization problem can be formulated as the maximization of
L CIB = I ( 𝒳 ; Ƶ ) β I ( 𝓨 ; Ƶ )
and is solved using update Equation (2) with (3) substituting the scalar y by vector y. The number of output clusters | | has to be chosen to | | = m = 1 M | m | while the single link from the imaginary central quantizer to the receiver in Figure 4 has a channel capacity of C sum = m = 1 M C m . The actual transmission over the M links has to be coordinated such that each sensor m transmits a specific part of the bits corresponding to its link capacity C m .
In the special case of the measurement process being modeled as additive noise, the algorithm can be simplified to a scalar optimization problem where maximum ratio combining of all inputs y m delivers a scalar sufficient statistics
y ¯ = m = 1 M γ m · y m
of the desired relevant signal x with an overall SNR γ = m γ m . The solution of the fcCEO scenario serves as an upper bound in this paper.

5. Partially Cooperative Distributed Sensing

In order to investigate how the gap between non-cooperative and fully-cooperative distributed compression can be reduced, partially cooperating sensors shall now be considered. Partial cooperation means a limited exchange of instantaneous side-information among the sensors during runtime due to a rate-limitation of inter-sensor links. Non-cooperative CEO and fully-cooperative CEO problems represent the extreme cases for zero rate and unlimited rate inter-sensor links, respectively. The rate limitation requires the compression of instantaneous side-information before forwarding it to other sensors.
In this paper, only deterministic mappings are considered for this compression, while indexes z m are still obtained by stochastic mappings. This is motivated by the fact that deterministic mappings do not require further lossy compression and the resulting side-information indices s m can be exploited at other sensors by choosing a particular mapping p ( z m | y m ) from a list of possible mappings designed offline in advance. As a consequence, the compression rates for instantaneous side-information can only be adjusted by changing the cardinalities | 𝕊 m | . For all results presented below, inter-sensor links are modeled as bit pipes being able to deliver s m reliably.
The GDIB algorithm to solve the non-cooperative CEO problem is based on the inner bound (5) of the CEO rate-region. Moreover, a greedy optimization approach is optimal due to the supermodularity of the compression rates in (5). Both require the Markovian structure in (4). However, cooperation among sensors changes the Markovian structure and implies different statistical dependencies among involved random variables. As (4) does not hold anymore in pcCEO scenarios, the inner bound on the rate-region in (5) cannot be utilized to find solutions of the pcCEO scenario. To the knowledge of the authors, tight bounds on the rate-region are not available for the cooperative case. Therefore, a heuristic approach based on the greedy optimization structure of the GDIB algorithm will be applied to solve the pcCEO scenario, which is not proven to be optimal. Nevertheless, the numerical evaluation of the found solutions demonstrate their usefulness. However, the computation of required pmfs becomes more challenging and results in recursive calculations given in Appendix A.1 and Appendix A.2 because the Markovian structure of (4) does not hold anymore in pcCEO scenarios.
This paper introduces three different inter-sensor communication protocols for exchanging this instantaneous side-information: successive broadcasting, a successive point-to-point transmission and a two-phase transmission. The first two protocols perform the exchange of instantaneous side-information s m with other sensors and the forwarding of compressed versions of z m to the common receiver in the same time slot. Contrarily, the two-phase transmission protocol separates the exchange of instantaneous side-information among sensors and the communication with the common receiver into two distinct phases. The latter starts after the exchange among sensors has been completed such that all sensors have (approximately) the same amount of side-information.

5.1. Successive Broadcasting Protocol

The system model for the successive broadcasting protocol is illustrated in Figure 5. In the same time slot, sensor m 1 not only forwards a compressed version of the quantization index z m 1 to the common receiver, but also broadcasts instantaneous side-information s m 1 to all other sensors. However, due to the greedy optimization structure, only subsequent sensors can exploit this instantaneous side-information. Thus, sensor m can exploit indices s < m of all previously transmitting sensors in order to select its quantization index z m as well as a new instantaneous side-information index s m . This scenario leads to the Markov model
p ( x , y , z , s ) = m = 1 M p ( z m | y m , s < m ) p ( s m | y m , s < m ) p ( y m | x ) p ( x ) .
The indices s m and z m are obtained by deterministic mappings p ( s m | y m , s < m ) { 0 , 1 } and stochastic mappings p ( z m | y m , s < m ) [ 0 , 1 ] , respectively. The design of these mappings can be performed offline leveraging the IB principle as illustrated in Figure 6. It combines the observation y m and the instantaneous side-information s < m to indexes s m and z m while maintaining as much information as possible about the relevant signal x.

5.1.1. Generation of Broadcast Side-Information

The design of p ( s m | y m , s < m ) is inspired by the general GDIB algorithm, i.e., the optimization is done in a greedy manner. Again, there emerges one optimization problem for each sensor:
L BC - SIDE ( 1 ) = I ( 𝒳 ; 𝒮 1 ) β I ( 𝒴 1 ; 𝒮 1 )
L BC - SIDE ( M 1 ) = I ( 𝒳 ; 𝒮 M 1 | 𝓢 < M 1 ) β I ( 𝒴 M 1 ; 𝒮 M 1 | 𝓢 < M 1 ) .
The optimization problem of the first sensor equals the individual scalar optimization without any side-information at all, as described in Section 2. Subsequent sensors combine the instantaneous side-information of all previously transmitting sensors s < m with its observation y m . The relevant mutual information and the compression rate of sensor m are conditioned on 𝓢 < m since broadcasting instantaneous side-information ensures all successive sensors to have access to s < m allowing Wyner–Ziv coding for generating s m . Each optimization problem given in (12) and (13) can be solved by taking the derivative with respect to the mapping p ( s m | y m , s < m ) and equating it to zero. This results in the implicit update equation
p ( s m | y m , s < m ) = e d β m ( y m , s m , s < m ) s m e d β m ( y m , s m , s < m )
with
d β m ( y m , s m , s < m ) 1 β m D K L [ p ( x | y m , s < m ) p ( x | s m ) ] log p ( s m | s < m ) .
As in the general GDIB algorithm, the implicit update equation in (14) can be solved using a Blahut–Arimoto like algorithm resulting in local optimal solutions.

Algorithmic pcCEO Solution for the Successive Broadcasting Protocol

After designing the mapping for the instantaneous side-information, the mapping p ( z m | y m , s < m ) can be optimized, again by means of the IB principle. Therefore, the original GDIB algorithm is modified to exploit the broadcasted instantaneous side-information, defining the GDIB-BC algorithm. The optimization problem for each sensor is given as
L GDIB - BC ( 1 ) = I ( 𝒳 ; Ƶ 1 ) β 1 I ( 𝒴 1 ; Ƶ 1 )
L GDIB - BC ( M ) = I ( 𝒳 ; Ƶ M | Ƶ < M ) β M I ( 𝒴 M , 𝓢 < M ; Ƶ M | Ƶ < M ) .
The main difference to the original GDIB optimization problem in (6) and (7) lies in the definition of the compression rate I ( 𝒴 m , 𝓢 < m ; Ƶ m | Ƶ < m ) which emerges from the combination of the observation y m and the instantaneous side-information s < m . Taking the derivative of the optimization problem for sensor m with respect to the mapping p ( z m | y m , s < m ) and equating it to zero delivers
p ( z m | y m , s < m ) = e d β m ( y m , z m , s < m ) z m e d β m ( y m , z m , s < m )
with
d β m ( y m , z m , s < m ) E Ƶ < m | y m , s < m [ 1 β m · D K L [ p ( x | y m , s < m , z < m ) p ( x | z m ) ] log p ( z m | z < m ) ] .
Again, the implicit update equation in (18) can be solved using a Blahut–Arimoto like algorithm. The extended Blahut–Arimoto like algorithm to design the mapping p ( z m | y m , s < m ) of sensor m for a specific Lagrange parameter β m and instantaneous side-information s < m is given in Algorithm 1. The input pmf p ( y m 1 , s < m 1 , z < m 1 , x ) can be computed during the optimization of previous sensors. Lines 3 to 5 determine the required pmfs for the calculation of the KL-divergence of (19) in lines 6 to 9. The statistical distance of (19) is determined in lines 10 to 14. It is used to update the quantizer mapping p ( z m | y m , s < m ) of sensor m. This procedure is repeated until no significant changes of the desired mappings occur anymore. The algorithm returns the updated mapping p ( z m | y m , s < m ) as well as the pmf p ( y m , s < m , z < m , x ) , which is used as an input for the successive sensor.
Algorithm 1: Extended Blahut–Arimoto algorithm for broadcast cooperating sensors.
Entropy 24 00438 i001
The parameter β m , which determines the compression rate at sensor m, has to be adjusted such that I ( 𝒴 m , 𝓢 < m ; Ƶ m | Ƶ < m ) C m is fulfilled. Similar to the original GDIB algorithm, the GDIB-BC algorithm has to be performed for each sensor and all possible optimization orders.

5.1.2. Evolution of Instantaneous Side-Information

Figure 7 illustrates the amount of instantaneous side-information available at the different sensors in a network of size M = 6 considering the broadcast of side-information. It depicts the relevant mutual information I ( 𝒳 ; 𝓢 m ) versus the sensor number m for different cardinalities | 𝕊 m | and SNRs γ m . The relevant signal is chosen to be a uniformly distributed 4-ASK signal leading to | 𝕏 | = 4 . As expected, the amount of available instantaneous side-information increases with each additional sensor for all | 𝕊 m | and γ m . To be more specific, the resolution and the quality of instantaneous side-information available at sensor m increases with growing m. In the considered symmetric scenario, the amount of information I ( 𝒳 ; 𝒮 m | 𝓢 < m ) a sensor can contribute to I ( 𝒳 ; 𝓢 m ) gets smaller for each additional sensor and the slopes of the curves decrease. Since one bit is not enough to represent the information of | 𝕏 | = 4 , the largest gain can be observed between | 𝕊 m |   = 2 and | 𝕊 m |   = 4 . Increasing the cardinality further to | 𝕊 m |   = 8 results only in a small additional improvement. Certainly, this observation depends on the relevant signal 𝒳 and can not be generalized.

5.1.3. Performance for Different Network Sizes

Figure 8 and Figure 9 illustrate the overall performance of the GDIB-BC approach when broadcasting instantaneous side-information for different network sizes. The relevant mutual information I ( 𝒳 ; Ƶ ) is depicted versus the number of sensors M in the network. The gray colored area represents the non-achievable region, since I ( 𝒳 ; Ƶ ) cannot exceed I ( 𝒳 ; 𝓨 ) due to the data-processing inequality. Both figures consider a scenario where all sensors in the network share the same channel to the common receiver with a fixed sum-rate C sum in an orthogonal way and a round robin fashion. Consequently, larger network sizes correspond to smaller individual capacities C m = C sum M for each forward link. The performance of partially cooperating sensors broadcasting instantaneous side-information (pcCEO-BC) is compared to the non-cooperative case (CEO) of Section 3 and the fully cooperative case (fcCEO) of Section 4. As already mentioned, these two scenarios provide upper and lower bounds. In general, it can be observed that increasing the number of sensors in the network also increases the overall relevant mutual information I ( 𝒳 ; Ƶ ) . This holds even for the case without cooperation, since each sensor applies Wyner–Ziv coding and exploits the mapping of previously designed quantizers as statistical side-information [14]. Independent of the cardinality | 𝕊 m | , the performance of the pcCEO-BC scenario is superior to the case without cooperation among sensors. This difference grows for larger network sizes because the amount of information s < m has about the relevant variable x increases. As expected from Figure 7, increasing the cardinality | 𝕊 m | not only improves the relevant information I ( 𝒳 ; 𝓢 m ) , but also the overall performance measured by I ( 𝒳 ; Ƶ ) . However, it can be observed that even for large | 𝕊 m | there remains a gap to the fcCEO upper bound, especially for smaller network sizes or lower SNRs. This gap can be explained by the successive transmission protocol resulting in a gradually increasing amount of instantaneous side-information at the sensors. For instance, the first sensors does not profit at all from the partial cooperation in contrast to the fcCEO scenario where all sensors exploit almost the same amount of side-information.
Considering the pmfs in Algorithm 1, it becomes obvious that larger networks might suffer from the curse of dimensionality. More precisely, pmfs like p ( y m , s < m , z < m , x ) can become very large during the optimization for larger network sizes. Moreover, the mapping p ( z m | y m , s < m ) also depends on the network size, i.e., this problem does not only occur during the optimization, but also when storing the already optimized mapping. This numerical issue is the reason why there is no result for | 𝕊 m |   = 8 and a network size of M = 6 in Figure 8 and Figure 9. In this case, it requires 2024 GiB (1 GiB = 1024 MiB, 1 MiB = 1024 KiB, 1 KiB = 1024 byte) just for storing a single instance of the pmf p ( y m , s < m , z < m , x ) .

5.2. Successive Point-to-Point Protocol

For larger network sizes, broadcasting side-information might not be feasible anymore, since the dimensions of the mappings p ( z m | y m , s < m ) and p ( s m | y m , s < m ) as well as intermediate pmfs used within the optimization become huge. In order to relax this curse of dimensionality, the successive way of cooperation is exploited and the instantaneous side-information of sensor m shall only be forwarded to the direct successor m + 1 as depicted in Figure 10. Hence, a sequential chain is established from the first to the last sensor leading to the Markov Model:
p ( x , y , z , s ) = m = 1 M p ( z m | y m , s m 1 ) p ( s m | y m , s m 1 ) p ( y m | x ) p ( x ) .
Again, the instantaneous side-information is obtained by a deterministic mapping optimized by means of the information bottleneck principle, illustrated in Figure 11. With each step in the sequential chain, the information s m has about the relevant signal x increases.

5.2.1. Generation of Point-to-Point Side-Information

Similar to the broadcast case, the design of p ( s m | y m , s m 1 ) is inspired by the original GDIB algorithm. The optimization problem can be formulated in a greedy manner as
L PTP - SIDE ( 1 ) = I ( 𝒳 ; 𝒮 1 ) β I ( 𝒴 1 ; 𝒮 1 )
L PTP - SIDE ( M 1 ) = I ( 𝒳 ; 𝒮 M 1 ) β I ( 𝒴 M 1 , 𝒮 M 2 ; 𝒮 M 1 ) ,
where Equation (21) equals the individual scalar optimization without any side-information. Subsequent sensors combine the instantaneous side-information s m 1 sent by the previous sensor with its observation y m . In contrast to the broadcast case, the relevant mutual information is not conditioned on 𝓢 < m as in (12) and (13) because sensor m will only have access to s m 1 and not to indices of any other sensor. Therefore, Wyner–Ziv coding cannot be applied for exchanging instantaneous side-information with the successive point-to-point protocol. The optimization problems can be solved by taking the derivative with respect to the mapping p ( s m | y m , s m 1 ) and equating it to zero, resulting in the implicit update equation
p ( s m | y m , s m 1 ) = e d β m ( y m , s m , s m 1 ) s m e d β m ( y m , s m , s m 1 )
with
d β m ( y m , s m , s m 1 ) 1 β m · D K L [ p ( x | y m , s m 1 ) p ( x | s m ) ] log p ( s m ) .
As in the broadcast case, using a Blahut–Arimoto like algorithm to solve the update Equation (23) results in local optimal solutions.

5.2.2. Algorithmic pcCEO Solution Applying the Successive Point-to-Point Protocol

After the optimization of the mapping for instantaneous side-information, the mapping p ( z m | y m , s m 1 ) can be designed by means of the information bottleneck principle. Inspired by the original GDIB algorithm, the optimization problem can be formulated as
L GDIB - PTP ( 1 ) = I ( 𝒳 ; Ƶ 1 ) β 1 I ( 𝒴 1 ; Ƶ 1 )
L GDIB - PTP ( M ) = I ( 𝒳 ; Ƶ M | Ƶ < M ) β M I ( 𝒴 M , 𝒮 M 1 ; Ƶ M | Ƶ < M ) .
The main difference to the original GDIB optimization problem in (6) and (7) now lies in the compression rate I ( 𝒴 m , 𝒮 m 1 ; Ƶ m | Ƶ < m ) , which emerges from the combination of the instantaneous side-information s m 1 of the previous sensor m 1 and the observation y m of sensor m. Taking the derivative with respect to the mapping p ( z m | y m , s m 1 ) and equating it to zero, the optimization problem for sensor m can be solved, leading to the implicit update equation
p ( z m | y m , s m 1 ) = e d β m ( y m , z m , s m 1 ) z m e d β m ( y m , z m , s m 1 )
with
d β m ( y m , z m , s m 1 ) E Ƶ < m | y m , s m 1 [ 1 β m · D K L [ p ( x | y m , s m 1 , z < m ) p ( x | z m ) ] log p ( z m | z < m ) ] .
Thus, the mapping p ( z m | y m , s m 1 ) can be optimized using a Blahut–Arimoto-like algorithm. The specific algorithm for a given sensor m and a Lagrange parameter β m is given in Algorithm 2. The input pmfs p ( z i | y i , s i 1 ) i < m and p ( s i | z i , x ) i < m as well as p ( z < m 1 , x ) are calculated in advance by previous sensor optimizations. Lines 3 to 7 calculates required pmfs as given the Appendix A.2. The KL-divergence is calculated in lines 8 to 11. Using this, the statistical distance d β m ( z m , y m , s m 1 ) of (28) can be calculated in lines 12 to 16, which is then used to update the quantizer mapping p ( z m | y m , s m 1 ) . The algorithm stops if this mapping does not change significantly anymore during subsequent iterations. Finally, the output pmfs p ( s m | z m , x ) and p ( z < m , x ) need to be calculated in lines 21 to 25 for their usage in the optimization of the next sensor.
Similar to the original GDIB algorithm, the optimization needs to be done for all possible optimization orders. A simple bisection search can be applied to find the rate-fulfilling parameter β m , such that I ( 𝒴 m , 𝒮 m 1 ; Ƶ m | Ƶ < m ) C m holds.

5.2.3. Evolution of Instantaneous Side-Information

Figure 12 illustrates the amount of instantaneous side-information I ( 𝒳 ; 𝒮 m ) at a specific sensor m in a network of size M = 6 using the successive point-to-point transmission protocol for different cardinalities | 𝕊 m | . Obviously, I ( 𝒳 ; 𝒮 m ) increases with each further sensor. The main difference to the broadcast case is that the instantaneous side-information provided to sensor m is represented by a single highly compressed index s m 1 with cardinality | 𝕊 m 1 | . While the resolution | 𝕊 < m | of the available instantaneous side-information s < m increases with m in the broadcast case, it remains the same for the successive point-to-point protocol. Therefore, a higher cardinality | 𝕊 m | is required compared to the broadcast case to avoid additional compression losses.
Figure 12. Evolution of I ( 𝒳 ; 𝒮 m ) for sensor m in a network with M = 6 sensors and different cardinalities | 𝕊 m | for the successive point-to-point transmission protocol; | 𝕏 | = 4 , | 𝕐 m |   = 64 .
Figure 12. Evolution of I ( 𝒳 ; 𝒮 m ) for sensor m in a network with M = 6 sensors and different cardinalities | 𝕊 m | for the successive point-to-point transmission protocol; | 𝕏 | = 4 , | 𝕐 m |   = 64 .
Entropy 24 00438 g012
Algorithm 2: Extended Blahut–Arimoto algorithm for the successive point-to-point protocol.
Entropy 24 00438 i002

5.2.4. Performance for Different Network Sizes

Figure 13 and Figure 14 illustrate the overall performance of the pcCEO system with point-to-point exchanged instantaneous side-information where all sensors share the same channel to the common receiver with a fixed sum-rate C sum = m = 1 M C m in an orthogonal way and a round robin fashion. Again, the black curves represent the non-cooperative CEO scenario and the fcCEO scenario. Hence, they serve as lower and upper bound, respectively. In general, the curves are very similar to those for broadcasting instantaneous side-information in Figure 8 and Figure 9. Independent of the SNR or the sum-rate C sum , the relevant mutual information I ( 𝒳 ; Ƶ ) increases for larger networks and even a single bit as instantaneous side-information | 𝕊 m |   = 2 leads to slight improvements compared to the non-cooperative case. However, there still remains a gap to the fcCEO scenario even for large | 𝕊 m | , which results from the successive communication strategy, since sensors at the beginning of the optimization chain can exploit no or little instantaneous side-information.

5.2.5. Performance for Different Sum-Rates

Figure 15 illustrates the influence of the sum-rate C sum = m = 1 M C m for a scenario with M = 5 sensors. Naturally, larger sum-rates correlate with higher individual link capacities. Again, CEO and fcCEO scenarios provide lower and upper bounds, respectively. For a cardinality of | 𝕊 m |   = 2 , only a small gain compared to the non-cooperative CEO scenario can be observed. However, the gain gets more and more significant with increasing | 𝕊 m | . Comparing the results to the upper fcCEO bound illuminates the loss due to limited available side-information at early transmitting sensors. The largest difference can be observed for sum-rates between 2 C sum 4 bit/s/Hz.

5.2.6. Asymmetric Scenarios

A very important part is the investigation of asymmetric scenarios. As the achievable relevant information I ( 𝒳 ; Ƶ ) of a non-cooperative CEO scenario is very sensitive to the optimization order, i.e., the Wyner–Ziv coding strategy, in asymmetric scenarios [14] the question arises if the exchange of instantaneous side-information can improve the robustness against bad optimization orders. Therefore, the same two asymmetric setups as in [14] are analyzed. Scenario 1 considers the case where sensors with low SNRs γ m have low link capacities C m while sensors with high SNRs γ m have high link capacities C m . Scenario 2 considers the opposite case, where sensors with low SNRs have high link capacities and vice versa.
Figure 16 illustrates the relevant mutual information I ( 𝒳 ; Ƶ ) for all M ! = 24 sensor permutations for a network of M = 4 sensors. The dots represent the results from [14] for a non-cooperative CEO scenario. Blue dots show Scenario 1 while the red dots represent Scenario 2. The results for the pcCEO scenario with successive point-to-point side-information exchange is depicted as bars.
Comparing the non-cooperative case with the successive point-to-point exchange of side-information for Scenario 1, we observe a slight increase of the overall relevant mutual information I ( 𝒳 ; Ƶ ) for partial cooperation and this particular scenario. Moreover, the influence of the Wyner–Ziv coding strategy (optimization order) becomes smaller due to cooperation. The performance for Scenario 2 is worse than the performance for Scenario 1, again for both the cooperative and the non-cooperative case. In this scenario, accurate measurements have to be strongly compressed in order to forward them to the common receiver while unreliable measurements cannot contribute much to the overall performance although they can be forwarded to the common receiver at high rates. However, the loss due to bad optimization orders is much lower for partial cooperation. A sensor with a bad forward channel and a high SNR can still forward its information to the next sensor, which might have a better forward channel. Therefore, exchanging instantaneous side-information can improve the robustness against bad optimization orders.

5.3. Two-Phase Transmission Protocol with Artificial Side-Information

Previous subsections revealed that partial cooperation by exchanging instantaneous side-information improves the overall performance. However, a gap to the fcCEO scenario still remains, and we claimed that the successive exchange of instantaneous side-information is the reason for this difference. Due to the sequential forwarding protocols considered so far, early sensors have no or little instantaneous side-information. They hardly profit from the cooperation as opposed to the full cooperation case where all sensors have access to the complete information. In order to substantiate this statement, a third transmission protocol consisting of two phases is considered. Inspired by the fcCEO scenario, the first cooperation phase is used to exchange instantaneous side-information between all sensors, while the transmission phase is used to forward the information to the common receiver in the usual way. The difference to the fcCEO scenario is that only compressed versions of the observations can be exchanged during the cooperation phase.
For simplicity, we assume that each sensor obtains the same instantaneous side-information represented by s * , independent of its position in the optimization chain, see Figure 17. Moreover, we pursue the EXIT chart philosophy [33], where extrinsic information is artificially created to analyze the information exchange between decoders in concatenated coding schemes. In the pcCEO context, the artificial side-information can be interpreted as extrinsic information about the relevant signal x being generated by adding AWGN to x. The noise variance is adapted to obtain a specific SNR γ extr or equivalently a desired mutual side-information I ( 𝒳 ; 𝒮 * ) . It has to be emphasized that γ extr can be chosen independently from the measurement SNRs at the sensors in order to obtain general conclusions. Since the instantaneous side-information is created artificially, s * is assumed to be independent of the indexes y m given the relevant signal x, i.e., p ( y m , s * | x ) = p ( y m | x ) p ( s * | x ) holds. This simplifies the Markovian structure of the optimization problem which equals the one of the original CEO problem. With the same argumentation as in the original CEO problem, we claim that the supermodularity holds and the greedy optimization structure is optimal. This model leads to the modified optimization problem
L GDIB - TP ( 1 ) = I ( 𝒳 ; Ƶ 1 ) β 1 I ( 𝒴 1 , 𝒮 * ; Ƶ 1 )
L GDIB - TP ( M ) = I ( 𝒳 ; Ƶ M | Ƶ < M ) β M I ( 𝒴 M , 𝒮 * ; Ƶ M | Ƶ < M ) .
The optimization problem can be solved using the same strategy as described in previous Section 5.1 and Section 5.2. This leads to the implicit update equation
p ( z m | y m , s * ) = e d β m ( y m , z m , s * ) z m e d β m ( y m , z m , s * )
with
d β m ( y m , z m , s * ) E Ƶ < m | y m , s * [ 1 β m · D K L [ p ( x | y m , s * , z < m ) p ( x | z m ) ] log p ( z m | z < m ) ] .

5.3.1. Performance of Two-Phase Transmission

Figure 18 illustrates the same experiment as in Figure 8 or Figure 13, but for the two-phase transmission protocol. The extrinsic information is chosen independent of the measurement SNR and has its own SNR γ ext represented by different colors in Figure 18. The cardinality of the extrinsic information is chosen as | 𝕊 * |   = 512 to not introduce any compression losses. As before, the black dashed line represents the fcCEO scenario. The curve for γ extr = γ m = 8 dB represents the case where each sensor forwards instantaneous side-information whose quality corresponds to its measurement SNR. We observe the same performance as for the fcCEO scenario. This demonstrates that the remaining performance gap to the fcCEO scenario disappears completely for appropriate cooperation among sensors. Naturally, decreasing the SNR of the extrinsic information γ ext or equivalently I ( 𝒳 ; 𝒮 * ) leads to a lower overall performance I ( 𝒳 ; Ƶ ) .

5.3.2. Influence of Extrinsic Information

The influence of extrinsic information is depicted in Figure 19 for γ m = 8 dB and γ m = 3 dB. Therefore, the overall relevant mutual information I ( 𝒳 ; Ƶ ) is depicted versus the mutual information of the extrinsic information I ( 𝒳 ; 𝒮 * ) for different network sizes. As before, all sensors share the same forward channel with C sum = 2.5 bit/s/Hz and C m = C sum M . Providing no extrinsic information, i.e., I ( 𝒳 ; 𝒮 * ) = 0 delivers the same result as the non-cooperative CEO scenario of Section 3. Naturally, enhancing the quality of the extrinsic information increases the overall relevant mutual information I ( 𝒳 ; Ƶ ) up to the maximum of 2 bit/s/Hz.

6. Conclusions

This paper extends the non-cooperative CEO scenario allowing partial cooperation among sensors in the network. Therefore, it extends the algorithmic solution introduced in [14] for three different inter-sensor communication protocols: successive broadcasting, successive point-to-point communication and a two-phase transmission protocol. The first two protocols perform the exchange of instantaneous side-information and forwarding information to the common receiver at the same time step. Therefore, successive broadcasting exploits the instantaneous side-information of all previous sensors within the optimization chain. Since this may cause dimensionality problems during the optimization, the successive point-to-point transmission protocol forwards the instantaneous side-information only to the next sensor. It turns out that allowing this partial communication outperforms the non-cooperative compression where no communication among sensors is possible. Moreover, cooperative compression shows a larger robustness to suboptimal Wyner–Ziv coding strategies in asymmetric scenarios. However, a small performance gap to the fcCEO scenario still remains for the proposed successive broadcasting and successive point-to-point transmission protocols. This gap can be closed by a third protocol separating the cooperation from the forwarding phase and allowing each sensor to access the maximal available side-information. Although no formal conclusion about the optimality of the pcCEO can be drawn, the closeness to the fcCEO scenario in the investigated simulations reveals that solutions found by the proposed greedy algorithms are at least close to optimal.

Author Contributions

Conceptualization, S.S. and V.K.; formal analysis, S.S. and V.K.; investigation, S.S., A.D.A. and V.K.; methodology, S.S.; software, S.S. and A.D.A.; supervision, V.K.; validation, S.S.; visualization, S.S.; writing—original draft, S.S.; writing—review and editing, S.S. and V.K. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the University of Rostock, and in part by the German Research Foundation (DFG) in the funding programme Open Access Publishing under Grant 325496636. Parts of the computation were done by using a compute cluster funded by DFG (grant: 440623123).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Appendix A.1. Optimization for Broadcasting Side-Information

According to Section 5.1, partial cooperation by broadcasting side-information requires the maximization of
L GDIB - BC ( m ) = I ( 𝒳 ; Ƶ m | Ƶ < m ) β m I ( 𝒴 m , 𝓢 < m ; Ƶ m | Ƶ < m ) .
The variational problem can be solved by taking the derivative with respect to the mapping p ( z m | y m , s < m ) and equating it to zero. In the following, the derivatives of both mutual information are given.

Appendix A.1.1. Derivative of I(𝒳; Ƶm|Ƶ<m)

The relevant mutual information in (A1) can be rewritten such that the desired mapping occurs explicitly.
I ( 𝒳 ; Ƶ m | Ƶ < m ) = E 𝒳 , Ƶ m , Ƶ < m [ log p ( z m | x , z < m ) p ( z m | z < m ) ] = z m z < m x y m s < m p ( z m | y m , s < m ) p ( y m , s < m , z < m , x ) · log a 𝕐 m b 𝕊 < m p ( z m | a , b ) p ( a , b | z < m , x ) z m z < m x y m s < m p ( z m | y m , s < m ) p ( y m , s < m , z < m , x ) · log a 𝕐 m b 𝕊 < m p ( z m | a , b ) p ( a , b | z < m )
The derivative of (A2) delivers
I ( 𝒳 ; Ƶ m | Ƶ < m ) p ( z m | y m , s < m ) = x z < m p ( y m , s < m , z < m , x ) · log p ( z m | x , z < m ) + x z < m s < m y m p ( z m | y m , s < m ) p ( y m , s < m , z < m , x ) = p ( z m , z < m , x ) · p ( y m , s < m | x , z < m ) p ( z m | x , z < m ) x z < m p ( y m , s < m , z < m , x ) · log p ( z m | z < m ) x z < m s < m y m p ( z m | y m , s < m ) p ( y m , s < m , z < m , x ) = p ( z m , z < m , x ) · p ( y m , s < m | z < m ) p ( z m | z < m ) = z < m p ( y m , s < m , z < m ) x p ( x | y m , s < m , z < m ) log p ( z m | x , z < m ) p ( z m | z < m ) .
Exploiting Bayes’ theorem, the argument of the logarithmic function can be rewritten
p ( z m | x , z < m ) p ( z m | z < m ) = p ( x | z m ) p ( x | z < m ) = p ( x | z m ) p ( x | y m , s < m , z < m ) · p ( x | y m , s < m , z < m ) p ( x | z < m ) .
The last ratio in (A4) can be dropped because it does not depend on p ( z m | y m , s < m ) and its contribution can be incorporated into the Lagrange multiplier β m . The insertion of the first ratio into (A3) yields the contribution of the derivative of the relevant mutual information
I ( 𝒳 ; Ƶ m | Ƶ < m ) p ( z m | y m , s < m ) z < m p ( y m , s < m , z < m ) x p ( x | y m , s < m , z < m ) log p ( x | y m , s < m , z < m ) p ( x | z m ) = z < m p ( y m , s < m , z < m ) D K L [ p ( x | y m , s < m , z < m ) p ( x | z m ) ] .

Appendix A.1.2. Derivative of I(𝒴m, 𝓢<m; Ƶm|Ƶ<m)

With the definition of the conditional compression rate
I ( 𝒴 m , 𝓢 < m ; Ƶ m | Ƶ < m ) = E 𝒴 m , Ƶ m , Ƶ < m , 𝓢 < m [ log p ( z m | y m , s < m , z < m ) p ( z m | z < m ) ] = z m z < m s < m y m p ( z m | y m , s < m ) p ( y m , s < m , z < m ) log p ( z m | y m , s < m ) z m z < m s < m y m p ( z m | y m , s < m ) p ( y m , s < m , z < m ) log a 𝕐 m b 𝕊 < m p ( z m | a , b ) p ( a , b | z < m )
its derivative becomes
I ( 𝒴 m , 𝓢 < m ; Ƶ m | Ƶ < m ) p ( z m | y m , s < m ) = z < m p ( y m , s < m , z < m ) log p ( z m | y m , s < m ) + z < m p ( y m , s < m , z < m ) p ( z m | y m , s < m ) p ( z m | y m , s < m ) z < m p ( y m , s < m , z < m ) log p ( z m | z < m ) z < m y m s < m p ( z m | y m , s < m ) p ( y m , s < m , z < m ) = p ( z m , z < m ) · p ( y m , s < m | z < m ) p ( z m | z < m ) = z < m p ( y m , s < m , z < m ) log p ( z m | y m , s < m ) p ( z m | z < m ) .

Appendix A.1.3. Fusion of Derived Parts

Combining the result in (A5) and (A7) delivers the complete derivative
z < m p ( y m , s < m , z < m ) D K L [ p ( x | y m , s < m , z < m ) p ( x | z m ) ]   β m z < m p ( y m , s < m , z < m ) log p ( z m | y m , s < m ) +   β m z < m p ( y m , s < m , z < m ) · log p ( z m | z < m ) = 0 .
Following the idea of Blahut and Arimoto [34], p ( x | z m ) and p ( z m | z < m ) are assumed to be independent of p ( z m | y m , s < m ) . With this trick, (A8) can be resolved with respect to the desired mapping of sensor m leading to the implicit solution
p ( z m | y m , s < m ) = e d β m ( y m , z m , s < m ) z m e d β m ( y m , z m , s < m )
with
d β m ( y m , z m , s < m ) z < m p ( z < m | y m , s < m ) 1 β m D K L [ p ( x | y m , s < m , z < m ) p ( x | z m ) ] log p ( z m | z < m ) = E Ƶ < m | y m , s < m 1 β m D K L [ p ( x | y m , s < m , z < m ) p ( x | z m ) ] log p ( z m | z < m ) .

Appendix A.1.4. Calculating Required pmfs

This section covers the calculation of the required pmfs for the previously described algorithm. The first term in the KL divergence in (A10) is defined by
p ( x | y m , s < m , z < m ) = p ( x , y m , s < m , z < m ) x p ( x , y m , s < m , z < m )
where p ( x , y m , s < m , z < m ) is determined recursively as
p ( x , y m , s < m , z < m ) = y m 1 p ( z m 1 | y m 1 , s < m 1 ) · p ( s m 1 | y m 1 , s < m 1 ) p ( y m | x ) p ( x , y m 1 , s < m 1 , z < m 1 ) .
In (A12), the pmf p ( z m 1 | y m 1 , s < m 1 ) represents the quantizer mapping of the previous sensor m 1 while p ( s m 1 | y m 1 , s < m 1 ) denotes its the mapping for the instantaneous side-information. Hence, both have already been determined when optimizing sensor m 1 leading to a recursive computation. The second term in the KL divergence in (A10) is calculated by
p ( x | z m ) = p ( z m , x ) x p ( z m , x )
where p ( z m , x ) can be calculated by
p ( z m , x ) = s < m y m p ( z m | y m , s < m ) p ( x , y m , s < m , z < m )
with p ( z m | y m , s < m ) being the quantizer mapping of the current sensor. The joint pmf p ( x , y m , s < m , z < m ) has already been calculated in (A12). The argument of the logarithm in (A10) can be derived as
p ( z m | z < m ) = x p ( z m , x ) z m x p ( z m , x ) .
Finally, the pmf to calculate the conditional expectation in (A10) is determined as
p ( z < m | y m , s < m ) = x p ( x , y m , s < m , z < m ) z < m x p ( x , y m , s < m , z < m ) .
Note that all above equations simplify to the scalar IB equations given in Section 2 when optimizing the first sensor.

Appendix A.2. Optimization for Point-to-Point Exchange of Side-Information

According to Section 5.2, partial cooperation using a point-to-point communication protocol requires the maximization of
L GDIB - PTP ( m ) = I ( 𝒳 ; Ƶ m | Ƶ < m ) β m I ( 𝒴 m , 𝒮 m 1 ; Ƶ m | Ƶ < m ) .
The variational problem can be solved by taking the derivative with respect to the mapping p ( z m | y m , s m 1 ) and equating it to zero. In the following, the derivatives of both mutual information are given.

Appendix A.2.1. Derivative of I(𝒳; Ƶm|Ƶ<m)

The relevant mutual information in (A17) can be rewritten such that the desired mapping occurs explicitly.
I ( 𝒳 ; Ƶ m | Ƶ < m ) = E 𝒳 , Ƶ m , Ƶ < m [ log p ( z m | x , z < m ) p ( z m | z < m ) ] = z m x y m s m 1 p ( z m | y m , s m 1 ) p ( y m , s m 1 , z < m , x ) · log a 𝕐 m b 𝕊 m 1 p ( z m | a , b ) p ( a , b | z < m , x ) z m x y m s m 1 p ( z m | y m , s m 1 ) p ( y m , s m 1 , z < m , x ) · log a 𝕐 m b 𝕊 m 1 p ( z m | a , b ) p ( a , b | z < m )
The derivative of (A18) delivers
I ( 𝒳 ; Ƶ m | Ƶ < m ) p ( z m | y m , s m 1 ) = x z < m p ( y m , s m 1 , z < m , x ) · log p ( z m | x , z < m ) + x z < m s m 1 y m p ( z m | y m , s m 1 ) p ( y m , s m 1 , z < m , x ) = p ( z m , z < m , x ) · p ( y m , s m 1 | x , z < m ) p ( z m | x , z < m ) x z < m p ( y m , s m 1 , z < m , x ) · log p ( z m | z < m ) x z < m s m 1 y m p ( z m | y m , s m 1 ) p ( y m , s m 1 , z < m , x ) = p ( z m , z < m , x ) · p ( y m , s m 1 | z < m ) p ( z m | z < m ) = x z < m p ( y m , s m 1 , z < m , x ) · log p ( z m | x , z < m ) p ( z m | z < m ) = z < m p ( y m , s m 1 , z < m ) x p ( x | y m , s m 1 , z < m ) log p ( z m | x , z < m ) p ( z m | z < m ) .
Exploiting Bayes’ theorem the argument of the logarithmic function can be rewritten
p ( z m | x , z < m ) p ( z m | z < m ) = p ( x | z m ) p ( x | z < m ) = p ( x | z m ) p ( x | y m , s m 1 , z < m ) p ( x | y m , s m 1 , z < m ) p ( x | z < m ) .
The last ratio in (A20) can be dropped because it does not depend on p ( z m | y m , s m 1 ) and its contribution can be incorporated into the Lagrange multiplier β m . The insertion of the first ratio into (A19) yields the contribution of the derivative of the relevant mutual information
I ( 𝒳 ; Ƶ m | Ƶ < m ) p ( z m | y m , s m 1 ) z < m p ( y m , s m 1 , z < m ) x p ( x | y m , s m 1 , z < m ) log p ( x | y m , s m 1 , z < m ) p ( x | z m ) = z < m p ( y m , s m 1 , z < m ) D K L [ p ( x | y m , s m 1 , z < m ) p ( x | z m ) ] .

Appendix A.2.2. Derivative of I(𝒴m, 𝒮m−1; Ƶm|Ƶ<m)

With the definition of the conditional compression rate
I ( 𝒴 m , 𝒮 m 1 ; Ƶ m | Ƶ < m ) = E 𝒴 m , Ƶ m , Ƶ < m , 𝒮 m 1 [ log p ( z m | y m , s m 1 , z < m ) p ( z m | z < m ) ] = z m z < m s m 1 y m p ( z m | y m , s m 1 ) p ( y m , s m 1 , z < m ) log p ( z m | y m , s m 1 ) z m z < m s m 1 y m p ( z m | y m , s m 1 ) p ( y m , s m 1 , z < m ) log a 𝕐 m b 𝕊 m 1 p ( z m | a , b ) p ( a , b | z < m )
its derivative becomes
I ( 𝒴 m , 𝒮 m 1 ; Ƶ m | Ƶ < m ) p ( z m | y m , s m 1 ) = z < m p ( y m , s m 1 , z < m ) log p ( z m | y m , s m 1 ) + z < m p ( y m , s m 1 , z < m ) p ( z m | y m , s m 1 ) p ( z m | y m , s m 1 ) z < m p ( y m , s m 1 , z < m ) log p ( z m | z < m ) z < m y m s m 1 p ( z m | y m , s m 1 ) p ( y m , s m 1 , z < m ) = p ( z m , z < m ) · p ( y m , s m 1 | z < m ) p ( z m | z < m ) = z < m p ( y m , s m 1 , z < m ) log p ( z m | y m , s m 1 ) p ( z m | z < m ) .

Appendix A.2.3. Fusion of Derived Parts

Combining the result in (A21) and (A23) delivers the complete derivative
z < m p ( y m , s m 1 , z < m ) D K L [ p ( x | y m , s m 1 , z < m ) p ( x | z m ) ]   β m z < m p ( y m , s m 1 , z < m ) log p ( z m | y m , s m 1 ) +   β m z < m p ( y m , s m 1 , z < m ) · log p ( z m | z < m ) = 0 .
Following the idea of Blahut and Arimoto [34], p ( x | z m ) and p ( z m | z < m ) are assumed to be independent of p ( z m | y m , s m 1 ) . With this trick, (A24) can be resolved with respect to the desired mapping of sensor m leading to the implicit solution
p ( z m | y m , s m 1 ) = e d β m ( y m , z m , s m 1 ) z m e d β m ( y m , z m , s m 1 )
with
d β m ( y m , z m , s m 1 ) z < m p ( z < m | y m , s m 1 ) 1 β m D K L [ p ( x | y m , s m 1 , z < m ) p ( x | z m ) ] log p ( z m | z < m ) = E Ƶ < m | y m , s m 1 1 β m D K L [ p ( x | y m , s m 1 , z < m ) p ( x | z m ) ] log p ( z m | z < m ) .

Appendix A.2.4. Calculating Required pmfs

This section covers the calculation of the required pmfs for the previously described algorithm. The first term in the KL divergence in (A26) is defined by
p ( x | y m , s m 1 , z < m ) = p ( x , y m , s m 1 , z < m ) x p ( x , y m , s m 1 , z < m )
with
p ( x , y m , s m 1 , z < m ) = p ( z m 1 , s m 1 | x , z < m 1 ) p ( x , y m , z < m 1 ) .
The first term on the right hand side in (A28) can be calculated by
p ( z m 1 , s m 1 | x , z < m 1 ) = y m 1 s m 2 p ( z m 1 | y m 1 , s m 2 ) p ( s m 1 | y m 1 , s m 2 ) p ( y m 1 | x ) p ( s m 2 | x , z < m 1 )
where p ( z m 1 | y m 1 , s m 2 ) represents the mapping of previously designed quantizer m 1 , p ( s m 1 | y m 1 , s m 2 ) denotes its mapping of instantaneous side-information and p ( y m 1 | x ) statistically describes the corresponding known measurement channel. The last term p ( s m 2 | x , z < m 1 ) has already been calculated when optimizing previous sensors (in this case the pre-predecessor) in a recursive way
p ( s m | x , z < m ) = y m s m 1 p ( s m | y m , s m 1 ) p ( y m | x , z m ) p ( s m 1 | x , z < m 1 )
with
p ( y m | x , z m ) = p ( y m , x , z m ) y m p ( y m , x , z m )
and
p ( y m , x , z m ) = s m 1 p ( z m | y m , s m 1 ) p ( y m | x ) p ( s m 1 | x ) p ( x ) .
The second term on the right hand side in (A28) can be determined by
p ( x , y m , z < m 1 ) = p ( y m | x ) p ( z < m 1 , x ) .
Again, the last term in (A33) is obtained as side product when optimizing the previous sensor by
p ( z m , x ) = y m s m 1 p ( z m | y m , s m 1 ) p ( z m 1 , s m 1 | x , z < m 1 ) p ( x , y m , z < m 1 )
where p ( z m 1 , s m 1 | x , z < m 1 ) and p ( x , y m , z < m 1 ) are already defined in (A29) and (A33), respectively. The second term in the KL divergence in (A26) is calculated by
p ( x | z m ) = p ( z m , x ) x p ( z m , x )
where p ( z m , x ) is already defined in (A34). The term in the logarithm in (A26) can be expressed as
p ( z m | z < m ) = x p ( z m , x ) x z m p ( z m , x ) .
Finally, the required pmf to calculate the conditional expectation in (A26) can be determined by
p ( z < m | y m , s m 1 ) = x p ( x , y m , s m 1 , z < m ) x z < m p ( x , y m , s m 1 , z < m )
with p ( x , y m , s m 1 , z < m ) being already defined in (A28). Note that all above equations simplify to the scalar IB equations given in Section 2 when optimizing the first sensor. Moreover, when optimizing the second sensor, there is no pre-predecessor m 2 and its impact on the above equations can be omitted.

References

  1. Oohama, Y. The rate-distortion function for the quadratic Gaussian CEO problem. IEEE Trans. Inf. Theory 1998, 44, 1057–1070. [Google Scholar] [CrossRef]
  2. Viswanathan, H.; Berger, T. The Quadratic Gaussian CEO Problem. IEEE Trans. Inf. Theory 1997, 43, 1549–1559. [Google Scholar] [CrossRef]
  3. Chen, J.; Zhang, X.; Berger, T.; Wicker, S.B. An upper bound on the sum-rate distortion function and its corresponding rate allocation schemes for the CEO problem. IEEE J. Sel. Areas Commun. 2004, 22, 977–987. [Google Scholar] [CrossRef]
  4. Prabhakaran, V.; Tse, D.; Ramachandran, K. Rate region of the quadratic Gaussian CEO problem. In Proceedings of the International Symposium on Information Theory (ISIT 2004), Chicago, IL, USA, 27 June–2 July 2004; p. 119. [Google Scholar] [CrossRef]
  5. Oohama, Y. Rate-distortion theory for Gaussian multiterminal source coding systems with several side informations at the decoder. IEEE Trans. Inf. Theory 2005, 51, 2577–2593. [Google Scholar] [CrossRef]
  6. Wagner, A.; Tavildar, S.; Viswanath, P. Rate Region of the Quadratic Gaussian Two-Encoder Source-Coding Problem. IEEE Trans. Inf. Theory 2008, 54, 1938–1961. [Google Scholar] [CrossRef]
  7. Ugur, Y.; Aguerri, I.E.; Zaidi, A. Vector Gaussian CEO problem under logarithmic loss. In Proceedings of the 2018 IEEE Information Theory Workshop (ITW), Guangzhou, China, 25–29 November 2018; pp. 1–5. [Google Scholar]
  8. Uğur, Y.; Aguerri, I.E.; Zaidi, A. Vector Gaussian CEO Problem Under Logarithmic Loss and Applications. IEEE Trans. Inf. Theory 2020, 66, 4183–4202. [Google Scholar] [CrossRef] [Green Version]
  9. Courtade, T.A.; Weissman, T. Multiterminal Source Coding Under Logarithmic Loss. IEEE Trans. Inf. Theory 2014, 60, 740–761. [Google Scholar] [CrossRef] [Green Version]
  10. Berger, T.; Zhang, Z.; Viswanathan, H. The CEO Problem [Multiterminal Source Coding]. IEEE Trans. Inf. Theory 1996, 42, 887–902. [Google Scholar] [CrossRef]
  11. Eswaran, K.; Gastpar, M. Remote Source Coding under Gaussian Noise: Dueling Roles of Power and Entropy Power. arXiv 2018, arXiv:1805.06515v2. [Google Scholar]
  12. Zaidi, A.; Aguerri, I.E. Distributed Deep Variational Information Bottleneck. In Proceedings of the 2020 IEEE 21st International Workshop on Signal Processing Advances in Wireless Communications (SPAWC), Atlanta, GA, USA, 26–29 May 2020; pp. 1–5. [Google Scholar] [CrossRef]
  13. Aguerri, I.E.; Zaidi, A. Distributed Variational Representation Learning. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 120–138. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Steiner, S.; Kuehn, V.; Stark, M.; Bauch, G. Reduced-Complexity Optimization of Distributed Quantization Using the Information Bottleneck Principle. IEEE Open J. Commun. Soc. 2021, 2, 1267–1278. [Google Scholar] [CrossRef]
  15. Steiner, S.; Kuehn, V. Distributed Compression using the Information Bottleneck Principle. In Proceedings of the ICC 2021—IEEE International Conference on Communications, Montreal, QC, Canada, 14–23 June 2021; pp. 1–6. [Google Scholar] [CrossRef]
  16. Estella Aguerri, I.; Zaidi, A. Distributed information bottleneck method for discrete and Gaussian sources. In Proceedings of the International Zurich Seminar on Information and Communication (IZS 2018) Proceedings, Zurich, Switzerland, 21–23 February 2018; ETH Zurich: Zurich, Switzerland, 2018; pp. 35–39. [Google Scholar]
  17. Uğur, Y.; Aguerri, I.E.; Zaidi, A. A generalization of blahut-arimoto algorithm to compute rate-distortion regions of multiterminal source coding under logarithmic loss. In Proceedings of the 2017 IEEE Information Theory Workshop (ITW), Kaohsiung, Taiwan, 6–10 November 2017; pp. 349–353. [Google Scholar]
  18. Prabhakaran, V.; Ramchandran, K.; Tse, D. On the Role of Interaction Between Sensors in the CEO Problem. In Proceedings of the 42nd Annual Allerton Conference on Communication, Control, and Computing, Monticello, IL, USA, 29 September–1 October 2004. [Google Scholar]
  19. Draper, S.; Wornell, G. Side Information Aware Coding Strategies for Sensor Networks. IEEE J. Sel. Areas Commun. 2004, 22, 966–976. [Google Scholar] [CrossRef]
  20. Simeone, O. Source and Channel Coding for Homogeneous Sensor Networks with Partial Cooperation. IEEE Trans. Wirel. Commun. 2009, 8, 1113–1117. [Google Scholar] [CrossRef]
  21. Permuter, H.; Steinberg, Y.; Weissman, T. Problems we can solve with a helper. In Proceedings of the 2009 IEEE Information Theory Workshop on Networking and Information Theory, Volos, Greece, 10–12 June 2009; pp. 266–270. [Google Scholar] [CrossRef]
  22. Tishby, N.; Pereira, F.C.; Bialek, W. The Information Bottleneck Method. In Proceedings of the 37th Annual Allerton Conference on Communication, Control, and Computing, Monticello, IL, USA, 22–24 September 1999; pp. 368–377. [Google Scholar]
  23. Slonim, N. The Information Bottleneck Theory and Applications. Ph.D. Thesis, Hebrew University of Jerusalem, Jerusalem, Israel, 2002. [Google Scholar]
  24. Hassanpour, S.; Wuebben, D.; Dekorsy, A. Overview and Investigation of Algorithms for the Information Bottleneck Method. In Proceedings of the SCC 2017—11th International ITG Conference on Systems, Communications and Coding, Hamburg, Germany, 6–9 February 2017. [Google Scholar]
  25. Lewandowsky, J.; Bauch, G. Information-Optimum LDPC Decoders Based on the Information Bottleneck Method. IEEE Access 2018, 6, 4054–4071. [Google Scholar] [CrossRef]
  26. Zeitler, G. Low-precision analog-to-digital conversion and mutual information in channels with memory. In Proceedings of the 48th Annual Allerton Conference on Communication, Control and Computing, Monticello, IL, USA, 29 September–1 October 2010; pp. 745–752. [Google Scholar]
  27. Meidlinger, M.; Matz, G. On Irregular LDPC Codes with Quantized Message Passing Decoding. In Proceedings of the 2017 IEEE 18th International Workshop on Signal Processing Advances in Wireless Communications (SPAWC), Sapporo, Japan, 3–6 July 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1–5. [Google Scholar] [CrossRef]
  28. Romero, F.; Kurkoski, B. LDPC Decoding Mappings That Maximize Mutual Information. IEEE J. Sel. Areas Commun. 2016, 34, 2391–2401. [Google Scholar] [CrossRef]
  29. Zeitler, G. Low-Precision Quantizer Design for Communication Problems. Ph.D. Thesis, Technische Universitaet Muenchen, Muenchen, Germany, 2012. [Google Scholar]
  30. Chen, D.; Kuehn, V. Alternating information bottleneck optimization for the compression in the uplink of C-RAN. In Proceedings of the 2016 IEEE International Conference on Communications (ICC), Kuala Lumpur, Malaysia, 23–27 May 2016; pp. 1–7. [Google Scholar] [CrossRef]
  31. Lewandowsky, J.; Stark, M.; Bauch, G. Information Bottleneck Graphs for Receiver Design. In Proceedings of the 2016 IEEE International Symposium on Information Theory (ISIT), Barcelona, Spain, 10–15 July 2016; pp. 2888–2892. [Google Scholar] [CrossRef]
  32. Fujishige, S. Submodular Functions and Optimization; Elsevier: Amsterdam, The Netherlands, 2005. [Google Scholar]
  33. ten Brink, S. Convergence Behavior of Iteratively Decoded Parallel Concatenated Codes. IEEE Trans. Commun. 2001, 49, 1727–1737. [Google Scholar] [CrossRef]
  34. Cover, T.; Thomas, J. Elements of Information Theory, 2nd ed.; Wiley & Sons: New York, NY, USA, 2006. [Google Scholar]
Figure 1. Illustration of a remote sensing problem for a single sensor.
Figure 1. Illustration of a remote sensing problem for a single sensor.
Entropy 24 00438 g001
Figure 2. (a) Illustration of the IB setup, (b) Exemplary IB graph.
Figure 2. (a) Illustration of the IB setup, (b) Exemplary IB graph.
Entropy 24 00438 g002
Figure 3. Non-cooperative distributed sensing system with M sensors, a common receiver and individual link capacities C m .
Figure 3. Non-cooperative distributed sensing system with M sensors, a common receiver and individual link capacities C m .
Entropy 24 00438 g003
Figure 4. Model of a centralized compression approach representing the fully cooperative Chief Executive Officer scenario.
Figure 4. Model of a centralized compression approach representing the fully cooperative Chief Executive Officer scenario.
Entropy 24 00438 g004
Figure 5. Partially cooperative CEO scenario using broadcast exchange of side-information among sensors.
Figure 5. Partially cooperative CEO scenario using broadcast exchange of side-information among sensors.
Entropy 24 00438 g005
Figure 6. Graphical illustration of IB fusion of involved inputs to determine instantaneous side-information s m (a) and the quantizer sensor output z m (b) for a broadcast exchange of instantaneous side-information.
Figure 6. Graphical illustration of IB fusion of involved inputs to determine instantaneous side-information s m (a) and the quantizer sensor output z m (b) for a broadcast exchange of instantaneous side-information.
Entropy 24 00438 g006
Figure 7. Available mutual information I ( 𝒳 ; 𝓢 m ) for sensor m in a network with M = 6 sensors and different cardinalities | 𝕊 m | using the successive broadcasting protocol; | 𝕏 | = 4 , | 𝕐 m |   = 64 .
Figure 7. Available mutual information I ( 𝒳 ; 𝓢 m ) for sensor m in a network with M = 6 sensors and different cardinalities | 𝕊 m | using the successive broadcasting protocol; | 𝕏 | = 4 , | 𝕐 m |   = 64 .
Entropy 24 00438 g007
Figure 8. Relevant mutual information I ( 𝒳 ; Ƶ ) versus the network size for a fixed sum-rate of C sum = 2.5 bit/s/Hz and C m = C sum M using the successive broadcasting protocol with different cardinalities | 𝕊 m | ; | 𝕏 | = 4 , | 𝕐 m |   = 64 , | m |   = 4 .
Figure 8. Relevant mutual information I ( 𝒳 ; Ƶ ) versus the network size for a fixed sum-rate of C sum = 2.5 bit/s/Hz and C m = C sum M using the successive broadcasting protocol with different cardinalities | 𝕊 m | ; | 𝕏 | = 4 , | 𝕐 m |   = 64 , | m |   = 4 .
Entropy 24 00438 g008
Figure 9. Relevant mutual information I ( 𝒳 ; Ƶ ) versus the network size for a fixed sum-rate of C sum = 4 bit/s/Hz and C m = C sum M using the successive broadcasting protocol with different cardinalities | 𝕊 m | ; | 𝕏 | = 4 , | 𝕐 m |   = 64 , | m |   = 4 .
Figure 9. Relevant mutual information I ( 𝒳 ; Ƶ ) versus the network size for a fixed sum-rate of C sum = 4 bit/s/Hz and C m = C sum M using the successive broadcasting protocol with different cardinalities | 𝕊 m | ; | 𝕏 | = 4 , | 𝕐 m |   = 64 , | m |   = 4 .
Entropy 24 00438 g009
Figure 10. Partially cooperative CEO scenario using successive point-to-point transmission of side-information.
Figure 10. Partially cooperative CEO scenario using successive point-to-point transmission of side-information.
Entropy 24 00438 g010
Figure 11. Graphical illustration of IB fusion of two inputs to determine instantaneous side-information s m (a) and the quantizer sensor output z m (b) for a successive point-to-point transmission of instantaneous side-information.
Figure 11. Graphical illustration of IB fusion of two inputs to determine instantaneous side-information s m (a) and the quantizer sensor output z m (b) for a successive point-to-point transmission of instantaneous side-information.
Entropy 24 00438 g011
Figure 13. Relevant mutual information I ( 𝒳 ; Ƶ ) versus the network size for a fixed sum-rate of C sum = 2.5 bit/s/Hz and C m = C sum M using the successive point-to-point transmission protocol with different cardinalities | 𝕊 m | ; | 𝕏 | = 4 , | 𝕐 m |   = 64 , | m |   = 4 .
Figure 13. Relevant mutual information I ( 𝒳 ; Ƶ ) versus the network size for a fixed sum-rate of C sum = 2.5 bit/s/Hz and C m = C sum M using the successive point-to-point transmission protocol with different cardinalities | 𝕊 m | ; | 𝕏 | = 4 , | 𝕐 m |   = 64 , | m |   = 4 .
Entropy 24 00438 g013
Figure 14. Relevant mutual information I ( 𝒳 ; Ƶ ) versus the network size for a fixed sum-rate of C sum = 4 bit/s/Hz and C m = C sum M using the successive point-to-point transmission protocol with different cardinalities | 𝕊 m | ; | 𝕏 | = 4 , | 𝕐 m |   = 64 , | m |   = 4 .
Figure 14. Relevant mutual information I ( 𝒳 ; Ƶ ) versus the network size for a fixed sum-rate of C sum = 4 bit/s/Hz and C m = C sum M using the successive point-to-point transmission protocol with different cardinalities | 𝕊 m | ; | 𝕏 | = 4 , | 𝕐 m |   = 64 , | m |   = 4 .
Entropy 24 00438 g014
Figure 15. Relevant mutual information I ( 𝒳 ; Ƶ ) versus sum-rate C sum with C m = C sum M using the successive point-to-point transmission protocol with different cardinalities | 𝕊 m | ; M = 5 , | 𝕏 | = 4 , | 𝕐 m |   = 64 , | m |   = 4 .
Figure 15. Relevant mutual information I ( 𝒳 ; Ƶ ) versus sum-rate C sum with C m = C sum M using the successive point-to-point transmission protocol with different cardinalities | 𝕊 m | ; M = 5 , | 𝕏 | = 4 , | 𝕐 m |   = 64 , | m |   = 4 .
Entropy 24 00438 g015
Figure 16. Relevant mutual information I ( 𝒳 ; Ƶ ) for non-symmetric scenario with M = 4 sensors, SNRs γ m = [2,4,6,8] dB and | 𝕏 | = 4 , | 𝕐 m |   = 64 , | m |   = 4 using the successive point-to-point transmission protocol with | 𝕊 m |   = 8 .
Figure 16. Relevant mutual information I ( 𝒳 ; Ƶ ) for non-symmetric scenario with M = 4 sensors, SNRs γ m = [2,4,6,8] dB and | 𝕏 | = 4 , | 𝕐 m |   = 64 , | m |   = 4 using the successive point-to-point transmission protocol with | 𝕊 m |   = 8 .
Entropy 24 00438 g016
Figure 17. Partially cooperative CEO scenario using the two-phase transmission protocol.
Figure 17. Partially cooperative CEO scenario using the two-phase transmission protocol.
Entropy 24 00438 g017
Figure 18. Relevant mutual information I ( 𝒳 ; Ƶ ) versus the network size for a fixed sum-rate of C sum = 2.5 bit/s/Hz and C m = C sum M using a two-phase transmission protocol for artificially decoupled extrinsic information with different γ ext ; γ m = 8 dB, | 𝕏 | = 4 , | 𝕐 m |   = 64 , | m |   = 4 , | 𝕊 * |   = 512 .
Figure 18. Relevant mutual information I ( 𝒳 ; Ƶ ) versus the network size for a fixed sum-rate of C sum = 2.5 bit/s/Hz and C m = C sum M using a two-phase transmission protocol for artificially decoupled extrinsic information with different γ ext ; γ m = 8 dB, | 𝕏 | = 4 , | 𝕐 m |   = 64 , | m |   = 4 , | 𝕊 * |   = 512 .
Entropy 24 00438 g018
Figure 19. Relevant mutual information I ( 𝒳 ; Ƶ ) versus extrinsic mutual information I ( 𝒳 ; 𝒮 * ) for different network sizes and a fixed sum-rate of C sum = 2.5 bit/s/Hz and C m = C sum M and | 𝕏 | = 4 , | 𝕐 m |   = 64 , | m |   = 4 .
Figure 19. Relevant mutual information I ( 𝒳 ; Ƶ ) versus extrinsic mutual information I ( 𝒳 ; 𝒮 * ) for different network sizes and a fixed sum-rate of C sum = 2.5 bit/s/Hz and C m = C sum M and | 𝕏 | = 4 , | 𝕐 m |   = 64 , | m |   = 4 .
Entropy 24 00438 g019
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Steiner, S.; Aminu, A.D.; Kuehn, V. Distributed Quantization for Partially Cooperating Sensors Using the Information Bottleneck Method. Entropy 2022, 24, 438. https://doi.org/10.3390/e24040438

AMA Style

Steiner S, Aminu AD, Kuehn V. Distributed Quantization for Partially Cooperating Sensors Using the Information Bottleneck Method. Entropy. 2022; 24(4):438. https://doi.org/10.3390/e24040438

Chicago/Turabian Style

Steiner, Steffen, Abdulrahman Dayo Aminu, and Volker Kuehn. 2022. "Distributed Quantization for Partially Cooperating Sensors Using the Information Bottleneck Method" Entropy 24, no. 4: 438. https://doi.org/10.3390/e24040438

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop