Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
Retrieval and Ranking of Combining Ontology and Content Attributes for Scientific Document
Previous Article in Journal
The Linear Relationship Model with LASSO for Studying Stock Networks
Previous Article in Special Issue
A Model for Tacit Communication in Collaborative Human-UAV Search-and-Rescue
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Generalized Information-Theoretic Framework for the Emergence of Hierarchical Abstractions in Resource-Limited Systems

by
Daniel T. Larsson
1,*,
Dipankar Maity
2 and
Panagiotis Tsiotras
3
1
D. Guggenheim School of Aerospace Engineering, Georgia Institute of Technology, Atlanta, GA 30332-0150, USA
2
Department of Electrical and Computer Engineering, The University of North Carolina at Charlotte, Charlotte, NC 28223-0001, USA
3
D. Guggenheim School of Aerospace Engineering, Institute for Robotics and Intelligent Machines, Georgia Institute of Technology, Atlanta, GA 30332-0150, USA
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(6), 809; https://doi.org/10.3390/e24060809
Submission received: 7 March 2022 / Revised: 5 June 2022 / Accepted: 6 June 2022 / Published: 9 June 2022
(This article belongs to the Special Issue Information Theory in Motion Planning and Control)

Abstract

:
In this paper, a generalized information-theoretic framework for the emergence of multi-resolution hierarchical tree abstractions is developed. By leveraging ideas from information-theoretic signal encoding with side information, this paper develops a tree search problem which considers the generation of multi-resolution tree abstractions when there are multiple sources of relevant and irrelevant, or possibly confidential, information. We rigorously formulate an information-theoretic driven tree abstraction problem and discuss its connections with information-theoretic privacy and resource-limited systems. The problem structure is investigated and a novel algorithm, called G-tree search, is proposed. The proposed algorithm is analyzed and a number of theoretical results are established, including the optimally of the G-tree search algorithm. To demonstrate the utility of the proposed framework, we apply our method to a real-world example and provide a discussion of the results from the viewpoint of designing hierarchical abstractions for autonomous systems.

1. Introduction

Driven by the human ability to discern pertinent details from immense amounts of perceptual information, the process of identifying task-relevant structures from data has long been considered a cornerstone to the development of intelligent systems [1,2,3,4]. To this end, researchers in the autonomous systems community have spend a great deal of effort studying abstractions, which is a problem generally viewed as an information-removal procedure to discard details that are not relevant for a given task [1,2,3,4,5]. The central motivation for the employment of abstractions is to simplify the problem domain by removing details that can be safely ignored, thereby creating a new representation of the problem for which reasoning and decision making requires fewer computational resources [1,2,3]. Despite their importance, autonomous systems thus far seldom design abstractions on their own, instead relying on system designers and prior domain knowledge to provide hand-crafted rules for the emergence of abstractions as a function of task in various domains [3,5]. In spite of these shortcomings, abstractions have seen wide-spread use in a number of autonomous systems applications.
Perhaps the most notable field of research where abstractions have seen particular success is within the planning community [1]. Examples of work that employ the power of abstract representations in planning for autonomous systems include [6,7,8,9,10,11,12,13,14]. The idea of utilizing abstractions in planning is to form reduced graphs on which classical search algorithms, such as A * and Dijkstra, are implemented. By reducing the number of vertices in the graph, the computational burden of executing these search algorithms is reduced. However, while the cited works all leverage graph abstractions to ease the computational cost of planning, the methods by which they generate these abstractions differ. For example, in [7,8,9,10,11] the environment abstractions are created via the wavelet transform. In contrast, the works of [12,13,14] generate abstractions of the environment in the form of multi-resolution quadtree and octree data structures. Notably, the work of [12,13] develops a framework that incorporates sensor uncertainty in robotic systems by merging ideas from multi-resolution planning and probabilistic tree structures introduced in [15]. Today, the use of probabilistic trees in robotics is ubiquitous, and has led to the development of open-source software packages for their implementation [16].
Motivated by the possibly dynamic nature of the environment as well as sensing limitations inherent to autonomous systems, the abstractions employed in all the aforementioned works maintain high resolution nearest the autonomous agent (e.g., robotic ground vehicle), while aggregating other portions of the environment at various resolution levels. In this way, the region nearest to the vehicle is considered the most relevant, and thus preserved through the process of abstraction. To strike a balance between path-optimality (system performance) and the computational cost of planning, agents recursively re-plan as they traverse the world.
The design of abstractions has also been considered by information theorists in the context of optimal signal encoding for communication over capacity-limited channels [17]. In order to formulate mathematical optimization problems that yield optimal encoders it is required to identify the relevant structure of the original signal necessary to guarantee that a satisfactory system performance can be achieved. To this end, the framework of rate-distortion theory approaches the optimal encoder problem by measuring the degree of compression via the mutual information between the compressed representation and the original signal, whereas the performance of the system is quantified by a user-provided distortion function [17]. In this way, the distortion function implicitly specifies which aspects of the original signal are relevant, and should be retained, in order to guarantee low distortion. A notable drawback to the rate-distortion framework is, however, the need to specify the distortion function, which may be difficult and non-intuitive for a given task [18].
In contrast, the information bottleneck (IB) method developed in [18] approaches the optimal encoder problem to preserve relevant information more directly. That is, the IB method considers an optimal encoder problem where the degree of achieved compression is captured by the mutual information between the compressed and original signals, and the model quality is measured by the mutual information between the compressed representation and an auxiliary variable which is assumed to contain task-relevant information. The IB approach is entirely data-driven, requiring only the joint distribution of the original signal and relevant information (i.e., the data) in order to be applied.
Owing to its general statistical formulation, the IB method, or some variation of it, has been considered in a number of studies [19,20,21,22,23,24,25,26]. Among these works, reference [19] develops an approach to obtain deterministic encoders to an IB-like problem motivated by reducing the number of clusters in the compressed space as opposed to designing encoders for communication. Consequently, the deterministic IB [19] measures the degree of achieved compression not by the mutual information between the original and compressed representations, as in communication systems, but by the entropy of the reduced space. The work of [20] considers the IB problem with side-information, allowing for both relevant and irrelevant structures to be provided to aid the identification of task-relevant information during the creation of signal encoders. The authors of [21,23] consider a multivariate extension of the IB principle, employing the use of Bayesian networks to specify the compression-relevance relations between the random variables to be maintained through the abstraction process. It should be noted that, while it does not directly employ the IB principle in its formulation, the empirical coordination problem [27,28] considers an information-theoretic compression problem over a graph, where interconnections between vertices represent communication links that agents may use to correlate their sequence of outcomes. Much like the multi-IB method [21,23], the network (communication) topology specifies the statistical dependencies that are possible in the empirical coordination problem. Observe, however, that the objective of the empirical coordination problem is to characterize the set of achievable joint distributions that are possible with various network topologies and communication (code) rates between vertices, whereas the multi-IB problem is a generalization of the encoder-design problem considered by the IB method to multivariate settings where the Bayesian networks are used to specify the relationships between source, reproduction and prediction (relevance) variables.
Other variants of the IB principle include the work in [22], where the authors consider the development of a bottom-up, agglomerative, hard clustering approach that employs the IB objective in determining which clusters to myopically merge at each step of the proposed algorithm. In related work inspired by the AIB problem, the authors of [29] exploit the structure of the AIB merging rule to design algorithms that form compressed representations of images by performing a sequence of greedy merges based on minimizing the stage-wise loss of relevant information at each iteration. Crucially, however, the algorithms developed in [29] do not consider the IB problem as they aim to design a sequence of myopic merges so as to minimize the loss of only relevant information, as compared with the much more challenging IB problem of simultaneously balancing information retention and information-theoretic compression. Moreover, in contrast to the work presented in this paper, the algorithms developed in [29] are not accompanied by theoretical performance guarantees that certify the optimality of the abstractions, nor are the methods readily extendable to cases where information from multiple sources must be considered in the design of compressed representations. Finally, the research conducted by the authors of [24] considers the IB problem in the setting of jointly-Gaussian data. More specifically, the authors of [24] established that when the original signal and the auxiliary (relevant) variable are jointly Gaussian, the solution to the IB problem is a noisy linear projection. For completeness, we note that when the data are not jointly Gaussian, or are described by a general probability density function, a solution to the IB problem is difficult to obtain. However, a number of studies have proposed methods leveraging variational inference in order to obtain approximate solutions to the IB problem in these cases [30,31,32,33].
Employing a unified viewpoint between abstractions in autonomy and those driven by information-theoretic principles, the authors of [34,35,36,37] developed frameworks for the emergence of abstractions in autonomous systems via methods inspired by information-theoretic signal compression. For example, the work of [34] employs the use of the IB principle to generate multi-resolution quadtree abstractions for planning, developing a framework that couples environment resolution, information and path value. Moreover, the research conducted in [35] utilizes environment abstractions to reduce the computational cost of evaluating mutual-information objective functions in active sensing applications. Of the reviewed works, those most closely related to the developments in this paper is that of [36,37], where the authors develop algorithms to select multi-resolution trees that are optimal with respect to the IB objective in both the soft-constrained (Lagrangian) [36] and hard-constrained [37] settings of the IB problem.
Inspired by the recent developments in information-theoretic driven approaches for generating abstractions for autonomous agents for the purposes of planning, the contribution of this paper is the development of a generalized information-theoretic framework that allows for multi-resolution tree abstractions to be obtained when multiple sources of relevance and irrelevance are specified. The incorporation of irrelevant information allows for connections between our framework and notions of information-theoretic privacy. Moreover, our generalized approach allows for abstractions to be refined by removing aspects of the relevant variables that are correlated with the irrelevant information structure, thus allowing for more compressed representations to emerge. This is especially critical in resource-constrained systems, which must make the best use of scarce on-board memory and bandwidth-limited communication channels.
The remainder of the paper is organized as follows. We begin in Section 3 with a brief overview of information-theoretic signal compression and detail the connection between hierarchical trees and signal encoders. Section 4 contains our formal problem statement. We propose and discuss solution approaches in Section 5. In Section 6, we present a discussion and comparison between the information-bottleneck method and the information-bottleneck problem with side-information (IBSI) in the setting of hierarchical tree abstractions. Examples and results are discussed in Section 7 before concluding remarks in Section 8. Proofs for the theoretical results presented in this paper are provided in the appendices.

2. Notation

Let R denote the set of real numbers and, for any integer n > 0 , let R n denote the n-dimensional Euclidean space. The set of non-negative real numbers is denoted by R + = { x R : x 0 } . For any vector x R n , [ x ] i is the i th element of the vector x for i { 1 , , n } . For any integer n > 0 , the collection of all non-negative n-dimensional vectors is denoted R + n = { x R n : [ x ] i 0 , 1 i n } . Given any two vectors x , y R n , the notation x y is understood component-wise; that is x y implies [ x ] i [ y ] i for all i { 1 , , n } . Unless otherwise stated, all logarithms are base e.

3. Preliminaries

The development of a framework for generating information-theoretic multi-resolution abstractions requires the introduction of concepts from both information theory and graph theory in order to rigorously define trees and encoder problems. We begin by introducing necessary topics from information theory before proceeding to introduce hierarchical trees and their connection to multi-resolution representations of the environment. In the interest of succinctness, we only introduce the relevant topics from information theory necessary for the developments of our framework, and refer the interested readers to [17,38] for a more comprehensive exposition of information theoretic principles and classical signal compression frameworks. We close this section by elucidating how multi-resolution trees can be viewed as deterministic encoders having a special structure, thereby allowing us to employ information-theoretic concepts from signal encoding theory to formulate the tree abstraction problem we consider in this paper.
Information-theoretic frameworks for compression model signals according to their statistical structure. Consequently, we require the introduction of a probability space. To this end, let ( Ω , F , P ) be a probability space with finite sample space Ω , σ -algebra F and probability measure P . We define the random variables X : Ω R , Y : Ω R , and T : Ω R , where the random variable X has probability distribution (mass function) assigned according to p ( x ) = P ( { ω Ω : X ( ω ) = x } ) , with the mass functions for Y and T defined analogously.

3.1. Mutual Information

Given two distributions p ( x ) and ν ( x ) over the same set of outcomes, the Kullback-Leibler (KL) divergence between the distributions p ( x ) and ν ( x ) is
D KL ( p ( x ) , ν ( x ) ) x p ( x ) log p ( x ) ν ( x ) .
The KL-divergence is non-negative and equals zero if and only if p ( x ) = q ( x ) for all outcomes x [17]. The mutual information between two random variables X and T is defined in terms of the KL-divergence as
I ( T ; X ) D KL ( p ( t , x ) , p ( t ) p ( x ) ) = t , x p ( t , x ) log p ( t , x ) p ( t ) p ( x ) .
The mutual information is symmetric (i.e., I ( T ; X ) = I ( X ; T ) ), non-negative, and equals zero if and only if p ( t , x ) = p ( t ) p ( x ) . The mutual information plays an important role in signal compression theory where it represents the code rate, or average number of bits per source symbol. Consequently, if T is a compressed representation (or reproduction) of X, then lower values of I ( T ; X ) correspond to greater degrees of achieved compression. In the more general setting, the mutual information is a measure of the degree of statistical correlation between the random variables X and T, where I ( T ; X ) = 0 if and only if X and T are independent. The mutual information also satisfies
I ( T ; X ) = H ( X ) H ( X | T ) = H ( T ) H ( T | X ) ,
where H ( X ) is the Shannon entropy (the Shannon entropy of the random variable X is H ( X ) = x p ( x ) log p ( x ) ) of the random variable X and H ( T | X ) is the conditional entropy, measuring the average uncertainty in T when given knowledge of X. Lastly, the Jensen-Shannon divergence between a collection of probability distributions { p 1 ( x ) , , p n ( x ) } with weights Π R + n is given by
JS Π ( p 1 ( x ) , , p n ( x ) ) = i = 1 n [ Π ] i D KL ( p i ( x ) , p ¯ ( x ) ) ,
where 0 [ Π ] i 1 for all i { 1 , , n } , i = 1 n [ Π ] i = 1 and p ¯ ( x ) = i = 1 n [ Π ] i p i ( x ) [22,39,40].

3.2. Trees and Trees as Encoders

Our goal in this paper is to leverage information-theoretic signal compression principles in order to generate abstractions for autonomous systems in the form of multi-resolution tree structures. However, existing frameworks for signal encoding, such as rate-distortion theory [17] or the information bottleneck (IB) method [18], do not impose any structural constraints on the resulting encoder in order to guarantee that the solution corresponds to a tree representation. The added constraint poses a significant challenge, as existing methods do not consider such limitations on the set of feasible encoders. To tackle this problem, we will elucidate how trees can be viewed as encoders with a specific structure.
We assume that the environment W R d is a d-dimensional grid-world and that there is an integer > 0 such that the environment is contained within a hypercube of side length 2 . A hierarchical, multi-resolution depiction of W can be represented as a tree (a tree is a connected acyclic graph [41]) T = ( N ( T ) , E ( T ) ) where N ( T ) is a collection of nodes and E ( T ) is a collection of edges that describe the nodal interconnections [41]. We will henceforth limit the discussion to the case when the tree structure is that of a quadtree, however it should be noted that the theory developed in this paper applies straightforwardly to general tree structures. Given an environment W , we will take T Q to denote the set of all feasible quadtree representations of W , and let T W T Q be the tree whose leafs define the finest resolution depiction of W . An example is shown in Figure 1. In the sequel, we follow the notation and definitions of [36,37]. To this end, we let N k ( T q ) denote all the nodes of the tree T q T Q at depth k { 0 , , } , and, for any t N ( T W ) , C ( t ) will denote the set of children of t. The set of leafs of T q T Q is given by N leaf ( T q ) and the interior node set is N int ( T q ) = N ( T q ) N leaf ( T q ) .
Given a formal definition of a tree, we are now ready to discuss the connection between hierarchical trees and signal encoders. To this end, it was noted in [36,37] that hierarchical tree abstractions of W can be viewed as deterministic encoders having a specific structure. To this end, notice from Figure 1 that by changing the tree T T Q we alter the multi-resolution representation of the environment W . Moreover, any tree T T Q can be created by aggregating finest resolution cells to some parent node in the tree in such a way that the resulting tree is in the space T Q .
To make the connection to an information-theoretic framework for compression more precise, we let X : Ω R be the random variable corresponding to the uncompressed signal. In our setting, the uncompressed signal can be the original map of the environment, and therefore the outcomes of X are the finest resolution grid cells of W . For example, in the full-resolution ( 4 × 4 ) environment T W depicted in Figure 1, we have X : Ω { x 1 , , x 16 } . Notice that each tree T q T Q defines a compressed random variable T q : Ω R whose outcomes are the elements of the set N leaf ( T q ) . The relationship between X and T q can be characterized by a deterministic encoder p q ( t | x ) where p q ( t | x ) = 1 if and only if the finest resolution cell x N leaf ( T W ) is aggregated to the node t N leaf ( T q ) in the tree T q . However, it is important to note that not all deterministic encoders correspond to valid tree representations of W , which is a challenge we will discuss in the development of our proposed solution approach.
The observation that a tree can be represented as a deterministic encoder, allows us to express information-theoretic quantities as a function of the tree, described next. Consider, for example, the case when a given joint distribution p ( x , y ) is provided, describing how the finest resolution cells are correlated with a specified random variable Y, which we assume contains task-relevant information. Imagine now that we wish to compress the signal X in the form of a hierarchical tree so that the resulting tree is maximally retentive regarding the relevant variable Y. The resulting joint distribution p q ( t , x , y ) can be computed according to p q ( t , x , y ) = p q ( t | x ) p ( x , y ) , which is a function of the tree T q T Q , where we have employed the fact that T q is conditionally independent of Y when given X. From this, we note that the distributions p q ( t , y ) , p q ( t , x ) , p q ( t ) , p ( x ) and p ( y ) can be obtained via the appropriate marginalization of the joint distribution p q ( t , x , y ) . Therefore, we can write the mutual information as a function of the tree as
I X ( T q ) I ( T q ; X ) = t , x p q ( t , x ) log p q ( t , x ) p q ( t ) p ( x ) ,
and
I Y ( T q ) I ( T q ; Y ) = t , y p q ( t , y ) log p q ( t , y ) p q ( t ) p ( y ) ,
where I X ( T q ) quantifies the degree of compression and I Y ( T q ) quantifies the amount of relevant information contained in the tree T q . In this setting, we note that the distributions p ( x ) and p ( y ) do not depend on the tree T q , as they can be obtained directly from the input distribution p ( x , y ) . In an analogous manner, one may also define the amount of irrelevant information, represented by a random variable Z, contained in the tree T q T Q as
I Z ( T q ) I ( T q ; Z ) = t , z p q ( t , z ) log p q ( t , z ) p q ( t ) p ( z ) ,
where we assume that p ( x , y , z ) is provided and p q ( t , x , y , z ) = p q ( t | x ) p ( x , y , z ) .
The expressions (5)–(7) may be generalized to the case where we have a collection { Y 1 , , Y n } of relevant and { Z 1 , , Z m } of irrelevant variables, respectively, as follows. Given the joint distribution p ( x , y 1 , , y n , z 1 , , z m ) specifying the correlations between relevant and irrelevant variables, the information of each variable contained in the tree T q T Q is given by
I Y i ( T q ) = I ( T q ; Y i ) = t , y i p q ( t , y i ) log p q ( t , y i ) p q ( t ) p ( y i ) , i { 1 , , n } ,
and
I Z j ( T q ) = I ( T q ; Z j ) = t , z j p q ( t , z j ) log p q ( t , z j ) p q ( t ) p ( z j ) , j { 1 , , m } ,
where p q ( t , x , y 1 , , y n , z 1 , , z m ) = p q ( t | x ) p ( x , y 1 , , y n , z 1 , , z m ) . Having related trees to signal encoders and showing how mutual information terms can be written as a function of the tree T q T Q , we now turn to formally state the problem we consider for the remainder of the paper.

4. Problem Formulation

In Section 3, we discussed the relation between trees and signal encoders and showed how the observation that a tree T q T Q can be represented as a deterministic encoder p q ( t | x ) allows us to quantify the information contained in the tree. With these observations, we can now formally state the problem we consider in this paper.
Problem 1.
Given the environment W , vectors β R + n and γ R + m , a scalar α 0 and the joint distribution p ( x , y 1 , , y n , z 1 , , z m ) , consider the problem of maximizing
max T q T Q i = 1 n [ β ] i I Y i ( T q ) j = 1 m [ γ ] j I Z j ( T q ) α I X ( T q ) .
It should be noted that Problem 1 cannot be solved by applying existing algorithms (e.g., the Blahut-Arimoto algorithm [17,18] or the iterative IB method [18,39]) from signal encoding theory as the set of feasible solutions is discrete, in addition to the presence of the constraint that p q ( t | x ) must correspond to a tree T q T Q . The added constraint poses significant technical challenges, as it is not obvious how this constraint can be represented mathematically so as to render Problem 1 solvable via numerical methods. Moreover, as a result of the discrete nature of T Q , it follows that Problem 1 cannot be solved via standard (sub-)gradient approaches from optimization theory, as it belongs to a class of combinatorial optimization problems. Despite these challenges, in the next section we propose a novel and tractable numerical algorithm to find a solution to Problem 1 with theoretical guarantees.
Before proceeding, we provide a few comments regarding the relation of Problem 1 to other areas of research. Namely, Problem 1 is similar to problems considered in the information-theoretic security community [42,43,44] where { Z 1 , , Z m } are viewed as private variables whose information content we wish not to disclose to an un-trusted party. In this setting, the value of I Z j ( T ) represents the amount of private information disclosed by the tree T T Q and the vector of weights γ R + m encodes the relative cost of private information disclosure, allowing for the privacy variables to be distinguished in their importance of revelation. Alternatively, one may interpret the privacy aspects of Problem 1 via conditional entropy. Using (3) and (7), we can write I Z j ( T q ) = H ( Z j ) H ( Z j | T q ) and note that H ( Z j ) is constant, given the data p ( x , y 1 , , y n , z 1 , , z m ) . Consequently, performing the maximization in Problem 1 encourages solutions T q T Q for which H ( Z j | T q ) is as large as possible, amounting to trees that attempt to make Z j and T q independent since H ( Z j | T q ) H ( Z j ) . Then, Fano’s inequality ([17], pp. 37–41) implies that the lower bound of the error probability of any estimator designed to infer Z j from T q increases as a function of H ( Z j | T q ) . It follows that when H ( Z j | T q ) large, the probability of error when estimating the value of Z j from T q increases [17,42]. Consequently, information regarding Z j remains protected.
It should be noted that the incorporation of additional irrelevant variables when designing abstractions has been considered in other works. Previous approaches that introduce irrelevance variables when forming abstractions, such as the IBSI method [20], employ the viewpoint that the information provided via { Z 1 , , Z m } is general task-irrelevant information, with no motivation from an information-theoretic security standpoint. In the IBSI approach, the incorporation of irrelevant information helps improve the quality of abstractions with respect to the task-relevant variable, as aspects of the task-relevant variable that are correlated with the irrelevant information can be discarded when forming the compressed representations. In summary, we note that, while our formulation given by Problem 1 can be interpreted from an information-theoretic security standpoint, the main motivation for our approach is not one of security. Rather, it is the development of a general information-theoretic framework that allows for both relevant and irrelevant information to be specified and balanced versus compression in the design of multi-resolution tree abstractions for autonomous systems. However, as the discussion above shows, the proposed framework could also be useful in obscuring private information contained in quadtree abstractions.

5. Solution Approach

In this section, we discuss an approach to find a solution to Problem 1 and introduce a tractable numerical algorithm that searches for an optimal tree as a function of the weight parameters β R + n , γ R + m , and α 0 . In what follows, it will be convenient to write the objective of Problem 1 in terms of the function J : T Q × R + n × R + m × R + R , defined by
J ( T q ; β , γ , α ) = i = i n [ β ] i I Y i ( T q ) j = 1 m [ γ ] j I Z j ( T q ) α I X ( T q ) .
Then our problem is one of selecting a tree T q ˜ T Q such that
T q ˜ argmax T q T Q J ( T q ; β , γ , α ) .
The evaluation of the objective (10) for a given tree T q T Q may be computationally expensive, as it requires the computation of each joint distribution p ( t , y i ) , i { 1 , , n } , and p ( t , z j ) , j { 1 , , m } , as well as the evaluation of the mutual information terms (8) and (9), each of which requires summation over the sample spaces of Ω T q , Ω Y i and Ω Z j for each i { 1 , , n } and j { 1 , , m } . Such a computation is especially burdensome if the sample spaces have a large number of elements. Instead, we seek an easier, less computationally costly incremental approach toward evaluating the objective (10) for any T q T Q .
To this end, we write the objective (10) for any T q T Q as
J ( T q ; β , γ , α ) = J ( T 0 ; β , γ , α ) + u = 0 q 1 [ J ( T u + 1 ; β , γ , α ) J ( T u ; β , γ , α ) ] ,
where { T 0 , , T q 1 } T Q is a collection of trees in the space T Q . While the relation (12) is valid for any tree T q T Q and any collection { T 0 , , T q 1 } T Q , it was noted in [36,37] that when the tree T 0 T Q and the sequence { T 0 , , T q 1 } T Q is selected in a specific way, the objective (12) reduces to a special form. Specifically, if we select T 0 T Q as the tree consisting of a single node where all finest resolution cells are aggregated, and the sequence { T 0 , , T q 1 } T Q is constructed by expanding a leaf node of T u to create T u + 1 for u { 0 , , q 1 } , then (12) can be expressed in terms of the local changes made in moving from the tree T i to T i + 1 . Formally, when the tree T 0 T Q is selected to be the root tree (the root tree R W is the tree R W T Q such that N int ( R W ) = ), and the sequence { T u } u = 0 q is constructed so that N ( T u + 1 ) N ( T u ) = C ( t ) = { t 1 , , t 4 } for some t N leaf ( T u ) for all u { 0 , , q 1 } , the objective (12) takes the form
J ( T q ; β , γ , α ) = s N int ( T q ) Δ J ( s ; β , γ , α ) ,
where
Δ J ( t ; β , γ , α ) = i = 1 n [ β ] i Δ I Y i ( t ) j = 1 m [ γ ] j Δ I Z j ( t ) α Δ I X ( t ) ,
and Δ I Y i ( t ) = I Y i ( T u + 1 ) I Y i ( T u ) , Δ I Z j ( t ) = I Z j ( T u + 1 ) I Z j ( T u ) , Δ I X ( t ) = I X ( T u + 1 ) I X ( T u ) are given by
Δ I Y i ( t ) = p ( t ) JS Π ( p ( y i | t 1 ) , , p ( y i | t 4 ) ) , i { 1 , , n } ,
Δ I Z j ( t ) = p ( t ) JS Π ( p ( z j | t 1 ) , , p ( z j | t 4 ) ) , j { 1 , , m } ,
Δ I X ( t ) = p ( t ) H ( Π ) ,
p ( y i | t ) = u = 1 4 [ Π ] u p ( y i | t u ) ,
p ( z j | t ) = u = 1 4 [ Π ] u p ( z j | t u ) ,
p ( t ) = u = 1 4 p ( t u ) ,
Π = p ( t 1 ) p ( t ) , p ( t 2 ) p ( t ) , p ( t 3 ) p ( t ) , p ( t 4 ) p ( t ) .
The relations (15)–(21) are computed via direct calculation in terms of the difference in mutual information between two encoders corresponding to the trees T u + 1 , T u T Q that satisfy N ( T u + 1 ) N ( T u ) = C ( t ) for some t N leaf ( T u ) . Observe that the condition N ( T u + 1 ) N ( T u ) = C ( t ) for some t N leaf ( T u ) implies that the trees T u + 1 and T u differ only by a single nodal expansion. An example is shown in Figure 2. To show that the term J ( T 0 ; β , γ , α ) = 0 in (12) when the tree T 0 is taken to be the root tree, we note from (3) that 0 I ( T 0 ; Y i ) H ( T 0 ) , 0 I ( T 0 ; Z j ) H ( T 0 ) and 0 I ( T 0 ; X ) H ( T 0 ) . Then, since the root tree has only a single leaf node, it follows that the distribution p 0 ( t ) is deterministic. As a result, H ( T 0 ) = 0 and thus J ( T 0 ; β , γ , α ) = 0 .
It is important to note that the incremental relations (15)–(21) depend only on the node t N leaf ( T u ) expanded in moving from the tree T u to T u + 1 , and not on any other nodes in the tree. As a result, the evaluation of the incremental changes in information are dependent only on the changes induced by expanding the node t N leaf ( T u ) , thereby alleviating the need to sum over all the outcomes of the random variable T u as otherwise required in order to evaluate the mutual information. Furthermore, the observation that the objective and information terms can be decomposed into an incremental form according to (13)–(21) allows for tractable algorithms to be designed in order to obtain a solution to Problem 1. Lastly, it is important to note that there is no loss of generality in using the expression (13). To see why this is the case, we present the following definition.
Definition 2
([41]). A tree G = ( N ( G ) , E ( G ) ) is asubtreeof the tree H = N ( H ) , E ( H ) , denoted by G H , if N ( G ) N ( H ) and E ( G ) E ( H ) .
Note that the root tree is a subtree of every tree in the space T Q . As a result, one can always express the cost (10) as (13), since each tree T q T Q can be obtained by starting at the root tree T 0 T Q and creating a sequence { T u } u = 0 q such that N ( T u + 1 ) N ( T u ) = C ( t ) for some t N leaf ( T u ) and all u { 0 , , q 1 } . Next, we leverage the structure of our problem to design a tractable algorithm in order to find the solution to Problem 1.

5.1. The Generalized Tree Search Algorithm (G-Tree Search)

In this section, we show how the structural properties of Problem 1 discussed in the previous section can be exploited in order to yield a tractable algorithm to find a multi-resolution tree that is a solution to (11). Specifically, among all trees T T Q , we seek those trees that ensure no improvement of (10) is possible, as these trees provide the best trade-off between relevant information retention, irrelevant information removal, and compression. The following definition establishes the notion of optimality we employ throughout this paper.
Definition 3.
A tree T T Q isoptimal with respect to Jif J ( T ˜ ; β , γ , α ) J ( T ; β , γ , α ) for all trees T ˜ T Q .
To differentiate between candidate solutions, we specify additional properties considered favorable for an optimal multi-resolution tree. One such property is that the tree be minimal, which is defined as follows.
Definition 4.
A tree T T Q isminimal with respect to Jif J ( T ˜ ; β , γ , α ) < J ( T ; β , γ , α ) for all trees T ˜ T Q such that T ˜ T .
A tree that is both optimal and minimal will be called an optimal minimal tree. Importantly, an optimal minimal tree is guaranteed to not contain any redundant nodal expansions. In other words, removing any portion of an optimal minimal tree is guaranteed to result in a pruned tree that is strictly worse with respect to the objective function. In contrast, if an optimal tree is not minimal, then some portion(s) of the tree can be pruned with no loss in the objective value, indicating that the non-minimal tree contains redundant nodal expansions. Thus, of all optimal trees, the minimal solution is preferred as it contains the fewest number of leaf nodes among solution candidates and also requires the least amount of resources to store in memory. Our goal is then to design an algorithm that returns, as a function of β R + n , γ R + m and α 0 , an optimal minimal tree.
In theory, one may take a number of approaches to find a solution (not necessarily optimal) to Problem 1. One approach is the brute-force method of generating each tree in the space T Q and picking one that satisfies (11); a process which is akin to grid-search methods in optimization theory. However, such an exhaustive approach does not scale well to large environments. Alternatively, one may notice that the node-wise structure of the cost (14) renders the implementation of a greedy approach straightforward. Specifically, given any tree T u T Q one may expand the leaf node t N leaf ( T u ) that results in the greatest change in the cost Δ J ( t ; β , γ , α ) . By expanding a node t N leaf ( T u ) , we remove { t } and add its children C ( t ) to the leaf set to generate the tree T u + 1 , leaving other nodes unchanged. One may continue this process until a tree is reached for which no further improvement is possible, as quantified by the one-step incremental objective value Δ J ( t ; β , γ , α ) . This myopic steepest-ascent-like approach is not guaranteed to find an optimal solution, however, as the process may fail to identify expansions that are suboptimal with respect to the one-step objective Δ J ( t ; β , γ , α ) , but lead to higher-valued expansions in future iterations. Consequently, we seek to incorporate the value of expansions-to-come when deciding whether or not to expand a leaf node of the current tree.
To this end, we introduce a generalized tree search algorithm we call G-tree search. The G-tree search algorithm works from top-down, starting at the root tree R W T Q and utilizing the function G : N ( T W ) × R + n × R + m × R + R + , defined as
G ( t ; β , γ , α ) = max { Δ J ( t ; β , γ , α ) + t C ( t ) G ( t ; β , γ , α ) , 0 } , if t N int ( T W ) , 0 , otherwise ,
in order to decide which nodes to expand. Specifically, given any tree T u T Q , G-tree search will inspect the G-values, computed according to (22), for each node t N leaf ( T u ) and expand a node t N leaf ( T u ) for which G ( t ; β , γ , α ) > 0 . Once a node t N leaf ( T u ) is selected for expansion, a new tree T u + 1 T Q is defined by removing the node t and adding its children, C ( t ) , to the set of leafs, leaving the other nodes in the tree T u unchanged. In this way, the tree T u + 1 is related to T u via N leaf ( T u + 1 ) = ( N leaf ( T u ) { t } ) C ( t ) . The process then repeats until we find a tree T q ˜ T Q for which there does not exist t N leaf ( T q ˜ ) such that G ( t ; β , γ , α ) > 0 . Note that by designing the algorithm in this way, the constraint T T Q is naturally enforced. The G-tree search method is detailed in Algorithm 1. Note that the pseudo-code for a greedy tree search is identical to that of G-tree search in Algorithm 1 with each G ( t ; β , γ , α ) replaced by Δ J ( t , β , α , γ ) . We will discuss the shortcomings of the greedy approach in more detail in Section 6.1.
Algorithm 1 The G-tree Search Algorithm.
Entropy 24 00809 i001
     A few comments are in order regarding the G-tree search method. First, the routine ComputeGvalues ( · ) populates the G-values, as follows. The routine utilizes the joint distribution p ( x , y 1 , , y n , z 1 , , z m ) in order to compute the values of Δ I Y i ( t ) , Δ I Z j ( t ) and Δ I X ( t ) for all i { 1 , , n } , j { 1 , , m } and t N int ( T W ) . Given the weights ( β , γ , α ) R + n × R + m × R + one may compute Δ J ( t ; β , γ , α ) and apply the rule (22) to obtain the G-values via a recursion that begins at the leafs of T W . The pseudo-code for the ComputeGvalues procedure is shown in Algorithm 2. Lastly, the function UpdateInformation ( t ) updates the information contained in the tree at the current time-step of the solution. It does so by utilizing the values of Δ I Y i ( t ) , Δ I Z j ( t ) and Δ I X ( t ) for each i { 1 , , n } and j { 1 , , m } , which were computed in the process of evaluating the nodal G-values described above. The information contained in the tree T u + 1 is then given by I Y i ( T u + 1 ) = I Y i ( T u ) + Δ I Y i ( t ) , I Z j ( T u + 1 ) = I Z j ( T u ) + Δ I Z j ( t ) and I X ( T u + 1 ) = I X ( T u ) + Δ I X ( t ) where i { 1 , , n } and j { 1 , , m } . Recall that starting the algorithm at the root tree T 0 T Q implies, for all i and j, we have I Y i ( T 0 ) = 0 , I Z j ( T 0 ) = 0 and I X ( T 0 ) = 0 .
Algorithm 2 The ComputeGvalues routine.
Entropy 24 00809 i002

5.2. Theoretical Analysis of the G-Tree Search Algorithm

In this section, we discuss the theoretical properties of the G-tree search algorithm introduced in Section 5.1. Our main result is that the G-tree search algorithm returns an optimal minimal tree. In our analysis, we will oftentimes refer to the part of a tree T q T Q that is descendant (or rooted) at some node t N ( T q ) . To make this notion precise, we have the following definition.
Definition 5
([36]). Let t N ( T q ) be a node in the tree T q T Q .The subtree of T q T Q rooted at nodet is denoted by T q ( t ) and has node set
N T q ( t ) = t N ( T q ) : t i D i ,
where D 1 = t , D i + 1 = A D i , and
A D i = t N ( T W ) : t t ^ D i C t ^ .
An example of a subtree is shown in Figure 3. Each time the G-tree search visits and expands a node t N int ( T W ) , the algorithm can be viewed as determining the part of the subtree rooted at t for which a net increase in the objective can be achieved. For example, consider the case when the algorithm is provided with a tree T q T Q . In order to determine whether or not expanding some t N leaf ( T q ) will lead to a tree of greater objective value than T q , the algorithm must determine if expanding the node t leads to future expansions that result with a tree T q ˜ T Q for which J ( T q ˜ ; β , γ , α ) > J ( T q ; β , γ , α ) . Of course, if Δ J ( t ; β , γ , α ) > 0 for some t N leaf ( T q ) , then it is clear that expanding the node t leads to a tree that improves the value of the objective. However, when Δ J ( t ; β , γ , α ) 0 , the decision of whether or not to expand t N leaf ( T q ) is not so clear, as the algorithm must then consider if, by continuing the expansion process along the children of t, can result in a tree that improves of the overall objective value. In essence, we are interested in investigating how the G-function in (22) relates to the incremental objective value of a subtree rooted at any t N int ( T W ) and to show that, if there exists a subtree rooted at t that results in a overall improvement of the objective, then G ( t ; β , γ , α ) > 0 . To answer this question, we have the following results.
Lemma 6.
Let t N int ( T W ) , β R + n , γ R + m and α 0 . Then, G ( t ; β , γ , α ) s N int ( T q ( t ) ) Δ J ( s ; β , γ , α ) for all T q T Q .
Proof. 
The proof is presented in Appendix A. □
Corollary 7.
Let t N int ( T W ) , β R + n , γ R + m and α 0 . If there exists a tree T q T Q such that s N int ( T q ( t ) ) Δ J ( s ; β , γ , α ) > 0 then G ( t ; β , γ , α ) > 0 .
Proof. 
The result is immediate from Lemma 6. □
As a result of Lemma 6 and Corollary 7, we are guaranteed that, if there is a subtree rooted at t for which an increase in the overall objective is possible, then G ( t ; β , γ , α ) > 0 . Furthermore, we are guaranteed that the value of the G-function (22) is bounded below by the incremental value of the objective contributed by any subtree rooted at t. The converse to Lemma 6 and Corollary 7 is also of important; namely if we know G ( t ; β , γ , α ) > 0 for some t N int ( T W ) , then it is of interest in establishing whether or not this implies that there is a subtree rooted at t for which a net increase in the objective is possible. This leads us to the following results.
Lemma 8.
Let t N int ( T W ) , β R + n , γ R + m and α 0 . If G ( t ; β , γ , α ) > 0 , then there exists a tree T q T Q such that G ( t ; β , γ , α ) = s N int ( T q ( t ) ) Δ J ( s ; β , γ , α ) .
Proof. 
The proof is presented in Appendix B. □
Corollary 9.
Let t N int ( T W ) , β R + n , γ R + m and α 0 . If G ( t ; β , γ , α ) > 0 , then there exists a tree T q T Q such that s N int ( T q ( t ) ) Δ J ( s ; β , γ , α ) > 0 .
Proof. 
The proof follows from Lemma 8. □
Importantly, Lemma 8 establishes the connection between the G-function and the incremental objective value, as well as the existence of a subtree rooted at t N int ( T W ) for which a net positive objective increment is possible, in the case when G ( t ; β , γ , α ) > 0 . Moreover, Lemma 8 and Corollary 9 together guarantee that if the G-value of a node is strictly positive, then there exists a subtree rooted at the node t N int ( T W ) such that expanding t (and possibly continuing the expansions process along the children of t) will result in a tree that has strictly greater objective value. Also, observe that by combining the results of Corollaries 7 and 9 we obtain the following lemma.
Lemma 10.
Let t N int ( T W ) , β R + n , γ R + m and α 0 . Then G ( t ; β , γ , α ) > 0 if and only if there exists a tree T q T Q such that s N int ( T q ( t ) ) Δ J ( s ; β , γ , α ) > 0 .
Proof. 
The result is a consequence of Corollaries 7 and 9. □
Lemma 10 is important, as it provides necessary and sufficient conditions linking the existence of a subtree rooted at any node t N int ( T W ) to the value of the nodal G-function value. We are now in a position to prove the optimality of the solution returned by G-tree search, as stated by the following theorem.
Theorem 11.
Assume β R + n , γ R + m and α 0 . Then, the G-tree search algorithm returns an optimal minimal tree with respect to J.
Proof. 
The proof is given in Appendix C. □
As a consequence of Theorem 11, we can guarantee that for any set of parameters ( β , γ , α ) R + n × R + m × R + , the G-tree search algorithm will return a tree that is the optimal and minimal solution to Problem 1.

5.3. Complexity Analysis

While Theorem 11 establishes that the G-tree search algorithm introduced in Section 5.1 returns an optimal minimal tree that satisfies (11), it is also important to characterize the number of operations required to execute the algorithm, in the worst case. To this end, we note that a tree T W T Q corresponding to some grid in the d-dimensional space with side-length 2 has N = k = 0 2 d k total nodes, where N = 2 d k = 0 2 d ( k ) = | N leaf ( T W ) | ( k = 0 2 d ( k ) ) | N leaf ( T W ) | ( M M 2 d / ( 2 d 1 ) ) 2 | N leaf ( T W ) | for any integer d > 0 . As a result, the number of nodes in the tree is on the order of the number of leaf nodes of T W . Thus, executing the G-tree search algorithm in Algorithm 1 once the G-function is known, requires order | N leaf ( T W ) | operations, as the search may visit, in the worst case, every node in the tree. Now, note that for a given number of relevant variables n > 0 and irrelevant variables m 0 , the computation of the G-function requires on the order of n + m + 2 operations per node in the tree, corresponding to the calculation of Δ I Y i ( t ) , Δ I Z j ( t ) and Δ I X ( t ) and G-function values. Thus, visiting each node in the tree requires on the order of ( n + m + 2 ) | N leaf ( T W ) | operations. Consequently, updating the G-values and running the G-tree search requires on the order of ( n + m + 3 ) | N leaf ( T W ) | operations in the worst case for a given setting of the problem.

6. The IB and IBSI Principles as Special Cases

In this section, we show how obtaining multi-resolution trees via the information bottleneck (IB) [18] and the information bottleneck with side-information (IBSI) [20] are special cases of Problem 1. Moreover, we establish a relation between these two approaches, showing how a tree solution to the IBSI problem can be obtained from the solution of the IB problem, the latter of which does not consider the removal of irrelevant information from the abstractions. Facilitating a theoretical connection between these two approaches helps us understand the impact of irrelevant information on the resulting tree solutions, and furthermore allows us to incorporate irrelevant information after-the-fact, which is useful in applications when only relevant details are known ahead of time. We begin with a brief overview of the IB and IBSI problems in the context of our problem.

6.1. Multi-Resolution Trees via the IB Principle and Drawbacks of Myopic Tree Search

Recall from Section 3 that the IB problem furnishes an approach to design compressed representations of the original signal that are maximally informative regarding task-relevant information. The IB method was first introduced in [18] and considers the optimization problem
min p ( t | x ) I ( T ; X ) β I ( T ; Y ) ,
where the minimization is over all conditional distributions p ( t | x ) , and β 0 trades the importance of compression and relevant information retention. It is important to note that the original formulation of the IB problem does not impose any constraints on the encoder p ( t | x ) beyond those required to ensure that p ( t | x ) is a valid probability distribution. However, as discussed in Section 3, multi-resolution trees can be viewed as deterministic encoders that have special structure. Consequently, one may employ the IB principle in order to generate multi-resolution abstractions that compress the environment and retain task-relevant information, which is given by the problem
max T T Q I Y ( T ) 1 β I X ( T ) ,
where (24) is obtained by multiplying (23) by the constant M M 1 / β for β > 0 and restricting the search to the space T Q . Recently, the problem (24) was considered by the authors of [36], who introduce an algorithm called Q-tree search that returns, as a function of β > 0 , an optimal solution to (24). Interestingly, the Q-tree search algorithm emerges as a special case of the more general G-tree search method developed in this paper by defining
Δ L ( t ; β ) = Δ J ( t ; 1 , 0 , M M 1 / β ) ,
and by taking
Q ( t ; β ) = G ( t ; 1 , 0 , M M 1 / β ) .
As a result, we see that the Q-tree search method in [36] is a special case of G-tree search, where there is only a single relevant variable with unit weight, there is no irrelevant information, and α = M M 1 / β .
The single variable case also provides some intuition into the differences between G-tree search and an approach that relies on a one-step steepest-ascent (greedy) method to find a solution to (24). The pseudo-code for a greedy tree search is the same as that of G-tree search in Algorithm 1 with each G ( t ; β , γ , α ) replaced with Δ J ( t ; β , γ , α ) . Implementing a greedy approach, a visited node is expanded only if the one-step cost Δ J ( t ; β , γ , α ) > 0 . In the single variable case, this means from (25) that
Δ L ( t ; β ) = p ( t ) JS Π ( p ( y | t 1 ) , , p ( y | t 4 ) ) 1 β H ( Π ) > 0 ,
or, equivalently (it can be shown that Δ L ( t ; β ) 0 as p ( t ) 0 . See ([36], Proposition 1) for more details), JS Π ( p ( y | t 1 ) , , p ( y | t 4 ) ) 1 β H ( Π ) > 0 , where t i C ( t ) for all i { 1 , , 4 } . By changing β > 0 , we alter the preference between trees that represent highly compressed versions of X (low β ) and trees that are more informative regarding Y (large β ). From (27), we see that the critical value of β > 0 for the node t N int ( T W ) , denoted β cr : N int ( T W ) [ 0 , ] , is given by
β cr ( t ) = H ( Π ) JS Π ( p ( y | t 1 ) , , p ( y | t 4 ) ) , if JS Π ( p ( y | t 1 ) , , p ( y | t 4 ) ) > 0 , , otherwise ,
where C ( t ) = { t 1 , , t 4 } . The node-wise critical β -values, β cr ( t ) , determine at what point the node will be expanded, should it be visited by the greedy search algorithm. Namely, if β β cr ( t ) then the node t N ( T W ) will not be expanded, even if the algorithm were to visit the node during the search. Moreover, we see from (28) that among those nodes for which H ( Π ) is constant, those that have a greater diversity among the distributions { p ( y | t 1 ) , , p ( y | t 4 ) } will have lower critical β -values than those nodes with less diversity in the conditional distributions { p ( y | t 1 ) , , p ( y | t 4 ) } . Intuitively, what this means is that nodes that provide more Y-information for a fixed amount of X-information will be expanded at lower values of β compared with those that provide less Y-information (e.g., more homogeneous cells), should these nodes be reached.
On the other hand, (28) also shows why a greedy approach may fail to find an optimal solution, even in the single variable setting. To see why this is the case, consider the 4 × 4 environment in Figure 4. In this example, Y : Ω { 0 , 1 } is the relevant variable, where the grid shading shows the distribution p ( y | x ) . The environment is symmetric in the sense that each of the quadrants contains one cell for which p ( y | x ) = c for some 0 < c 1 . If we assume p ( x ) to be a uniform distribution, then when t is taken to be the root node of T W , one will find that JS Π ( p ( y | t 1 ) , , p ( y | t 4 ) ) = 0 , H ( Π ) > 0 , where { t 1 , , t 4 } = C ( t ) . Consequently, from (28), we have β cr ( t ) = . As a result, a greedy implementation will never expand the root node and one will always recover the trivial abstraction, even though there is relevant information in the environment. In contrast, the G-tree search method does not suffer from this drawback, as it incorporates the reward-to-come of future expansions, thereby allowing it to find trees that recover all the relevant information in the environment for finite values of β .

6.2. Multi-Resolution Trees via the IB and IBSI Principles

In a similar manner to the IB method considered in Section 6.1, the IBSI approach [20] considers the problem
min p ( t | x ) I ( T ; X ) β I ( T ; Y ) + γ I ( T ; Z ) ,
where the minimization is over all conditional distributions p ( t | x ) , and γ 0 weights the relative importance of relevant information retention and irrelevant information removal. Constrained to the space of multi-resolution tree abstractions, we obtain the IBSI problem over the space of trees given by
max T T Q I Y ( T ) 1 β [ γ I Z ( T ) I X ( T ) ] .
The IBSI problem over the space of trees is therefore a special case of Problem 1 where the relevant information has unit weight, the irrelevant information has weight M M γ / β and α = M M 1 / β for β > 0 . In (30), the scalar γ 0 specifies the relative importance of relevant information retention and irrelevant information removal, whereas β > 0 balances the importance of compression. We can solve (30) via G-tree search by defining
Δ M ( t ; β , γ ) = Δ J ( t ; 1 , M M γ / β , M M 1 / β ) ,
and taking the function S : N ( T W ) × ( 0 , ) × R + R + to be given by the rule
S ( t ; β , γ ) = G ( t ; 1 , M M γ / β , M M 1 / β ) .
We will call the special case of the G-tree search algorithm implemented with the G-function in (32) S-tree search. As a result of (26) and (32), we see that multi-resolution trees via the IB or IBSI principles are obtained by employing the G-tree search. Moreover, the resulting abstractions are guaranteed to be minimal and optimal with respect to their objectives, as specified by Theorem 11.
While (26) and (32) show how trees via the IB and IBSI principles can be obtained as special cases of G-tree search, these relations do not provide us with an understanding of how the introduction of irrelevant information changes the solution of the problem. Thus, to better understand the impact of the presence of irrelevant information on the resulting tree abstractions, we present the following results.
Lemma 12.
Let t N ( T W ) . Then S ( t ; β , γ ) Q ( t ; β ) for all γ 0 and β > 0 .
Proof. 
The proof is presented in Appendix D. □
Corollary 13.
Let γ 0 , β > 0 and assume T q Q * , T q S * T Q are the trees returned by Q-tree search and S-tree search respectively. Then T q S * T q Q * .
Proof. 
The proof is presented in Appendix E. □
Corollary 13 essentially states that trees that emerge as a solution to (30) contain no more leaf nodes than those that solve (24), since T q S * T q Q * . Consequently, for fixed β > 0 , the presence of irrelevant information works to reduce the number of leaf nodes of the resulting abstraction. This is consistent with the original motivation for the inclusion of irrelevant information discussed in [20]. Namely, the purpose of introducing irrelevant information is so as the improve the quality of the abstractions (or reduce the rate of the code in communication systems) by removing the aspects of the relevant information that are correlated with the irrelevance variable Z [20]. In this way, a higher degree of compression can be achieved since we may remove the irrelevant components of the relevant information, a process which is not considered in the IB framework. When applied to multi-resolution tree abstractions, as in our case, this is manifested as a reduction in the number of leaf nodes, as established by Corollary 13.
Having established the influence of irrelevant information on the resulting abstractions, it is also practical to derive an explicit relation between the Q- and S-functions, so that a multi-resolution abstraction that considers irrelevant information can be obtained from the Q-tree search solution. For this result, we define the function P : N ( T W ) × ( 0 , ) × [ 0 , ) R according to the rule
P ( t ; β , γ ) = Q ( t ; β ) γ Δ I Z ( t ) t B t c Q ( t ; β ) + t B t P ( t ; β , γ ) Q ( t ; β ) ,
when t N int ( T W ) , where P ( t ; β , γ ) = 0 for t N leaf ( T W ) , and B t = { t C ( t ) : P ( t ; β , γ ) > 0 } , B t c = C ( t ) B t .
Lemma 14.
Let γ 0 and β > 0 . Then P ( t ; β , γ ) Q ( t ; β ) for all t N ( T W ) .
Proof. 
The proof is presented in Appendix F. □
We then have the following result.
Proposition 15.
Let γ 0 , β > 0 . If the function Q ( t ; β ) is known for every t N ( T W ) then
S ( t ; β , γ ) = max { P ( t ; β , γ ) , 0 } .
Proof. 
The proof is presented in Appendix G. □
Proposition 15 allows us to obtain a multi-resolution tree that incorporates irrelevant information from knowledge of only the Q-function employed by Q-tree search, as well as the values of Δ I Z ( t ) for each t N int ( T W ) , but does not require Δ I X ( t ) or Δ I Y ( t ) for t N int ( T W ) . Furthermore, Proposition 15 allows us to incorporate irrelevant information after initially designing a tree that is maximally task-relevant via the IB principle (24). We now turn our attention to demonstrating the utility of the G-tree search method on a non-trivial numerical example.

7. Numerical Example and Discussion

In this section, we demonstrate the utility of our approach on a real-world example. We consider the image shown in Figure 5a, which is of size 256 × 256 . The image in Figure 5a is then segmented, so that each pixel in the original image is classified into one of six distinct categories; the segmented image together with the original image is shown in Figure 5b. Segmented images such as the one shown in Figure 5b arise frequently in autonomous driving scenarios, where it is of interest to remove irrelevant details from the representation so as to focus available resources on only those aspects of the image that are considered important (e.g., the location of the obstacles or the shape of the road). In the segmented image shown in Figure 5c, we see that the task of maintaining relevant information regarding the road corresponds to retaining the red color while the remaining colors, such light green and yellow, are not relevant to the task of identifying the road and should be removed from the representation.
The input data are provided to the G-tree search algorithm as follows. We consider each finest-resolution pixel as an outcome of the uncompressed random variable X. Since the agent may not, in general, have the resources (time, computational, etc.) in order to determine the color (or category) information of each pixel with certainty, we model each color in Figure 5c as a random variable. To this end, for each color in Figure 5c we introduce a random variable, where colors that are assumed to be relevant are denoted as Y i and those considered irrelevant as Z j . For example, if we would like to generate abstractions where red and blue are relevant (and therefore should be retained) and yellow as irrelevant (and should be removed), we may define Y 1 to correspond to the category (or color) red and Y 2 to blue, whereas Z 1 may represent yellow.
Strictly speaking, knowledge of the distributions p ( y i | x ) , p ( z j | x ) and p ( x ) is sufficient to apply our method, as from relations (15)–(21) we see that these distributions allow the determination of Δ I Y i ( t ) , Δ I Z j ( t ) and Δ I X ( t ) for all i, j and t. The conditional distributions p ( y i | x ) and p ( z j | x ) are obtained from the image segmentation step, where p ( y i | x ) is the probability that the cell x has the color corresponding to Y i , with an analogous interpretation for p ( z j | x ) and Z j .In this example, p ( x ) is assumed to be uniform, although any valid distribution is permissible in our framework (the G-tree search approach can handle any valid distribution p ( x ) without modification. The use of a non-uniform p ( x ) will lead to region-specific abstraction, where the G-tree search algorithm refines in regions only where p ( x ) > 0 . For more information, the interested reader is referred to [36]). The joint distributions p ( x , y i ) and p ( x , z j ) are then assigned according to p ( x , y i ) = p ( y i | x ) p ( x ) and p ( x , z j ) = p ( z j | x ) p ( x ) , respectively. In the more general setting where the input is the joint probability mass function p ( x , y 1 , , y n , z 1 , , z m ) , the distributions p ( x , y i ) , p ( x , z j ) and p ( x ) can be obtained via marginalization, and the conditional distributions p ( y i | x ) and p ( z j | x ) required to compute (15)–(17) are acquired by applying standard rules for conditional probability.
In order to provide a basis for the discussion that follows, we show in Figure 6 a selection of abstractions obtained by trading relevant information and compression (i.e., the IB problem setting) in the case where red is the relevant variable. A number of observations can be deduced from the abstractions in Figure 6. Firstly, it should be noted that G-tree search finds a tree that retains all the available red information and contains only about 1.42 % of the nodes of the finest-resolution space. Next, notice that by changing the parameter α , we change the relative importance of compression and information retention. Consequently, at larger values of α , we obtain abstractions that contain less red information but contain fewer leaf nodes (achieve a greater degree of compression) as compared to the abstractions that arise as α is decreased.
Furthermore, observe from Figure 6 that regions in the image that contain both no red information and are homogeneous in red color remain aggregated even at high values of α . This occurs for two reasons. Firstly, observe from (15) that if a node t N int ( T W ) has children C ( t ) = { t 1 , , t 4 } for which p ( y | t 1 ) = = p ( y | t 4 ) for all y, then Δ I Y ( t ) = 0 as JS Π ( p ( y | t 1 ) , , p ( y | t 4 ) ) = 0 . Consequently, regions that either contain no red or that are homogeneous in red color contain no relevant information. Intuitively, if a region in the finest resolution is homogeneous in the color red, then no information is lost by aggregating homogeneously-colored finest resolution cells (i.e., given the aggregated cell we can perfectly predict the color of the descendant nodes). Thus, nodes t N int ( T W ) for which all descendant nodes are homogeneous in red color provide no additional relevant information, and thus one can see from (22) that G ( t ; β , γ , α ) = 0 for these nodes. Notice that the reason regions with no or homogeneous relevant information remain aggregated is due to Theorem 11. To see why, consider the scenario when compression is ignored α = 0 . In this case, regions that contain no relevant information may be expanded at no cost, but would not contribute to an increase in the objective value as seen by relation (13). However, such expansions would lead to a non-minimal tree to be returned by the G-tree search algorithm, which is precluded by Theorem 11. As a result, G-tree search will return the tree with the least number of leaf-nodes that attains the optimal objective function value. This implies that the tree returned by G-tree search maintains regions with no relevant information aggregated.
Next, we generate multi-resolution tree abstractions by employing G-tree search to not only retain relevant information, but also remove information that is considered irrelevant. To this end, we continue our example of considering red as the relevant variable of interest, now letting light green and yellow be irrelevant variables and represented by Z 1 and Z 2 , respectively. Example abstractions obtained in this case are shown in Figure 7. Notice that the case shown in Figure 7c corresponds to the standard IB problem with red as relevant and no penalty on compression.
A number of observations can be made from the sample abstractions shown in Figure 7. First, notice that, in comparison with the abstractions shown in Figure 6 which only consider the retention of the color red, the abstractions in Figure 7 aggregate cells along the boundary of red and the irrelevant information (light green and yellow) so as to obscure this information from the abstraction, while being as predictive regarding red (the relevant information) as possible. Moreover, observe that at greater values of the vector γ , regions of yellow and light green are shown in lower resolution as compared with the resolution of these areas at lower values of γ . Notice also that, as the irrelevant information is ignored ( [ γ ] 1 = [ γ ] 2 = 0 ), we recover an abstraction (Figure 7c) that is equivalent to the tree in Figure 6c returned by the standard IB approach, which does not consider the removal of the irrelevant information content. To better illustrate the differences between the standard IB case shown in Figure 6 and the generalized tree search scenario in Figure 7, we show the normalized information retained by each color for various settings of the weight parameters in Figure 8. The results shown in Figure 8 are obtained by setting β = 1 , α = 0 and by varying the vector γ R + 2 .
Figure 8 shows the normalized information retained in the solutions returned by G-tree search for two cases: (i) the standard IB problem with red as relevant, and (ii) the generalized tree search with red as relevant and light green as well as yellow as irrelevant. In the standard IB problem, decreasing the value of α 0 leads to abstractions that are more informative regarding the relevant information, at the cost of obtaining a tree T T Q that achieves a lower degree of compression. Consequently, one moves from right to left in Figure 8 (left) as the value of α 0 is increased. In contrast, in the generalized setting of maintaining red information while removing light green and yellow, increasing the weights of the irrelevant information leads to abstractions that achieve more compression, since the importance of information removal increases with larger values of γ . Thus, keeping all other weights constant, we move from right to left in Figure 8 (right) as γ R + 2 is increased.
We also see from Figure 8 that, compared with the IB tree solutions, the trees obtained from the G-tree search approach in case (ii) retain less information regarding light green and yellow. One may also observe from Figure 8 that when the generalized-tree search algorithm is tasked with retaining red information while removing light green and yellow, less red information is retained. This occurs as the importance of information removal necessitates an abstract representation in order to obscure, or remove, the irrelevant details. However, it is only regions that contain both relevant and irrelevant information that are of interest to the algorithm in this case, since regions that contain no relevant information are not refined even in the absence of irrelevant information content. In other words, one may view the relevant information as driving refinement, while irrelevant information promoting aggregation. It is therefore regions that contain both irrelevant and relevant information that becomes the focus of G-tree search. We can observe this trend in the abstractions shown in Figure 7. Specifically, notice that regions not containing any relevant information (i.e., regions with no red) are left unchanged and aggregated in Figure 7a–c. In contrast, when comparing the results of Figure 6, where irrelevant information is not taken into account, to those of Figure 7, we see that the areas containing both relevant and irrelevant information are aggregated as the relative importance of information removal is increased. This occurs for the aforementioned reasons, namely, we must sacrifice some relevant information in order to obscure, or remove, the irrelevant details. At the same time, those regions containing red and no irrelevant colors are maintained with relatively high resolution (e.g., the middle of the image where red boarders with darker green), since these regions contain relevant information with no irrelevant details.
We conclude this section by briefly showcasing the versatility of the G-tree search algorithm to remove redundancies from segmented images. Since the G-tree search algorithm allows any integer number n 0 of relevant random variables to be defined, it is possible to allow each color in Figure 5c to be a distinct relevant variable. In this case, G-tree search will find trees for which the distinct colors are as distinguishable as possible while balancing the degree of compression achieved by the abstraction. Interestingly, if one were to take [ β ] i = 1 for all i { 1 , , 6 } and α = 0 , then G-tree search will find a multi-resolution tree that retains all the color information, while removing as much redundancy as possible, as seen in Figure 9. Remarkably, the abstraction in Figure 9 contains only 5 % of the nodes of the finest-resolution representation in Figure 5c while retaining all the color (semantic) information.
The ability to compress the environment in this way while losing no information regarding the information content of the image represents a drastic savings in the required on-board memory needed to store the depiction of the environment.

8. Conclusions

In this paper, we developed a generalized information-theoretic framework for the emergence of multi-resolution abstractions for autonomous agents. To achieve our goal, we formulated the problem of selecting a multi-resolution tree by considering an objective that aims to maximally retain task-relevant information, while simultaneously removes task-irrelevant, or confidential, information and achieves as much compression as possible. Motivated by its use in signal compression theory, we employ the mutual information in order to measure the degree of achieved compression as well as the amount of relevant and irrelevant information retained in the resulting abstract representation. We rigorously investigate the mathematical properties and structure of the problem and discuss the connections between hierarchical tree abstractions and deterministic signal encoders. Moreover, it is shown that the problem we consider has a special structure, whereby the relevant and irrelevant information contained in a hierarchical tree can be expressed in terms of the incremental information contributions of the non-leaf (interior) nodes of the tree. This special, incremental, structure of the problem facilitates the design of the G-tree search algorithm, which searches over the space of multi-resolution abstractions for a solution that maximizes an information-theoretic objective. We detail our proposed algorithm and prove a number of theoretical results, including that the proposed G-tree search algorithm is guaranteed to return an optimal multi-resolution tree. The complexity of the proposed G-tree search algorithm is analyzed and it is shown that multi-resolution tree abstractions via the information bottleneck (IB) method and the information bottleneck problem with side-information (IBSI) are recovered as special cases of our formulation. A non-trivial numerical example is presented to demonstrate the utility of the proposed approach.

Author Contributions

Conceptualization, D.T.L. and P.T.; methodology, D.T.L. and D.M.; software, D.T.L.; validation, D.T.L.; formal analysis, D.T.L.; investigation, D.T.L.; resources, P.T.; data curation, D.T.L.; writing—original draft preparation, D.T.L.; writing—review and editing, D.T.L., D.M. and P.T.; visualization, D.T.L.; supervision, D.M. and P.T.; project administration, P.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Office of Naval Research awards N00014-18-1-2375 and N00014-18-1-2828 and by Army Research Laboratory under DCIST CRA W911NF-17-2-0181.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Appendix A

Proof of Lemma 6. 
The proof is given by induction. Consider a node t N 1 ( T W ) and a tree T q T Q . For the tree T q and the node t, there are two options: Either N int ( T q ( t ) ) = { t } or N int ( T q ( t ) ) = . If N int ( T q ( t ) ) = { t } , then
s N int ( T q ( t ) ) Δ J ( s ; β , γ , α ) = Δ J ( t ; β , γ , α ) max { Δ J ( t ; β , γ , α ) , 0 } = G ( t ; β , γ , α ) .
If, instead, N int ( T q ( t ) ) = , then
s N int ( T q ( t ) ) Δ J ( s ; β , γ , α ) = 0 max { Δ J ( t ; β , γ , α ) , 0 } = G ( t ; β , γ , α ) .
Assume the result holds for all t N k ( T W ) and some k { 1 , , 1 } , and consider any node t N k 1 ( T W ) and tree T q T Q . If N int ( T q ( t ) ) , then
s N int ( T q ( t ) ) Δ J ( s ; β , γ , α ) = Δ J ( t ; β , γ , α ) + t C ( t ) s N int ( T q ( t ) ) Δ J ( s ; β , γ , α ) , Δ J ( t ; β , γ , α ) + t C ( t ) G ( t ; β , γ , α ) ,
where the first step follows from writing the set N int ( T q ( t ) ) = { t } t C ( t ) N int ( T q ( t ) ) , and the second step is a result of invoking the induction hypothesis, since t C ( t ) N k ( T W ) . We then have
s N int ( T q ( t ) ) Δ J ( s ; β , γ , α ) Δ J ( t ; β , γ , α ) + t C ( t ) G ( t ; β , γ , α ) , max { Δ J ( t ; β , γ , α ) + t C ( t ) G ( t ; β , γ , α ) , 0 } = G ( t ; β , γ , α ) .
If N int ( T q ( t ) ) = , then
s N int ( T q ( t ) ) Δ J ( s ; β , γ , α ) = 0 max { Δ J ( t ; β , γ , α ) + t C ( t ) G ( t ; β , γ , α ) , 0 } , = G ( t ; β , γ , α ) .
Thus, s N int ( T q ( t ) ) Δ J ( s ; β , γ , α ) G ( t ; β , γ , α ) for all T q T Q and any t N int ( T W ) . □

Appendix B

Proof of Lemma 8. 
The proof is given by induction. First, let t N 1 ( T W ) be a node such that G ( t ; β , γ , α ) > 0 . Next, consider any T q T Q such that N int ( T q ( t ) ) = { t } . For any such tree, we have
s N int ( T q ( t ) ) Δ J ( s ; β , γ , α ) = Δ J ( t ; β , γ , α ) .
Since G ( t ; β , γ , α ) > 0 , it follows from the definition of G ( t ; β , γ , α ) , that
G ( t ; β , γ , α ) = Δ J ( t ; β , γ , α ) = s N int ( T q ( t ) ) Δ J ( s ; β , γ , α ) ,
and thus, there exits a tree T q such that G ( t ; β , γ , α ) = s N int ( T q ( t ) ) Δ J ( s ; β , γ , α ) .
Assume now that the result holds for all t N k ( T W ) and some k { 1 , , 1 } , and consider any t N k 1 ( T W ) for which G ( t ; β , γ , α ) > 0 . From the definition of G ( t ; β , γ , α ) , we have that
G ( t ; β , γ , α ) = max { Δ J ( t ; β , γ , α ) + t C ( t ) G ( t ; β , γ , α ) , 0 } , = Δ J ( t ; β , γ , α ) + t C ( t ) G ( t ; β , γ , α ) ,
since G ( t ; β , γ , α ) > 0 . Now, let U = { t C ( t ) : G ( t ; β , γ , α ) > 0 } . From the induction hypothesis, it follows that there exists a tree T q ^ T Q such that G ( t ; β , γ , α ) = s N int ( T q ^ ( t ) ) Δ J ( s ; β , γ , α ) for each t U C ( t ) . Consider then any tree T q T Q such that the subtree rooted at t has interior node set
N int ( T q ( t ) ) = { t } t U N int ( T q ^ ( t ) ) .
Then,
G ( t ; β , γ , α ) = Δ J ( t ; β , γ , α ) + t U G ( t ; β , γ , α ) , = Δ J ( t ; β , γ , α ) + t U s N int ( T q ^ ( t ) ) Δ J ( s ; β , γ , α ) = s N int ( T q ( t ) ) Δ J ( s ; β , γ , α ) .
Thus, if G ( t ; β , γ , α ) > 0 then there exits a tree T q T Q such that G ( t ; β , γ , α ) = s N int ( T q ( t ) ) Δ J ( s ; β , γ , α ) . □

Appendix C

In the following proof, we let T q ˜ T Q be the tree that is minimal and optimal with respect to J and take T G T Q denote the tree returned by G-tree search. We will show that T q ˜ = T G for all β R + n , γ R + m and α 0 .
Proof of Theorem 11. 
We begin by showing G ( t ; β , γ , α ) = 0 for all t N leaf ( T q ˜ ) . Consider the case when T q ˜ T W (notice that if T q ˜ = T W then G ( t ; β , γ , α ) = 0 for all t N leaf ( T q ˜ ) by definition). The proof is given by contradiction. Assume that there exists a node t N leaf ( T q ˜ ) for which G ( t ; β , γ , α ) > 0 . Then, from Lemma 10, it follows that there exists a tree T q ^ T Q such that s N int ( T q ^ ( t ) ) Δ J ( s ; β , γ , α ) > 0 . Now, consider the tree T q T Q with interior node set
N int ( T q ) = N int ( T q ˜ ) N int ( T q ^ ( t ) ) .
It then follows that
J ( T q ; β , γ , α ) J ( T q ˜ ; β , γ , α ) = s N int ( T q ^ ( t ) ) Δ J ( s ; β , γ , α ) > 0 ,
and thus there exists a tree T q ^ T Q such that J ( T q ^ ; β , γ , α ) > J ( T q ˜ ; β , γ , α ) . However, T q ˜ is optimal, leading to a contradiction. Thus, G ( t ; β , γ , α ) = 0 for all t N leaf ( T q ˜ ) . As a result, G-tree search will terminate at the leafs of T q ˜ should it reach them during the expansion process. However, the algorithm may terminate prior to reaching the leafs of T q ˜ , and so we conclude T G T q ˜ .
Next, we establish that G ( t ; β , γ , α ) > 0 for all t N int ( T q ˜ ) . We consider the case when N int ( T q ˜ ) (if N int ( T q ˜ ) = then there does not exist t N int ( T q ˜ ) such that G ( t ; β , γ , α ) = 0 ). The proof is given by contradiction. To this end, assume that there exists a node t N int ( T q ˜ ) such that G ( t ; β , γ , α ) = 0 . Since t N leaf ( T q ˜ ) and N int ( T q ˜ ) , it follows that N int ( T q ˜ ( t ) ) . Now, consider the tree T q T Q with interior node set
N int ( T q ) = N int ( T q ˜ ) N int ( T q ˜ ( t ) ) .
Then, since T q T q ˜ and T q ˜ is minimal, it follows that
0 < J ( T q ˜ ; β , γ , α ) J ( T q ; β , γ , α ) = s N int ( T q ˜ ( t ) ) Δ J ( s ; β , γ , α ) .
Consequently, there exists a tree T q ˜ T Q such that s N int ( T q ˜ ( t ) ) Δ J ( s ; β , γ , α ) > 0 . However, this is a contradiction, since G ( t ; β , γ , α ) = 0 . Thus, G ( t ; β , γ , α ) > 0 for all t N int ( T q ˜ ) . As a result, G-tree search will not terminate prior to reaching the leafs of T q ˜ , and so T q ˜ T G . We have therefore established that T G T q ˜ T G . Thus, T q ˜ = T G . □

Appendix D

Proof of Lemma 12. 
The proof is given by induction. Let t N 1 ( T W ) . Notice that for node t, C ( t ) N leaf ( T W ) . Consequently, S ( t ; β , γ ) = 0 and Q ( t ; β ) = 0 for all t C ( t ) . From the definitions of S and Q we then have
S ( t ; β , γ ) = max { Δ M ( t ; β , γ ) , 0 } max { Δ L ( t ; β ) , 0 } = Q ( t ; β ) ,
where the inequality follows from the observation that Δ M ( t ; β , γ ) = Δ L ( t ; β ) γ Δ I Z ( t ) Δ L ( t ; β ) as γ Δ I Z ( t ) 0 .
Now assume that the result holds for all t N k ( T W ) , k { 1 , , 1 } and consider any t N k 1 ( T W ) . For node t we have, by definition and the induction hypothesis, that
S ( t ; β , γ ) = max { Δ M ( t ; β , γ ) + t C ( t ) S ( t ; β , γ ) , 0 } , max { Δ M ( t ; β , γ ) + t C ( t ) Q ( t ; β ) , 0 } ,
since C ( t ) N k ( T W ) . Then, since Δ M ( t ; β , γ ) = Δ L ( t ; β ) γ Δ I Z ( t ) Δ L ( t ; β ) , we obtain
S ( t ; β , γ ) max { Δ M ( t ; β , γ ) + t C ( t ) Q ( t ; β ) , 0 } , max { Δ L ( t ; β ) + t C ( t ) Q ( t ; β ) , 0 } = Q ( t ; β ) ,
thereby establishing that S ( t ; β , γ ) Q ( t ; β ) . □

Appendix E

Proof of Corollary 13. 
Notice that Lemma 12 establishes that if, for some t N int ( T W ) , β > 0 , and γ 0 , we have S ( t ; β , γ ) > 0 , then Q ( t ; β ) > 0 . As a result, any node t N int ( T W ) expanded by S-tree search is also expanded by Q-tree search. Thus, T q S * T q Q * . □

Appendix F

Proof of Lemma 14. 
The proof is given by induction. First consider any t N 1 ( T W ) and observe that for all such nodes C ( t ) N leaf ( T W ) . Thus, by definition, Q ( t ; β ) = 0 and P ( t ; β , γ ) = 0 for all t C ( t ) . Therefore,
P ( t ; β , γ ) = Q ( t ; β ) γ Δ I Z ( t ) Q ( t ; β ) ,
which follows since γ Δ I Z ( t ) 0 .
Assume the result holds for all t N k ( T W ) and some k { 1 , , 1 } . Consider now any t N k 1 ( T W ) . Then, for the node t, we have by definition
P ( t ; β , γ ) = Q ( t ; β ) γ Δ I Z ( t ) t B t c Q ( t ; β ) + t B t P ( t ; β , γ ) Q ( t ; β ) .
Since B t C ( t ) N k ( T W ) , it follows from the induction hypothesis that
P ( t ; β , γ ) Q ( t ; β ) γ Δ I Z ( t ) t B t c Q ( t ; β ) ,
since the hypothesis furnishes that P ( t ; β , γ ) Q ( t ; β ) 0 for all t N k ( T W ) . From the non-negativity of the Q-function and that γ Δ I Z ( t ) 0 , we arrive at
P ( t ; β , γ ) Q ( t ; β ) γ Δ I Z ( t ) t B t c Q ( t ; β ) Q ( t ; β ) ,
and thus P ( t ; β , γ ) Q ( t ; β ) for all t N ( T W ) . □

Appendix G

Proof of Proposition 15. 
The proof is given by induction. Consider first any t N 1 ( T W ) . For the node t, it follows from the definition of S ( t ; β , γ ) , that
S ( t ; β , γ ) = max { Δ M ( t ; β , γ ) , 0 } = max { Δ L ( t ; β ) γ Δ I Z ( t ) , 0 } ,
since S ( t ; β , γ ) = 0 for all C ( t ) N leaf ( T W ) , and Δ M ( t ; β , γ ) = Δ I Y ( t ) 1 β Δ I X ( t ) γ Δ I Z ( t ) = Δ L ( t ; β ) γ Δ I Z ( t ) . Furthermore, notice that, by definition of P ( t ; β , γ ) , we have
P ( t ; β , γ ) = Q ( t ; β ) γ Δ I Z ( t ) .
There are two cases to consider: namely, Q ( t ; β ) > 0 and Q ( t ; β ) = 0 . In the first case (i.e., Q ( t ; β ) > 0 ), we have that Q ( t ; β ) = Δ L ( t ; β ) . Consequently,
S ( t ; β , γ ) = max { Δ L ( t ; β ) γ I Z ( t ) , 0 } = max { Q ( t ; β ) γ I Z ( t ) , 0 } , = max { P ( t ; β , γ ) , 0 } .
For the second case, it follows from Lemma 12 that S ( t ; β , γ ) = 0 . Thus,
S ( t ; β , γ ) = 0 = max { P ( t ; β , γ ) , 0 } ,
which holds since, by Lemma 14, P ( t ; β , γ ) 0 when Q ( t ; β ) = 0 . Consequently, S ( t ; β , γ ) = max { P ( t ; β , γ ) , 0 } for all t N 1 ( T W ) .
Assume now that the result holds for all t N k ( T W ) and some k { 1 , , 1 } , and consider any t N k 1 ( T W ) . For the node t, we have, by definition, that
S ( t ; β , γ ) = max { Δ M ( t ; β , γ ) + t C ( t ) S ( t ; β , γ ) , 0 } .
Notice that C ( t ) N k ( T W ) , and thus, by the induction hypothesis, we have
S ( t ; β , γ ) = max { Δ M ( t ; β , γ ) + t C ( t ) max { P ( t ; β , γ ) , 0 } , 0 } .
Next, define the set B t = { t C ( t ) : P ( t ; β , γ ) > 0 } , and write
S ( t ; β , γ ) = max { Δ M ( t ; β , γ ) + t B t P ( t ; β , γ ) , 0 } = max { Δ L ( t ; β ) γ Δ I Z ( t ) + t B t P ( t ; β , γ ) , 0 } , = max { Δ L ( t ; β ) + t C ( t ) Q ( t ; β ) t C ( t ) Q ( t ; β ) γ Δ I Z ( t ) + t B t P ( t ; β , γ ) , 0 } .
Notice that
Δ L ( t ; β ) + t C ( t ) Q ( t ; β ) t C ( t ) Q ( t ; β ) γ Δ I Z ( t ) + t B t P ( t ; β , γ ) = Δ L ( t ; β ) + t C ( t ) Q ( t ; β ) t S t c Q ( t ; β ) γ Δ I Z ( t ) + t B t [ P ( t ; β , γ ) Q ( t ; β ) ] ,
where S t c = C ( t ) B t . We now consider the cases when Q ( t ; β ) > 0 and Q ( t ; β ) = 0 . If Q ( t ; β ) > 0 then Q ( t ; β ) = Δ L ( t ; β ) + t C ( t ) Q ( t ; β ) , and so we have
Δ L ( t ; β ) + t C ( t ) Q ( t ; β ) t C ( t ) Q ( t ; β ) γ Δ I Z ( t ) + t B t P ( t ; β , γ ) = Δ L ( t ; β ) + t C ( t ) Q ( t ; β ) t S t c Q ( t ; β ) γ Δ I Z ( t ) + t B t [ P ( t ; β , γ ) Q ( t ; β ) ] , = Q ( t ; β ) t S t c Q ( t ; β ) γ Δ I Z ( t ) + t B t [ P ( t ; β , γ ) Q ( t ; β ) ] , = P ( t ; β , γ ) ,
where the final equality follows from the definition of P ( t ; β , γ ) . Consequently,
S ( t ; β , γ ) = = max { Δ L ( t ; β ) + t C ( t ) Q ( t ; β ) t C ( t ) Q ( t ; β ) γ Δ I Z ( t ) + t B t P ( t ; β , γ ) , 0 } , = max { P ( t ; β , γ ) , 0 } .
In the second case, when Q ( t ; β ) = 0 , we have from Lemma 12 that S ( t ; β , γ ) = 0 . Furthermore, from Lemma 14 we have P ( t ; β , γ ) 0 . Therefore,
S ( t ; β , γ ) = 0 = max { P ( t ; β , γ ) , 0 } .
As a result, S ( t ; β , γ ) = max { P ( t ; β , γ ) , 0 } for all t N ( T W ) . □

References

  1. Holte, R.C.; Choueiry, B.Y. Abstraction and reformulation in artificial intelligence. Philos. Trans. R. Soc. B 2003, 358, 1197–1204. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Zucker, J.D. A grounded theory of abstraction in artificial intelligence. Philos. Trans. R. Soc. B 2003, 358, 1293–1309. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Ponsen, M.; Taylor, M.E.; Tuyls, K. Abstraction and generalization in reinforcement learning: A summary and framework. In Proceedings of the International Workshop on Adaptive and Learning Agents, Budapest, Hungary, 12 May 2009. [Google Scholar]
  4. Niv, Y. Learning task-state representations. Nat. Neurosci. 2019, 22, 1544–1553. [Google Scholar] [CrossRef] [PubMed]
  5. Brooks, R.A. Intelligence without representation. Artif. Intell. 1991, 47, 139–159. [Google Scholar] [CrossRef]
  6. Behnke, S. Local multiresolution path planning. In RoboCup 2003: Robot Soccer World Cup VII; Springer: Berlin/Heidelberg, Germany, 2004. [Google Scholar]
  7. Cowlagi, R.V.; Tsiotras, P. Multiresolution motion planning for autonomous agents via wavelet-based cell decompositions. IEEE Trans. Syst. Man Cybern. Part B Cybern. 2012, 42, 1455–1469. [Google Scholar] [CrossRef] [PubMed]
  8. Cowlagi, R.V.; Tsiotras, P. Multi-resolution path planning: Theoretical analysis, efficient implementation, and extensions to dynamic environments. In Proceedings of the IEEE Conference on Decision and Control, Atlanta, GA, USA, 15–17 December 2010; pp. 1384–1390. [Google Scholar]
  9. Cowlagi, R.V.; Tsiotras, P. Multiresolution path planning with wavelets: A local replanning approach. In Proceedings of the American Control Conference, Seattle, WA, USA, 11–13 June 2008; pp. 1220–1225. [Google Scholar]
  10. Tsiotras, P.; Bakolas, E. A hierarchical on-line path planning scheme using wavelets. In Proceedings of the European Control Conference, Kos, Greece, 2–5 July 2007. [Google Scholar]
  11. Tsiotras, P.; Jung, D.; Bakolas, E. Multiresolution hierarchical path-planning for small UAVs using wavelet decompositions. J. Intell. Robot. Syst. 2012, 66, 505–522. [Google Scholar] [CrossRef]
  12. Hauer, F.; Kundu, A.; Rehg, J.M.; Tsiotras, P. Multi-scale perception and path planning on probabilistic obstacle maps. In Proceedings of the IEEE International Conference on Robotics and Automation, Seattle, WA, USA, 26–30 May 2015. [Google Scholar]
  13. Hauer, F.; Tsiotras, P. Reduced complexity multi-scale path-planning on probabilitic maps. In Proceedings of the IEEE Conference on Robotics and Automation, Stockholm, Sweden, 16–21 May 2016. [Google Scholar]
  14. Kambhampati, S.; Davis, L.S. Multiresolution path planning for mobile robots. IEEE J. Robot. Autom. 1986, RA-2, 135–145. [Google Scholar] [CrossRef] [Green Version]
  15. Kraetzschmar, G.K.; Gassull, G.P.; Uhl, K. Probabilistic quadtrees for variable-resolution mapping of large environments. IFAC Proc. Vol. 2004, 37, 675–680. [Google Scholar] [CrossRef]
  16. Hornung, A.; Wurm, K.M.; Bennewitz, M.; Stachniss, C.; Burgard, W. Octomap: An efficient probabilistic 3D mapping framework based on octrees. Auton. Robot. 2013, 34, 189–206. [Google Scholar] [CrossRef] [Green Version]
  17. Cover, T.M.; Thomas, J.A. Elements of Information Theory, 2nd ed.; John Wiley & Sons: Hoboken, NJ, USA, 2006. [Google Scholar]
  18. Tishby, N.; Pereira, F.C.; Bialek, W. The information bottleneck method. In Proceedings of the Allerton Conference on Communication, Control and Computing, Monticello, IL, USA, 22–24 September 1999. [Google Scholar]
  19. Strouse, D.; Schwab, D.J. The deterministic information bottleneck. Neural Comput. 2017, 29, 1611–1630. [Google Scholar] [CrossRef] [Green Version]
  20. Chechik, G.; Tishby, N. Extracting relevant structures with side information. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 9–12 December 2002. [Google Scholar]
  21. Friedman, N.; Mosenzon, O.; Slonim, N.; Tishby, N. Multivariate information bottleneck. In Proceedings of the Uncertainty in Artificial Intelligence, Seattle, WA, USA, 2–5 August 2001. [Google Scholar]
  22. Slonim, N.; Tishby, N. Agglomerative information bottleneck. In Proceedings of the Advances in Neural Information Processing Systems, Denver, CO, USA, 29 November–4 December 1999. [Google Scholar]
  23. Slonim, N.; Friedman, N.; Tishby, N. Multivariate information bottleneck. Neural Comput. 2006, 18, 1739–1789. [Google Scholar] [CrossRef]
  24. Chechik, G.; Gloverson, A.; Tishby, N.; Weiss, Y. Information bottleneck for Gaussian variables. J. Mach. Learn. Res. 2005, 6, 165–188. [Google Scholar]
  25. Estella Aguerri, I.; Zaidi, A. Distributed information bottleneck method for discrete and Gaussian sources. In Proceedings of the International Zurich Seminar on Information and Communication, Zurich, Switzerland, 21–23 February 2018. [Google Scholar]
  26. Aguerri, I.E.; Zaidi, A. Distributed variational representation learning. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 120–138. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Cover, T.M.; Permuter, H.H. Capacity of coordinated actions. In Proceedings of the IEEE International Symposium on Information Theory, Nice, France, 24–30 June 2007. [Google Scholar]
  28. Cuff, P.W.; Permuter, H.H.; Cover, T.M. Coordination capacity. IEEE Trans. Inf. Theory 2010, 56, 4181–4206. [Google Scholar] [CrossRef] [Green Version]
  29. Bardera, A.; Rigau, J.; Boada, I.; Feixas, M.; Sbert, M. Image segmentation using information bottleneck method. IEEE Trans. Image Process. 2009, 18, 1601–1612. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  30. Kolchinsky, A.; Tracey, B.D.; Wolpert, D.H. Nonlinear information bottleneck. Entropy 2019, 21, 1181. [Google Scholar] [CrossRef] [Green Version]
  31. Alemi, A.A.; Fischer, I.; Dillon, J.V.; Murphy, K. Deep variational information bottleneck. In Proceedings of the International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
  32. Chalk, M.; Marre, O.; Tkacik, G. Relevant sparse codes with variational information bottleneck. In Proceedings of the Advances in Neural Information Processing Systems, Barcelona, Spain, 6–10 December 2016. [Google Scholar]
  33. Zaidi, A.; Estella-Aguerri, I.; Shamai, S. On the information bottleneck problems: Models, connections, applications and information theoretic views. Entropy 2020, 22, 151. [Google Scholar] [CrossRef] [Green Version]
  34. Larsson, D.T.; Maity, D.; Tsiotras, P. Information-theoretic abstractions for planning in agents with computational constraints. IEEE Robot. Autom. Lett. 2021, 6, 7651–7658. [Google Scholar] [CrossRef]
  35. Nelson, E.; Corah, M.; Michael, N. Environment model adaptation for mobile robot exploration. Auton. Robot. 2018, 42, 257–272. [Google Scholar] [CrossRef]
  36. Larsson, D.T.; Maity, D.; Tsiotras, P. Q-tree search: An information-theoretic approach toward hierarchical abstractions for agents with computational limitations. IEEE Trans. Robot. 2020, 36, 1669–1685. [Google Scholar] [CrossRef]
  37. Larsson, D.T.; Maity, D.; Tsiotras, P. Information-theoretic abstractions for resource-constrained agents via mixed-integer linear programming. In Proceedings of the Proceedings of the Workshop on Computation-Aware Algorithmic Design for Cyber-Physical Systems, Nashville, TN, USA, 18 May 2021. [Google Scholar]
  38. Gallager, R.G. Information Theory and Reliable Communication; Wiley: New York, NY, USA, 1968. [Google Scholar]
  39. Slonim, N. The Information Bottleneck: Theory and Applications. Ph.D. Thesis, The Hebrew University, Jerusalem, Israel, 2002. [Google Scholar]
  40. Lin, J. Divergence measures based on the Shannon entropy. IEEE Trans. Inf. Theory 1991, 37, 145–151. [Google Scholar] [CrossRef] [Green Version]
  41. Bondy, J.A.; Murty, U.S.R. Graph Theory with Applications; Elsevier Science: New York, NY, USA, 1976. [Google Scholar]
  42. Nekouei, E.; Tanaka, T.; Skoglund, M.; Johansson, K.H. Information-theoretic approaches to privacy in estimation and control. Annu. Rev. Control. 2019, 47, 412–422. [Google Scholar] [CrossRef]
  43. Makhdoumi, A.; Salamatian, S.; Fawaz, N.; Médard, M. From the information bottleneck to the privacy funnel. In Proceedings of the IEEE Information Theory Workshop, Hobart, NSW, Australia, 2–5 November 2014. [Google Scholar]
  44. Du Pin Calmon, F.; Fawaz, N. Privacy against statistical inference. In Proceedings of the Allerton Conference on Communication, Control and Computing, Monticello, IL, USA, 1–5 October 2012. [Google Scholar]
Figure 1. (top) The tree T W together with 4 × 4 grid world representation. (bottom) Multi-resolution abstraction of the world W in the form of a quadtree. Notice that the tree T T Q is formed by aggregating finest resolution cells that are leafs of T W to their parent nodes, which are leafs of T . Aggregated nodes are shown in grey shading. In both figures, black filled nodes are those nodes that are part of the set N int ( · ) , whereas nodes with no fill comprise the set N leaf ( · ) .
Figure 1. (top) The tree T W together with 4 × 4 grid world representation. (bottom) Multi-resolution abstraction of the world W in the form of a quadtree. Notice that the tree T T Q is formed by aggregating finest resolution cells that are leafs of T W to their parent nodes, which are leafs of T . Aggregated nodes are shown in grey shading. In both figures, black filled nodes are those nodes that are part of the set N int ( · ) , whereas nodes with no fill comprise the set N leaf ( · ) .
Entropy 24 00809 g001
Figure 2. Two trees that differ by only a single leaf node expansion. In moving from tree (a) to (b), the node t ^ is expanded, adding its children as leafs to create the tree shown in (b). Interior nodes are shown in black, whereas leaf nodes are white. (a) Some tree T q T Q for which t ^ N leaf ( T u ) . (b) The tree T q ^ T Q which is created by expanding the node t ^ N leaf ( T u ) .
Figure 2. Two trees that differ by only a single leaf node expansion. In moving from tree (a) to (b), the node t ^ is expanded, adding its children as leafs to create the tree shown in (b). Interior nodes are shown in black, whereas leaf nodes are white. (a) Some tree T q T Q for which t ^ N leaf ( T u ) . (b) The tree T q ^ T Q which is created by expanding the node t ^ N leaf ( T u ) .
Entropy 24 00809 g002
Figure 3. Example of a tree T q T Q and the subtree T q ( t ) rooted at t. (a) The tree T q T Q together with a node t N int ( T q ) shown. (b) the subtree T q ( t ) rooted at t with original tree T q shown in background.
Figure 3. Example of a tree T q T Q and the subtree T q ( t ) rooted at t. (a) The tree T q T Q together with a node t N int ( T q ) shown. (b) the subtree T q ( t ) rooted at t with original tree T q shown in background.
Entropy 24 00809 g003
Figure 4. Tree (left) and grid (right) of a 4 × 4 example where greedy (myopic) tree search fails. Shading of red scales with p ( y | x ) (all red shades are equal). Note that, in this environment, parents of leaf nodes contain relevant information, whereas the root node does not, as each of the quadrants is equal in their prediction of Y due to environment symmetry.
Figure 4. Tree (left) and grid (right) of a 4 × 4 example where greedy (myopic) tree search fails. Shading of red scales with p ( y | x ) (all red shades are equal). Note that, in this environment, parents of leaf nodes contain relevant information, whereas the root node does not, as each of the quadrants is equal in their prediction of Y due to environment symmetry.
Entropy 24 00809 g004
Figure 5. (a) Original 256 × 256 image. (b) Segmented 256 × 256 image with original image in background. (c) Segmented 256 × 256 image passed to G-tree search.
Figure 5. (a) Original 256 × 256 image. (b) Segmented 256 × 256 image with original image in background. (c) Segmented 256 × 256 image passed to G-tree search.
Entropy 24 00809 g005
Figure 6. Sample abstractions obtained via the generalized tree search algorithm by defining the color red as relevant and ignoring other colors. In this scenario, the G-tree search method reduces to the IB problem (see Section 6). We assume the (scalar) weighing parameter β = 1 and change only α in (10) to generate the abstractions shown. (a) α = 0.14 , M M I Y ( T ) / I ( X ; Y ) = 0.9346 , 0.647 % of leaf nodes. (b) α = 0.05 , M M I Y ( T ) / I ( X ; Y ) = 0.9874 , 1.23 % of leaf nodes. (c) α = 0.01 , M M I Y ( T ) / I ( X ; Y ) = 1 , 1.42 % of leaf nodes.
Figure 6. Sample abstractions obtained via the generalized tree search algorithm by defining the color red as relevant and ignoring other colors. In this scenario, the G-tree search method reduces to the IB problem (see Section 6). We assume the (scalar) weighing parameter β = 1 and change only α in (10) to generate the abstractions shown. (a) α = 0.14 , M M I Y ( T ) / I ( X ; Y ) = 0.9346 , 0.647 % of leaf nodes. (b) α = 0.05 , M M I Y ( T ) / I ( X ; Y ) = 0.9874 , 1.23 % of leaf nodes. (c) α = 0.01 , M M I Y ( T ) / I ( X ; Y ) = 1 , 1.42 % of leaf nodes.
Entropy 24 00809 g006
Figure 7. Multi-resolution trees returned by G-tree search in the case when red is relevant and yellow and light green are considered irrelevant for the environment shown in Figure 5. (a) Solution for α = 0 , β = 1 , [ γ ] 1 = [ γ ] 2 = 1.2 . (b) Solution for α = 0 , β = 1 , [ γ ] 1 = [ γ ] 2 = 0.84 . (c) Solution for α = 0 , β = 1 , [ γ ] 1 = [ γ ] 2 = 0 .
Figure 7. Multi-resolution trees returned by G-tree search in the case when red is relevant and yellow and light green are considered irrelevant for the environment shown in Figure 5. (a) Solution for α = 0 , β = 1 , [ γ ] 1 = [ γ ] 2 = 1.2 . (b) Solution for α = 0 , β = 1 , [ γ ] 1 = [ γ ] 2 = 0.84 . (c) Solution for α = 0 , β = 1 , [ γ ] 1 = [ γ ] 2 = 0 .
Entropy 24 00809 g007
Figure 8. Normalized information retained vs. degree of achieved compression for each color in Figure 5c. Two cases are shown: (i) the standard IB problem with red as relevant (left), and (ii) generalized tree search with red as relevant and light green and yellow as irrelevant (right). The bar color corresponds with the color in Figure 5c. Data are normalized by the information of each color contained in the tree recovered when executing G-tree search with weights β = 1 , α = 0 , and γ = 0 (i.e., the tree that retains all the relevant information).
Figure 8. Normalized information retained vs. degree of achieved compression for each color in Figure 5c. Two cases are shown: (i) the standard IB problem with red as relevant (left), and (ii) generalized tree search with red as relevant and light green and yellow as irrelevant (right). The bar color corresponds with the color in Figure 5c. Data are normalized by the information of each color contained in the tree recovered when executing G-tree search with weights β = 1 , α = 0 , and γ = 0 (i.e., the tree that retains all the relevant information).
Entropy 24 00809 g008
Figure 9. Multi-resolution abstraction of the environment that retains all color information. The image contains only 5 % of the nodes compared to the original.
Figure 9. Multi-resolution abstraction of the environment that retains all color information. The image contains only 5 % of the nodes compared to the original.
Entropy 24 00809 g009
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Larsson, D.T.; Maity, D.; Tsiotras, P. A Generalized Information-Theoretic Framework for the Emergence of Hierarchical Abstractions in Resource-Limited Systems. Entropy 2022, 24, 809. https://doi.org/10.3390/e24060809

AMA Style

Larsson DT, Maity D, Tsiotras P. A Generalized Information-Theoretic Framework for the Emergence of Hierarchical Abstractions in Resource-Limited Systems. Entropy. 2022; 24(6):809. https://doi.org/10.3390/e24060809

Chicago/Turabian Style

Larsson, Daniel T., Dipankar Maity, and Panagiotis Tsiotras. 2022. "A Generalized Information-Theoretic Framework for the Emergence of Hierarchical Abstractions in Resource-Limited Systems" Entropy 24, no. 6: 809. https://doi.org/10.3390/e24060809

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop