Abstract
Joining multiple decision-makers together is a powerful way to obtain more sophisticated decision-making systems, but requires to address the questions of division of labor and specialization. We investigate in how far information constraints in hierarchies of experts not only provide a principled method for regularization but also to enforce specialization. In particular, we devise an information-theoretically motivated on-line learning rule that allows partitioning of the problem space into multiple sub-problems that can be solved by the individual experts. We demonstrate two different ways to apply our method: (i) partitioning problems based on individual data samples and (ii) based on sets of data samples representing tasks. Approach (i) equips the system with the ability to solve complex decision-making problems by finding an optimal combination of local expert decision-makers. Approach (ii) leads to decision-makers specialized in solving families of tasks, which equips the system with the ability to solve meta-learning problems. We show the broad applicability of our approach on a range of problems including classification, regression, density estimation, and reinforcement learning problems, both in the standard machine learning setup and in a meta-learning setting.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
1 Introduction
Intelligent agents are often conceptualized as decision-makers that learn probabilistic models of their environment and optimize utilities or disutilities like cost or loss functions [91]. In the general case we can think of a utility function as a black-box oracle that provides a numerical score that rates any proposed solution to a supervised, unsupervised or reinforcement learning problem. In context of decision-making, naïvely enumerating all possibilities and searching for an optimal solution is usually prohibitively expensive. Instead, intelligent agents must invest their limited resources in such a way that they achieve an optimal trade-off between expected utility and resource costs in order to enable efficient learning and acting. This trade-off is the central issue in the fields of bounded or computational rationality with repercussions across other disciplines including economics, psychology, neuroscience and artificial intelligence [3, 16, 23, 24, 26, 35, 55, 66, 78]. The information-theoretic approach to bounded rationality is a particular instance of bounded rationality where the resource limitations are modeled by information constraints [18, 28, 54, 59, 63, 64, 73, 85, 92] closely related to Jaynes’ maximum entropy or minimum relative entropy principle [41]. At the heart of information-theoretic models of bounded rationality lies the information utility trade-off for lossy compression, abstraction and hierarchy formation [23]. The optimal utility information trade-off leads to an optimal arrangement of decision-makers and encourages the emergence of specialized agents which in turn facilitates an optimal division of labor reducing computational effort [29, 35].
In the context of machine learning, hierarchical inference can also be regarded as an example of bounded rational decision-making with information constraints, where different models correspond to different experts that are bound by their respective priors over parameters and try to optimize the marginal likelihood given data [23]. The priors that are acquired by the different expert models can be regarded as efficient abstractions that facilitate generalization to novel problems. Hierarchical inference models have been used successfully, for example, to model human behavior when learning with very limited resources from a handful of examples while excelling at adapting quickly [12, 22, 40, 42]. We can study sample-efficient adaptation to new problems as an instance of “learning to learn” or meta-learning [15, 74, 84] which is an ongoing and active field of research [11, 19, 45, 65, 71, 90, 94]. While we can define meta-learning in different ways [13, 27, 37, 52, 88], a common denominator is that systems capable of meta-learning happen to learn on two levels, each with a different time scale: slow learning across different tasks (the meta-level), and fast learning to adapt to each task individually. Understanding efficient meta-learning still poses a challenging machine learning problem, as applying pre-trained models to new tasks naïvely usually leads to poor performance, as with each new incoming batch of data the agent has to perform expensive and slow re-learning of its policy.
In this study, we aim to harness the efficient adaptation and the ability of meta-learning of hierarchical inference processes for the learning of decision-making hierarchies formulated in terms of arbitrary utility functions—see Fig. 1 and Table 1. The formulation in terms of general utility functions makes this approach applicable to a broad range of machine learning problems that can be formulated as optimization problems, including supervised and unsupervised learning, and reinforcement learning. To this end, we extend our previous work on specialization in hierarchical decision-making systems introduced in Hihn et al. [36] to the problem of meta-learning. After introducing information-theoretic constraints for learning and decision-making in Sect. 2, we explain our hierarchical online learning approach to classification, regression, unsupervised learning and reinforcement learning in Sect. 3 for the case of within task specialization. We extend our mixture-of-experts learning experiments from Hihn et al. [36] for supervised and reinforcement learning in Sects. 4.1 and 4.3 and devise a novel application to density estimation in Sect. 4.2. The extended experiments in the classification and reinforcement learning setting provide new insights into how the number of experts influences the information processing and classification error and how expert policies partition the reinforcement learning problem space amongst themselves. In Sects. 5 and 6 we extend the approach from state-based specialization to the case of across task specialization. We show that this task specialization gives rise to a novel meta-learning approach where the meta-learner assigns previously unseen tasks to experts specialized on similar tasks. In order to split the task space and to assign the partitions to experts, we learn to represent tasks through a common latent embedding, which a gating network uses to distribute tasks among the experts-similar to the posterior over models in a hierarchical inference setup. In Sect. 7, we discuss novel aspects of the current study in the context of previous literature and conclude with a final summary.
2 Information Constraints in Learning and Decision-making
2.1 Decision-Making with Information Constraints
Given a utility function \(\mathbf {U}(x, y)\) that indicates the desirability of each action \(y \in \mathcal {Y}\) taken in each context \(x \in \mathcal {X}\), a fully rational agent picks action \(y^*_{x} = \mathop {\hbox {arg max}}\limits \nolimits _{y} \mathbf {U}(x, y)\). A bounded rational decision-maker with information constraints [64] is modeled by an upper bound B on the Kullback–Leibler divergence \(\text {D}_\text {KL}(p(y\vert x)\vert \vert p(y)) = \sum _{y}{p(y\vert x) \log {\frac{p(y\vert x)}{p(y)}}}\) between the agent’s prior p(y) and posterior policy \(p(y\vert x)\) to express the limitation on the decision-maker’s information processing resources for reducing uncertainty when optimizing \(\mathbf {U}(x, y)\). This results in the following optimization problem:
by introducing a Lagrange multiplier \(\beta \in \mathbb {R}^+\) that is determined by B. For \(\beta \rightarrow \infty \) we recover the maximum utility solution and for \(\beta \rightarrow 0\) the agent can only act according to the prior. In the case of a known state distribution p(x), the optimal prior is given by the marginal \(p(y) = \sum _{x}{p(x) p(y\vert x)}\) and the expected Kullback–Leibler divergence becomes equal to the mutual information I(X; Y).
When aggregating bounded-rational agents into hierarchical decision-making systems that split the action space into soft partitions [23], an expert selection policy \(p(m\vert x)\) can be introduced that selects an expert m for a given state x that chooses their action according to a stochastic policy \(p(y\vert x,m)\). Similar to Eq. (1), such a hierarchical decision-making system with information constraints can be expressed by the following optimization problem:
where \(\beta _1\) is the resource parameter for the expert selection stage and \(\beta _2\) for the experts, and \(I(\cdot ;\cdot |M)\) denotes the conditional mutual information. The optimal solution is a set of coupled equations
with the marginals \(p(m) =\sum _{x} p(x) p(m\vert x)\) and \(p(y\vert m) =\sum _{x} p(x\vert m)p(y\vert x,m)\), and the free-energy difference \(\Delta \text {F}_{\text {par}}(x,m) = \mathbb {E}_{p(y|x,m)}[\mathbf {U}] - \frac{1}{\beta }\text {D}_\text {KL}(p(y|x,m)\vert \vert p(y|m))\). If we assume \(\mathbf {U}(x,y)\) to represent the log-likelihood and m to represent different statistical models with parameters y to explain observables x, then Eq. (2) describes the problem of hierarchical inference in terms of an optimization problem to produce the posteriors p(y|x, m) and p(m|x).
2.2 Learning with Information Constraints
In learning problems, the true utility \(\mathbf {U}(x, y)\) is unknown, instead we are given a function \(\mathbf {U}_D(x, y)\) after having observed n data samples \(D = \{(x_i,y_i)\}_{i=1}^n\) and we know that \(\mathbf {U}_D(x, y)\rightarrow \mathbf {U}(x, y)\) for \(n\rightarrow \infty \). Assuming that we know p(x), we are looking for the best strategy p(y|x) that optimizes \(\mathbf {U}(x, y)\). If we were to treat this as an inference problem, we could index different candidate solutions \(p(y|x,\theta )\) with a parameter \(\theta \) and place a prior \(\pi (\theta )\) over this parameter. From PAC-Bayes analyses it is well-known that the Gibbs distribution \(\pi _D(\theta )\propto \pi (\theta )\exp \left( \gamma \mathbf {U}_D(\theta )\right) \) minimizes the generalization error [57, 58]
where \(U_n(\theta ) = \frac{1}{n}\sum _{x \in D}\sum _y p(y|x,\theta ) \mathbf {U}_D(x, y)\) and \(U(\theta ) = \sum _{x,y} p(x) p(y|x,\theta ) \mathbf {U}(x, y)\). In a classification problem, for example, we could choose the utility \(\mathbf {U}(x,y)=\mathbb {I}_{y=h(x)}\) where \(\mathbb {I}\) is the indicator function and h(x) is the mapping from x to the true labels \(y_{true}\). Just like the prior \(\pi (\theta )\) can be seen to regularize the search for the parameter \(\theta \), we can therefore regard the \(\text {D}_\text {KL}\) as a principled regularizer in the space of probability distributions over \(\theta \). Regularization techniques [38, 46, 79] are typically introduced to overcome the problem of overfitting (i.e., large generalization error), such that we can regard information constraints in \(\theta \)-space as a particular instance of regularization in the space of hypotheses.
Instead of studying information constraints in hypothesis space governing the update from prior \(\pi (\theta )\) to posterior \(\pi _D(\theta )\), we could also directly study information constraints in the output space governing the update between the prior and posterior predictive distributions \(p(y|x) = \sum _\theta \pi (\theta )p(y|x,\theta )\) and \(p_D(y|x) = \sum _\theta \pi _D(\theta )p(y|x,\theta )\). If the update from \(\pi (\theta )\) to \(\pi _D(\theta )\) is bound by \(\text {D}_\text {KL}(\pi _D(\theta )||\pi (\theta ))\le C_1\), then the update from p(y|x) to \(p_D(y|x)\) will be bound by \(\text {D}_\text {KL}(p_D(y|x)||p(y|x))\le C_2\). This suggests that one could try to use the \(\text {D}_\text {KL}\) in output space directly as a regularizer. Instead of limiting our search for distributions p(y|x) by imposing a prior in \(\theta \)-space, we limit our search for p(y|x) directly through a prior p(y) for some suitable \(C_2\).
Such output regularization techniques have indeed been proposed in the literature. In an unregularized classifier, for example, the probabilities assigned to incorrect class labels would often be pushed close to zero without \(\text {D}_\text {KL}\) regularization, such that \(p_{\theta }(y|x)\) collapses to a delta function over the correct class label, which is a sign of overfitting [83]. To avoid this, it has been suggested [69] to view the probabilities assigned to incorrect labels as knowledge the learner has extracted from the dataset, which can be achieved by encouraging the learner to produce output distributions that balance high entropy and minimal loss:
where x, y are training data, \(\mathcal {L}(x, y)\) is a error function and \(H(p) = -\sum _{x_\in \mathcal {X}}p(x)\log p(x)\) is the entropy of p. This technique introduced by Pereyra et al. [69] has immediate connections to our approach through the following observation: adding the \(\text {D}_\text {KL}\) between the agent’s posterior \(p_\theta (y|x)\) and prior p(y) recovers confidence penalty, if the agent’s prior policy is uniform. A similar idea has also been suggested in the context of using label smoothing as an output regularization technique [60].
In reinforcement learning encouraging the agent to learn policies with high entropy is a widely applied technique known as maximum entropy reinforcement learning (RL) [34]. Maximum entropy RL typically penalizes deviation from a fixed uniformly distributed prior to promote exploration, but in a more general setting we can discourage deviation from an arbitrary prior policy by optimizing for
where \(\beta \) trades off between reward and entropy, such that \(\beta \rightarrow \infty \) recovers the standard RL value function and \(\beta \rightarrow 0\) recovers the value function under a random policy. While the entropy term is often regarded as a simple method to encourage exploration, we could similarly regard it as an information constraint for regularization to prevent overfitting as in the case of supervised learning.
3 Within Task Specialization in Hierarchical Multi-agent Policies
In the following we introduce the building blocks of our novel gradient based algorithm to learn the components of a hierarchical multi-agent policy with information constraints. In particular, we leverage the hierarchical model introduced earlier to learn a utility driven partitioning of the state and action spaces. We will demonstrate experimentally how limiting the amount of information each agent can process leads to specialization. First we will show how to transform this principle into a general on-line learning algorithm and afterwards we will derive applications to supervised, unsupervised, and reinforcement learning. In Table 1 we summarize the different setups in supervised, unsupervised and reinforcement learning with according utility functions.
The model consists of two stages: an expert selection stage followed by an action selection stage. The first stage learns a soft partitioning of the state space and assigns each partition optimally to the experts according to a parametrized stochastic policy \(p_\theta (m\vert x)\) with parameters \(\theta \) such that under an information-theoretic constraint we can maximize the free energy \(\Delta \text {F}_{\text {par}}(x,m)\). We start by rewriting Eq. (2) as:
where we define the objective J(x, m, y) as
and \(\theta , \vartheta \) are the parameters of the selection policy and the expert policies. Note that each expert policy has a distinct set of parameters \(\vartheta = \{\vartheta _m\}_m\), but we drop the m index for readability.
As outlined in Sect. 2.1, the optimal prior to find an optimal utility information trade-off is the marginal of the posterior policy given by \(p(y) = \sum _{x} p(x) p(y\vert x)\). It would be prohibitive to compute the prior in each step, as it would require marginalizing over large action and state spaces. Instead, we approximate p(m) and \(p(y\vert m)\) by exponential running mean averages of the posterior policies with momentum terms \(\lambda _1\) and \(\lambda _2\):
3.1 Specialization in Supervised Learning
We set the negative loss as the utility \(\mathbf {U}(y, {\hat{y}}) = -\mathcal {L}(y, {\hat{y}})\), where \({\hat{y}}\) represents the expert’s response (predicted class label, regressed value) and y is the true label. In our implementation we use the cross-entropy loss \(\mathcal {L}(y, {\hat{y}}) = \sum _i y_i \log \frac{1}{{\hat{y}}_i} = -\sum _i y_i \log {\hat{y}}_i\) as a performance measure for classification tasks, and for regression the mean squared error \(\mathcal {L}(y, {\hat{y}}) = \sum _i \vert \vert {\hat{y}}_i-y_i\vert \vert ^2_2 \) between the prediction \({\hat{y}}\) and the ground truth values y. The selection policy thus optimizes
where \({\hat{f}}(m,x) {:}{=}\mathbb {E}_{p_\vartheta ({\hat{y}}\vert m,x)}\big [-\mathcal {L}({\hat{y}},y) - \frac{1}{\beta _2}\log \frac{p_\vartheta ({\hat{y}}\vert x,m)}{p({\hat{y}}\vert m)}\big ]\) is the free energy of expert m. Note that this introduces a double expectation, which we can estimate by Monte Carlo sampling. The experts thus simply optimize their free energy objective defined by
3.2 Specialization in Unsupervised Learning
Unsupervised learning [8] deals with learning from data in the face of missing labels, such as clustering [93] and density estimation [77] algorithms. The density estimation method we propose is similar to RBF Networks [76] or Gaussian Mixture Models [10]. In the following we will show how our method can handle unsupervised learning, where we interpret each expert as a Normal-Wishart distribution. We model this by learning a distribution \(p(\varvec{\mu },\varvec{\varLambda } \vert \varvec{\omega }, \lambda , \varvec{W}, \nu )\) over means \(\varvec{\mu }\) and covariance matrices \(\varvec{\varLambda }\) as Normal-Wishart distributions:
where \(\mathcal {W}\) is Wishart a distribution, \(\mathcal {N}\) a Normal distribution, \(\varvec{\omega } \in \mathbb {R}^D\) is the mean of the normal distribution, \(\varvec{W} \in \mathbb {R}^{D\times D}\) is the scale matrix, \(\nu > D - 1\) is the degree of freedom, \(\lambda > 0\) is a scaling factor, and D denotes the dimensionality of the data. Sampling is straightforward: we first sample \(\varvec{\varLambda }\) from a Wishart distribution with parameters \(\mathbf {W}\) and \(\nu \). Next we sample \(\varvec{\mu }\) from a multivariate normal distribution with mean \(\varvec{\omega }\) and variance \((\lambda \varvec{\varLambda })^{-1}\). We assume the data x follows a normal distribution \(x \thicksim \mathcal {N}(\varvec{\mu }, (\lambda \varvec{\varLambda })^{-1})\). The parameters \(\nu \), \(\lambda \) are hyper-parameters we set beforehand, such that we are interested in finding the parameters \(\varvec{\mu }^*\) and \(\varvec{W}^*\) maximizing the likelihood of the data:
Thus, in this setting the expert’s task is to find parameters \(\omega ^*\) and \(\varvec{W}^*\) in order to select a tuple \((\varvec{\mu }, \varvec{\varLambda })\) that models the likelihood of the data well. The objective of the selector is to assign data to the experts that not only have a set of parameters that yield high likelihood on the assigned data, but also have low statistical complexity as measured by the \(\text {D}_\text {KL}\) between the expert’s posterior and prior distributions. We can now define the free energy difference for each expert as
where \(p(\varvec{\mu },\varvec{\varLambda })\) is the expert’s posterior Normal-Wishart distribution over the parameters \(\mu \) and \(\lambda \) and \(p_0(\varvec{\omega }_0,\varvec{\varLambda }_0)\) is the expert’s prior, p and \(p_0\) are the experts posterior and prior distribution and \(\ell (x\vert \varvec{\mu }, (\lambda \varvec{\varLambda })^{-1})\) is the Gaussian log likelihood
of a data point x given the distribution parameters \(\varvec{\mu }, (\lambda \varvec{\varLambda })^{-1}\). This serves as the basis for the selector’s task of assigning data to the expert with maximum free energy by optimizing
We can compute the \(\text {D}_\text {KL}\) between two Normal-Wishart distributions p and q as
where C is a term that does not depend on the parameters we optimize, so we can omit it, as we are only interested in relative changes in the \(\text {D}_\text {KL}\) caused by changes to \(\varvec{W}\) and \(\varvec{\omega }\) (see “Appendix B” for details on the derivation).
3.3 Specialization in RL Agents
In reinforcement learning we model sequential decision problems by defining a Markov Decision Process (MDP) as a tuple \((\mathcal {S}, \mathcal {A}, P, r)\), where \(\mathcal {S}\) is the set of states, \(\mathcal {A}\) the set of actions, \(P: \mathcal {S} \times \mathcal {A} \times \mathcal {S} \rightarrow [0,1]\) is the transition probability, and \(r: \mathcal {S} \times \mathcal {A} \rightarrow \mathbb {R}\) is a reward function. The aim is to find the parameter \(\theta ^* = \mathop {\hbox {arg max}}\limits \nolimits _{\theta } J(p_\theta )\) of a policy \(p_\theta \) that maximizes the expected discounted reward \(J(p_\theta )= \mathbb {E}_{p_\theta }\left[ \sum _{t=0}^T \gamma ^t r(x_t, a_t)\right] \). In case of an infinite horizon, we have \(T\rightarrow \infty \). We define \(r(\tau ) = \sum _{t=0}^T \gamma ^t r(x_t, a_t)\) as the cumulative reward of trajectory \(\tau = \{(x_t, a_t)\}_{i=0}^T\), where we generate the trajectory according to the policy \(p(a_t|x_t)\) and the environmental dynamics \(P(x_{t+1}\vert x_t, a_t)\). Reinforcement learning [81] models assume that an agent interacts with an environment over a number of discrete time steps t. At each time step t, the agent finds itself in a state \(x_t\) and selects an action \(a_t\) according to the policy \(p(a_t\vert x_t)\). In return, the environment transitions to the next state \(x_{t+1}\) and generates a scalar reward \(r_t\). Here, we consider policy gradient methods [82] which are a popular choice to tackle continuous reinforcement learning problems. The main idea is to directly manipulate the parameters \(\theta \) of the policy in order to maximize the objective \(J(p_\theta )\) by taking steps in the direction of the gradient \(\nabla _\theta J(p_\theta )\).
In the following we will derive our algorithm for specialization in hierarchical reinforcement learning agents. Note that in the reinforcement learning setup the reward function r(x, a) defines the utility \(\mathbf {U}(x,a)\). In maximum entropy RL (see e.g., Haarnoja et al. [34]) the regularization penalizes deviation from a fixed uniformly distributed prior, but in a more general setting we can discourage deviation from an arbitrary prior policy by optimizing for:
where \(\beta \) trades off between reward and entropy, such that \(\beta \rightarrow \infty \) recovers the standard RL value function and \(\beta \rightarrow 0\) recovers the value function under a random policy.
To optimize the objective (18) we define two separate kinds of value function, \(V_\phi \) for the selector and one value function \(V_\varphi \) for each expert. Thus, each expert is an actor-critic with separate actor and critic networks. Similarly, the selector has an actor-critic architecture, where the actor network selects experts and the critic learns to predict the expected free energy of the experts depending on a state variable. The selector’s policy is represented by \(p_\theta \), while each expert’s policy is represented by a distribution \(p_\vartheta \).
3.3.1 Value Functions
In standard reinforcement learning the discounted reward is defined as
which is usually learned through a parameterized value function \(V_\psi \) by regressing
Here \(\psi \) are some arbitrary parameters of the value representation, \(V_\psi (x_t)\) is the predicted value estimate for state \(x_t\), and \(\mathcal {D}\) is a set of trajectories \(\tau \) up to horizon T collected by roll-outs of the policies.
Similar to the standard discounted reward \(R_t\), we can now define a discounted free energy \(F_t\) as
where \(f(x,m,a) = r(x, a) - \frac{1}{\beta _2} \log \frac{p_{\vartheta }(a\vert x,m)}{p(a\vert m)}\). Accordingly, we can learn a value function \(V_\varphi \) for each expert by parameterizing the value function with a neural network and performing regression on \(F_T\). Similarly, we can define a discounted free energy \({\bar{F}}_t\) for the selector
with \({\bar{f}}(x,m) = \mathbb {E}_{p_\vartheta (a\vert x,m)}[r(x,a) - \frac{1}{\beta _2} \log \frac{p(a\vert x, m)}{p(a\vert m)}]\) that is learned through the selector’s value function \(V_\phi \) by regressing \({\bar{F}}_t\).
3.3.2 Policy Learning
In standard reinforcement learning a common technique to update a parametric policy representation \(p_\omega (a|x)\) with parameters \(\omega \) is to use policy gradients that optimize the cumulative reward
expected under the critic’s prediction \(V_\psi (x)\), by following the gradient
This policy gradient formulation [82] is prone to producing high variance gradients. A common technique to reduce the variance is to formulate the updates using the advantage function instead of the reward [5]. The advantage function \(A(a_t,s_t)\) is a measure of how well a certain action a performs in a state x compared to the average performance in that state, i.e., \(A(a,x) = Q(x,a) - V_\psi (x)\). Here, V(x) is the value function and is a measure of how well the agent performs in state x, and Q(x, a) is an estimate of the cumulative reward achieved in state x when the agent executes action a. Thus, the advantage is an estimate of how advantageous it is to pick a in state x in relation to a baseline performance \(V_\psi (x)\). Instead of learning the value and the Q function, we can define the advantage function solely based on the critic’s estimate \(V_\psi (x)\) in the following way
giving the following gradient estimates for the policy parameters
where \(\alpha \) is a learning rate and \(\mathcal {D}\) is a set of trajectories \(\tau \) produced by the policies.
Similar to the standard policy update based on the advantage function, the expert selection stage can be formulated by optimizing the expected advantage \(\mathbb {E}_p(a|x,m) \left[ A_m(x,a)\right] \) for expert m with
Accordingly, we can define an expected advantage function \(\mathbb {E}_p(m|x) \left[ {\bar{A}}(x,m)\right] \) for the selector with
We estimate the double expectation by Monte Carlo sampling, where in practice we use a single (x, m, a) tuple for \({\hat{f}}(x,m)\), which enables us to employ our algorithm in an on-line optimization fashion.
4 Experiments and Results: Within Task Specialization
In the following we will show applications to different learning tasks where the overall complexity of the problem exceeds the processing power of the individual (linear) experts. In particular, we will look at classification, density estimation, and reinforcement learning. See “Appendix A” for implementation details. In Sect. 7.2 we discuss how to choose proper values for \(\beta _1\) and \(\beta _2\) and their effects on learning.
4.1 Classification with Linear Decision-Makers
When dealing with complex data for classification (or regression) it is often beneficial to combine multiple classifiers (or regressors). In the framework of ensemble learning, for example, multiple classifiers join together to stabilize learning and improve the results [47], such as Mixture-of-Experts [96] and Multiple Classifier systems [9]. The method we propose when applied to classification problems falls within this scope of algorithms. The application to regression is an example of local linear regression [6].
We evaluate our method on three synthetic datasets for classification—see Fig. 2. Our method is able to partition the problem set into subspaces (third column from the left) and fit a linear decision-maker on each subset, achieving acceptable classification accuracy on synthetic non-linear datasets. The results on the ”Half Moons” dataset show an example where the quality of the classification does not improve with the number of experts, essentially because a satisfactory performance can already be achieved with a single expert. We can see that a single expert is classifying the data reasonably well and adding more experts improves accuracy only marginally, whereas in the remaining two datasets the accuracy increases with the number of experts. Regarding the information processed by each expert \(I(X;Y\vert M)\), a single expert on the ”Half Moons” achieves a competitive score compared to the system with two and four experts, which in turn results in high accuracy. This also manifests in the selection prior p(m) which shows for this dataset a non-uniform division of labor between the experts. In contrast to this, the results on ”Circles” and ”Blobs” show how adding more experts is beneficial if the underlying dataset has a highly non-linear structure. In both settings the information processed by a single expert is close to zero bits and classification accuracy is at chance level. Adding more experts allows for specialization and thus increased processing power \(I(X;Y\vert M)\) which in turn achieves higher classification accuracy.
In Fig. 3 we compare the performance of the “Circles” dataset to different baselines constructed from decision tree learning: a single decision tree with varying depth, multiple decision trees as part of a random forest, and multiple decision trees within Adaboost. In Fig. 4 we report how information processing in the selection and the action stage influences classification accuracy. We can see that under a certain threshold the accuracy is random at best and increases with processing power, which is not surprising. Additionally, we can see that a high selection processing power compensates low processing power in the decision-makers up to a certain degree. If the processing power of the experts is too low, adding more experts does not increase the system’s performance indefinitely, as it becomes harder for the selection stage to pick a certain expert. This happens because the experts can only specialize on a small subspace of the task due to their low information-processing capabilities.
4.2 Unsupervised Learning
We report unsupervised learning results in Fig. 5, where we show how the system deals with unlabeled data. In this experiment the synthetic dataset contains four clusters and our algorithm is able to perform the clustering as we add more and more experts to partition the state space further. If we provide more experts (in this case 8) than there are clusters (in this case 4), the selector neglects the additional experts and only assigns data to four of them. This indicates that the optimization process aims to find the optimal number of experts in this case.
4.3 Reinforcement Learning with Linear Decision-Makers
In the following we will show how our hierarchical system is able to solve a continuous reinforcement learning problem using an optimal arrangement of linear control policies. We evaluate on a task known as Acrobot [80], more commonly referred to as the inverted double pendulum. The task is to swing a double-linked pendulum up and keep it balanced as long as possible, where the agent receives a reward of 10 minus a distance penalty between its current state and the goal state. Each episode terminates as soon as the environment reaches a predefined terminal state (hanging downwards) or after 1000 time steps. To balance the pendulum the agent is able to apply a force to the base joint in the range of \(a \in [-1, +1]\), thus creating a movement to the left or to the right. This environment poses a non-linear control problem and thus a single linear controller cannot give an optimal solution. We show how using our approach enables a committee of linear experts to efficiently solve this non-linear task. We report the results in Fig. 6. We allowed for five experts (\(\beta _2 = 2.5\)), but our system learns that three experts are sufficient to solve the task. The priors for each expert (lower right figure, each color represents an expert) center on −1, 0, and 1, which correspond to swinging the double pendulum to the left, no force, and swinging to the right. The remaining two experts overlap accordingly. We can see that the average information processing in the five expert setup decreases, while in the selection it increases to \(\log 3\). Both indicate that the system has learned an optimal arrangement of three experts and is thus able to achieve maximum reward and eventually catches up to the performance of a non-linear neural network controller trained with TRPO [75] that does not have to struggle with the restriction to linear controllers as our algorithm. Our method successfully learned a partitioning of the double-pendulum state space without having any prior information about any of the system dynamics or the state space.
In Fig. 7 we show how a model with two linear experts and selector learns to control the cart pole problem. Our method recovers the well known solution in which the pole can be balanced by two linear controllers, where one (dark purple) focuses on keeping the pole upright and the other (dark yellow) on moving the cart such that the other linear controller can take over.
5 Across Task Specialization for Hierarchical Meta-Learning
The methods presented in Sect. 3 can be easily extended to achieve meta-learning by changing the way the selection mechanism is trained. Instead of assigning individual states that occur within a task, the selector assigns a complete dataset of a task to an expert. To do so, we must find a feature vector z(d) of the dataset d. This feature vector must fulfill the following desiderata: (1) invariance against permutation of data in d, (2) high representational capacity, (3) efficient computability, and (4) constant dimensionality regardless of sample size K. In the following we propose methods to extract such features for image classification, regression, and reinforcement learning problems—see Fig. 13. While the experts are trained on the training dataset \(D_\text {meta-train}\), their performance used to optimize the selection policy is based on the validation dataset \(D_\text {meta-val}\). The validation dataset contains previously unseen samples that are similar to those in the training set, thus providing a measure for the generalization of each expert. In effect, the selector operates on complete datasets, while the experts operate on single samples.
5.1 Specialization in Meta-Supervised Learning
In a supervised learning task we are usually interested in a dataset consisting of multiple input and output pairs \(D = \{(x_i, y_i)\}^N_{i=1}\) and the learner’s task is to find a function f(x) that maps from input to output, for example through a deep neural network. To do this, we split the dataset into training and test sets and fit a set of parameters \(\theta \) on the training data and evaluate on test data using the learned function \(f_\theta (x)\). In meta-learning, we are instead working with meta-datasets \(\mathcal {D}\), each containing regular datasets split into training and test sets. We thus have different sets for meta-training, meta-validation, and meta-test, i.e., \(\mathcal {D} = \{D_\text {meta-train}, D_\text {meta-val},D_\text {meta-test}\}\). The goal is to train a learning procedure (the meta-learner) that can take as input one of its training sets \(D_\text {meta-train}\) and produce a classifier (the learner) that achieves low prediction error on its corresponding test set \(D_\text {meta-test}\). The meta-learning is then updated using performance measure based on the learner’s performance on \(D_{\text {meta-val}}\), compare Algorithm 1. This may not always be the case, but our work (among others, e.g., Finn et al. [19]) follow this paradigm. The rationale being that the meta-learner is trained such that it implicitly optimizes the base learner’s generalization capabilities. The dataset generating distribution \(p(\mathcal {D})\) is unknown to the learner but remains fixed over course of training. The case where \(p(\mathcal {D})\) is changing is study in the field of (meta) continual learning, but is not the focus of this work.
For image classification, we propose to pass the images with positive labels in the dataset through a convolutional autoencoder and use the outputs of the bottleneck layer. Convolutional autoencoders are generative models that learn to reconstruct their inputs by minimizing the Mean-Squared-Error between the input and the reconstructed image (see e.g., Vincent et al. [89]. In this way we get similar embeddings z(d) for similar inputs belonging to the same class. We compute the latent representation for each positive sample in d and pass it through a pooling function h(z(d)) to find a single embedding for the complete dataset—see Fig. 8 for an overview of our proposed model. We found that max pooling yields the best results, while one could use others, such as mean or min pooling. Yao et al. [94] propose a similar feature set. For regression, we transform the training data into a feature vector z(d) by binning the points into N bins according to their x value and collecting the y value. If more than one point falls into the same bin we average the y values, thus providing invariance against the order of the data in \(D_{\text {meta-train}}\). We use this feature vector to assign each dataset to an expert according to \(p_\theta (m\vert h(z(D_{\text {meta-train}})))\), which we abbreviate to \(p_\theta (m\vert D_\text {meta-train})\).
In contrast to the objective defined by Eq. (10), the selection policy now selects experts based on their free-energy that is computed over datasets \(D_{\text {meta-val}}\) and the selection policy depends on the training datasets \(D_{\text {meta-train}}\)
where \({\hat{f}}(m,D_{\text {meta-val}}) {:}{=}\mathbb {E}_{p_\vartheta ({\hat{y}}\vert m,x)}\big [-\mathcal {L}({\hat{y}},y) - \frac{1}{\beta _2}\log \frac{p_\vartheta ({\hat{y}}\vert x,m)}{p({\hat{y}}\vert m)}\big ]\) is the free energy of expert m on dataset \(D_{\text {meta-val}}\), \(\mathcal {L}({\hat{y}},y)\) is loss function, and \((x,y) \in D_\text {meta-val}\). The experts optimize their free energy objective on the training dataset \(D_{\text {meta-train}}\) defined by
where \((x, y) \in D_\text {meta-train}\).
5.2 Specialization in Meta-Reinforcement Learning
In meta reinforcement learning we extend the problem to a set of tasks \(t_i \in T\), where a MDP \(t_i = (\mathcal {S}, \mathcal {A}, P_i, r_i)\) defines each task \(t_i\). We are now interested in finding a set of policies \(\varTheta \) that maximizes the average cumulative reward across all tasks in T and generalizes well to new tasks sampled from a different set of tasks \(T'\).
In this setting we use a dynamic recurrent neural network (RNN) with independent recurrent units [53] to classify trajectories and a second RNN to learn the value function (see “Appendix A” for details). We feed the RNN with L \((s_t, a_t, r_t, s_{t+1})\) tuples to describe the underlying Markov Decision Process describing the task. At \(t = 0\) we sample the expert according to the learned prior distribution p(m), as there is no other information available until we have collected L samples at which point an expert is selected—see Algorithm 2. Lan et al. [50] propose a similar feature set. Importantly, the expert policies are trained on the meta-training environments, but evaluated on unseen but similar validation environments. In this setting we define the discounted free energy \({\bar{F}}_t\) for the selector as
with \({\bar{f}}(x,m) = \mathbb {E}_{p_\vartheta (a\vert x,m)}\left[ r_{\text {meta-train}}(x,a) - \frac{1}{\beta _2} \log \frac{p(a\vert x, m)}{p(a\vert m)}\right] \), where \(r_{\text {meta-val}}\) is a reward function defined by a validation environment (see Fig. 10 for details).
6 Experimental Results: Across Task Specialization and Meta-Learning
In this section we present our experimental results in the meta-learning domain. To show the flexibility of our method we evaluate on regression, classification and reinforcement learning problems. In regression, we evaluate how our method adapts to changing sine functions, for classification we look at the Omniglot dataset [49]. To evaluate on reinforcement learning we introduce a range of versions of the double pendulum task [80]. We provide all experimental details such as network architectures and hyper-parameters in “Appendix B”.
6.1 Sinusoid Regression
We adopt this task from Finn et al. [19]. In this K-shot problem, each task consists of learning to predict a function of the form \(y = a \cdot \sin (x + b)\), with both \(a \in [0.1, 5]\) and \(b \in [0, 2\pi ]\) chosen uniformly, and the goal of the learner is to find y given x based on only K pairs of (x, y). Given that the underlying function changes in each iteration it is impossible to solve this problem with a single learner. As a loss function we use Mean-Squared-Error and the dataset embedding is described in Algorithm 1. Each expert is a shallow neural network consisting of a single hidden layer connected to an output layer (see “Appendix A” for details). Our results show that by combing expert networks, we are able to reduce the generalization error iteratively as we add more experts to our system—see Fig. 9 for \(K = 5\) and \(K = 10\) settings. In Fig. 9 we show how the system is able to capture the underlying problem structure as we add more experts and in Fig. 11 we visualize how the selector’s partition of the problem space looks like. In “Appendix A”, Fig. 15 we show additional results and give an overview of our algorithm in Algorithm 1.
6.2 Few-Shot Classification
A special case of meta-learning for classification are K-Shot N-way tasks, where a training set consists of K labeled examples of each of the N classes (\(K\cdot N\) examples per dataset) and a corresponding test set is also provided. In our study, we focus on the following variation of K-Shot N-Way tasks: 2K samples (K positive and K negative examples) define a training dataset which the meta-learner must assign to an expert learner that has a specialized policy to classify the 2K samples. To evaluate experts performance we use a meta-validation set consisting of a different set of 2K samples. Note that we draw the negative examples from any of the remaining classes.
The Omniglot dataset [48] consists of over 1600 characters from 50 alphabets (see Fig. 12 for examples ). As each character has merely 20 samples each drawn by a different person, this forms a difficult learning task and is thus often referred to as the ”transposed MNIST” dataset. The Omniglot dataset is a standard meta-learning benchmarking dataset [19, 71, 90].
We consider three experimental setups: (1) how does a learner with only a single hidden layer perform when trained naïvely compared to with sophisticated methods such as MAML [19] and Matching Nets [90] as a baseline? (2) does the system benefit from adding more experts and if so, at what rate? and (3) how does our method compare to the aforementioned algorithms? Regarding (1) we note that introducing constraints by reducing the representational power of the models does not facilitate specialization is it would by explicit information-processing constraints. In the bottom row of Fig. 9 we address question (2). We can interpret this curve as the rate-utility curve showing the trade-off between information processing and expected utility (transparent area represents one standard deviation), where increasing I(X; M) improves adaptation. The improvement gain grows logarithmically, which is consistent with what rate-distortion theory would suggest. In Table 2 we present empirical results addressing question 3).
We train the learner on a subset of the dataset (\(\approx 80\%\), \(\approx \) 1300 classes) and evaluate on the remaining \(\approx 300\) classes, thus investigating the ability to generalize to unseen data. In each round we build the datasets \(D_{\text {meta-train}}\) and \(D_{\text {meta-test}}\) by selecting a target class \(c_t\) and sample K positive and K negative samples. To generate negative samples we draw K images randomly out of the remaining \(N-1\) classes. We present the selection network with the feature representation of the K positive training samples (see Fig. 8), but evaluate the experts’ performance on the 2K test samples in \(D_{\text {meta-test}}\). We can interpret the free energy of the experts in this setting as a measure of how well the expert is able to generalize to new samples of the target class. Using this optimization scheme, we train the expert networks to become experts in recognizing a subset of classes. We train the experts using the 2K samples from the training dataset that the selection network assigned to an expert—see Table 2 for results. We followed the design of Vinyals et al. [90] to design our experts but reduce the number of blocks to one to introduce effects of resource limitation, whereas in the original study the authors used four blocks. The single convolutional block consists of 32 \(3 \times 3\) filters with strided convolutions followed by a batch normalization layer and a ReLu non-linearity. The output is fed into a softmax layer giving a distribution over the classes (see also “Appendix A”). This reduced representational capacity drives specialization, as the expert can not reliably classify all characters from the training data, but a subset is feasible (see also Fig. 12). To evaluate our method we compare different ensemble sizes against three baselines: pre-training, MAML [19] and Matching Networks [90]. In the pre-training setting we train a single convolutional neural network on batches drawn from the training data and evaluate on validation data by allowing 10 gradient steps for fine-tuning. Note, that when using the architecture of 4 blocks as suggested in the original paper [19, 90], we are able to achieve \(\ge \)95% accuracy on the test data in both MAML and matching nets, but not on the pre-training setting.
6.3 Meta Reinforcement Learning
In meta reinforcement learning the goal is to find a policy that performs well over a set of tasks. We create a set of RL tasks by extending the “Inverted Double Pendulum problem” [80] implemented in OpenAI Gym [14] by allowing a broad range of task parameters. Each time we create a new environment we sample from a specified distribution, where we modify inertia, motor torques, reward function, goal position and invert the control signal—see Table 10 for details. We create one environment per training episode, where during a single training episode parameters remain unchanged. We measure the free energy of an expert on a second task with parameters also drawn from T.
To evaluate the agents’ meta-learning capabilities we define a second set of tasks \(T'\) where the parameter distributions are different, providing new but similar reinforcement learning problems. In each episode we sample M environments from T and update the system batch-wise. After training we evaluate on tasks from \(T'\), thus testing the agents generalization. We trained the system for 1000 Episodes with 64 tasks from T and evaluate for 100 system updates on tasks from \(T'\). We report the results in Fig. 10, where we can see improving performance as we add more experts, where the mutual information characterizing the selection stage indicates that the selector is able to identify suitable experts for the tasks.
7 Discussion
7.1 Analyzing the Discovered Meta-Knowledge
In contrast to monolithic approaches that train a single agent as a black-box we can analyze and interpret the meta-knowledge discovered by our hierarchical method. In the following we will discuss this in the supervised and reinforcement learning setting.
In meta-regression the problem space defined by a set of sine functions \(y = a\cdot \sin (x+b)\) is split among the ensemble of expert regressors based on only \(K \in \{1, 5, 10\}\). As expected, the assignment becomes more accurate the more samples we have—see Fig. 11 where we report how the selection network partitions the sine task space. The shape of the emerged clusters indicates that the selection is mainly based on the amplitude a of the current sine function, indicating that from an adaptation point-of-view it is more efficient to group sine functions based on amplitude a instead of phase b. We can also see that an expert specializes on the low values for b as it covers the upper region of the \(a\times b\) space. The selection network splits this region among multiple experts if we increase the set of experts to 8 or more.
We can also analyze the class assignment policy learned by the selector network. The assignment map in Fig. 12 suggests that the selector learns to group characters by their alphabet based only on features. The selector’s policy spreads different characters from the same alphabet (e.g., alphabets 0, 36, and 46) across multiple experts while assigning similar characters from different alphabets to the same experts. This specializations gives rise to the meta-learning ability as it is able to adapt expert parameters within only a few gradient steps. We generated this plot by passing images of characters through the system (i.e., computing their latent embedding and assigning them to experts) after training is complete to obtain an overview of the class distribution.
For reinforcement learning we demonstrate this idea by looking at how the selection networks splits the state space into to linearly controllable regions in Fig. 7. We can see that one expert receives the states around zero (dark purple) and the other experts sees only states near the boundary (dark yellow). We derive from this that the purple expert specializes on balancing the pole and the yellow expert specializes on moving the cart such that the other expert can easily balance the pole. This is consistent with the fact that a linear policy can balance the pole when it is close to standing upwards.
7.2 Critical Issues
A limitation of our method is low sample efficiency in the RL domain. To alleviate this problem one could imagine to combine our system with model-based RL methods which promise to reduce the number of samples the agent needs to see in order to learn efficiently. Another research direction would be to investigate our systems performance in continual adaptation tasks, such as in the work of Yao et al. [94]. There the agent is continuously exposed to datasets (e.g., additional classes and samples). The restriction to binary meta classification tasks is another limitation of our method, which we leave for feature work.
Another open question remains the tuning of \(\beta \) values. As the utility function can in principle be unbounded whereas the information-processing quantities are obviously bounded, the agent may be faced with learning two values that differ greatly in magnitude. This becomes especially challenging in reinforcement learning scenarios, where a value function of varying magnitude has to be learned. This poses a difficult learning problem and there have been several proposals to tackle this. One method is dubbed ‘Pop-Art’ by van Hasselt et al. [87], that treats the incoming utility values as a stream of data and normalize the values to a given range. In the reinforcement learning setup we also tried a cooling schedule for \(\beta \), as suggested by Fox et al. [20]. In their work the authors propose to change the weight of the entropy penalty in MaxEnt RL as time progresses, thus encouraging exploration in the beginning and penalizing it the more information an agent has gathered. We did not observe any significant performance gains.
The specific value of \(\beta \) depends on the scale of the utility function. As the value of \(\beta \) strongly influences the outcome of the experiments, it must be chosen with care and comes with the same limitations as any other regularization technique. If it is chosen to small, the regularization term dominates the utility term and the experts are not able to learn anything. On the other hand, if it is set to a large value, the regularization term vanishes, such that no specialization can occur and thus the selector has a hard time assigning data to experts. To remedy this, there are in principle two ways to choose \(\beta \): one is to set a information-processing limit for each expert and then (manually or with some optimization technique) tune beta such hat this constraint is satisfied. This has the advantage that this value can be interpreted, e.g., “the experts can process 1 bit of information, i.e., distinguishing two options”. The other way is to run a grid search over a pre-defined range and chose the one that fits best. In this work, we used the second strategy.
7.3 Related Work
Investigating information-theoretic cost functions and constraints in learning systems has recently enjoyed increasing interest in machine learning [21, 32, 33, 86]. The method we propose in this study falls into a wider class of algorithms that aim to deal more efficiently with learning and decision-making problems [17, 25, 28,29,30,31, 35, 51, 56, 62, 68, 73].
Applying such constraints to methods for reinforcement learning is often motivated by the aim of stabilizing learning and reducing sample complexity. One such approach is Trust Region Policy Optimization (TRPO) introduced by Schulman et al. [75], where a \(\text {D}_\text {KL}\) penalty between the old and the new policy is imposed to limit update steps, providing a theoretical guarantee for monotonic policy improvement. In our approach we define this region by \(\text {D}_\text {KL}\) between the agent’s posterior and prior policy, thus allowing to learn this region and to adapt it over time. This basic idea has been extend to meta-learning by [72], which we use to compare our method against in meta-rl experiments. Daniel et al. follow a similar idea with relative entropy policy search methods [17]. The algorithm builds on learning a gating policy that can decide which sub-policy to choose. The authors impose a \(\text {D}_\text {KL}\) constraint between the data distribution and the next policy to achieve this. Both these approaches smoothen the updates as large steps are discouraged. Our method follows the same principle, but we enforce small updates by discouraging deviation from the agent’s prior and by limiting the representational power (e.g., by linear decision makers).
The hierarchical structure we employ is related to Mixture of Experts (MoE) models. Jacobs et al. [39] introduced MoE as tree structured models for complex classification and regression problems, where the underlying approach is the divide and conquer paradigm. As in our approach, three main building blocks define MoEs: gates, experts, and a probabilistic weighting to combine expert predictions. Learning proceeds by finding a soft partitioning of the input space and assigning partitions to experts performing well on the partition. In this setting, the model response is then a sum of the experts’ outputs, weighted by how confident the gate is in the expert’s opinion. Yuksel et al. [96] provide an extensive overview of recent developments in MoEs. The approach we propose allows learning such models, but also has applications to more general decision-making settings such as reinforcement learning. Ghosh et al. [25] recently applied the divide-and-conquer principle to reinforcement learning. They argue that dividing a central policy into sub-policies improves the learning phase by improving sample efficiency. To evaluate this approach they assume pre-defined partitions on the action and state space on which they train local policies. The information-theoretic constraints during training enforce similarity between the local policies such that a single central policy arises as weighted combination of all local policies. In contrast, in our approach all posterior expert policies remain close to their priors thereby minimizing their informational surprise. This mechanism leads to the emergence of specialized policies. In effect, this enforces the local policies to be as diverse as possible. Crucially, in our method the partitioning is not predefined but a result of the optimization process itself.
Untangling the underlying structure in control systems usually requires a-priori knowledge of the system dynamics, e.g., [2, 70, 95]. The algorithm proposed by Abramova et al. [2] splits the state space of the inverted pendulum into predefined bins to fit a linear control policy to stabilize individually. Then, the authors suggest to control the whole system by learning a selection policy over the given linear controllers. In contrast to this, our approach relies only on the reward signal to learn the selection policy and the linear control policy simultaneously. This fact alone poses a difficult learning problem as both system parts have to adjust to one another on different timescales. Other decentralized approaches (e.g., Allamraju and Chowdhary [4]) have trained separate decentralized models to fuse them into a single model. In contrast, our method learns sub-policies that act on their own.
Most other methods for meta-learning such as the work of [19] and [71] find an initial parametrization of a single learner, such that the agent can adapt quickly to new problems. This initialization represents prior knowledge and can be regarded as an abstraction over related tasks and our method takes this idea one step further by finding a possibly disjunct set of such compressed task properties. Another way of thinking of such abstractions by lossy compression is to go from a task-specific posterior to a task-agnostic prior strategy. By having a set of priors the task specific information is available more locally then with a single prior, as in MAML [19] and the work of [71]. In principle, this can help to adapt within fewer iterations. Thus our method can be seen as the general case of such monolithic meta-learning algorithms. Instead of learning similarities within a problem, we can also try to learn similarities between different problems (e.g., different classification datasets), as is described in the work of [94]. In this way, the partitioning is governed by different tasks, where our study however focuses on discovering meta-information within the same task family, where the meta-partitioning is determined solely by the optimization process and can thus potentially discover unknown dynamics and relations within a task family.
8 Conclusion
In summary, we introduce and evaluate a promising novel on-line learning paradigm for hierarchical multi-agent systems. The main idea of our approach is an optimal soft partitioning by considering the agents’ information constraints. The partitioning is automatic without relying on any prior information about the underlying problem structure or control dynamics in the case of model free learning. This makes our model abstract and principled. We apply it on a variety of tasks including multi-agent decision-making, mixture-of-expert regression, and divide-and-conquer reinforcement learning. We have extended this idea to a novel information-theoretic approach to meta-learning. In effect, the hierarchical structure equips the system with optimal initializations covering the input space which facilitates quick adaptation to new similar tasks. To build this hierarchical structure, we have proposed feature extraction models for classification, regression and reinforcement learning, that are able to extract task relevant information efficiently, invariant to the order of the inputs. The main strength of our approach is that it follows from simple principles that give rise to a large range of applications. Moreover, we can interpret the system performance by studying the information processing both at the selection stage and at the expert level, as we have shown by analyzing the discovered meta-knowledge. This can help to alleviate the problems inherent to black-box-approaches, for example based on deep neural networks.
References
Abadi M, Barham P, Chen J, Chen Z, Davis A, Dean J, Devin M, Ghemawat S, Irving G, Isard M et al (2016) Tensorflow: a system for large-scale machine learning. In: 12th USENIX symposium on operating systems design and implementation (OSDI 16), pp 265–283
Abramova E, Dickens L, Kuhn D, Faisal A (2012) Hierarchical, heterogeneous control of non-linear dynamical systems using reinforcement learning. In: European workshop on reinforcement learning at ICML
Aldrich H (1999) Organizations evolving. Sage, London
Allamraju R, Chowdhary G (2017) Communication efficient decentralized Gaussian process fusion for multi-UAS path planning. In: Proceedings of the 2017 American control conference (ACC). IEEE, pp 4442–4447
Arulkumaran K, Deisenroth MP, Brundage M, Bharath AA (2017) Deep reinforcement learning: a brief survey. IEEE Signal Process Mag 34(6):26–38
Atkeson CG, Moore AW, Schaal S (1997) Locally weighted learning for control. In: Lazy learning. Springer, Berlin, pp 75–113
Balasundaram S, Meena Y (2019) Robust support vector regression in primal with asymmetric huber loss. Neural Process Lett 49(3):1399–1431
Barlow HB (1989) Unsupervised learning. Neural Comput 1(3):295–311
Bellmann P, Thiam P, Schwenker F (2018) Multi-classifier-systems: architectures, algorithms and applications. In: Computational intelligence for pattern recognition, Springer, Berlin, pp 83–113
Biernacki C, Celeux G, Govaert G (2000) Assessing a mixture model for clustering with the integrated completed likelihood. IEEE Trans Pattern Anal Mach Intell 22(7):719–725
Botvinick M, Ritter S, Wang JX, Kurth-Nelson Z, Blundell C, Hassabis D (2019) Reinforcement learning, fast and slow. Trends in cognitive sciences
Braun DA, Mehring C, Wolpert DM (2010) Structure learning in action. Behav Brain Res 206(2):157–165
Brazdil P, Carrier CG, Soares C, Vilalta R (2008) Metalearning: applications to data mining. Springer, Berlin
Brockman G, Cheung V, Pettersson L, Schneider J, Schulman J, Tang J, Zaremba W (2016) Openai gym. arXiv preprint arXiv:1606.01540
Caruana R (1997) Multitask learning. Mach Learn 28(1):41–75
Damasio A (2009) Neuroscience and the emergence of neuroeconomics. In: Neuroeconomics. Elsevier, pp 207–213
Daniel Christian, Neumann Gerhard, Peters Jan (2012) Hierarchical relative entropy policy search. In: Artificial Intelligence and Statistics, pages 273–281
Edward V, Noah G, Griffiths TL, Tenenbaum JB (2014) One and done? Optimal decisions from very few samples. Cognit Sci 38(4):599–637
Finn C, Abbeel P, Levine S (2017) Model-agnostic meta-learning for fast adaptation of deep networks. In: Proceedings of the 34th international conference on machine learning, vol 70, pp 1126–1135. JMLR. org
Fox R, Pakman A, Tishby N (2016) Taming the noise in reinforcement learning via soft updates. In: Proceedings of the thirty-second conference on uncertainty in artificial intelligence, pp 202–211
Galashov A, Jayakumar SM, Hasenclever L, Tirumala D, Schwarz J, Desjardins G, Czarnecki WM, Teh YW, Pascanu R, Heess N (2019) Information asymmetry in KL-regularized RL. In: Proceedings of the international conference on representation learning
Genewein T, Hez E, Razzaghpanah Z, Braun DA (2015) Structure learning in bayesian sensorimotor integration. PLoS Comput Biol 11(8):e1004369
Genewein T, Leibfried F, Grau-Moya J, Braun DA (2015) Bounded rationality, abstraction, and hierarchical decision-making: an information-theoretic optimality principle. Front Robot AI 2:27
Gershman SJ, Horvitz EJ, Tenenbaum JB (2015) Computational rationality: a converging paradigm for intelligence in brains, minds, and machines. Science 349(6245):273–278
Ghosh D, Singh A, Rajeswaran A, Kumar V, Levine S (2018) Divide-and-conquer reinforcement learning. In: Proceedings of the international conference on representation learning
Gigerenzer G, Brighton H (2009) Homo heuristicus: why biased minds make better inferences. Top Cognit Sci 1(1):107–143
Giraud-Carrier C (2008) Metalearning-a tutorial. In: Tutorial at the 7th international conference on machine learning and applications (ICMLA), San Diego, California, USA
Gottwald S, Braun DA (2019) Bounded rational decision-making from elementary computations that reduce uncertainty. Entropy 21(4)
Gottwald S, Braun DA (2019) Systems of bounded rational agents with information-theoretic constraints. Neural Comput 31(2):440–476
Grau-Moya J, Krüger M, Braun DA (2017) Non-equilibrium relations for bounded rational decision-making in changing environments. Entropy 20(1):1
Grau-Moya Jordi, Leibfried Felix, Genewein Tim, Braun Daniel A (2016) Planning with information-processing constraints and model uncertainty in markov decision processes. In: Proceeedings of the Joint European Conference on Machine Learning and Knowledge Discovery in Databases, pages 475–491. Springer
Grau-Moya J, Leibfried F, Vrancx P (2019) Soft q-learning with mutual-information regularization. In: Proceedings of the international conference on learning representations
Grover A, Ermon S (2019) Uncertainty autoencoders: Learning compressed representations via variational information maximization. In: Proceedings of the the 22nd international conference on artificial intelligence and statistics, pp 2514–2524
Haarnoja T, Tang H, Abbeel P, Levine S (2017) Reinforcement learning with deep energy-based policies. In: Proceedings of the 34th international conference on machine learning, vol 70, pp 1352–1361. JMLR. org
Hihn H, Gottwald S, Braun DA (2018) Bounded rational decision-making with adaptive neural network priors. In: IAPR workshop on artificial neural networks in pattern recognition. Springer, pp 213–225
Hihn H, Gottwald S, Braun DA (2019) An information-theoretic on-line learning principle for specialization in hierarchical decision-making systems. In: Proceedings of the 2019 IEEE conference on decision-making and control (CDC)
Hutter F, Kotthoff L, Vanschoren J, Automated machine learning. Springer, Berlin
Ioffe S, Szegedy C (2015) Batch normalization: accelerating deep network training by reducing internal covariate shift. arXiv preprint arXiv:1502.03167
Jacobs RA, Jordan MI, Nowlan SJ, Hinton GE (1991) Adaptive mixtures of local experts. Neural Comput 3(1):79–87
Jankowski N, Duch W, Grkabczewski K (2011) Meta-learning in computational intelligence, vol 358. Springer, Berlin
Jaynes ET (1996) Probability theory: the logic of science. Washington Universityn St. Louis, MO
Kemp C, Perfors A, Tenenbaum JB (2007) Learning overhypotheses with hierarchical bayesian models. Dev Sci 10(3):307–321
Kingma Diederik P, Ba Jimmy (2014) Adam: A method for stochastic optimization. In: Proceedings of the International Conference on Representation Learning
Kingma DP, Welling M (2013) Auto-encoding variational bayes. In: Proceedings of the international conference on representation learning
Koch G, Zemel R, Salakhutdinov R (2015) Siamese neural networks for one-shot image recognition. In: ICML deep learning workshop, vol 2
Kukačka J, Golkov V, Cremers D (2017) Regularization for deep learning: a taxonomy. arXiv preprint arXiv:1710.10686
Kuncheva LI (2004) Combining pattern classifiers: methods and algorithms. Wiley, London
Lake B, Salakhutdinov R, Gross J, Tenenbaum J (2011) One shot learning of simple visual concepts. In: Proceedings of the annual meeting of the cognitive science society, vol 33
Lake BM, Salakhutdinov R, Tenenbaum JB (2015) Human-level concept learning through probabilistic program induction. Science 350(6266):1332–1338
Lan L, Li Z, Guan X, Wang P (2019) Meta reinforcement learning with task embedding and shared policy. In: Proceedings of the international joint conference on artificial intelligence
Leibfried F, Braun DA (2015) A reward-maximizing spiking neuron as a bounded rational decision maker. Neural Comput 27(8):1686–1720
Lemke C, Budka M, Gabrys B (2015) Metalearning: a survey of trends and technologies. Artif Intell Rev 44(1):117–130
Li S, Li W, Cook C, Zhu C, Gao Y (2018) Independently recurrent neural network (INDRNN): building a longer and deeper RNN. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 5457–5466
Lindig-Leon Cecilia, Gottwald Sebastian, Braun Daniel Alexander (2019) Analyzing abstraction and hierarchical decision-making in absolute identification by information-theoretic bounded rationality. Front Neurosci 13:1230
Manson SM (2006) Bounded rationality in agent-based models: experiments with evolutionary programs. Int J Geogr Inf Sci 20(9):991–1012
Martius G, Der R, Ay N (2013) Information driven self-organization of complex robotic behaviors. PloS one 8(5):e63400
McAllester DA (1999) Pac-bayesian model averaging. In: Proceedings of the twelfth annual conference on Computational learning theory, pp 164–170
McAllester DA (2003) Pac-bayesian stochastic model selection. Mach Learn 51(1):5–21
McKelvey RD, Palfrey TR (1995) Quantal response equilibria for normal form games. Games Econ Behav 10(1):6–38
Müller R, Kornblith S, Hinton GE (2019) When does label smoothing help? In: Advances in neural information processing systems, pp 4694–4703
Nagabandi A, Clavera I, Liu S, Fearing RS, Abbeel P, Levine S, Finn C (2018) Learning to adapt in dynamic, real-world environments through meta-reinforcement learning. In: International conference on learning representations
Neumann G, Daniel C, Kupcsik A, Deisenroth M, Peters J (2013) Information-theoretic motor skill learning. In: Proceedings of the AAAI workshop on intelligent robotic systems
Ortega P, Braun D (2011) Information, utility and bounded rationality. Lect Notes Artif Intell 6830:269–274
Ortega PA, Braun DA (2013) Thermodynamics as a theory of decision-making with information-processing costs. Proc R Soc Lond A: Math Phys Eng Sci 469(2153)
Ortega PA, Wang JX, Rowland M, Genewein T, Kurth-Nelson Z, Pascanu R, Heess N, Veness J, Pritzel A, Sprechmann P et al (2019) Meta-learning of sequential strategies. arXiv preprint arXiv:1905.03030
Payne JW, Payne JW, Bettman JR, Johnson EJ (1993) The adaptive decision maker. Cambridge University Press, Cambridge
Pedregosa F, Varoquaux G, Gramfort A, Michel V, Thirion B, Grisel O, Blondel M, Prettenhofer P, Weiss R, Dubourg V, Vanderplas J, Passos A, Cournapeau D, Brucher M, Perrot M, Duchesnay E (2011) Scikit-learn: machine learning in python. J Mach Learn Res 12:2825–2830
Peng Z, Genewein T, Leibfried F, Braun DA (2017) An information-theoretic on-line update principle for perception-action coupling. In: Proceedings of the 2017 IEEE/RSJ international conference on intelligent robots and systems (IROS). IEEE, pp 789–796
Pereyra G, Tucker G, Chorowski J, Kaiser Ł, Hinton G (2017) Regularizing neural networks by penalizing confident output distributions. In: Proceedings of the international conference on learning representations (ICLR) 2017
Randløv J, Barto AG, Rosenstein MT (2000) Combining reinforcement learning with a local control algorithm. In: Proceedings of the international conference on machine learning
Ravi S, Larochelle H (2017) Optimization as a model for few-shot learning. In: Proceedings of the international conference on learning representations
Rothfuss J, Lee D, Clavera I, Asfour T, Abbeel P (2018) Promp: proximal meta-policy search. In: International conference on learning representations
Schach S, Gottwald S, Braun DA (2018) Quantifying motor task performance by bounded rational decision theory. Front Neurosci, 12
Schmidhuber J, Zhao J, Wiering M (1997) Shifting inductive bias with success-story algorithm, adaptive levin search, and incremental self-improvement. Mach Learn 28(1):105–130
Schulman J, Levine S, Abbeel P, Jordan M, Moritz P (2015) Trust region policy optimization. In: Proceedings of the international conference on machine learning, pp 1889–1897
Schwenker F, Kestler HA, Palm G (2001) Three learning phases for radial-basis-function networks. Neural Netw 14(4–5):439–458
Silverman BW (2018) Density estimation for statistics and data analysis. Routledge, London
Simon HA (1955) A behavioral model of rational choice. Q J Econ 69(1):99–118
Srivastava N, Hinton G, Krizhevsky A, Sutskever I, Salakhutdinov R (2014) Dropout: a simple way to prevent neural networks from overfitting. J Mach Learn Res 15(1):1929–1958
Sutton RS (1996) Generalization in reinforcement learning: successful examples using sparse coarse coding. In: Advances in neural information processing systems, pp 1038–1044
Sutton RS, Barto AG (2018) Reinforcement learning: an introduction. MIT Press, Cambridge
Sutton RS, McAllester DA, Singh SP, Mansour Y (2000) Policy gradient methods for reinforcement learning with function approximation. In: Advances in neural information processing systems, pp 1057–1063
Szegedy C, Vanhoucke V, Ioffe S, Shlens J, Wojna Z (2016) Rethinking the inception architecture for computer vision. In: 2016 IEEE conference on computer vision and pattern recognition (CVPR), pp 2818–2826
Thrun S, Pratt L (2012) Learning to learn. Springer, Berlin
Tishby N, Polani D (2011) Information theory of decisions and actions. In: Perception-action cycle: models architectures, and hardware. Springer, Berlin
Tschannen M, Djolonga J, Rubenstein PK, Gelly S, Lucic M (2020) On mutual information maximization for representation learning. In: Proceedings of the international conference on representation learning
van Hasselt HP, Guez A, Hessel M, Mnih V, Silver D (2016) Learning values across many orders of magnitude. In: Advances in neural information processing systems, pp 4287–4295
Vilalta R, Drissi Y (2002) A perspective view and survey of meta-learning. Artif Intell Rev 18(2):77–95
Vincent P, Larochelle H, Bengio Y, Manzagol P-A (2008) Extracting and composing robust features with denoising autoencoders. In: Proceedings of the 25th international conference on machine learning, pp 1096–1103. ACM
Vinyals O, Blundell C, Lillicrap T, Wierstra D et al (2016) Matching networks for one shot learning. In: Advances in neural information processing systems, pp 3630–3638
Von Neumann J, Morgenstern O (2007) Theory of games and economic behavior (commemorative edition). Princeton University Press, Princeton
Wolpert DH (2006) Information theory—the bridge connecting bounded rational game theory and statistical physics. In: Complex engineered systems. Springer, Berlin, pp 262–290
Xu R, Wunsch D (2008) Clustering, vol 10. Wiley, London
Yao H, Wei Y, Huang J, Li Z (2019) Hierarchically structured meta-learning. In: Proceedings of the international conference on machine learning, pp 7045–7054
Yoshimoto J, Nishimura M, Tokita Y, Ishii S (2005) Acrobot control by learning the switching of multiple controllers. Artif Life Robot 9(2):67–71
Yuksel SE, Wilson JN, Gader PD (2012) Twenty years of mixture of experts. IEEE Trans Neural Netw Learn Syst 23(8):1177–1193
Funding
Open Access funding enabled and organized by Projekt DEAL. This research was supported by the European Research Council, grant number ERC-StG-2015-ERC, Project ID: 678082, “BRISC: Bounded Rationality in Sensorimotor Coordination”.
Author information
Authors and Affiliations
Contributions
H.H. and D.A.B conceived the project, H.H. designed and implemented the algorithms and experiments and evaluated the results. Both authors discussed the results and wrote the manuscript. Both authors read and approved the manuscript.
Corresponding author
Ethics declarations
Conflict of interest
The authors declare no conflicts of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendices
Appendix A: Experimental Details
We implemented all experiments in Tensorflow and ran experiments working with images on a single NVIDIA Tesla V100-SXM2 32GB GPU. We implemented our experiments in TensorFlow [1], Scikit-learn [67] and OpenAI Gym [14].
1.1 Non-meta-Learning Setting
For classification (Sect. 4.1) we use a linear predictor as experts, such that the expert’s response is a linear combination of the inputs weighted by some learned parameters \(\omega \): \(y = \omega ^Tx\) and the selector is a two-layer MLP with 10 units each and tanh non-linearities.
In reinforcement learning (Sect. 4.3) the selection network is a two layer network with 32 units per layer and tanh non-linearities. Experts are two layer networks with tanh non-linearities that learn log-variance and mean of a Gaussian distribution to predict the control signal. The critic networks have the same architecture but learn directly the value function. As the action space is continuous in the interval [−1,1] we learn \(\mu \) and \(\log (\sigma )\) of a Gaussian by parameterizing the distribution with a neural network. We sample actions by re-parameterizing the distribution to \(p(a) = \mu + \sigma \epsilon \), where \(\epsilon \thicksim \mathcal {N}(0,1)\), so that the distribution is differentiable w.r.t. the network outputs (“re-parametrization trick“ introduced by Kingma and Welling [44]).
We train all networks using Adam [43] with a learning rate of \(3\cdot 10^{-4}\) with Mini-Batch sizes of 32 and sample 1024 data from each task (“Half Moon” etc.) for 10,000 episodes. We average the results presented are over 10 random seeds.
1.2 Meta-Learning Setting
The features that we use in the meta-learning setting for the selector in the different learning scenarios are depicted in Fig. 13. For regression (Sect. 6.1) we use a two layer selection network with 16 units each followed by tanh non-linearities. The experts are shallow neural networks with a single hidden layer that learn log-variance and mean of a Gaussian distribution which they use for prediction. We use the “Huber Loss” instead of MSE as it is more robust [7]. We optimize all networks using Adam [43]. We set \(\beta _1 = 25\) and \(\beta _2 = 1.25\).
For Omniglot (Sect. 6.2) we followed the design of Vinyals et al. [90] but reduce the number of blocks to one. We used a single convolutional block consisting of 32 \(3 \times 3\) filters with strided convolutions followed by a batch normalization layer and a ReLu non-linearity. The output is fed into a softmax layer giving a distribution over classes. During training we used a meta-batch size of 16. The convolutional autoencoder is a 3 layer network consisting of 16, 16, and 4 filters each with size \(3 \times 3\) with strided convolutions followed by a leaky ReLu non-linearity. The layers are mirrored by de-convolotional layers to reconstruct the image. This results in an image embedding with dimensionality 64. The selection network is a two layer network with 32 units, followed by a ReLu non-linearity, a dropout layer [79] per layer and is fed into a softmax normalization to produce a distribution over the experts. For 8 and more experts we add a third layer with 32 units. To improve generalization we add a MaxNorm regularization on the weights. We augment the dataset by rotating each image in 90, 180, and 270 degrees resulting in 80 images per class. We also normalize the images two be in (0,1) range. We evaluate our method by resetting the system to the state after training and allow for 10 gradient updates and report the final accuracy. We train all networks using Adam [43] with a learning rate of \(3 \cdot 10^{-4}\). In this experiment we set \(\beta _1 = 20.0\) and \(\beta _2 = 2.5\) for 2 and 4 experts and \(\beta _1 = 50.0\) and \(\beta _2 = 1.25\) for 8 and 16 experts.
In meta reinforcement learning (Sect. 6.3) the selector’s actor and critic net are build of RNNs with 200 hidden units each. The critic is trained to minimize the Huber loss between the prediction and the cumulative reward. The experts are two layer networks with 64 units each followed by ReLu non-linearities and used to learn the parameters of a Gaussian distribution. The critics have the same architecture (except for the output dimensionality). The actors learning rate is set to \(10^{-4}\) and the critics to \(10^{-3}\). We optimize all networks using Adam [43]. We set \(\beta _1 = 25.0\) and \(\beta _2 = 2.5\).
To evaluate MAML on 2-way N-shot omniglot dataset we used a inner learning rate of \(\alpha = 0.05\) and one inner update step per iteration for all settings. We used a single convolutional block followed by a fully connected layer with 64 units and a ReLU non-linearity. For matching networks we used the same architecture. Note, that we reduce the number of layers to make the tests comparable. Using the suggested architectures [19, 90] we achieve classification accuracy \(\ge 95\%\).
Appendix B: DKL Between Two Normal Wishart Distributions
KL divergence from Q to P is defined as \(\text {D}_\text {KL}(P \Vert Q) = \int p(m) \log \frac{p(m)}{q(m)}dx\) and a Normal-Wishart distribution is defined as
where \(\varvec{\omega } \in \mathbb {R}^D, \varvec{W} \in \mathbb {R}^{D\times D},\nu> D - 1, \lambda > 0\) are the parameters of the distribution. We optimize over \(\varvec{\omega }\) and \(\varvec{W}\), which makes part of the \(\text {D}_\text {KL}\) terms constant. So we have:
Now let
We can find the \(\text {D}_\text {KL}\) as follows:
where \(\mathbb {E}_{p(\varvec{\varLambda })}[X] = \nu \varvec{W}\) and \(\text {D}_\text {KL}\) between two Normal Distributions \(p(\varvec{\mu } \vert \varvec{\varLambda })\) and \(q(\varvec{\mu } \vert \varvec{\varLambda })\) is
so we have
The \(\text {D}_\text {KL}\) between two Wishart Distributions \(p(\varvec{\varLambda })\) and \(q(\varvec{\varLambda })\) is
where \(\varGamma _D(m)\) is the multivariate Gamma distribution and \( \psi _D(m)\) is the derivative of the log of the multivariate Gamma distribution, each parameterized by the dimensionality of \(\varvec{\mu }\) and \(\varvec{W}\) denoted by D. As we keep \(\nu \) and \(\lambda \) fixed, we can use an estimate, which is only off by constant factor C:
where \(C = \frac{D}{2} \left( \frac{\lambda _{q}}{\lambda _{p}} - \log \frac{\lambda _{q}}{\lambda _{p}} - 1 \right) + \log \frac{\varGamma _D\left( \frac{\nu _q}{2} \right) }{\varGamma _D\left( \frac{\nu _p}{2} \right) } + \tfrac{\nu _p - \nu _q }{2} \psi _D\left( \frac{\nu _p}{2}\right) \)
Appendix C: Additional Results
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Hihn, H., Braun, D.A. Specialization in Hierarchical Learning Systems. Neural Process Lett 52, 2319–2352 (2020). https://doi.org/10.1007/s11063-020-10351-3
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11063-020-10351-3