# an_explicitly_relational_neural_network_architecture__8d3c3f88.pdf An Explicitly Relational Neural Network Architecture Murray Shanahan 1 2 Kyriacos Nikiforou 1 Antonia Creswell 1 Christos Kaplanis 1 David Barrett 1 Marta Garnelo 1 With a view to bridging the gap between deep learning and symbolic AI, we present a novel endto-end neural network architecture that learns to form propositional representations with an explicitly relational structure from raw pixel data. In order to evaluate and analyse the architecture, we introduce a family of simple visual relational reasoning tasks of varying complexity. We show that the proposed architecture, when pre-trained on a curriculum of such tasks, learns to generate reusable representations that better facilitate subsequent learning on previously unseen tasks when compared to a number of baseline architectures. The workings of a successfully trained model are visualised to shed some light on how the architecture functions. 1. Introduction When humans face novel problems, they are able to draw effectively on past experience with other problems that are superficially very different, but that have similarities on a more abstract, structural level. This ability is essential for lifelong, continual learning, and confers on humans a degree of data efficiency, powers of transfer learning, and a capacity for out-of-distribution generalisation that contemporary machine learning has yet to match (Garnelo et al., 2016; Lake et al., 2017; Marcus, 2018; Smith, 2019). A case may be made that all these issues are different facets of the same underlying challenge, namely the challenge of devising systems that learn to construct general-purpose, reusable representations (Mc Carthy, 1987; Bengio et al., 2013). A representation is general-purpose and reusable to the extent that it contains information whose domain of application exceeds the context within which it was acquired. 1Deep Mind, London, UK 2Imperial College London, London, UK. Correspondence to: Murray Shanahan . Proceedings of the 37 th International Conference on Machine Learning, Online, PMLR 119, 2020. Copyright 2020 by the author(s). Representations that are general-purpose and reusable improve data efficiency because a system that already knows how to build representations relevant to a novel task (despite its novelty) doesn t have to learn that task from scratch. Ideally, a system that efficiently exploits general-purpose, reusable representations in this way should be the very same system that learned how to construct them in the first place. Moreover, in learning to solve a novel task using such representations, we should expect the system to learn further representations that are themselves general-purpose and reusable. So, with the exception of the very first representations the system learns, all learning in such a system would in effect be transfer learning, and the process of learning would be inherently cumulative, continual, and lifelong. One approach to building such a system is to take inspiration from the paradigm of classical, symbolic AI (Garnelo & Shanahan, 2019). Building on the mathematical foundations of first-order predicate calculus, a typical symbolic AI system works by applying logic-like rules of inference to language-like propositional representations whose elements are objects and relations. Thanks to their declarative character and compositional structure, these representations lend themselves naturally to generality and reusability. However, in contrast to contemporary deep learning systems, the representations deployed in classical AI are not usually learned from data but hand-crafted (Harnad, 1990). The aim of the present work is to get the best of both worlds with an endto-end differentiable neural network architecture that builds in propositional, relational priors in much the same way that a convolutional network builds in spatial and locality priors. The architecture introduced here builds on recent work with non-local network architectures that learn to discover and exploit relational information (Wang et al., 2018), notably relation nets (Santoro et al., 2017; Palm et al., 2018) and architectures based on multi-head attention (Vaswani et al., 2017; Santoro et al., 2018; Zambaldi et al., 2019). However, these architectures generate representations that lack explicit structure. There is, in general, no straightforward mapping from the parts of a representation to the usual elements of a symbolic medium such as predicate calculus: propositions, relations, and objects. To the extent that these elements are present, they are smeared across the embedding vector, which makes representations hard to interpret and makes it An Explicitly Relational Neural Network Architecture A B C means C=AB softmax(Q1KT) softmax(Q2KT) m x g n x g Figure 1. The Predi Net architecture. WK and WS are shared across heads, whereas WQ1 and WQ2 are local to each head. more difficult for downstream processing to take advantage of compositionality. Here we present an architecture, which we call a Predi Net, that learns representations whose parts map directly onto propositions, relations, and objects. To build a sound, scientific understanding of the proposed architecture, and to facilitate a detailed comparison with other architectures, the present study focuses on simple tasks requiring relatively little data and computation. We develop a family of small, simple visual datasets that can be combined into a variety of multi-task curricula and used to assess the extent to which an architecture learns representations that are general-purpose and reusable. We report the results of a number of experiments using these datasets that demonstrate the potential of an explicitly relational network architecture to improve data efficiency and generalisation, to facilitate transfer, and to learn reusable representations. The main contribution of the present paper is a novel architecture that learns to discover objects and relations in high-dimensional data, and to represent them in a form that is beneficial for downstream processing. The Predi Net architecture does not itself carry out logical inference, but rather extracts relational structure from raw data that has the potential to be exploited by subsequent processing. Here, for the purpose of evaluation, we graft a simple multi-layer perceptron output module to the Predi Net and train it on a simple set of spatial reasoning problems. The aim is to acquire a sufficient scientific understanding of the architecture and its properties in this minimalist setting before applying it to more complex problems using more sophisticated forms of downstream inference. 2. The Predi Net Architecture The idea that propositions are the building blocks of knowledge dates back to the ancient Greeks, and provides the foundation for symbolic AI, via the 19th century mathematical work of Boole and Frege (Russell & Norvig, 2009). An elementary proposition asserts that a relationship holds between a set of objects. Propositions can be combined using logical connectives (and, or, not, etc), and can participate in inference processes such as deduction. The task of the Predi Net is to (learn to) transform high-dimensional data such as images into propositional representations that are useful for downstream processing. A Predi Net module (Fig. 1) can be thought of as a pipeline comprising three stages: attention, binding, and evaluation. The attention stage selects pairs of objects of interest, the binding stage instantiates the first two arguments of a set of three-place predicates (relations) with selected object pairs, and the evaluation stage computes values for each predicate s remaining (scalar) argument such that the resulting proposition is true. More precisely, a Predi Net module comprises k heads, each of which computes j relations between pairs of objects (Fig. 1). The input to the Predi Net is a matrix, L, comprising n rows of feature vectors, where each feature vector has length m. In the present work, L is computed by a convolutional neural network (CNN). The CNN outputs a feature map consisting of n feature vectors that tile the input image. The last two elements of the feature vector are the xy co-ordinates of the associated patch in the image. So the length m of each feature vector corresponds to the number of filters in the final CNN layer plus 2 (for the co-ordinates), and the ith element of a feature vector (for i < m 2) is the output of the ith filter. For a given input L, each head h computes the same set of relations (using shared weights WS) but selects a different pair of objects, using dot-product attention based on key-query matching (Vaswani et al., 2017). Each head computes a separate pair of queries Qh 1 and Qh 2 (via W h Q1 and W h Q2). But the key space K (defined by WK) is shared, so that the set of entities that are candidates for attention is consistent across heads. The whole (flattened) image is used to generate queries, allowing attention masks An Explicitly Relational Neural Network Architecture Training object set: pentominoes Held-out object set 1: Held-out object set 2: stripes 5 possible row / column patterns 0 AAA 1 ABA 2 AAB 3 ABB 4 ABC column patterns (multi-task) match rows (c) (d) Figure 2. Relations Game object sets and tasks. (a) Example objects from the training set and held-out test sets. (b) There are five possible row / column patterns. In a multi-task setting, recognising each row pattern is a separate task. (c) Three examples tasks for the single-task setting. (d) An example target task (left) and curriculum (right) for the multi-task setting. The curriculum task ids (right) for each of the three examples (2, 4, and 3) correspond to the respective patterns in (b), and the task in each case is to confirm whether or not the column of objects in the image conforms to the designated pattern. The aim of the target task (left) is to test whether the two rows of objects have the same pattern according to (b). to depend on the image s full (non-local) content. Qh 1 = flatten(L)W h Q1 (1) Qh 2 = flatten(L)W h Q2 (2) K = LWK (3) Applying the resulting pair of attention masks directly to L yields a pair of objects Eh 1 and Eh 2 , each represented by a weighted sum of feature vectors. Eh 1 = softmax(Qh 1K )L (4) Eh 2 = softmax(Qh 2K )L (5) All j relations between Eh 1 and Eh 2 are then evaluated. There are many ways to compute a relationship between a pair of objects represented as feature vectors. We chose to compute the values of relations by taking vector differences, which has been shown to be effective in the context of relationally structured knowledge bases (Bordes et al., 2011; Socher et al., 2013). In the current architecture, Eh 1 and Eh 2 are subject to a linear mapping (via WS) into j 1D spaces, one per relation, and the resulting vector is passed through an element-wise comparator, yielding a vector of differences Dh. Dh = Eh 1 WS Eh 2 WS (6) The last two elements of Eh 1 and Eh 2 (the positions P h 1 and P h 2 , respectively) are concatenated to the vector Dh of differences to give the head s output Rh = (Dh, P h 1 , P h 2 ). Finally, the outputs of all k heads are concatenated, yielding the output of the Predi Net module, a vector R of length k(j + 4). In predicate calculus terms, the final output of a Predi Net module with k heads and j relations represents the conjunction of elementary propositions i=1 ψi(dh i , eh 1, eh 2) (7) where ψi(dh i , eh 1, eh 2) asserts that dh i is the distance between objects eh 1 and eh 2 in the 1D space defined by column i of the weight matrix WS, and the denotations of eh 1 and eh 2 are captured by the vectors Qh 1 and Qh 2 respectively, given the key-space defined by K. Equation 7 supplies a semantics for the Predi Net s final output vector R that maps each of its elements onto a welldefined logical formula, something that cannot be claimed for other architectures, such as the relation net or multi-head attention net. In the experiments reported here, only R is used for downstream processing, and this vector by itself doesn t have the logical structure described by Equation 7. However, the Predi Net module can easily be extended to deliver an additional output in explicitly propositional form, with a predicate-argument structure corresponding to the RHS of Equation 7. In the present paper, the pared-down vector form facilitates our experimental investigation, but in its explicitly propositional form, the Predi Net s output could be piped directly to (say) a Prolog interpreter (Fig.7), An Explicitly Relational Neural Network Architecture Central module MLP Task 1, task 2, Central module MLP 2: Pre-training Central module MLP Task 0 Frozen 3: Input & central nets pre-trained Central module (eg: Task 0 1: No pre-training 4: Input net pre-trained Figure 3. The four-stage experimental protocol for multi-task curriculum training. The same input module (CNN) and output module (MLP) are used for the Predi Net and all baseline architectures; only the central module varies. Task identifiers are appended to the central module s output vector. to an inductive logic programming system, to a statistical relational learning system, or indeed to another differentiable neural module. 3. Datasets and Tasks For deployment in a realistic setting, we anticipate embedding the Predi Net in a larger architecture specifically designed to exploit its advantages, while the aim of the present work is to acquire a basic understanding of its properties and behaviour in a more carefully controlled setting. Nevertheless, to demonstrate that it serves as a generic neural network module, we substituted a Predi Net for the central fully-connected layer in a strong baseline reinforcement learning agent (Fig. S20) and compared it to the original on a standard suite of 57 Atari games. The resulting network performed comparably, using 69% fewer parameters (Fig. S21), and generates interpretable attention masks (Fig. S22); cf. (Mott et al., 2019)). However, to facilitate a more in-depth scientific study, we needed small, simple datasets that allow the operation of the architecture to be examined in detail and the fundamental premises of its design to be assessed. Our experimental goals in the present paper are 1) to test the hypothesis that the Predi Net architecture learns representations that are general-purpose and reusable, and 2) insofar as this is true, to investigate why. Existing datasets for relational reasoning tasks, such as CLEVR (Johnson et al., 2017) and sort-of-CLEVR (Santoro et al., 2017), were ruled out because they include confounding complexities, such as occlusion and shadows or language input, and/or because they don t lend themselves to the fine-grained tasklevel splits we required. Consequently, we devised a new configurable family of simple classification tasks that we collectively call the Relations Game. A Relations Game task involves the presentation of an image containing a number of objects laid out on a 3 3 grid, and the aim (in most tasks) is to label the image as True or False according to whether a given relationship holds among the objects in the image. While the elementary propositions learned by the Predi Net only assert simple relationships between pairs of entities, Relations Game tasks generally involve learning compound relations involving multiple relationships among many objects. The objects in question are drawn from either a training set or one of two held-out sets (Fig. 2a). None of the shapes or colours in the training set occurs in either of the held-out sets. The training object set contains 8 uniformly coloured pentominoes and their rotations and reflections (37 shapes in all) with 25 possible colours. The first held-out object set contains 8 uniformly coloured hexominoes and their rotations and reflections (46 shapes in all) with 25 possible colours, and the second held-out object set contains only squares, but with a striped pattern of held-out colours. Each Relations Game task is tied to a given relation. Even with such a simple setup, the number of definable relations among all possible combinations of objects is astronomical (2(n+1)9 for n distinct objects), although only a few of them will make intuitive sense. For the present study, we defined a handful of intuitively meaningful relations and generated corresponding labelled datasets comprising 50% positive and 50% negative examples. A selection is shown in Fig. 2c. The between relation holds iff the image contains three objects in a line in which the outer two objects have the same shape, orientation, and colour. The occurs relation holds iff there is an object in the bottom row of three objects that has the same shape, orientation, and colour as the (single) object in the top row. The same relation holds iff the image contains two objects of the same shape, orientation, and colour. In each case, we balanced the set of negative examples to ensure that tricky images involving pairs of objects with the same colour but different shape or the same shape but different colour occur just as frequently as those with objects that differ in both colour and shape. An Explicitly Relational Neural Network Architecture 4. Experimental setup At the top level, each architecture we consider in this paper comprises 1) a single convolutional input layer (CNN), 2) a central module (which might be a Predi Net or a baseline), and 3) a small output multi-layer perceptron (MLP) (Fig. 3). A pair of xy co-ordinates is appended to each CNN feature vector, denoting its position in convolved image space and, where applicable, a one-hot task identifier is appended to the output of the central module. For most tasks, the final output of the MLP is a one-hot label denoting True or False. (The only exception is the four-label colour / shape task (Fig. 6).) The loss function used was softmax cross entropy. The Predi Net was evaluated by comparing it to several baselines: two MLP baselines (MLP1 and MLP2), a relation net baseline (Santoro et al., 2017) (RN), and a multi-head attention baseline (Vaswani et al., 2017; Zambaldi et al., 2019) (MHA). To facilitate a fair comparison, the top-level schematic is identical for the Predi Net and for all baselines (Fig. 3). All use the same input CNN architecture and the same output MLP architecture, and differ only in the central module. In MLP1, the central module is a single fully-connected layer with Re Lu activations, while in MLP2 it has two layers. In RN, the central module computes the set of all possible pairs of feature vectors, each of which is passed through a 2-layer MLP; the resulting vectors are then aggregated by taking their element-wise means to yield the output vector. Finally, MHA comprises multiple heads, each of which generates mappings from the input feature vectors to sets of keys K, queries Q, and values V , and then computes softmax(QK )V . Each head s output is a weighted sum of the resulting vectors, and the output of the MHA central module is the concatenation of all its heads outputs. The Predi Net used here comprises k = 32 heads and j = 16 relations (Fig. 1). All reported experiments were carried out using stochastic gradient descent, and all results shown are averages over 10 runs. Further experimental details are given in the Supplementary Material, which also shows results for experiments with different numbers of heads and relations, and with the Adam optimiser, all of which present qualitatively similar results. To assess the generality and reusability of the representations produced by the Predi Net, we adopted a four-stage experimental protocol wherein 1) the network is pre-trained on a curriculum of one or more tasks, 2) the weights in the input CNN and Predi Net are frozen while the weights in the output MLP are re-initialised with random values, and 3) the network is retrained on a new target task or set of tasks (Fig. 3). In step 3, only the weights in the output MLP change, so the target task can only be learned to the extent that the Predi Net delivers re-usable representations to it, representations the Predi Net has learned to produce without Table 1. Effectiveness in a single-task Relations Game setting. Relation Object set MLP1 MLP2 RN MHA Predi Net Hexominoes 96.1 0.66 96.4 0.58 73.2 5.12 94.7 13.58 100 0.0 Stripes 93.3 1.12 94.0 1.17 72.9 4.94 93.7 13.45 100 0.0 Hexominoes 98.7 0.45 98.8 0.38 70.8 1.44 89.2 10.19 99.2 0.44 Stripes 96.9 0.78 97.3 0.44 65.2 4.51 85.5 10.61 98.7 0.66 Hexominoes 88.0 1.43 94.8 3.15 61.6 1.13 88.4 16.35 98.5 0.81 Stripes 73.2 3.06 87.3 6.48 62.6 2.45 80.8 13.62 96.9 1.03 Hexominoes 81.5 2.21 84.4 3.62 55.0 0.89 54.7 0.76 95.4 0.98 Stripes 78.2 2.61 80.8 5.03 54.0 1.25 53.6 0.70 95.5 1.01 colour/shape Hexominoes 66.1 3.51 66.9 7.52 43.9 8.47 96.9 1.20 97.8 0.46 exposure to the target task. To assess this, we can compare the learning curves for the target task with and without pretraining. We expect pre-training to improve data efficiency, so we should see accuracy increasing more quickly with pre-training than without it. For evidence of transfer, and to confirm the hypothesis of reusability, we are also interested in the final performance on the target task after pre-training, given that the weights of the pre-trained input CNN and Predi Net are frozen. This measure indicates how well a network has learned to form useful representations. The more different the target task is from the pre-training curriculum, the more impressed we should be that the network is able to learn the target task. As a prelude to investigating the issues of generality and reusability, we studied the effectiveness of the Predi Net architecture in a single-task Relations Game setting. Results obtained on a selection of five tasks same , between , occurs , xoccurs , and colour / shape are summarised in Table 1. The first three tasks are as described in Fig. 2. The xoccurs relation is similar to occurs. It holds iff the object in the top row occurs in the bottom row and the other two objects in the bottom row are different. The colour / shape task involves four labels, rather than the usual two: same-shape / same-colour; different-colour / same-shape; same-colour / different shape; different-colour / different shape. In the dataset for this task, each image contains two objects randomly placed, and one of the four labels must be assigned appropriately. Table 1 shows the accuracy obtained by each of the five architectures after 100,000 batches when tested on the two held-out object sets. The Predi Net is the only architecture that achieves over 90% accuracy on all tasks with both held-out object sets after 100,000 batches. On the xoccurs task, the Predi Net outperforms the baselines by more than 10%, and on the colour / shape task (where chance is 25%), it out-performs all the baselines except MHA by 25% or more. Next, using the protocol outlined in Fig. 3, we compared the Predi Net s ability to learn re-usable representations with An Explicitly Relational Neural Network Architecture MLP1 MLP2 RN MHA Predi Net Pre-training task(s) No pre-training Input & central nets pre-trained Input net pre-trained Figure 4. Multi-task curriculum training. The target tasks are three column patterns (AAB, ABA, and ABB) and the sole curriculum task is the between relation. The green line indicates the reusability of the learned representations. The Predi Net out-performs all four of the baselines. each of the baselines. We looked at a number of combinations of target tasks and pre-training curriculum tasks. Fig. 4 depicts our findings for one these combinations in detail, specifically three target tasks corresponding to three of the five possible column patterns (ABA, AAB, and ABB (Fig. 2d)), and a pre-training curriculum comprising the single between task. The plots present learning curves for each of the five architectures at each of the four stages of the experimental protocol. In all cases, accuracy is shown for the stripes held-out object set (not the training set). Of particular interest are the (green) curves corresponding to Stage 3 of the experimental protocol. These show how well each architecture learns the target task(s) after the central module has been pre-trained on the curriculum task(s) and its weights are frozen. The Predi Net learns faster than any of the baselines, and is the only one to achieve an accuracy of 90%. The rapid reusability of the representations learned by both the MHA baseline and the Predi Net is noteworthy because the between relation by itself seems an unpromising curriculum for subsequently learning the AAB and ABB column patterns. As the (red) curve for Stage 4 of the protocol shows, the reusability of the Predi Net s representations cannot be accounted for by the pre-training of the input CNN alone. Fig. 5 shows a larger range of target task / curriculum task combinations, concentrating exclusively on the Stage 3 learning curves. Here a more complete picture emerges. In both Fig. 5a and Fig. 5d the target task is match rows (Fig. 2d), but they differ in their pre-training curricula. The curriculum for Fig. 5d is three of the five row patterns (ABA, AAB, and ABB). This is the only case where the Predi Net does not learn representations that are more useful for the target task than those of all the baselines, outperforming only two of the four. However, when the curriculum is the three analogous column patterns rather than row patterns, the performance of all four baselines collapses to chance, while the Predi Net does well, attaining similar performance as for the row-based curriculum (Fig. 5a). This suggests the Predi Net is able to learn representations that are orientation invariant, which aids transfer. This hypothesis is supported by Fig. 5e, where the target tasks are all five row patterns, while the curriculum is all five column patterns. None of the baselines is able to learn reusable representations in this context; all remain at chance, whereas the Predi Net achieves 85% accuracy. To better understand the operation of the Predi Net, we carried out a number of visualisations. One way to find out what the Predi Net s heads learn to attend is to submit images to a trained network and, for each head h, apply the two attention masks softmax(Qh 1K ) and softmax(Qh 2K ) to each of the n feature vectors in the convolved image L. The resulting matrix can then be plotted as a heat map to show how attention is distrubuted over the image. We did this for a number of networks trained in the single-task setting. Fig. 6a shows two examples, and the Supplementary Material contains a more extensive selection. As we might expect, most of the attention focuses on the centres of single objects, and many of the heads pick out pairs of distinct objects in various combinations. But some heads attend to halves or corners of objects. Although most attention is focal, whether directed at object centres or object parts, some heads exhibit diffuse attention, which is possible thanks to the soft key-query matching mechanism. So the Predi Net can (but isn t forced to) treat the background as a single entity, or to treat an identical pair of objects as a single entity. To gain some insight into how the Predi Net encodes relations, we carried out principal component analysis (PCA) An Explicitly Relational Neural Network Architecture d e Target Pre-training a match rows 3 colum patterns b 5 column patterns between c 3 column patterns between d match rows 3 row patterns e 5 row patterns 5 column patterns Figure 5. Reusability of representations learned with a variety of target and pre-training tasks. The Predi Net (purple line) out-performs the four baseline on three out of the four combinations. on each head of the central module s output vectors for a number of trained networks, again in the single-task setting (Fig. 6b). We chose the four-label colour / shape task to train on, and mapped 10,000 example images onto the first two principal components, colouring each with their ground-truth label. We found that, for some heads, differences in colour and shape appear to align along separate axes (Fig. 6b). This contrasts with the MHA baseline, whose heads don t seem to individually cluster the labels in a meaningful way. For the other baselines, which lack the multi-head organisation of the Predi Net and the MHA network, the only option is to carry out PCA on the whole output vector of the central module. Doing this, however, does not produce interpretable results for any of the architectures (Fig. S8). We also identified the heads in the Predi Net that attended to both objects in the image and found that they overlapped almost entirely with those that meaningfully clustered the labels (Fig. S10). Finally, still using the shape / colour task , we carried out an ablation study, which showed that the Predi Net is significantly more robust than the MHA network to pruning a random subset of heads at test time. Moreover, if pruned to leave only those heads that attended to the two objects, the performance of the full network could be captured with just a handful of heads (Fig. 6c). Taken together, these results are suggestive of something we might term relational disentangling in the Predi Net. Finally, to flesh out the claim that the Predi Net generates explicitly relational representations according to the semantics of Equation 7, we extended the Predi Net module to generate an additional output in the form of a Prolog program (Fig. 7). This involves assigning symbolic identifiers 1) to each of the Predi Net s j relations, and 2) to every object picked out by its k heads via the attention masks they compute. Then the corresponding j k propositions can be enumerated in Pro- log syntax. Assigning symbolic identifiers to the relations is trivial. But because attention masks can differ slightly even when they ostensibly pick out the same region of the input image, it s necessary to cluster them before assigning symbolic identifiers to the corresponding objects. We used mean shift clustering for this. Fig. 7 presents a sample of the Predi Net s output in Prolog form, along with an example of deductive inference carried out with this program. The example shown is not intended to be especially meaningful; without further analysis, we lack any intuitive understanding of the relations the Predi Net has discovered. But it demonstrates that the representations the Predi Net produces can be understood in predicate calculus terms, and that symbolic deductive inference is one way (though not the only way) in which they might be deployed downstream. 6. Related Work The need for good representations has long been recognised in AI (Mc Carthy, 1987; Russell & Norvig, 2009), and is fundamental to deep learning (Bengio et al., 2013). The importance of reusability and abstraction, especially in the context of transfer, is emphasised by Bengio, et al. (Bengio et al., 2013), who argue for feature sets that are invariant to the irrelevant features and disentangle the relevant features . Our work here shares this motivation. Other work has looked at learning representations that are disentangled at the feature level (Higgins et al., 2017a; 2018). The novelty of the Predi Net is to incorporate architectural priors that favour representations that are disentangled at the relational and propositional levels. Previous work with relation nets and multi-head attention nets has shown how non-local information can be extracted from raw pixel data and used to solve tasks that require relational reasoning. (Santoro et al., 2017; Palm et al., 2018; Santoro et al., 2018; Zambaldi et al., An Explicitly Relational Neural Network Architecture (a) Attention Raw image Raw image Predi Net Training set (pentominoes) Predi Net Test set (hexominoes) MHA Training set (pentominoes) Same colour Same shape Different colour Same shape Same colour Different shape Different colour Different shape Figure 6. (a) Attention heat maps for the first four heads of a trained Predi Net. Left: trained on the same task. Right: trained on the occurs task. (b) Principal component analysis. Left: PCA on the output of a selected head for a Predi Net trained on the colour / shape task for pentominoes images (training set). Centre: The same Predi Net applied to hexominoes (held-out test set). Right: PCA applied to a representative head of the MHA baseline with pentominoes (training set). (c) Ablation study. Accuracy for Predi Net and MHA on the colour / shape task when random subsets of the heads are used at test time. Predi Net* only samples from heads that attend to the two objects. holds(r0,V,X,Y) :- r0(V,X,Y). holds(r1,V,X,Y) :- r1(V,X,Y). holds(r2,V,X,Y) :- r2(V,X,Y). holds(r3,V,X,Y) :- r3(V,X,Y). holds(r4,V,X,Y) :- r4(V,X,Y). holds(r5,V,X,Y) :- r5(V,X,Y). holds(r6,V,X,Y) :- r6(V,X,Y). holds(r7,V,X,Y) :- r7(V,X,Y). small(V) :- V > -0.2, V < 0.2. Hand-written Prolog code ? holds(R,V,ob_2,X), small(V). Prolog query Predi Net output in Prolog form Prolog answers Input image Predi Net attention masks and object ids r0(-0.11, ob_0, ob_1). r1(-0.34, ob_0, ob_1). r2(-0.50, ob_0, ob_1). r3(-0.54, ob_0, ob_1). r4(0.80, ob_0, ob_1). r3(0.06, ob_2, ob_0). r4(-0.53, ob_2, ob_0). r5(-0.45, ob_2, ob_0). r6(1.00, ob_2, ob_0). r7(-0.59, ob_2, ob_0). R = r0, V = -0.15, X = ob_4 R = r0, V = -0.07, X = ob_0 R = r3, V = 0.06, X = ob_0 R = r4, V = -0.1, X = ob_4 ob_0 ob_1 ob_2 ob_3 (a) (b) (e) Figure 7. Predi Net output in propositional form. (a) A small Predi Net (8 heads, 8 relations) trained on the between task is given an image. (b) Mean shift clustering is applied to the set of all attention masks computed by the heads. Each of the resulting 6 clusters is assigned a symbolic identifier. (c) Each relation is also given a symbolic identifier, and all 64 propositions computed by the Predi Net are enumerated in Prolog syntax, in accordance with Equation 7. (A subset is shown.) (d) The results can be combined with further hand-written Prolog clauses. (Upper-case letters denote variables, while constants start with lower-case letters.) (e) Prolog queries can then be submitted. Here we are asking which relations r hold with a small value v between ob 2 and any other object x. (f) The query yields four answers. An Explicitly Relational Neural Network Architecture 2019) But unlike the Predi Net, these networks don t produce representations with an explicitly relational, propositional structure. By addressing the problem of acquiring structured representations, the Predi Net complements another thread of related work, which is concerned with learning how to carry out inference with structured representations, but which assumes the job of acquiring those representations is done elsewhere (Getoor & Taskar, 2007; Battaglia et al., 2016; Rockt aschel & Riedel, 2017; Evans & Grefenstette, 2018; Dong et al., 2019). In part, the present work is motivated by the conviction that curricula will be essential to lifelong, continual learning in a future generation of RL agents if they are to exhibit more general intelligence, just as they are for human children. Curricular pre-training has a decade-long pedigree in deep learning (Bengio et al., 2009). Closely related to curriculum learning is the topic of transfer (Bengio, 2012), a hallmark of general intelligence and the subject of much recent attention (Higgins et al., 2017b; Kansky et al., 2017; Schwarz et al., 2018). The Predi Net exemplifies a different (though not incompatible) viewpoint on curriculum learning and transfer from that usually found in the neural network literature. Rather than (or as well as) a means to guide the network, step by step, into a favourable portion of weight space, curriculum learning is here viewed in terms of the incremental accumulation of propositional knowledge. This necessitates the development of a different style of architecture, one that supports the acquisition of propositional, relational representations, which also naturally subserve transfer. Asai, whose paper was published while the present work was in progress, describes an architecture with some similarities to the Predi Net, but also some notable differences (Asai, 2019). For example, Asai s architecture assumes an input representation in symbolic form where the objects have already been segmented. By contrast, in the present architecture, the input CNN and the Predi Net s dot-product attention mechanism together learn what constitutes an object. 7. Conclusion and Further Work We have presented a neural network architecture capable, in principle, of supporting predicate logic s powers of abstraction without compromising the ideal of end-to-end learning, where the network itself discovers objects and relations in the raw data and thus avoids the symbol grounding problem entailed by symbolic AI s practice of hand-crafting representations (Harnad, 1990). Our empirical results support the view that a network architecturally constrained to learn explicitly propositional, relational representations will have beneficial data efficiency, generalisation, and transfer properties. Although, the present experiments don t use the fully propositional version of the Predi Net output, the concate- nated vector form inherits many of its beneficial properties, notably a degree of compositionality. In particular, one important respect in which the Predi Net differs from other network architectures is the extent to which it canalises information flow; at the core of the network, information is organised into small chunks which are processed in parallel channels that limit the ways the chunks can interact. We believe this pressures the network to learn representations where each separate chunk of information (such as a single value in the vector R ) has independent meaning and utility. (We see evidence of this in the relational disentanglement of Fig. 6.) The result should be a representation whose component parts are amenable to recombination, and therefore re-use in a novel task. But the findings reported here are just the first foray into unexplored architectural territory, and much work needs to be done to gauge the architecture s full potential. The focus of the present paper is the acquisition of propositional representations rather than their use. But thanks to the structural priors of its architecture, representations generated by a Predi Net module have a natural semantics compatible with predicate calculus (Equation 7), which makes them an ideal medium for logic-like downstream processes such as rule-based deduction, causal or counterfactual reasoning, and inference to the best explanation (abduction). One approach here would be to stack Predi Net modules and / or make them recurrent, enabling them to carry out the sort of iterated, sequential computations required for such processes (Palm et al., 2018; Dehghani et al., 2019). Another worthwhile direction for further research would be to develop reinforcement learning (RL) agents using the Predi Net architecture. One form of inference of particular interest in this context is model-based prediction, which can be used to endow an RL agent with look-ahead and planning abilities (Racani ere et al., 2017; Zambaldi et al., 2019). Our expectation is that RL agents in which explicitly propositional, relational representations underpin these capacities will manifest more of the beneficial data efficiency, generalisation, and transfer properties suggested by the present results. As a stepping stone to such RL agents, the Relations Game family of datasets could be extended into the temporal domain, and multi-task curricula developed to encourage the acquisition of temporal, as well as spatial, abstractions. Software and Data https://github.com/deepmind/ deepmind-research/tree/master/Predi Net. Acknowledgements Thanks to Irina Higgins, Neil Rabinowitz, David Reichert, David Raposo, Adam Santoro, and Daniel Zoran. An Explicitly Relational Neural Network Architecture Asai, M. Unsupervised grounding of plannable first-order logic representation from images. In International Conference on Automated Planning and Scheduling, 2019. Battaglia, P. W., Pascanu, R., Lai, M., Rezende, D. J., and Kavukcuoglu, K. Interaction networks for learning about objects, relations and physics. In Advances in Neural Information Processing Systems, pp. 4502 4510, 2016. Bengio, Y. Deep learning of representations for unsupervised and transfer learning. In Proc. ICML Workshop on Unsupervised and Transfer Learning, volume 27, pp. 17 37, 2012. Bengio, Y., Louradour, J., Collobert, R., and Weston, J. Curriculum learning. In Proc. 26th International Conference on Machine Learning, pp. 41 48, 2009. Bengio, Y., Courville, A., and Vincent, P. Representation learning: A review and new perspectives. IEEE transactions on pattern analysis and machine intelligence, 35(8): 1798 1828, 2013. Bordes, A., Weston, J., Collobert, R., and Bengio, Y. Learning structured embeddings of knowledge bases. In Proc. AAAI, pp. 301 306, 2011. Dehghani, M., Gouws, S., Vinyals, O., Uszkoreit, J., and Kaiser, L. Universal transformers. In Proc. International Conference on Learning Representations, 2019. Dong, H., Mao, J., Lin, T., Wang, C., Li, L., and Zhou, D. Neural logic machines. In International Conference on Learning Representations, 2019. Espeholt, L., Soyer, H., Munos, R., Simonyan, K., Mnih, V., Ward, T., Doron, Y., Firoiu, V., Harley, T., Dunning, I., Legg, S., and Kavukcuoglu, K. IMPALA: Scalable distributed deep-RL with importance weighted actor-learner architectures. In Proceedings of the 35th International Conference on Machine Learning, pp. 1407 1416, 2018. Evans, R. and Grefenstette, E. Learning explanatory rules from noisy data. Journal of Artificial Intelligence Research, 61:1 64, 2018. Garnelo, M. and Shanahan, M. Reconciling deep learning with symbolic artificial intelligence: representing objects and relations. Current Opinion in Behavioral Sciences, 29:17 23, 2019. Garnelo, M., Arulkumaran, K., and Shanahan, M. Towards deep symbolic reinforcement learning. ar Xiv preprint: 1609.05518, 2016. Getoor, L. and Taskar, B. (eds.). Introduction to Statistical Relational Learning. MIT Press, 2007. Harnad, S. The symbol grounding problem. Physica D: Nonlinear Phenomena, 42(1-3):335 346, 1990. Higgins, I., Matthey, L., Pal, A., Burgess, C., Glorot, X., Botvinick, M., Mohamed, S., and Lerchner, A. betavae: Learning basic visual concepts with a constrained variational framework. In International Conference on Learning Representations, 2017a. Higgins, I., Pal, A., Rusu, A., Matthey, L., Burgess, C., Pritzel, A., Botvinick, M., Blundell, C., and Lerchner, A. DARLA: improving zero-shot transfer in reinforcement learning. In Proc. 34th International Conference on Machine Learning, pp. 1480 1490, 2017b. Higgins, I., Sonnerat, N., Matthey, L., Pal, A., Burgess, C. P., Bosnjak, M., Shanahan, M., Botvinick, M., Hassabis, D., and Lerchner, A. Scan: Learning hierarchical compositional visual concepts. In International Conference on Learning Representations, 2018. Johnson, J., Hariharan, B., van der Maaten, L., Fei-Fei, L., Zitnick, C. L., and Girshick, R. B. Clevr: A diagnostic dataset for compositional language and elementary visual reasoning. In Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1988 1997, 2017. Kansky, K., Silver, T., M ely, D. A., Eldawy, M., L azaro Gredilla, M., Lou, X., Dorfman, N., Sidor, S., Phoenix, S., and George, D. Schema networks: Zero-shot transfer with a generative causal model of intuitive physics. In Proc. 34th International Conference on Machine Learning, volume 70, pp. 1809 1818, 2017. Lake, B. M., Ullman, T. D., Tenenbaum, J. B., and Gershman, S. J. Building machines that learn and think like people. Behavioral and Brain Sciences, 40:e253, 2017. Marcus, G. Deep learning: a critical appraisal. ar Xiv preprint ar Xiv:1801.00631, 2018. Mc Carthy, J. Generality in artificial intelligence. Communications of the ACM, 30(12):1030 1035, 1987. Mott, A., Zoran, D., Chrzanowski, M., Wierstra, D., and Rezende, D. Towards interpretable reinforcement learning using attention augmented agents. In Advances in Neural Information Processing Systems 32, pp. 12329 12338, 2019. Palm, R. B., Paquet, U., and Winter, O. Recurrent relational networks. In Advances in Neural Information Processing Systems, pp. 3368 3378, 2018. Racani ere, S., Weber, T., Reichert, D. P., Buesing, L., Guez, A., Rezende, D., Badia, A. P., Vinyals, O., Heess, N., Li, Y., Pascanu, R., Battaglia, P., Hassabis, D., Silver, D., and An Explicitly Relational Neural Network Architecture Wierstra, D. Imagination-augmented agents for deep reinforcement learning. In Advances in Neural Information Processing Systems, pp. 5694 5705, 2017. Rockt aschel, T. and Riedel, S. End-to-end differentiable proving. In Advances in Neural Information Processing Systems, pp. 3788 3800, 2017. Russell, S. and Norvig, P. Artificial Intelligence: A Modern Approach. Prentice Hall Press, 3rd edition, 2009. Santoro, A., Raposo, D., Barrett, D. G., Malinowski, M., Pascanu, R., Battaglia, P., and Lillicrap, T. A simple neural network module for relational reasoning. In Advances in Neural Information Processing Systems, pp. 4974 4983, 2017. Santoro, A., Faulkner, R., Raposo, D., Jack, R., Chrzanowski, M., Weber, T., Vinyals, O., Pascanu, R., and Lillicrap, T. Relational recurrent neural networks. In Advances in Neural Information Processing Systems, pp. 7299 7310, 2018. Schwarz, J., Czarnecki, W., Luketina, J., Grabska Barwinska, A., Teh, Y. W., Pascanu, R., and Hadsell, R. Progress & compress: A scalable framework for continual learning. In Proc. 35th International Conference on Machine Learning, volume 80, pp. 4528 4537, 2018. Smith, B. C. The Promise of Artificial Intelligence: Reckoning and Judgment. MIT Press, 2019. Socher, R., Chen, D., Manning, C., and Ng, A. Reasoning with neural tensor networks for knowledge base completion. In Advances in Neural Information Processing Systems, pp. 926 934, 2013. Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, Ł., and Polosukhin, I. Attention is all you need. In Advances in Neural Information Processing Systems, pp. 5998 6008, 2017. Wang, X., Girshick, R., Gupta, A., and He, K. Non-local neural networks. In Proc. IEEE Conference on Computer Vision and Pattern Recognition, 2018. Zambaldi, V., Raposo, D., Santoro, A., Bapst, V., Li, Y., Babuschkin, I., Tuyls, K., Reichert, D., Lillicrap, T., Lockhart, E., Shanahan, M., Langston, V., Pascanu, R., Botvinick, M., Vinyals, O., and Battaglia, P. Deep reinforcement learning with relational inductive biases. In International Conference on Learning Representations, 2019.