30
Neural Consciousness Flow Xiaoran Xu 1 , Wei Feng 1 , Zhiqing Sun 2 , Zhi-Hong Deng 2 1 Hulu LLC, Beijing, China {xiaoran.xu, wei.feng}@hulu.com 2 Peking University, Beijing, China {1500012783, zhdeng}@pku.edu.cn Abstract The ability of reasoning beyond data fitting is substantial to deep learning systems in order to make a leap forward towards artificial general intelligence. A lot of efforts have been made to model neural-based reasoning as an iterative decision- making process based on recurrent networks and reinforcement learning. Instead, inspired by the consciousness prior proposed by Yoshua Bengio [1], we explore reasoning with the notion of attentive awareness from a cognitive perspective, and formulate it in the form of attentive message passing on graphs, called neural consciousness flow (NeuCFlow). Aiming to bridge the gap between deep learning systems and reasoning, we propose an attentive computation framework with a three-layer architecture, which consists of an unconsciousness flow layer,a consciousness flow layer, and an attention flow layer. We implement the NeuCFlow model with graph neural networks (GNNs) and conditional transition matrices. Our attentive computation greatly reduces the complexity of vanilla GNN-based methods, capable of running on large-scale graphs. We validate our model for knowledge graph reasoning by solving a series of knowledge base completion (KBC) tasks. The experimental results show NeuCFlow significantly outperforms previous state-of-the-art KBC methods, including the embedding-based and the path-based. The reproducible code can be found by the link 1 below. 1 Introduction To discover the mystery of consciousness, several competing theories [2, 3, 4, 5] have been proposed by neuroscientists. Despite their contradictory claims, they share a common notion that consciousness is a cognitive state of experiencing one’s own existence, i.e. the state of awareness. Here, we do not refer to those elusive and mysterious meanings attributed to the word "consciousness". Instead, we focus on the basic idea, awareness or attentive awareness, to derive a neural network-based attentive computation framework on graphs, attempting to mimic the phenomenon of consciousness to some extent. The first work to bring the idea of attentive awareness into deep learning models, as far as we know, is Yoshua Bengio’s consciousness prior [1]. He points out the process of disentangling higher-level abstract factors from full underlying representation and forming a low-dimensional combination of a few selected factors or concepts to constitute a conscious thought. Bengio emphasizes the role of attention mechanism in expressing awareness, which helps focus on a few elements of state representation at a given moment and combining them to make a statement, an action or policy. Two recurrent neural networks (RNNs), the representation RNN and the consciousness RNN, are used to summarize the current and recent past information and encode two types of state, the unconscious state denoted by a full high-dimensional vector before applying attention, and the conscious state by a derived low-dimensional vector after applying attention. 1 https://github.com/netpaladinx/NeuCFlow Preprint. Under review.

Neural Consciousness Flow - GitHub Pages

  • Upload
    others

  • View
    8

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Neural Consciousness Flow - GitHub Pages

Neural Consciousness Flow

Xiaoran Xu1, Wei Feng1, Zhiqing Sun2, Zhi-Hong Deng21Hulu LLC, Beijing, China

{xiaoran.xu, wei.feng}@hulu.com2Peking University, Beijing, China

{1500012783, zhdeng}@pku.edu.cn

Abstract

The ability of reasoning beyond data fitting is substantial to deep learning systemsin order to make a leap forward towards artificial general intelligence. A lot ofefforts have been made to model neural-based reasoning as an iterative decision-making process based on recurrent networks and reinforcement learning. Instead,inspired by the consciousness prior proposed by Yoshua Bengio [1], we explorereasoning with the notion of attentive awareness from a cognitive perspective, andformulate it in the form of attentive message passing on graphs, called neuralconsciousness flow (NeuCFlow). Aiming to bridge the gap between deep learningsystems and reasoning, we propose an attentive computation framework witha three-layer architecture, which consists of an unconsciousness flow layer, aconsciousness flow layer, and an attention flow layer. We implement the NeuCFlowmodel with graph neural networks (GNNs) and conditional transition matrices.Our attentive computation greatly reduces the complexity of vanilla GNN-basedmethods, capable of running on large-scale graphs. We validate our model forknowledge graph reasoning by solving a series of knowledge base completion(KBC) tasks. The experimental results show NeuCFlow significantly outperformsprevious state-of-the-art KBC methods, including the embedding-based and thepath-based. The reproducible code can be found by the link1 below.

1 Introduction

To discover the mystery of consciousness, several competing theories [2, 3, 4, 5] have been proposedby neuroscientists. Despite their contradictory claims, they share a common notion that consciousnessis a cognitive state of experiencing one’s own existence, i.e. the state of awareness. Here, we do notrefer to those elusive and mysterious meanings attributed to the word "consciousness". Instead, wefocus on the basic idea, awareness or attentive awareness, to derive a neural network-based attentivecomputation framework on graphs, attempting to mimic the phenomenon of consciousness to someextent.

The first work to bring the idea of attentive awareness into deep learning models, as far as we know,is Yoshua Bengio’s consciousness prior [1]. He points out the process of disentangling higher-levelabstract factors from full underlying representation and forming a low-dimensional combinationof a few selected factors or concepts to constitute a conscious thought. Bengio emphasizes therole of attention mechanism in expressing awareness, which helps focus on a few elements of staterepresentation at a given moment and combining them to make a statement, an action or policy. Tworecurrent neural networks (RNNs), the representation RNN and the consciousness RNN, are used tosummarize the current and recent past information and encode two types of state, the unconsciousstate denoted by a full high-dimensional vector before applying attention, and the conscious state bya derived low-dimensional vector after applying attention.

1https://github.com/netpaladinx/NeuCFlow

Preprint. Under review.

Page 2: Neural Consciousness Flow - GitHub Pages

Inspired by the consciousness prior, we develop an attentive message passing mechanism. We modelquery-dependent states as motivation to drive iterative sparse access to an underlying large graph andnavigate information flow via a few nodes to reach a target. Instead of using RNNs, we use two GNNs[6, 7] with node state representations. Nodes sense nearby topological structures by exchangingmessages with neighbors, and then use aggregated information to update their states. However, thestandard message passing runs globally and uniformly. Messages gathered by a node can come frompossibly everywhere and get further entangled by aggregation operations. Therefore, we need to drawa query-dependent or context-aware local subgraph to guide message passing. Nodes within such asubgraph are densely connected, forming a community to further exchange and share information,reaching some resonance, and making subsequent decisions collectively to expand the subgraphand navigate information flow. To support such attentive information flow, we design an attentionflow layer above two GNNs. One GNN uses the standard message passing over a full graph, calledunconsciousness flow layer, while the other GNN runs on a subgraph built by attention flow, calledconsciousness flow layer. These three flow layers constitute our attentive computation framework.

We realize the connection between attentive awareness and reasoning. A reasoning process isunderstood as a sequence of obvious or interpretable steps, either deductive, inductive, or abductive,to derive a less obvious conclusion. From the aspect of awareness, reasoning requires computationto be self-attentive or self-aware during processing in a way different from fitting by a black box.Therefore, interpretability must be one of the properties of reasoning. Taking KBC tasks as anexample, many embedding-based models [8, 9, 10, 11, 12, 13] can do a really good job in linkprediction, but lacking interpretation makes it hard to argue for their reasoning ability. People whoaim at knowledge graph reasoning mainly focus on the path-based models using RL [14, 15, 16, 17]or logic-like methods [18, 19] to explicitly model a reasoning process to provide interpretationsbeyond predictions. Here, instead, we apply a flow-based attention mechanism, proposed in [20],as an alternative to RL for learning composition structure. In a manner of flowing, attention canpropagate to cover a broader scope and increase the chance to hit a target. It maintains an end-to-enddifferentiable style, contrary to the way RL agents learn to choose a discrete action.

Other crucial properties of reasoning include relational inductive biases and iterative processing.Therefore, GNNs [6, 7] are a better choice compared to RNNs for encoding structured knowledgeexplicitly. Compared with the majority of previous GNN literature, focusing on the computation side,making neural-based architectures more composable and complex, we put a cognitive insight intoit under the notion of attentive awareness. Specifically, we design an attention flow layer to chainattention operations directly with transition matrices, parallel to the message-passing pipeline to getless entangled with representation computation. This gives our model the ability to select edges stepby step during computation and attend to a query-dependent subgraph, making a sharper predictiondue to the disentanglement. These extracted subgraphs can reduce the computation cost greatly. Inpractice, we find our model can be applied to very large graphs with millions of nodes, such as theYAGO3-10 dataset, even running on a single laptop.

Our contributions are three-fold: (1) We propose an attentive computation framework on graphs,combining GNNs’ representation power with explicit reasoning pattern, motivated by the cognitivenotion of attentive awareness. (2) We exploit query-dependent subgraph structure, extracted byan attention flow mechanism, to address two shortcomings of most GNN implementations: thecomplexity and the non-context-aware aggregation schema. (3) We design a specific architecture forKBC tasks and demonstrate our model’s strong reasoning capability compared to the state of the art,showing that a compact query-dependent subgraph is better than a path as a reasoning pattern.

2 Related Work

KBC and knowledge graph reasoning. Early work for KBC, including TransE [8] and its analogues[21, 22, 23], DistMult [9], ConvE [10] and ComplEx [11], focuses on learning embeddings of entitiesand relations. Some recent work of this line [12, 13] achieves high accuracy, yet unable to explicitlydeal with compositional relationships that is crucial for reasoning. Another line aims to learn inferencepaths [14, 24, 25, 26, 27, 28] for knowledge graph reasoning, such as DeepPath [15], MINERVA[16], and M-Walk [17], using RL to learn multi-hop relational paths over a graph towards a targetgiven a query. However, these approaches, based on policy gradients or Monte Carlo tree search,often suffer from low sample efficiency and sparse rewards, requiring a large number of rollouts or

2

Page 3: Neural Consciousness Flow - GitHub Pages

Figure 1: Illustration for the three-layer attentive computation framework. The bottom is a unifiedunconsciousness flow layer, the middle contains small disentangled subgraphs to run attentive messagepassing separately, constituting a consciousness flow layer, and the top is an attention flow layer forextracting local subgraph structures.

running many simulations, and also the sophisticated reward function design. Other efforts includelearning soft logical rules [18, 19] or compostional programs [29] to reason over knowledge graphs.

Relational reasoning by GNNs and attention mechanisms. Relational reasoning is regardedas the key component of humans’ capacity for combinatorial generalization, taking the form ofentity- and relation-centric organization to reason about the composition structure of the world[30, 31, 32, 33, 34]. A multitude of recent implementations [7] encode relational inductive biasesinto neural networks to exploit graph-structured representation, including graph convolution networks(GCNs) [35, 36, 37, 38, 39, 40, 41, 42] and graph neural networks [6, 43, 44, 45, 46], and overcomethe difficulty to achieve relational reasoning for traditional deep learning models. These approacheshave been widely applied to accomplishing real-world reasoning tasks (such as physical reasoning[45, 47, 48, 49, 50, 51], visual reasoning [44, 51, 52, 53, 54], textual reasoning [44, 55, 56], knowledgegraph reasoning [41, 57, 58], multiagent relationship reasoning [59, 60], and chemical reasoning [46]),solving algorithmic problems (such as program verification [43, 61], combinatorial optimization[62, 63, 64], state transitions [65], and bollean satisfiability [66]), or facilitating reinforcementlearning with the structured reasoning or planning ability [67, 68, 49, 50, 69, 70, 71]. Variants ofGNN architectures have been developed with different focuses. Relation networks [44] use a simplebut effective neural module to equip deep learning models with the relational reasoning ability, and itsrecurrent versions [55, 56] do multi-step relational inference for long periods; Interaction networks[45] provide a general-purpose learnable physics engine, and two of its variants are visual interactionnetworks [51] learning directly from raw visual data, and vertex attention interaction networks [60]with an attention mechanism; Message passing neural networks [46] unify various GCNs and GCNsinto a general message passing formalism by analogy to the one in graphical models.

Despite the strong representation power of GNNs, recent work points out its drawbacks that limit itscapability. The vanilla message passing or neighborhood aggregation schema cannot adapt to stronglydiverse local subgraph structure, causing performance degeneration when applying a deeper versionor running more iterations [72], since a walk of more steps might drift away from local neighborhoodwith information washed out via averaging. It is suggested that covariance rather than invarianceto permutations of nodes and edges is preferable [73], since being fully invariant by summing oraveraging messages may worsen the representation power, lacking steerability. In this context, ourmodel expresses permutation invariance under a constrained compositional transformation accordingto the group of possible permutations within each extracted query-dependent subgraph rather than theunderlying full graph. Another drawback is the heavy computation complexity. GNNs are notoriousfor its poor scalability due to its quadratic complexity in the number of nodes when graphs are fullyconnected. Even scaling linearly with the number of edges by exploiting structure sparsity can stillcause trouble on very large graphs, making selective or attentive computation on graphs so desirable.

Neighborhood attention operation can alleviate some limitation on GNNs’ representation power byspecifying different weights to different nodes or nodes’ features [74, 60, 53, 75]. These approachesoften use multi-head self-attention to focus on specific interactions with neighbors when aggregatingmessages, inspired by [76, 77, 78] originally for capturing long range dependencies. We notice thatmost graph-based attention mechanisms attend over neighborhood in a single-hop fashion, and [60]claims that the multi-hop architecture does not help in experiments, though they expect multiple hopsto offer the potential to model high-order interaction. However, a flow-based design of attention in[20] shows a promising way to characterize long distance dependencies over graphs, breaking theisolation of attention operations and stringing them in chronological order by transition matrices, likethe spread of a random walk, parallel to the message-passing pipeline.

3

Page 4: Neural Consciousness Flow - GitHub Pages

Query:(head, rel, ?)

AggrOp

...

Pooling or returning the last

... AggrOp

...

AttdOp

...

Unconsciousness Flow

Consciousness Flow

Sparse Transition

OneBatch

At tent i on Flow

...

Figure 2: The neural consciousness flow architecture.

It is natural to extend relational reasoning to graph structure inference or graph generation, such asreasoning about a latent interaction graph explicitly to acquire knowledge of observed dynamics [48],or learning generative models of graphs [79, 80, 81, 82]. Soft plus hard attention mechanisms may bea better alternative to probabilistic models that is hard to train with latent discrete variables or mightdegenerate multi-step predictions due to the inaccuracy (biased gradients) of back-propagation.

3 NeuCFlow Model

3.1 Attentive computation framework

We extend Bengio’s consciousness prior to graph-structured representation. Conscious thoughtsare modeled by a few selected nodes and their edges, forming a context-aware subgraph, cohesivewith sharper semantics, disentangled from the full graph. The underlying full graph forms the initialrepresentation, entangled but rich, to help shape potential high-level subgraphs. We use attention flowto navigate conscious thoughts, capturing a step-by-step reasoning pattern. The attentive computationframework, as illustrated in Figure 1, consists of: (1) an unconsciousness flow (U-Flow) layer, (2) aconsciousness flow (C-Flow) layer, and (3) an attention flow (A-Flow) layer, with four guidelines todesign a specific implementation as follows:

• U-Flow corresponds to a low-level computation graph for full state representation learning.

• C-Flow contains high-level disentangled subgraphs for context-aware representation learning.

• A-Flow is conditioned by both U-Flow and C-Flow, and also motivate C-Flow but not U-Flow.

• Information can be accessed by C-Flow from U-Flow with the help of A-Flow.

3.2 Model architecture design for knowledge graph reasoning

We choose KBC tasks to do KG reasoning. We let 〈V, E〉 denote a KG where V is a set of nodes (orentities) and E is a set of edges (or relations). A KG is viewed as a directed graph with each edgerepresented by a triple 〈head, rel, tail〉, where head is the head entity, tail is the tail entity, and relis their relation type. The aim of a KBC task is to predict potential unknown links, i.e., which entityis likely to be the tail given a query 〈head, rel, ?〉 with the head and the relation type specified.

The model architecture has three core components as shown in Figure 2. We here use the term"component" instead of "layer" to differentiate our flow layers from the referring normally used inneural networks, as each flow layer is more like a block containing many neural network layers.

U-Flow component. We implement this component over the full graph using the standard messagepassing mechanism [46]. If the graph has an extremely large number of edges, we sample a subset

4

Page 5: Neural Consciousness Flow - GitHub Pages

of edges, Eτsmpl ⊂ E , randomly each step when running message passing. For each batch of inputqueries, we let the representation computed by the U-Flow component be shared across these differentqueries, which means U-Flow is query-independent, with its state representation tensors containing nobatch dimension, so that its complexity does not scale with the batch size and the saved computationresources can be allocated to sampling more edges. In U-Flow, each node v has a learnable embeddingev and a dynamical state hτv for step τ , called unconscious node states, where the initial h0

v := evfor all v ∈ V . Each edge type r also has a learnable embedding er, and edge 〈v′, r, v〉 can produce amessage, denoted by mτ

〈v′,r,v〉, at step τ . The U-Flow component includes:

• Message function: mτ〈v′,r,v〉 = ψunc(h

τv′ , er, h

τv), where 〈v′, r, v〉 ∈ Eτsmpl.

• Message aggregation: µτv = 1√Nτv

∑v′,r mτ

〈v′,r,v〉, where 〈v′, r, v〉 ∈ Eτsmpl.

• Node state update function: hτ+1v = hτv + δunc(µ

τv , h

τv , ev), where v ∈ V .

We compute messages only for the sampled edges, 〈v′, r, v〉 ∈ Eτsmpl, each step. Functions ψunc andδunc are implemented by a two-layer MLP (using leakyReLu for the first layer and tanh for thesecond layer) with input arguments concatenated respectively. Messages are aggregated by dividingthe sum by the square root of the number of sampled neighbors that send messages, preserving thescale of variance. We use a residual adding to update each node state instead of a GRU or a LSTM.After running U-Flow for T steps, we return a pooling result or simply the last, hv := hT

v , to feedinto downstream components.

C-Flow component. C-Flow is query-dependent, which means that conscious node states, denotedby htv, have a batch dimension representing different input queries, making the complexity scalewith the batch size. However, as C-Flow uses attentive message passing, running on small localsubgraphs each conditioned by a query, we leverage the sparsity to record htv only for the visitednodes v ∈ Vtvisit. For example, when t = 0, for query 〈head, rel, ?〉, we start from node head, withV0visit = {vhead} being a singleton, and thus record h0

vheadonly. When computing messages, denoted

by mt〈v′,r,v〉, in C-Flow, we use a sampling-attending procedure, explained in Section 3.3, to further

control the number of computed edges. The C-Flow component has:

• Message function: mt〈v′,r,v〉 = ψcon(h

tv′ , cr,h

tv), where 〈v′, r, v〉 ∈ Ettopks(at+1) | topka(at), and

cr = [er,qhead,qrel].

• Message aggregation: µtv =1√Ntv

∑v′,r mt

〈v′,r,v〉, where 〈v′, r, v〉 ∈ Ettopks(at+1) | topka(at).

• Node state attending function: ηt+1v = at+1

v A · hv , where at+1v = at+1[v] and v ∈ Vt+1

visit.

• Node state update function: ht+1v = htv + δcon(µ

tv,h

tv, c

t+1v ), where ct+1

v = [ηt+1v ,qhead,qrel].

C-Flow and U-Flow share the embeddings er. A query is represented by its head and relationembeddings, qhead := ehead and qrel := erel, participating in computing messages and updatingnode states. We here select a subset of edges, Ettopks(at+1) | topka(at), rather than sampling, accordingto edges between the attended nodes at step t and the seen nodes at step t+ 1, defined in Section 3.3,as shown in Figure 3. We introduce the node state attending function to pass an unconscious state hvto C-Flow adjusted by a scalar attention at+1

v and a learnable matrix A. We initialize h0v := hv for

v ∈ V0visit, treating the rest as zero states.

A-Flow component. Attention flow is represented by a series of probability distributions changingacross steps, denoted as at, t = 1, 2 . . . , T . The initial distribution a0 is a one-hot vector witha0[vhead] = 1. To spread attention, we need to compute transition matrices Tt each step. Given thatA-Flow is conditioned by both U-Flow and C-Flow, we model the transition from v′ to v by twotypes of interaction: conscious-to-conscious, htv′ ∼ htv, and conscious-to-unconscious, htv′ ∼ hv.The former favors previously visited nodes, while the latter is useful to attend to unseen nodes.

Tt[:, v′] = softmaxv∈N tv′

(∑rαcc(h

tv′ , cr,h

tv) +

∑rαcu(h

tv′ , cr, hv)

)where αcc = MLP(htv′ , cr)

TΘccMLP(htv, cr) and αcu = MLP(htv′ , cr)TΘcuMLP(hv, cr), and

Θcc and Θcu are two learnable matrices. Each MLP uses one single layer with the leakyReLu

5

Page 6: Neural Consciousness Flow - GitHub Pages

All Candidate Nodes

Sampled Nodes

Seen Nodes

Attended Nodes

All Candidate Nodes

Sampled Nodes

Step t Step t+1

Visited Nodes Visited Nodes

Seen Nodes

Attended Nodes

Figure 3: The iterative sampling-attending procedure for attentive complexity reduction, balancingthe coverage as well as the complexity.

activation. To reduce the complexity for computing Tt, we select attended nodes, v′ ∈ topka(at),which is the set of nodes with the k-largest attention, and then sample v from v′ neighbors as nextnodes. Then, we compute a sparse Tt according to edges 〈v′, r, v〉 ∈ Esmpl | topka(at). Due to thefact that the attended nodes may not carry all attention, a small amount of attention can be lostduring transition, causing the total amount to decrease. Therefore, we use a renormalized version,at+1 = Ttat/‖Ttat‖. We use the final attention on the tail as the probability for prediction tocompute the training objective, as shown in Figure 2.

3.3 Complexity reduction by iterative sampling and attending

Previously, we use edge sampling, in a globally and uniformly random manner, to address thecomplexity issue in U-Flow, where we are not concerned about the batch size. Here, we need toconfront the complexity that scales with the batch size in C-Flow. Suppose that we run a normalmessage passing for T steps on a KG with |V| nodes and |E| edges for a batch of N queries. Then,the complexity is O(NTD(|V|+ |E|)) where D represents the number of representation dimensions.The complexity can be reduced to O(NTD(|V|+ |Esmpl|)) by using edges sampling. T is a smallpositive integer, often less than 10. D is normally between 50 and 200, and being too small forD would lead to underfitting. In U-Flow, we have N = 1, while in C-Flow, let us say N = 100.Then, to maintain the same complexity as U-Flow, we have to reduce the sampling rate by a factorof 100 on each query. However, the U-Flow’s edge sampling procedure is for the full graph, and itis inappropriate to apply to C-Flow on each query due to the reduced sample rate. Also, when |V|becomes as large as |Esmpl|, we also need to consider decreasing |V|.Good news is that C-Flow deals with a local subgraph for each query so that we only record a fewselected nodes, called visited nodes, denoted by Vtvisit. We can see that |Vtvisit| is much less than|V|. The initial V0

visit, when t = 0, contains only one node vhead, and then Vtvisit is enlarged eachstep by adding new nodes during spreading. When propagating messages, we only care about theone-step neighborhood each step. However, the spreading goes so rapidly that after only a few steps itcovers almost all nodes, causing the number of computed edges to increase dramatically. The key toaddress the problem is that we need to constrain the scope of nodes we jump from each step, i.e., thecore nodes that determine where we can go based on where we depart from. We call them attendednodes, which are in charge of the attending-from horizon, selected by topka(at) based on the currentattention at. Given the set of attended nodes, we still need edge sampling over their neighborhoodsin case of a hub node of extremely high degree. Here, we face a tricky problem that is to make atrade-off between the coverage and the complexity when sampling over the neighborhoods. Also, weneed to well maintain these coherent context-aware node states and avoid possible noises or driftingaway caused by sampling neighbors randomly. Therefore, we introduce an attending-to horizon insidethe sampling horizon. We compute A-Flow over the sampling horizon with a smaller dimensionto compute the attention, exchanged for sampling more neighbors to increase the coverage. Based

6

Page 7: Neural Consciousness Flow - GitHub Pages

Table 1: Statistics of the six KG datasets. A KG is built on all training triples including their inversetriples. Note that we do not count the inverse triples in FB15K, FB15K-237, WN18, WN18RR, andYAGO3-10 as shown below to be consistent with the statistics reported in other papers, though weinclude them in the training, validation and test set. PME (tr) means the proportion of multi-edgetriples in train; PME (te) means the proportion of multi-edge triples in test; AvgD (te) means theaverage length of shortest paths connecting each head-tail pair in test.

Dataset #Entities #Rels #Train #Valid #Test PME (tr) PME (te) AvgD (te)FB15K 14,951 1,345 483,142 50,000 59,071 81.2% 80.9% 1.22FB15K-237 14,541 237 272,115 17,535 20,466 38.0% 0% 2.25WN18 40,943 18 141,442 5,000 5,000 93.1% 94.0% 1.18WN18RR 40,943 11 86,835 3,034 3,134 34.5% 35.0% 2.87NELL995 74,536 200 149,678 543 2,818 100% 41.0% 2.06YAGO3-10 123,188 37 1,079,040 5,000 5,000 56.4% 56.0% 1.75

Table 2: Comparison results on the FB15K-237 and WN18RR datasets. Results of [♠] are takenfrom [83], results of [♣] from [10], results of [♥] from [17], results of [♦] from [12], and results of[4] from [16]. Some collected results only have a metric score while some including ours take theform of "mean (standard deviation)".

FB15K-237 WN18RRMetric (%) H@1 H@3 H@10 MRR H@1 H@3 H@10 MRRTransE [♠] - - 46.5 29.4 - - 50.1 22.6DistMult [♣] 15.5 26.3 41.9 24.1 39 44 49 43DistMult [♥] 20.6 (.4) 31.8 (.2) - 29.0 (.2) 38.4 (.4) 42.4 (.3) - 41.3 (.3)ComplEx [♣] 15.8 27.5 42.8 24.7 41 46 51 44ComplEx [♥] 20.8 (.2) 32.6 (.5) - 29.6 (.2) 38.5 (.3) 43.9 (.3) - 42.2 (.2)ConvE [♣] 23.7 35.6 50.1 32.5 40 44 52 43ConvE [♥] 23.3 (.4) 33.8 (.3) - 30.8 (.2) 39.6 (.3) 44.7 (.2) - 43.3 (.2)RotatE [♦] 24.1 37.5 53.3 33.8 42.8 49.2 57.1 47.6NeuralLP [♥] 18.2 (.6) 27.2 (.3) - 24.9 (.2) 37.2 (.1) 43.4 (.1) - 43.5 (.1)MINERVA [♥] 14.1 (.2) 23.2 (.4) - 20.5 (.3) 35.1 (.1) 44.5 (.4) - 40.9 (.1)MINERVA [4] - - 45.6 - 41.3 45.6 51.3 -M-Walk [♥] 16.5 (.3) 24.3 (.2) - 23.2 (.2) 41.4 (.1) 44.5 (.2) - 43.7 (.1)NeuCFlow 28.6 (.1) 40.3 (.1) 53.0 (.3) 36.9 (.1) 44.4 (.4) 49.7 (.8) 55.8 (.5) 48.2 (.5)

on the newly computed attention at+1, we select a smaller subset of nodes, topks(at+1), to receivemessages in C-Flow, called seen nodes, in charge of the attending-to horizon. The next attending-from horizon is chosen by topka(at+1) ⊂ topks(at+1), a sub-horizon of the current attending-tohorizon. All seen and attended nodes are stored as visited nodes along steps. We illustrate thissampling-attending procedure in Figure 3.

To compute our reduced complexity, we let Ne be the maximum number of sampled edges perattended node per step, Ns the maximum number of seen nodes per step, and Na the maximumnumber of attended nodes per step. We also denote the dimension number used in A-Flow as Da.For one batch, the complexity of C-Flow is O(NTD(Na +Ns +NaNs)) for the worst case, whereattended and seen nodes are fully connected, and O(NTD · c(Na +Ns)) in most cases, where c is asmall constant. The complexity of A-Flow is O(NTDaNaNe) where Da is much smaller than D.

4 Experiments

4.1 Datasets and experimental settings

Datasets. We evaluate our model using six large KG datasets2: FB15K, FB15K-237, WN18,WN18RR, NELL995, and YAGO3-10. FB15K-237 [84] is sampled from FB15K [8] with redundantrelations removed, and WN18RR [10] is a subset of WN18 [8] removing triples that cause testleakage. Thus, they are both considered more challenging. NELL995 [15] has separate datasets

2https://github.com/netpaladinx/NeuCFlow/tree/master/data

7

Page 8: Neural Consciousness Flow - GitHub Pages

Table 3: Comparison results on the FB15K and WN18 datasets. Results of [♠] are taken from [86],results of [♣] are from [10], results of [♦] are from [12], and results of [♥] are from [19]. Our resultstake the form of "mean (standard deviation)".

FB15K WN18Metric (%) H@1 H@3 H@10 MRR H@1 H@3 H@10 MRRTransE [♠] 29.7 57.8 74.9 46.3 11.3 88.8 94.3 49.5HolE [♠] 40.2 61.3 73.9 52.4 93.0 94.5 94.9 93.8DistMult [♣] 54.6 73.3 82.4 65.4 72.8 91.4 93.6 82.2ComplEx [♣] 59.9 75.9 84.0 69.2 93.6 93.6 94.7 94.1ConvE [♣] 55.8 72.3 83.1 65.7 93.5 94.6 95.6 94.3RotatE [♦] 74.6 83.0 88.4 79.7 94.4 95.2 95.9 94.9NeuralLP [♥] - - 83.7 76 - - 94.5 94NeuCFlow 72.6 (.4) 78.4 (.4) 83.4 (.5) 76.4 (.4) 91.6 (.8) 93.6 (.4) 94.9 (.4) 92.8 (.6)

Table 4: Comparison results on the YAGO3-10 dataset. Results of [♠] are taken from [10], andresults of [♣] are from [13].

YAGO3-10Metric (%) H@1 H@3 H@10 MRRDistMult [♠] 24 38 54 34ComplEx [♠] 26 40 55 36ConvE [♠] 35 49 62 44ComplEx-N3 [♣] - - 71 58NeuCFlow 48.4 59.5 67.9 55.3

for 12 query relations each corresponding to a single-query-relation KBC task. YAGO3-10 [85]contains the largest KG with millions of edges. Their statistics are shown in Table 1. We find somestatistical differences between train and test. In a KG with all training triples as its edges, a triple(head, rel, tail) is considered as a multi-edge triple if the KG contains other triples that also connecthead and tail ignoring the direction. We notice that FB15K-237 is a special case compared with theothers, as there are no edges in its KG directly linking any pair of head and tail in test. Therefore,when using training triples as queries to train our model, given a batch, for FB15K-237, we cut offfrom the KG all triples connecting the head-tail pairs in the given batch, ignoring relation types andedge directions, forcing the model to learn a composite reasoning pattern rather than a single-hoppattern, and for the rest datasets, we only remove the triples of this batch and their inverse from theKG before training on this batch.

Experimental settings. We use the same data split protocol as in many papers [10, 15, 16]. Wecreate a KG, a directed graph, consisting of all train triples and their inverse added for each datasetexcept NELL995, since it already includes reciprocal relations. Besides, every node in KGs has aself-loop edge to itself. We also add inverse relations into the validation and test set to evaluate thetwo directions. For evaluation metrics, we use HITS@1,3,10 and the mean reciprocal rank (MRR) inthe filtered setting for FB15K-237, WN18RR, FB15K, WN18, and YAGO3-10, and use the meanaverage precision (MAP) for NELL995’s single-query-relation KBC tasks. For NELL995, we followthe same evaluation procedure as in [15, 16, 17], ranking the answer entities against the negativeexamples given in their experiments. We run our experiments using a 12G-memory GPU, TITAN X(Pascal), with Intel(R) Xeon(R) CPU E5-2670 v3 @ 2.30GHz. Our code is written in Python basedon TensorFlow 2.0 and NumPy 1.16.

4.2 Baselines and comparison results

Baselines. We compare our model against embedding-based approaches, including TransE [8],TransR [22], DistMult [9], ConvE [10], ComplE [11], HolE [86], RotatE [12], and ComplEx-N3[13], and path-based approaches that use RL methods, including DeepPath [15], MINERVA [16],and M-Walk [17], and also that uses learned neural logic, NeuralLP [19]. For all the baselines, wequote the results from the corresponding papers instead of rerunning them. For our method, we run

8

Page 9: Neural Consciousness Flow - GitHub Pages

Table 5: Comparison results of MAP scores (%) on NELL995’s single-query-relation KBC tasks. Wetake our baselines’ results from [17]. All results take the form of "mean (standard deviation)" exceptfor TransE and TransR.

Tasks NeuCFlow M-Walk MINERVA DeepPath TransE TransRAthletePlaysForTeam 83.9 (0.5) 84.7 (1.3) 82.7 (0.8) 72.1 (1.2) 62.7 67.3AthletePlaysInLeague 97.5 (0.1) 97.8 (0.2) 95.2 (0.8) 92.7 (5.3) 77.3 91.2AthleteHomeStadium 93.6 (0.1) 91.9 (0.1) 92.8 (0.1) 84.6 (0.8) 71.8 72.2AthletePlaysSport 98.6 (0.0) 98.3 (0.1) 98.6 (0.1) 91.7 (4.1) 87.6 96.3TeamPlayssport 90.4 (0.4) 88.4 (1.8) 87.5 (0.5) 69.6 (6.7) 76.1 81.4OrgHeadQuarteredInCity 94.7 (0.3) 95.0 (0.7) 94.5 (0.3) 79.0 (0.0) 62.0 65.7WorksFor 86.8 (0.0) 84.2 (0.6) 82.7 (0.5) 69.9 (0.3) 67.7 69.2PersonBornInLocation 84.1 (0.5) 81.2 (0.0) 78.2 (0.0) 75.5 (0.5) 71.2 81.2PersonLeadsOrg 88.4 (0.1) 88.8 (0.5) 83.0 (2.6) 79.0 (1.0) 75.1 77.2OrgHiredPerson 84.7 (0.8) 88.8 (0.6) 87.0 (0.3) 73.8 (1.9) 71.9 73.7AgentBelongsToOrg 89.3 (1.2) - - - - -TeamPlaysInLeague 97.2 (0.3) - - - - -

the experiments three times in each hyperparameter setting on each dataset to report the means andstandard deviations of the results. We put the details of our hyperparameter settings in the appendix.

Comparison results and analysis. We first report the comparison on FB15K-23 and WN18RR inTable 2. NeuCFlow has a surprisingly good result, significantly outperforming all the comparedmethods in HITS@1,3 and MRR on both the two datasets. Compared to the best baseline, RotatE,published very recently, we only lose a few points in HITS@10 but gain a lot in HITS@1,3 and MRR.Based on the observation that NeuCFlow gains a larger amount of advantage when k in HITS@k getssmaller, we speculate that the reasoning ability acquired by NeuCFlow is to make a sharper predictionby exploiting graph-structured composition locally and conditionally, in contrast to embedding-basedmethods, which totally rely on vectorized representation. When a target becomes too vague topredict, reasoning may lose its great advantage, though still very competitive. However, path-basedbaselines, with a certain ability to do KG reasoning, perform worse than we expect. We argue that itis inappropriate to think of reasoning, a sequential decision process, as a sequence of nodes, i.e. apath, in KGs. The average length of the shortest paths between heads and tails in the test set in a KG,as shown in Table 1, suggests an extremely short path, making the motivation for using a path patternalmost pointless. The iterative reasoning pattern should be characterized in the form of dynamicallyvarying local graph-structured patterns, holding a bunch of nodes resonating with each other toproduce a decision collectively. Then, we run our model on larger KGs, including FB15K, WN18,and YAGO3-10, and summarize the comparison in Table 3,4, where NeuCFlow beats most well-known baselines and achieves a very competitive position against the best state-of-the-art methods.Moreover, we summarize the comparison on NELL995’s tasks in Table 5. NeuCFlow performs thebest on five tasks, also being very competitive against M-Walk, the best path-based method as far aswe know, on the rest. We find no reporting on the last two tasks from the corresponding papers.

4.3 Experimental analysis

Convergence analysis. During training we find that NeuCFlow converges surprisingly fast. We mayuse half of training examples to get the model well trained and generalize it to the test, sometimesproducing an even better metric score than trained for a full epoch, as shown in Figure 4(A). Comparedwith the less expensive computation using embedding-based models, although our model takes alarge number of edges to compute for each input query, consuming more time on one batch, it doesnot need a second epoch or even taking all training triples as queries in one epoch, thus saving a lotof training time. The reason may be that all queries are directly from the KG’s edge set and someof them have probably been exploited to construct subgraphs for many times during the training ofother queries, so that we might not have to train the model on each query explicitly as long as wehave other ways to exploit them.

Component analysis. If we do not run U-Flow, then the unconscious state hv is just the initialembedding of node v, and we can still run C-Flow as usual. We want to know whether the U-Flowcomponent is actually useful. Considering that long-distance message passing might bring in less

9

Page 10: Neural Consciousness Flow - GitHub Pages

0.5 1.0 1.5 2.0 2.5 3.0Epoch

40.0

42.5

45.0

47.5

50.0

52.5

55.0

57.5

60.0

Met

ric S

core

(%)

(A) Convergence Analysis (by evaluation on test during training)H@1H@3H@10MMR

H@1 H@3 H@10 MMR40.0

42.5

45.0

47.5

50.0

52.5

55.0

57.5

60.0

Met

ric S

core

(%)

(B) U-Flow Component AnalysisW/o U-FlowWith U-Flow

H@1 MMR40

42

44

46

48

50

52

54

Met

ric S

core

(%)

(C) Sampling Horizon AnalysisMax-sampled-edges-per-node = 20Max-sampled-edges-per-node = 50Max-sampled-edges-per-node = 100Max-sampled-edges-per-node = 200Max-sampled-edges-per-node = 400

H@1 MMR40

42

44

46

48

50

52

54

Met

ric S

core

(%)

(D) Attending-to Horizon AnalysisMax-seen-nodes-per-step = 20Max-seen-nodes-per-step = 50Max-seen-nodes-per-step = 100Max-seen-nodes-per-step = 200Max-seen-nodes-per-step = 400

H@1 MMR40

42

44

46

48

50

52

54

Met

ric S

core

(%)

(E) Attending-from Horizon AnalysisMax-attended-nodes-per-step = 5Max-attended-nodes-per-step = 10Max-attended-nodes-per-step = 20Max-attended-nodes-per-step = 40

H@1 MMR35.0

37.5

40.0

42.5

45.0

47.5

50.0

52.5

55.0

Met

ric S

core

(%)

(F) Searching Horizon Analysis#Steps-of-C-Flow = 2#Steps-of-C-Flow = 4#Steps-of-C-Flow = 6#Steps-of-C-Flow = 8

Figure 4: Experimental analysis on WN18RR: (A) During training we pick six model snapshotsat time points of 0.3, 0.5, 0.7, 1, 2, and 3 epochs and evaluate them on test; (B) The w/o U-Flowuses zero step to run U-Flow, while the with U-Flow uses two steps; (C)-(F) are for the sampling,attending and searching horizon analysis based on the standard hyperparameter settings listed in theappendix. The experimental analysis charts on FB15K-237 can be found in the appendix.

informative features, we compare running U-Flow for two steps against totally shutting it down. Theresult in Figure 4(B) shows that U-Flow brings a small gain in each metric on WN18RR.

Horizon analysis. The sampling, attending and searching horizons determine how large area theflow can spread over. They impact the computation complexity as well as the performance of themodel with different degrees depending on the properties of a dataset. Intuitively, enlarging theprobe scope by sampling more, attending more, or searching longer, may increase the chance to hita target. However, the experimental results in Figure 4(C)(D) show that it is not always the case.In Figure 4(E), we can see that increasing the maximum number of the attending-from nodes, i.e.attended nodes, per step is more important, but our GPU does not allow for a larger number toaccommodate more intermediate data produced during computation, otherwise causing the error ofResourceExhaustedError. Figure 4(F) shows the step number of C-Flow cannot get too small as two.

Attention flow analysis. If attention flow can really capture the way we reason about the world, itsprocess should be conducted in a diverging-converging thinking pattern. Intuitively, first, for thediverging thinking, we search and collect ideas as much as we can; then, for the converging thinking,we try to concentrate our thoughts on one point. To check whether the attention flow has such apattern, we measure the average entropy of attention distributions varying along steps and also theproportion of attention concentrated at the top-1,3,5 attended nodes. As we expect, attention indeedis more focused at the final step as well as at the beginning.

Time cost analysis. The time cost is affected not only by the scale of a dataset but also by thehorizon setting. For each dataset, we list the training time for one epoch corresponding to thestandard hyperparameter settings in the appendix. Note that there is always a trade-off between thecomplexity and the performance. We thus study whether we can reduce the time cost a lot at the priceof sacrificing a little performance. We plot the one-epoch training time in Figure 6(A)-(D), usingthe same settings as we do in the horizon analysis. We can see that Max-attended-nodes-per-stepand #Steps-of-C-Flow affect the training time significantly while Max-sampled-edges-per-node andMax-seen-nodes-per-step affect very slightly. Therefore, we can use smaller Max-sampled-edges-per-node and Max-seen-nodes-per-step in order to gain a larger batch size, making the computation moreefficiency as shown in Figure 6(E).

10

Page 11: Neural Consciousness Flow - GitHub Pages

0 1 2 3 4Step

0

2

4

6

8

10

Entro

py o

f Atte

ntio

n Di

strib

utio

n

(A) Attention Flow Analysis (on entroy)AthletePlaysForTeamAthletePlaysInLeagueAthleteHomeStadiumAthletePlaysSportTeamPlaysSportOrgHeadQuarteredInCityWorksForPersonBornInLocationPersonLeadsOrgOrgHiredPersonAgentBelongsToOrgTeamPlaysInLeague

0 1 2 3 4Step

1.00

0.75

0.50

0.25

0.00

0.25

0.50

0.75

1.00

Prop

ortio

n-of

-Top

1

(B) Attention Flow Analysis (on top1's proportion)

AthletePlaysForTeamAthletePlaysInLeagueAthleteHomeStadiumAthletePlaysSportTeamPlaysSportOrgHeadQuarteredInCityWorksForPersonBornInLocationPersonLeadsOrgOrgHiredPersonAgentBelongsToOrgTeamPlaysInLeague

0 1 2 3 4Step

1.00

0.75

0.50

0.25

0.00

0.25

0.50

0.75

1.00

Prop

ortio

n-of

-Top

3

(C) Attention Flow Analysis (on top3's proportion)

AthletePlaysForTeamAthletePlaysInLeagueAthleteHomeStadiumAthletePlaysSportTeamPlaysSportOrgHeadQuarteredInCityWorksForPersonBornInLocationPersonLeadsOrgOrgHiredPersonAgentBelongsToOrgTeamPlaysInLeague

0 1 2 3 4Step

1.00

0.75

0.50

0.25

0.00

0.25

0.50

0.75

1.00

Prop

ortio

n-of

-Top

5

(D) Attention Flow Analysis (on top5's proportion)

AthletePlaysForTeamAthletePlaysInLeagueAthleteHomeStadiumAthletePlaysSportTeamPlaysSportOrgHeadQuarteredInCityWorksForPersonBornInLocationPersonLeadsOrgOrgHiredPersonAgentBelongsToOrgTeamPlaysInLeague

Figure 5: Analysis of attention flow on NELL995 tasks: (A) records how the average entropy ofattention distributions varies along steps for each single-query-relation KBC task. (B)(C)(D) measurethe changing of the proportion of attention concentrated at the top-1,3,5 attended nodes per step foreach task.

0

2

4

6

8

10

Trai

ning

Tim

e fo

r One

Epo

ch (h

)

(A) Time Cost for Different Sampling HorizonsMax-sampled-edges-per-node = 20Max-sampled-edges-per-node = 50Max-sampled-edges-per-node = 100Max-sampled-edges-per-node = 200Max-sampled-edges-per-node = 400

0

2

4

6

8

10

Trai

ning

Tim

e fo

r One

Epo

ch (h

)

(B) Time Cost for Different Attending-to HorizonsMax-seen-nodes-per-step = 20Max-seen-nodes-per-step = 50Max-seen-nodes-per-step = 100Max-seen-nodes-per-step = 200Max-seen-nodes-per-step = 400

0

2

4

6

8

10

Trai

ning

Tim

e fo

r One

Epo

ch (h

)

(C) Time Cost for Different Attending-from HorizonsMax-attended-nodes-per-step = 5Max-attended-nodes-per-step = 10Max-attended-nodes-per-step = 20Max-attended-nodes-per-step = 40

0

2

4

6

8

10

Trai

ning

Tim

e fo

r One

Epo

ch (h

)

(D) Time Cost for Different Searching Horizons#Steps-of-C-Flow = 2#Steps-of-C-Flow = 4#Steps-of-C-Flow = 6#Steps-of-C-Flow = 8

0

2

4

6

8

10

Trai

ning

Tim

e fo

r One

Epo

ch (h

)

(E) Time Cost for Different Batch SizesBatch-size = 50Batch-size = 100Batch-size = 200Batch-size = 300

Figure 6: Analysis of time cost on WN18RR: (A)-(D) measure the training time for one epoch ondifferent horizon settings corresponding to Figure 4(C)-(F); (E) measures the training time for oneepoch for different batch sizes using the same horizon setting, which is Max-sampled-edges-per-node=20, Max-seen-nodes-per-step=20, Max-attended-nodes-per-step=20, and #Steps-of-C-Flow=8.The time cost analysis charts on FB15K-237 can be found in the appendix.

4.4 Visualization

To further demonstrate the reasoning ability acquired by our model, we show some visualizationresults of the extracted subgraphs on NELL995’s test data for 12 separate tasks. We avoid usingthe training data in order to show the generalization of our model’s learned reasoning ability onknowledge graphs. Here, we show the visualization result for the AthletePlaysForTeam task. The restcan be found in the appendix.

For the AthletePlaysForTeam taskQuery : ( concept_personnorthamerica_michael_turner , concept : a th le tep lays fo r team , concept_sports team_falcons )

Selected key edges :concept_personnorthamerica_michael_turner , concept : agentbe longstoorgan iza t ion , concept_spor ts league_nf lconcept_personnorthamerica_michael_turner , concept : athletehomestadium , concept_stadiumoreventvenue_georgia_domeconcept_spor ts league_nf l , concept : agentcompeteswithagent , concept_spor ts league_nf lconcept_spor ts league_nf l , concept : agentcompeteswithagent_inv , concept_spor ts league_nf lconcept_spor ts league_nf l , concept : teamplaysin league_inv , concept_sportsteam_sd_chargersconcept_spor ts league_nf l , concept : leaguestadiums , concept_stadiumoreventvenue_georgia_domeconcept_spor ts league_nf l , concept : teamplaysin league_inv , concept_sports team_falconsconcept_spor ts league_nf l , concept : agentbe longs toorgan iza t ion_ inv , concept_personnorthamer ica_michael_turnerconcept_stadiumoreventvenue_georgia_dome , concept : leaguestadiums_inv , concept_spor ts league_nf lconcept_stadiumoreventvenue_georgia_dome , concept : teamhomestadium_inv , concept_sports team_falconsconcept_stadiumoreventvenue_georgia_dome , concept : athletehomestadium_inv , concep t_a th le te_ joey_har r ing tonconcept_stadiumoreventvenue_georgia_dome , concept : athletehomestadium_inv , concept_ath le te_roddy_whi teconcept_stadiumoreventvenue_georgia_dome , concept : athletehomestadium_inv , concept_coach_deangelo_hal lconcept_stadiumoreventvenue_georgia_dome , concept : athletehomestadium_inv , concept_personnorthamer ica_michael_turnerconcept_spor ts league_nf l , concept : subpar to fo rgan iza t i on_ inv , concept_sports team_oakland_raidersconcept_sportsteam_sd_chargers , concept : teamplaysinleague , concept_spor ts league_nf lconcept_sportsteam_sd_chargers , concept : teamplaysagainstteam , concept_sports team_falconsconcept_sportsteam_sd_chargers , concept : teamplaysagainst team_inv , concept_sports team_falconsconcept_sportsteam_sd_chargers , concept : teamplaysagainstteam , concept_sports team_oakland_raidersconcept_sportsteam_sd_chargers , concept : teamplaysagainst team_inv , concept_sports team_oakland_raidersconcept_sportsteam_falcons , concept : teamplaysinleague , concept_spor ts league_nf lconcept_sportsteam_falcons , concept : teamplaysagainstteam , concept_sportsteam_sd_chargersconcept_sportsteam_falcons , concept : teamplaysagainst team_inv , concept_sportsteam_sd_chargersconcept_sportsteam_falcons , concept : teamhomestadium , concept_stadiumoreventvenue_georgia_dome

11

Page 12: Neural Consciousness Flow - GitHub Pages

concept_personnorthamerica_michael_turner

concept_stadiumoreventvenue_georgia_dome

concept_sportsleague_nfl

concept_coach_deangelo_hall concept_athlete_roddy_white

concept_athlete_joey_harrington

concept_athlete_quarterback_matt_ryan

concept_coach_jerious_norwood

concept_athlete_chris_redman

concept_city_atlanta

concept_sportsteam_falcons

concept_sportsteam_dallas_cowboys

concept_sportsteam_seahawks

concept_sportsteam_sd_chargers

concept_sportsteam_minnesota_vikingsconcept_sportsteam_broncos

concept_sportsteam_cleveland_browns

concept_sportsteam_titans

concept_sportsteam_buffalo_billsconcept_sportsteam_kansas_city_chiefs

concept_sportsteam_oakland_raiders

concept_sport_football

concept_awardtrophytournament_division

concept_sportsteam_packers

concept_sportsteam_colts

concept_sportsteam_eagles

concept_sportsteam_new_york_giants

concept_sportsteam_bears_29_17

concept_sportsteam_bills

concept_sportsteam_steelers

concept_sportsteam_buccaneers concept_sportsteam_rams

concept_sportsteam_saintsconcept_sportsteam_bucs concept_sportsteam_tampa

concept_sportsteam_texansconcept_sportsteam_new_york_jets

concept_sportsteam_ravens

concept_personnorthamerica_michael_turner

concept_stadiumoreventvenue_georgia_dome

concept_sportsleague_nfl

concept_coach_deangelo_hallconcept_athlete_roddy_white

concept_athlete_joey_harrington

concept_athlete_quarterback_matt_ryan

concept_coach_jerious_norwood

concept_sportsteam_falconsconcept_sportsteam_oakland_raiders

concept_sport_football

concept_sportsteam_sd_chargers

concept_sportsteam_minnesota_vikings

concept_sportsteam_dallas_cowboys

concept_sportsteam_new_york_giants

concept_sportsteam_kansas_city_chiefs

concept_sportsteam_tampa

Figure 7: AthletePlaysForTeam. The head is concept_personnorthamerica_michael_turner, thequery relation is concept:athleteplaysforteam, and the desired tail is concept_sportsteam_falcons.The left is a full subgraph derived with max_attended_nodes_per_step = 20, and the right is afurther extracted subgraph from the left based on attention. The big yellow node represents the head,and the big red node represents the tail. Colors indicate how important a node is attended to in a localsubgraph. Grey means less important, yellow means it is more attended during the early steps, andred means it is more attended when getting close to the final step.

concept_sportsteam_falcons , concept : teamplaysagainstteam , concept_sports team_oakland_raidersconcept_sportsteam_falcons , concept : teamplaysagainst team_inv , concept_sports team_oakland_raidersconcept_sportsteam_falcons , concept : a th le te ledspor ts team_ inv , concep t_a th le te_ joey_har r ing tonconcep t_a th le te_ joey_har r ing ton , concept : athletehomestadium , concept_stadiumoreventvenue_georgia_domeconcept_a th le te_ joey_har r ing ton , concept : a th le te ledspor ts team , concept_sports team_falconsconcep t_a th le te_ joey_har r ing ton , concept : a th le tep lays fo r team , concept_sports team_falconsconcept_ath le te_roddy_whi te , concept : athletehomestadium , concept_stadiumoreventvenue_georgia_domeconcept_ath le te_roddy_whi te , concept : a th le tep lays fo r team , concept_sports team_falconsconcept_coach_deangelo_hal l , concept : athletehomestadium , concept_stadiumoreventvenue_georgia_domeconcept_coach_deangelo_hal l , concept : a th le tep lays fo r team , concept_sports team_oakland_raidersconcept_spor ts league_nf l , concept : teamplaysin league_inv , concept_sportsteam_new_york_giantsconcept_sportsteam_sd_chargers , concept : teamplaysagainst team_inv , concept_sportsteam_new_york_giantsconcept_sportsteam_falcons , concept : teamplaysagainstteam , concept_sportsteam_new_york_giantsconcept_sportsteam_falcons , concept : teamplaysagainst team_inv , concept_sportsteam_new_york_giantsconcept_sportsteam_oakland_raiders , concept : teamplaysagainst team_inv , concept_sportsteam_new_york_giantsconcept_sportsteam_oakland_raiders , concept : teamplaysagainstteam , concept_sportsteam_sd_chargersconcept_sportsteam_oakland_raiders , concept : teamplaysagainst team_inv , concept_sportsteam_sd_chargersconcept_sportsteam_oakland_raiders , concept : teamplaysagainstteam , concept_sports team_falconsconcept_sportsteam_oakland_raiders , concept : teamplaysagainst team_inv , concept_sports team_falconsconcept_sportsteam_oakland_raiders , concept : agentcompeteswithagent , concept_sports team_oakland_raidersconcept_sportsteam_oakland_raiders , concept : agentcompeteswithagent_inv , concept_sports team_oakland_raidersconcept_sportsteam_new_york_giants , concept : teamplaysagainstteam , concept_sportsteam_sd_chargersconcept_sportsteam_new_york_giants , concept : teamplaysagainstteam , concept_sports team_falconsconcept_sportsteam_new_york_giants , concept : teamplaysagainst team_inv , concept_sports team_falconsconcept_sportsteam_new_york_giants , concept : teamplaysagainstteam , concept_sports team_oakland_raiders

In the above case, the query is (concept_personnorthamerica_michael_turner, concept:athleteplays-forteam, ?) and the desired answer is concept_sportsteam_falcons. From Figure 7, we can see ourmodel learns that (concept_personnorthamerica_michael_turner, concept:athletehomestadium, con-cept_stadiumoreventvenue_georgia_dome) and (concept_stadiumoreventvenue_georgia_dome, con-cept:teamhomestadium_inv, concept_sportsteam_falcons) are two important facts to support the an-swer of concept_sportsteam_falcons. Besides, other facts, such as (concept_athlete_joey_harrington,concept:athletehomestadium, concept_stadiumoreventvenue_georgia_dome) and (concept_athlete-_joey_harrington, concept:athleteplaysforteam, concept_sportsteam_falcons), provide a vivid exam-ple that a person or an athlete with concept_stadiumoreventvenue_georgia_dome as his or her homestadium might play for the team concept_sportsteam_falcons. We have such examples more thanone, like concept_athlete_roddy_white’s and concept_athlete_quarterback_matt_ryan’s. The entity

12

Page 13: Neural Consciousness Flow - GitHub Pages

concept_sportsleague_nfl cannot help us differentiate the true answer from other NFL teams, but itcan at least exclude those non-NFL teams. In a word, our subgraph-structured representation canwell capture the relational and compositional reasoning pattern.

5 Conclusion

We introduce an attentive message passing mechanism on graphs under the notion of attentive aware-ness, inspired by the phenomenon of consciousness, to model the iterative compositional reasoningpattern by forming a compact query-dependent subgraph. We propose an attentive computationframework with three flow-based layer to combine GNNs’ representation power with explicit rea-soning process, and further reduce the complexity when applying GNNs to large-scale graphs. It isworth mentioning that our framework is not limited to knowledge graph reasoning, but has a widerapplicability to large-scale graph-based computation with a few input-dependent nodes and edgesinvolved each time.

References[1] Yoshua Bengio. The consciousness prior. CoRR, abs/1709.08568, 2017.

[2] Stanislas Dehaene, Michel Kerszberg, and Jean Pierre Changeux. A neuronal model of a globalworkspace in effortful cognitive tasks. Proceedings of the National Academy of Sciences of theUnited States of America, 95 24:14529–34, 1998.

[3] Giulio Tononi, Mélanie Boly, Marcello Massimini, and Christof Koch. Integrated informationtheory: from consciousness to its physical substrate. Nature Reviews Neuroscience, 17:450–461,2016.

[4] David Rosenthal and Josh Weisberg. Higher-order theories of consciousness. Scholarpedia,3:4407, 2008.

[5] Robert Van Gulick. Higher-order global states (hogs): an alternative higher-order model.Higher-order theories of consciousness, pages 67–93, 2004.

[6] Franco Scarselli, Marco Gori, Ah Chung Tsoi, Markus Hagenbuchner, and Gabriele Monfardini.The graph neural network model. IEEE Transactions on Neural Networks, 20:61–80, 2009.

[7] Peter W. Battaglia, Jessica B. Hamrick, Victor Bapst, Alvaro Sanchez-Gonzalez, Vinícius Flo-res Zambaldi, Mateusz Malinowski, Andrea Tacchetti, David Raposo, Adam Santoro, RyanFaulkner, Çaglar Gülçehre, Francis Song, Andrew J. Ballard, Justin Gilmer, George E. Dahl,Ashish Vaswani, Kelsey R. Allen, Charles Nash, Victoria Langston, Chris Dyer, Nicolas Heess,Daan Wierstra, Pushmeet Kohli, Matthew Botvinick, Oriol Vinyals, Yujia Li, and Razvan Pas-canu. Relational inductive biases, deep learning, and graph networks. CoRR, abs/1806.01261,2018.

[8] Antoine Bordes, Nicolas Usunier, Alberto García-Durán, Jason Weston, and Oksana Yakhnenko.Translating embeddings for modeling multi-relational data. In NIPS, 2013.

[9] Bishan Yang, Wen tau Yih, Xiaodong He, Jianfeng Gao, and Li Deng. Embedding entities andrelations for learning and inference in knowledge bases. CoRR, abs/1412.6575, 2015.

[10] Tim Dettmers, Pasquale Minervini, Pontus Stenetorp, and Sebastian Riedel. Convolutional 2dknowledge graph embeddings. In AAAI, 2018.

[11] Théo Trouillon, Johannes Welbl, Sebastian Riedel, Éric Gaussier, and Guillaume Bouchard.Complex embeddings for simple link prediction. In ICML, 2016.

[12] Zhiqing Sun, Zhi-Hong Deng, Jian-Yun Nie, and Jian Tang. Rotate: Knowledge graph embed-ding by relational rotation in complex space. CoRR, abs/1902.10197, 2018.

[13] Timothée Lacroix, Nicolas Usunier, and Guillaume Obozinski. Canonical tensor decompositionfor knowledge base completion. In ICML, 2018.

13

Page 14: Neural Consciousness Flow - GitHub Pages

[14] Ni Lao, Tom Michael Mitchell, and William W. Cohen. Random walk inference and learning ina large scale knowledge base. In EMNLP, 2011.

[15] Wenhan Xiong, Thien Hoang, and William Yang Wang. Deeppath: A reinforcement learningmethod for knowledge graph reasoning. In EMNLP, 2017.

[16] Rajarshi Das, Shehzaad Dhuliawala, Manzil Zaheer, Luke Vilnis, Ishan Durugkar, AkshayKrishnamurthy, Alexander J. Smola, and Andrew McCallum. Go for a walk and arrive atthe answer: Reasoning over paths in knowledge bases using reinforcement learning. CoRR,abs/1711.05851, 2018.

[17] Yelong Shen, Jianshu Chen, Pu Huang, Yuqing Guo, and Jianfeng Gao. M-walk: Learning towalk over graphs using monte carlo tree search. In NeurIPS, 2018.

[18] William W. Cohen. Tensorlog: A differentiable deductive database. CoRR, abs/1605.06523,2016.

[19] Fan Yang, Zhilin Yang, and William W. Cohen. Differentiable learning of logical rules forknowledge base reasoning. In NIPS, 2017.

[20] Xiaoran Xu, Songpeng Zu, Chengliang Gao, Yuan Zhang, and Wei Feng. Modeling attentionflow on graphs. CoRR, abs/1811.00497, 2018.

[21] Zhen Wang, Jianwen Zhang, Jianlin Feng, and Zheng Chen. Knowledge graph embedding bytranslating on hyperplanes. In AAAI, 2014.

[22] Yankai Lin, Zhiyuan Liu, Maosong Sun, Yang Liu, and Xuan Zhu. Learning entity and relationembeddings for knowledge graph completion. In AAAI, 2015.

[23] Guoliang Ji, Shizhu He, Liheng Xu, Kang Liu, and Jian Zhao. Knowledge graph embeddingvia dynamic mapping matrix. In ACL, 2015.

[24] Matt Gardner, Partha Pratim Talukdar, Jayant Krishnamurthy, and Tom Michael Mitchell.Incorporating vector space similarity in random walk inference over knowledge bases. InEMNLP, 2014.

[25] Kelvin Guu, John Miller, and Percy S. Liang. Traversing knowledge graphs in vector space. InEMNLP, 2015.

[26] Yankai Lin, Zhiyuan Liu, and Maosong Sun. Modeling relation paths for representation learningof knowledge bases. In EMNLP, 2015.

[27] Kristina Toutanova, Victoria Lin, Wen tau Yih, Hoifung Poon, and Chris Quirk. Compositionallearning of embeddings for relation paths in knowledge base and text. In ACL, 2016.

[28] Rajarshi Das, Arvind Neelakantan, David Belanger, and Andrew McCallum. Chains of reasoningover entities, relations, and text using recurrent neural networks. In EACL, 2017.

[29] Chen Liang, Jonathan Berant, Quoc V. Le, Kenneth D. Forbus, and Ni Lao. Neural symbolicmachines: Learning semantic parsers on freebase with weak supervision. In ACL, 2016.

[30] Kenneth H. Craik. The nature of explanation. 1952.

[31] John R. Anderson. Acquisition of cognitive skill. 1982.

[32] Dedre Gentner and Arthur B. Markman. Structure mapping in analogy and similarity. 1997.

[33] John E. Hummel and Keith J. Holyoak. A symbolic-connectionist theory of relational inferenceand generalization. Psychological review, 110 2:220–64, 2003.

[34] Brenden M. Lake, Tomer D. Ullman, Joshua B. Tenenbaum, and Samuel J Gershman. Buildingmachines that learn and think like people. The Behavioral and brain sciences, 40:e253, 2017.

[35] Joan Bruna, Wojciech Zaremba, Arthur Szlam, and Yann LeCun. Spectral networks and locallyconnected networks on graphs. CoRR, abs/1312.6203, 2014.

14

Page 15: Neural Consciousness Flow - GitHub Pages

[36] Mikael Henaff, Joan Bruna, and Yann LeCun. Deep convolutional networks on graph-structureddata. CoRR, abs/1506.05163, 2015.

[37] David K. Duvenaud, Dougal Maclaurin, Jorge Aguilera-Iparraguirre, Rafael Gómez-Bombarelli,Timothy Hirzel, Alán Aspuru-Guzik, and Ryan P. Adams. Convolutional networks on graphsfor learning molecular fingerprints. In NIPS, 2015.

[38] Steven M. Kearnes, Kevin McCloskey, Marc Berndl, Vijay S. Pande, and Patrick Riley. Molec-ular graph convolutions: moving beyond fingerprints. Journal of computer-aided moleculardesign, 30 8:595–608, 2016.

[39] Michaël Defferrard, Xavier Bresson, and Pierre Vandergheynst. Convolutional neural networkson graphs with fast localized spectral filtering. In NIPS, 2016.

[40] Mathias Niepert, Mohammed Hassan Ahmed, and Konstantin Kutzkov. Learning convolutionalneural networks for graphs. In ICML, 2016.

[41] Thomas N. Kipf and Max Welling. Semi-supervised classification with graph convolutionalnetworks. CoRR, abs/1609.02907, 2017.

[42] Michael M. Bronstein, Joan Bruna, Yann LeCun, Arthur Szlam, and Pierre Vandergheynst.Geometric deep learning: Going beyond euclidean data. IEEE Signal Processing Magazine,34:18–42, 2017.

[43] Yujia Li, Daniel Tarlow, Marc Brockschmidt, and Richard S. Zemel. Gated graph sequenceneural networks. CoRR, abs/1511.05493, 2016.

[44] Adam Santoro, David Raposo, David G. T. Barrett, Mateusz Malinowski, Razvan Pascanu,Peter W. Battaglia, and Timothy P. Lillicrap. A simple neural network module for relationalreasoning. In NIPS, 2017.

[45] Peter W. Battaglia, Razvan Pascanu, Matthew Lai, Danilo Jimenez Rezende, and KorayKavukcuoglu. Interaction networks for learning about objects, relations and physics. InNIPS, 2016.

[46] Justin Gilmer, Samuel S. Schoenholz, Patrick F. Riley, Oriol Vinyals, and George E. Dahl.Neural message passing for quantum chemistry. In ICML, 2017.

[47] Michael Chang, Tomer Ullman, Antonio Torralba, and Joshua B. Tenenbaum. A compositionalobject-based approach to learning physical dynamics. CoRR, abs/1612.00341, 2017.

[48] Thomas N. Kipf, Ethan Fetaya, Kuan-Chieh Wang, Max Welling, and Richard S. Zemel. Neuralrelational inference for interacting systems. In ICML, 2018.

[49] Alvaro Sanchez-Gonzalez, Nicolas Heess, Jost Tobias Springenberg, Josh Merel, Martin A.Riedmiller, Raia Hadsell, and Peter W. Battaglia. Graph networks as learnable physics enginesfor inference and control. In ICML, 2018.

[50] Jessica B. Hamrick, Kelsey R. Allen, Victor Bapst, Tina Zhu, Kevin R. McKee, Joshua B.Tenenbaum, and Peter W. Battaglia. Relational inductive bias for physical construction inhumans and machines. CoRR, abs/1806.01203, 2018.

[51] Nicholas Watters, Daniel Zoran, Théophane Weber, Peter W. Battaglia, Razvan Pascanu, andAndrea Tacchetti. Visual interaction networks: Learning a physics simulator from video. InNIPS, 2017.

[52] David Raposo, Adam Santoro, David G. T. Barrett, Razvan Pascanu, Timothy P. Lillicrap, andPeter W. Battaglia. Discovering objects and their relations from entangled scene representations.CoRR, abs/1702.05068, 2017.

[53] Xiaolong Wang, Ross B. Girshick, Abhinav Gupta, and Kaiming He. Non-local neural networks.2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 7794–7803,2018.

15

Page 16: Neural Consciousness Flow - GitHub Pages

[54] Xinlei Chen, Li-Jia Li, Li Fei-Fei, and Abhinav Gupta. Iterative visual reasoning beyondconvolutions. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages7239–7248, 2018.

[55] Adam Santoro, Ryan Faulkner, David Raposo, Jack W. Rae, Mike Chrzanowski, ThéophaneWeber, Daan Wierstra, Oriol Vinyals, Razvan Pascanu, and Timothy P. Lillicrap. Relationalrecurrent neural networks. In NeurIPS, 2018.

[56] Rasmus Berg Palm, Ulrich Paquet, and Ole Winther. Recurrent relational networks. In NeurIPS,2018.

[57] Daniel Oñoro-Rubio, Mathias Niepert, Alberto García-Durán, Roberto Gonzalez, andRoberto Javier López-Sastre. Representation learning for visual-relational knowledge graphs.CoRR, abs/1709.02314, 2017.

[58] Takuo Hamaguchi, Hidekazu Oiwa, Masashi Shimbo, and Yuji Matsumoto. Knowledge transferfor out-of-knowledge-base entities: A graph neural network approach. 2017.

[59] Sainbayar Sukhbaatar, Arthur Szlam, and Rob Fergus. Learning multiagent communicationwith backpropagation. In NIPS, 2016.

[60] Yedid Hoshen. Vain: Attentional multi-agent predictive modeling. In NIPS, 2017.

[61] Miltiadis Allamanis, Marc Brockschmidt, and Mahmoud Khademi. Learning to representprograms with graphs. CoRR, abs/1711.00740, 2018.

[62] Irwan Bello, Hieu Quang Pham, Quoc V. Le, Mohammad Norouzi, and Samy Bengio. Neuralcombinatorial optimization with reinforcement learning. CoRR, abs/1611.09940, 2017.

[63] Alex Nowak, Soledad Villar, Afonso S. Bandeira, and Joan Bruna. A note on learning algorithmsfor quadratic assignment with graph neural networks. CoRR, abs/1706.07450, 2017.

[64] Elias Boutros Khalil, Hanjun Dai, Yuyu Zhang, Bistra N. Dilkina, and Le Song. Learningcombinatorial optimization algorithms over graphs. In NIPS, 2017.

[65] Daniel D. Johnson. Learning graphical state transitions. In ICLR, 2017.

[66] Daniel Selsam, Matthew Lamm, Benedikt Bünz, Percy S. Liang, Leonardo de Moura, andDavid L. Dill. Learning a sat solver from single-bit supervision. CoRR, abs/1802.03685, 2018.

[67] Jessica B. Hamrick, Andrew J. Ballard, Razvan Pascanu, Oriol Vinyals, Nicolas Heess,and Peter W. Battaglia. Metacontrol for adaptive imagination-based optimization. CoRR,abs/1705.02670, 2017.

[68] Razvan Pascanu, Yujia Li, Oriol Vinyals, Nicolas Heess, Lars Buesing, Sébastien Racanière,David P. Reichert, Théophane Weber, Daan Wierstra, and Peter W. Battaglia. Learning model-based planning from scratch. CoRR, abs/1707.06170, 2017.

[69] Tingwu Wang, Renjie Liao, Jimmy Ba, and Sanja Fidler. Nervenet: Learning structured policywith graph neural networks. In ICLR, 2018.

[70] Vinícius Flores Zambaldi, David Raposo, Adam Santoro, Victor Bapst, Yujia Li, IgorBabuschkin, Karl Tuyls, David P. Reichert, Timothy P. Lillicrap, Edward Lockhart, Mur-ray Shanahan, Victoria Langston, Razvan Pascanu, Matthew Botvinick, Oriol Vinyals, andPeter W. Battaglia. Relational deep reinforcement learning. CoRR, abs/1806.01830, 2018.

[71] Sam Toyer, Felipe W. Trevizan, Sylvie Thiébaux, and Lexing Xie. Action schema networks:Generalised policies with deep learning. In AAAI, 2018.

[72] Keyulu Xu, Chengtao Li, Yonglong Tian, Tomohiro Sonobe, Ken ichi Kawarabayashi, andStefanie Jegelka. Representation learning on graphs with jumping knowledge networks. InICML, 2018.

[73] Risi Kondor, Hy Truong Son, Horace Pan, Brandon M. Anderson, and Shubhendu Trivedi.Covariant compositional networks for learning graphs. CoRR, abs/1801.02144, 2018.

16

Page 17: Neural Consciousness Flow - GitHub Pages

[74] Petar Velickovic, Guillem Cucurull, Arantxa Casanova, Alejandro Romero, Pietro Lió, andYoshua Bengio. Graph attention networks. CoRR, abs/1710.10903, 2018.

[75] Wouter Kool. Attention solves your tsp , approximately. 2018.

[76] Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. Neural machine translation by jointlylearning to align and translate. CoRR, abs/1409.0473, 2015.

[77] Zhouhan Lin, Minwei Feng, Cícero Nogueira dos Santos, Mo Yu, Bing Xiang, Bowen Zhou,and Yoshua Bengio. A structured self-attentive sentence embedding. CoRR, abs/1703.03130,2017.

[78] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez,Lukasz Kaiser, and Illia Polosukhin. Attention is all you need. In NIPS, 2017.

[79] Yujia Li, Oriol Vinyals, Chris Dyer, Razvan Pascanu, and Peter W. Battaglia. Learning deepgenerative models of graphs. CoRR, abs/1803.03324, 2018.

[80] Nicola De Cao and Thomas Kipf. Molgan: An implicit generative model for small moleculargraphs. CoRR, abs/1805.11973, 2018.

[81] Jiaxuan You, Zhitao Ying, Xiang Ren, William L. Hamilton, and Jure Leskovec. Graphrnn:Generating realistic graphs with deep auto-regressive models. In ICML, 2018.

[82] Aleksandar Bojchevski, Oleksandr Shchur, Daniel Zügner, and Stephan Günnemann. Netgan:Generating graphs via random walks. In ICML, 2018.

[83] Dai Quoc Nguyen, Tu Dinh Nguyen, Dat Quoc Nguyen, and Dinh Q. Phung. A novel embeddingmodel for knowledge base completion based on convolutional neural network. In NAACL-HLT,2018.

[84] Kristina Toutanova and Danqi Chen. Observed versus latent features for knowledge base andtext inference. 2015.

[85] Farzaneh Mahdisoltani, Joanna Asia Biega, and Fabian M. Suchanek. Yago3: A knowledgebase from multilingual wikipedias. In CIDR, 2014.

[86] Maximilian Nickel, Lorenzo Rosasco, and Tomaso A. Poggio. Holographic embeddings ofknowledge graphs. In AAAI, 2016.

17

Page 18: Neural Consciousness Flow - GitHub Pages

6 Appendix

6.1 Hyperparameter settings

Table 6: The standard hyperparameter settings we use for each dataset plus their training time for oneepoch. For the experimental analysis, we only adjust one hyperparameter and keep the remainingfixed at the standard setting. For NELL995, the training time per epoch means the average time costof the 12 single-query-relation tasks.Hyperparameter FB15K-237 FB15K WN18RR WN18 YAGO3-10 NELL995batch_size 80 80 100 100 100 10n_dims_att 50 50 50 50 50 200n_dims 100 100 100 100 100 200max_sampled_edges_per_step 10000 10000 10000 10000 10000 10000max_attended_nodes_per_step 20 20 20 20 20 100max_sampled_edges_per_node 200 200 200 200 200 1000max_seen_nodes_per_step 200 200 200 200 200 1000n_steps_of_u_flow 2 1 2 1 1 1n_steps_of_c_flow 6 6 8 8 6 5learning_rate 0.001 0.001 0.001 0.001 0.0001 0.001optimizer Adam Adam Adam Adam Adam Adamgrad_clipnorm 1 1 1 1 1 1n_epochs 1 1 1 1 1 3Training time per epoch (h) 25.7 63.7 4.3 8.5 185.0 0.12

Our hyperparameters can be categorized into three groups:

• The normal hyperparameters, including batch_size, n_dims_att, n_dims, learning_rate,grad_clipnorm, and n_epochs. Here, we set a smaller dimension, n_dims_att, for theattention flow computation, as it uses more edges for computation than the message passinguses in the consciousness flow layer, and also intuitively, it does not need to propagate high-dimensional messages but only compute a scalar score for each of the sampled neighbornodes, in concert with the idea in the key-value mechanism [1]. We set n_epochs = 1 inmost cases, indicating that our model needs to be trained only for one epoch due to its fastconvergence.

• The hyperparameters that are in charge of controlling the sampling-attending hori-zon, including max_sampled_edges_per_step that controls the maximum number tosample edges per step per query for the message passing in the unconsciousnessflow layer, and max_sampled_edges_per_node, max_attended_nodes_per_step andmax_seen_nodes_per_step that control the maximum number to sample edges connected toeach current node per step per query, the maximum number of current nodes to attend fromper step per query, and the maximum number of neighbor nodes to attend to per step perquery in the consciousness flow layer.

• The hyperparameters that are in charge of controlling the searching horizon, includingn_steps_of_u_flow representing the number of steps to run the unconcsiousness flow, andn_steps_of_c_flow representing the number of steps to run the consciousness flow.

Note that we choose these hyperparameters not only by their performances but also the computationresources available to us. In some cases, to deal with a very large knowledge graph with limitedresources, we need to make a trade-off between the efficiency and the effectiveness. For example, eachof NELL995’s single-query-relation tasks has a small training set, though still with a large graph, sowe can reduce the batch size in favor of affording larger dimensions and a larger sampling-attendinghorizon without any concern for waiting too long to finish one epoch.

6.2 Other experimental analysis

See Figure 8,9.

18

Page 19: Neural Consciousness Flow - GitHub Pages

0.5 1.0 1.5 2.0 2.5 3.0Epoch

20

25

30

35

40

45

50

55

60

65

Met

ric S

core

(%)

(A) Convergence Analysis (by evaluation on test during training)H@1H@3H@10MMR

H@1 H@3 H@10 MMR25

30

35

40

45

50

55

Met

ric S

core

(%)

(B) U-Flow Component AnalysisW/o U-FlowWith U-Flow

H@1 MMR25.0

27.5

30.0

32.5

35.0

37.5

40.0

42.5

45.0

Met

ric S

core

(%)

(C) Sampling Horizon AnalysisMax-sampled-edges-per-node = 20Max-sampled-edges-per-node = 50Max-sampled-edges-per-node = 100Max-sampled-edges-per-node = 200

H@1 MMR25.0

27.5

30.0

32.5

35.0

37.5

40.0

42.5

45.0

Met

ric S

core

(%)

(D) Attending-to Horizon AnalysisMax-seen-nodes-per-step = 20Max-seen-nodes-per-step = 50Max-seen-nodes-per-step = 100Max-seen-nodes-per-step = 200

H@1 MMR25.0

27.5

30.0

32.5

35.0

37.5

40.0

42.5

45.0

Met

ric S

core

(%)

(E) Attending-from Horizon AnalysisMax-attended-nodes-per-step = 5Max-attended-nodes-per-step = 10Max-attended-nodes-per-step = 20

H@1 MMR15

20

25

30

35

40

45

Met

ric S

core

(%)

(F) Searching Horizon Analysis#Steps-of-C-Flow = 2#Steps-of-C-Flow = 4#Steps-of-C-Flow = 6

Figure 8: Experimental analysis on FB15K-237: (A) During training we pick six model snapshotsat time points of 0.3, 0.5, 0.7, 1, 2, and 3 epochs and evaluate them on test; (B) The w/o U-Flowuses zero step to run U-Flow, while the with U-Flow uses two steps; (C)-(F) are for the sampling,attending and searching horizon analysis based on the standard hyperparameter settings listed in theappendix.

0

5

10

15

20

25

30

35

Trai

ning

Tim

e fo

r One

Epo

ch (h

)

(A) Time Cost for Different Sampling HorizonsMax-sampled-edges-per-node = 20Max-sampled-edges-per-node = 50Max-sampled-edges-per-node = 100Max-sampled-edges-per-node = 200

0

5

10

15

20

25

30

35

Trai

ning

Tim

e fo

r One

Epo

ch (h

)

(B) Time Cost for Different Attending-to HorizonsMax-seen-nodes-per-step = 20Max-seen-nodes-per-step = 50Max-seen-nodes-per-step = 100Max-seen-nodes-per-step = 200

0

5

10

15

20

25

30

35

Trai

ning

Tim

e fo

r One

Epo

ch (h

)

(C) Time Cost for Different Attending-from HorizonsMax-attended-nodes-per-step = 5Max-attended-nodes-per-step = 10Max-attended-nodes-per-step = 20

0

5

10

15

20

25

30

35

Trai

ning

Tim

e fo

r One

Epo

ch (h

)

(D) Time Cost for Different Searching Horizons#Steps-of-C-Flow = 2#Steps-of-C-Flow = 4#Steps-of-C-Flow = 6

0

5

10

15

20

25

30

35

Trai

ning

Tim

e fo

r One

Epo

ch (h

)

(E) Time Cost for Different Batch SizesBatch-size = 50Batch-size = 100Batch-size = 200Batch-size = 300

Figure 9: Analysis of time cost on FB15K-237: (A)-(D) measure the training time for one epochon different horizon settings corresponding to Figure 8(C)-(F); (E) measures the training time forone epoch for different batch sizes using the same horizon setting, which is Max-sampled-edges-per-node=20, Max-seen-nodes-per-step=20, Max-attended-nodes-per-step=20, and #Steps-of-C-Flow=6.

6.3 Other visualization

For the AthletePlaysInLeague taskQuery : ( concept_personnorthamerica_matt_treanor , concept : a th le tep lays in league , concept_sports league_mlb )

Selected key edges :concept_personnorthamerica_matt_treanor , concept : a t h l e t e f l y o u t t o s p o r t s t e a m p o s i t i o n , concept_spor ts teampos i t ion_centerconcept_personnorthamerica_matt_treanor , concept : a t h l e t ep l ayssp o r t , concept_spor t_basebal lconcept_spor ts teampos i t ion_center , concept : a t h l e t e f l y o u t t o s p o r t s t e a m p o s i t i o n _ i n v , concept_personus_orlando_hudsonconcept_spor ts teampos i t ion_center , concept : a t h l e t e f l y o u t t o s p o r t s t e a m p o s i t i o n _ i n v , concept_ath lete_ben_hendr icksonconcept_spor ts teampos i t ion_center , concept : a t h l e t e f l y o u t t o s p o r t s t e a m p o s i t i o n _ i n v , concept_coach_j_j__hardyconcept_spor ts teampos i t ion_center , concept : a t h l e t e f l y o u t t o s p o r t s t e a m p o s i t i o n _ i n v , concept_ath lete_hunter_penceconcept_spor t_basebal l , concept : a t h l e t e p l ay ss po r t _ i n v , concept_personus_orlando_hudsonconcept_spor t_basebal l , concept : a t h l e t e p l ay ss po r t _ i n v , concept_ath lete_ben_hendr icksonconcept_spor t_basebal l , concept : a t h l e t e p l ay ss po r t _ i n v , concept_coach_j_j__hardyconcept_spor t_basebal l , concept : a t h l e t e p l ay ss po r t _ i n v , concept_ath lete_hunter_penceconcept_personus_orlando_hudson , concept : a th le tep lays in league , concept_sports league_mlbconcept_personus_orlando_hudson , concept : a t h l e t ep l aysspo r t , concept_spor t_basebal lconcept_athlete_ben_hendr ickson , concept : coachesinleague , concept_sports league_mlbconcept_athlete_ben_hendr ickson , concept : a t h l e t ep l aysspo r t , concept_spor t_basebal l

19

Page 20: Neural Consciousness Flow - GitHub Pages

concept_personnorthamerica_matt_treanor

concept_sportsteamposition_center

concept_sport_baseball

concept_personus_orlando_hudson

concept_athlete_ben_hendrickson

concept_athlete_hunter_pence

concept_athlete_gary_carter

concept_athlete_kevin_cash

concept_athlete_scott_linebrink

concept_athlete_jeff_kentconcept_athlete_freddy_sanchez

concept_athlete_mike_mussina

concept_athlete_matt_clement

concept_coach_ian_snell

concept_athlete_shin_soo_choo

concept_athlete_justin_verlander

concept_athlete_aaron_hill

concept_coach_alex_gordon

concept_athlete_pelfreyconcept_athlete_rajai_davis

concept_athlete_justin_morneau

concept_coach_j_j__hardy

concept_athlete_steve_carlton

concept_sportsleague_mlb

concept_sportsteam_twins

concept_athlete_kei_igawa

concept_coach_adam_laroche

concept_male_mike_stanton

concept_athlete_jake_peavy

concept_coach_seth_mcclung

concept_athlete_mark_hendrickson

concept_personmexico_jason_giambi

concept_athlete_russell_branyan

concept_personnorthamerica_matt_treanorconcept_sportsteamposition_center

concept_sport_baseball

concept_personus_orlando_hudson

concept_athlete_ben_hendrickson

concept_athlete_hunter_pence

concept_athlete_gary_carter

concept_athlete_kevin_cash

concept_sportsleague_mlb

concept_coach_j_j__hardy

Figure 10: AthletePlaysInLeague. The head is , the query relation is concept:athleteplaysinleague,and the desired tail is . The left is a full subgraph derived with max_attended_nodes_per_step =20, and the right is a further extracted subgraph from the left based on attention. The big yellow noderepresents the head, and the big red node represents the tail. Colors indicate how important a node isattended to in a local subgraph. Grey means less important, yellow means it is more attended duringthe early steps, and red means it is more attended when getting close to the final step.

concept_coach_j_j__hardy , concept : coachesinleague , concept_sports league_mlbconcept_coach_j_j__hardy , concept : a th le tep lays in league , concept_sports league_mlbconcept_coach_j_j__hardy , concept : a t h l e t ep l aysspo r t , concept_spor t_basebal lconcept_athlete_hunter_pence , concept : a th le tep lays in league , concept_sports league_mlbconcept_athlete_hunter_pence , concept : a t h l e t ep l aysspo r t , concept_spor t_basebal lconcept_sportsleague_mlb , concept : coachesinleague_inv , concept_ath lete_ben_hendr icksonconcept_sportsleague_mlb , concept : coachesinleague_inv , concept_coach_j_j__hardy

For the AthleteHomeStadium taskQuery : ( concept_ath le te_el i_manning , concept : athletehomestadium , concept_stadiumoreventvenue_giants_stadium )

Selected key edges :concept_ath le te_el i_manning , concept : personbe longstoorgan izat ion , concept_sportsteam_new_york_giantsconcept_ath le te_el i_manning , concept : a th le tep lays fo r team , concept_sportsteam_new_york_giantsconcept_ath le te_el i_manning , concept : a th le te ledspor ts team , concept_sportsteam_new_york_giantsconcept_ath le te_el i_manning , concept : a th le tep lays in league , concept_spor ts league_nf lconcept_ath le te_el i_manning , concept : fa the ro fperson_ inv , concept_male_archie_manningconcept_sportsteam_new_york_giants , concept : teamplaysinleague , concept_spor ts league_nf lconcept_sportsteam_new_york_giants , concept : teamhomestadium , concept_stadiumoreventvenue_giants_stadiumconcept_sportsteam_new_york_giants , concept : personbe longstoorgan iza t ion_ inv , concept_ath le te_e l i_manningconcept_sportsteam_new_york_giants , concept : a th le tep lays fo r team_ inv , concept_ath le te_e l i_manningconcept_sportsteam_new_york_giants , concept : a th le te ledspor ts team_ inv , concept_ath le te_e l i_manningconcept_spor ts league_nf l , concept : teamplaysin league_inv , concept_sportsteam_new_york_giantsconcept_spor ts league_nf l , concept : agentcompeteswithagent , concept_spor ts league_nf lconcept_spor ts league_nf l , concept : agentcompeteswithagent_inv , concept_spor ts league_nf lconcept_spor ts league_nf l , concept : leaguestadiums , concept_stadiumoreventvenue_giants_stadiumconcept_spor ts league_nf l , concept : a th le tep lays in league_ inv , concept_ath le te_e l i_manningconcept_male_archie_manning , concept : fa thero fperson , concept_ath le te_e l i_manningconcept_spor ts league_nf l , concept : leaguestadiums , concept_stadiumoreventvenue_paul_brown_stadiumconcept_stadiumoreventvenue_giants_stadium , concept : teamhomestadium_inv , concept_sportsteam_new_york_giantsconcept_stadiumoreventvenue_giants_stadium , concept : leaguestadiums_inv , concept_spor ts league_nf lconcept_stadiumoreventvenue_giants_stadium , concept : p roxy fo r_ inv , concep t_c i t y_eas t_ ru the r fo rdconcep t_c i t y_eas t_ ru the r fo rd , concept : p roxy for , concept_stadiumoreventvenue_giants_stadiumconcept_stadiumoreventvenue_paul_brown_stadium , concept : leaguestadiums_inv , concept_spor ts league_nf l

For the AthletePlaysSport taskQuery : ( concept_ath le te_vernon_wel ls , concept : a t h l e t ep l aysspo r t , concept_spor t_basebal l )

20

Page 21: Neural Consciousness Flow - GitHub Pages

concept_athlete_eli_manning

concept_sportsteam_new_york_giants

concept_sportsleague_nfl

concept_male_archie_manning

concept_athlete_joe_bradley

concept_sport_football

concept_awardtrophytournament_super_bowl

concept_stadiumoreventvenue_giants_stadium

concept_sportsteam_patsconcept_city_philadelphia

concept_athlete_barry_bonds

concept_sportsteamposition_center

concept_bone_knee

concept_athlete_blair_betts

concept_personus_jeremy_shockey

concept_athlete_phil_simms

concept_sport_basketball

concept_athlete_rich_auriliaconcept_athlete_justin_tuck

concept_coach_john_mcgraw

concept_athlete_lawrence_taylor

concept_athlete_steve_smith

concept_lake_new

concept_city_east_rutherford

concept_sportsleague_nhl

concept_sportsteam_new_york_jets

concept_city_new_york

concept_person_belichick

concept_athlete_john_cassell

concept_stadiumoreventvenue_paul_brown_stadiumconcept_sportsteam_colts

concept_stadiumoreventvenue_mcafee_coliseum

concept_awardtrophytournament_super_bowl_xlii

concept_stadiumoreventvenue_lucas_oil_stadium

concept_stadiumoreventvenue_meadowlands_stadium

concept_stadiumoreventvenue_gillette_stadium

concept_stadiumoreventvenue_izod_center

concept_stadiumoreventvenue_metrodome

concept_geopoliticalorganization_new_england

concept_geopoliticallocation_national

concept_athlete_ensberg

concept_stateorprovince_states

concept_sport_hockey

concept_stadiumoreventvenue_united_center

concept_athlete_eli_manning

concept_sportsteam_new_york_giants

concept_sportsleague_nfl

concept_male_archie_manning

concept_athlete_joe_bradley

concept_sport_football

concept_stadiumoreventvenue_giants_stadium

concept_lake_new

concept_city_east_rutherford

concept_sportsleague_nhl

concept_stadiumoreventvenue_paul_brown_stadium

concept_stadiumoreventvenue_mcafee_coliseum

concept_stadiumoreventvenue_lucas_oil_stadium

Figure 11: AthleteHomeStadium. The head is concept_athlete_eli_manning, the query relation isconcept:athletehomestadium, and the desired tail is concept_stadiumoreventvenue_giants_stadium.The left is a full subgraph derived with max_attended_nodes_per_step = 20, and the right is afurther extracted subgraph from the left based on attention. The big yellow node represents the head,and the big red node represents the tail. Colors indicate how important a node is attended to in a localsubgraph. Grey means less important, yellow means it is more attended during the early steps, andred means it is more attended when getting close to the final step.

Selected key edges :concept_ath le te_vernon_wel ls , concept : a th le tep lays in league , concept_sports league_mlbconcept_ath le te_vernon_wel ls , concept : coachwontrophy , concept_awardtrophytournament_wor ld_ser iesconcept_ath le te_vernon_wel ls , concept : agen tco l labora tesw i thagen t_ inv , concept_sportsteam_blue_jaysconcept_ath le te_vernon_wel ls , concept : personbe longstoorgan izat ion , concept_sportsteam_blue_jaysconcept_ath le te_vernon_wel ls , concept : a th le tep lays fo r team , concept_sportsteam_blue_jaysconcept_ath le te_vernon_wel ls , concept : a th le te ledspor ts team , concept_sportsteam_blue_jaysconcept_sportsleague_mlb , concept : teamplaysin league_inv , concept_sportsteam_dodgersconcept_sportsleague_mlb , concept : teamplaysin league_inv , concept_sportsteam_yankeesconcept_sportsleague_mlb , concept : teamplaysin league_inv , concep t_spor ts team_p i t t sburgh_p i ra tesconcept_awardtrophytournament_world_ser ies , concept : teamwontrophy_inv , concept_sportsteam_dodgersconcept_awardtrophytournament_world_ser ies , concept : teamwontrophy_inv , concept_sportsteam_yankeesconcept_awardtrophytournament_world_ser ies , concept : awardtrophytournament is thechampionshipgameof thenat ionalspor t ,

concept_spor t_basebal lconcept_awardtrophytournament_world_ser ies , concept : teamwontrophy_inv , concep t_spor ts team_p i t t sburgh_p i ra tesconcept_sportsteam_blue_jays , concept : teamplaysinleague , concept_sports league_mlbconcept_sportsteam_blue_jays , concept : teamplaysagainstteam , concept_sportsteam_yankeesconcept_sportsteam_blue_jays , concept : teamplayssport , concept_spor t_basebal lconcept_sportsteam_dodgers , concept : teamplaysagainstteam , concept_sportsteam_yankeesconcept_sportsteam_dodgers , concept : teamplaysagainst team_inv , concept_sportsteam_yankeesconcept_sportsteam_dodgers , concept : teamwontrophy , concept_awardtrophytournament_wor ld_ser iesconcept_sportsteam_dodgers , concept : teamplayssport , concept_spor t_basebal lconcept_sportsteam_yankees , concept : teamplaysagainstteam , concept_sportsteam_dodgersconcept_sportsteam_yankees , concept : teamplaysagainst team_inv , concept_sportsteam_dodgersconcept_sportsteam_yankees , concept : teamwontrophy , concept_awardtrophytournament_wor ld_ser iesconcept_sportsteam_yankees , concept : teamplayssport , concept_spor t_basebal lconcept_sportsteam_yankees , concept : teamplaysagainstteam , concept_spor ts team_p i t t sburgh_p i ra tesconcept_sportsteam_yankees , concept : teamplaysagainst team_inv , concept_spor ts team_p i t t sburgh_p i ra tesconcept_spor t_basebal l , concept : teamplaysspor t_ inv , concept_sportsteam_dodgersconcept_spor t_basebal l , concept : teamplaysspor t_ inv , concept_sportsteam_yankeesconcept_spor t_basebal l , concept : awardt rophytournament is thechampionshipgameof thenat ionalspor t_ inv ,

concept_awardtrophytournament_wor ld_ser iesconcept_spor t_basebal l , concept : teamplaysspor t_ inv , concep t_spor ts team_p i t t sburgh_p i ra tesconcept_spor ts team_pi t tsburgh_p i ra tes , concept : teamplaysagainstteam , concept_sportsteam_yankeesconcept_spor ts team_pi t tsburgh_p i ra tes , concept : teamplaysagainst team_inv , concept_sportsteam_yankeesconcept_spor ts team_pi t tsburgh_p i ra tes , concept : teamwontrophy , concept_awardtrophytournament_wor ld_ser iesconcept_spor ts team_pi t tsburgh_p i ra tes , concept : teamplayssport , concept_spor t_basebal l

21

Page 22: Neural Consciousness Flow - GitHub Pages

concept_athlete_vernon_wells

concept_sportsleague_mlb

concept_sportsteam_blue_jays

concept_awardtrophytournament_world_series

concept_sportsteamposition_center

concept_sportsteam_yankees

concept_sportsteam_pittsburgh_pirates

concept_sportsteam_dodgers

concept_sportsteam_twins

concept_sport_baseball

concept_sportsteam_white_sox

concept_sportsteam_new_york_metsconcept_sportsteam_red_sox

concept_sportsteam_phillies

concept_sportsteam_cleveland_indians_organizationconcept_sportsteam_los_angeles_dodgers

concept_sportsteam_red_sox_this_seasonconcept_sportsteam_orioles

concept_sportsteam_detroit_tigersconcept_sportsteam_philadelphia_athletics

concept_sportsteam_washington_nationals

concept_sportsteam_mariners

concept_sportsteam_louisville_cardinals

concept_sportsteam_st___louis_cardinals

concept_country_netherlandsconcept_country_china

concept_country___america

concept_sportsgame_n1951_world_series

concept_sportsgame_aldsconcept_stadiumoreventvenue_us_cellular_field

concept_sportsteam_eagles

concept_country_canada_canada

concept_awardtrophytournament_championships

concept_sport_hockey

concept_sport_golf

concept_sport_football

concept_sport_ski

concept_sport_basketballconcept_geopoliticallocation_national

concept_sport_skiingconcept_country_countries

concept_publication_people_

concept_beverage_teaconcept_sport_cricket

concept_beverage_beer

concept_person_william001

concept_sport_soccerconcept_bank_national

concept_bank_china_construction_bank

concept_stateorprovince_states

concept_country_new_zealandconcept_country_u_s_

concept_hobby_hobbies

concept_country_the_united_kingdom

concept_country_thailand

concept_country_france_france

concept_country_switzerland

concept_awardtrophytournament_prizes

concept_country_spain

concept_country_britain

concept_country_wales

concept_country_portugal

concept_country_malaysiaconcept_country_bulgaria

concept_athlete_vernon_wells

concept_sportsleague_mlb

concept_sportsteam_blue_jaysconcept_awardtrophytournament_world_series

concept_sportsteamposition_center

concept_sportsteam_yankeesconcept_sportsteam_pittsburgh_pirates

concept_sportsteam_dodgers

concept_sportsteam_twins

concept_sport_baseball

concept_sport_hockey

concept_sport_golf

concept_sport_football

concept_sport_ski

concept_country_new_zealand

Figure 12: AthletePlaysSport. The head is concept_athlete_vernon_wells, the query relation isconcept:athleteplayssport, and the desired tail is concept_sport_baseball. The left is a full subgraphderived with max_attended_nodes_per_step = 20, and the right is a further extracted subgraphfrom the left based on attention. The big yellow node represents the head, and the big red noderepresents the tail. Colors indicate how important a node is attended to in a local subgraph. Greymeans less important, yellow means it is more attended during the early steps, and red means it ismore attended when getting close to the final step.

For the TeamPlaysSport taskQuery : ( concept_sportsteam_red_wings , concept : teamplayssport , concept_sport_hockey )

Selected key edges :concept_sportsteam_red_wings , concept : teamplaysagainstteam , concept_sportsteam_montreal_canadiensconcept_sportsteam_red_wings , concept : teamplaysagainstteam_inv , concept_sportsteam_montreal_canadiensconcept_sportsteam_red_wings , concept : teamplaysagainstteam , concept_spor ts team_blue_jacketsconcept_sportsteam_red_wings , concept : teamplaysagainstteam_inv , concept_spor ts team_blue_jacketsconcept_sportsteam_red_wings , concept : worksfor_ inv , concep t_a th le te_ l i ds t romconcept_sportsteam_red_wings , concept : o rgan iza t ionh i redperson , concep t_a th le te_ l i ds t romconcept_sportsteam_red_wings , concept : a th le tep lays fo r team_ inv , concep t_a th le te_ l i ds t romconcept_sportsteam_red_wings , concept : a th le te ledspor ts team_ inv , concep t_a th le te_ l i ds t romconcept_sportsteam_montreal_canadiens , concept : teamplaysagainstteam , concept_sportsteam_red_wingsconcept_sportsteam_montreal_canadiens , concept : teamplaysagainst team_inv , concept_sportsteam_red_wingsconcept_sportsteam_montreal_canadiens , concept : teamplaysinleague , concept_spor ts league_nhlconcept_sportsteam_montreal_canadiens , concept : teamplaysagainstteam , concept_spor ts team_leafsconcept_sportsteam_montreal_canadiens , concept : teamplaysagainst team_inv , concept_spor ts team_leafsconcept_sportsteam_blue_jackets , concept : teamplaysagainstteam , concept_sportsteam_red_wingsconcept_sportsteam_blue_jackets , concept : teamplaysagainstteam_inv , concept_sportsteam_red_wingsconcept_sportsteam_blue_jackets , concept : teamplaysinleague , concept_spor ts league_nhlconcep t_a th le te_ l ids t rom , concept : worksfor , concept_sportsteam_red_wingsconcep t_a th le te_ l ids t rom , concept : o rgan iza t ionh i redperson_ inv , concept_sportsteam_red_wingsconcep t_a th le te_ l ids t rom , concept : a th le tep lays fo r team , concept_sportsteam_red_wingsconcep t_a th le te_ l ids t rom , concept : a th le te ledspor ts team , concept_sportsteam_red_wingsconcept_sportsteam_red_wings , concept : teamplaysinleague , concept_spor ts league_nhlconcept_sportsteam_red_wings , concept : teamplaysagainstteam , concept_spor ts team_leafsconcept_sportsteam_red_wings , concept : teamplaysagainstteam_inv , concept_spor ts team_leafsconcept_sports league_nhl , concept : agentcompeteswithagent , concept_spor ts league_nhlconcept_sports league_nhl , concept : agentcompeteswithagent_inv , concept_spor ts league_nhlconcept_sports league_nhl , concept : teamplaysin league_inv , concept_spor ts team_leafsconcept_sportsteam_leafs , concept : teamplaysinleague , concept_spor ts league_nhlconcept_sportsteam_leafs , concept : teamplayssport , concept_sport_hockey

For the OrganizationHeadQuarteredInCity taskQuery : ( concept_company_disney , concept : o rgan i za t i onheadquar te red inc i t y , concept_c i ty_burbank )

22

Page 23: Neural Consciousness Flow - GitHub Pages

concept_sportsteam_red_wings

concept_athlete_lidstrom

concept_sportsteam_blue_jackets

concept_sportsteam_montreal_canadiens

concept_sportsteam_anaheim_ducks

concept_sportsteam_columbus_blue_jackets

concept_athlete_chelios

concept_sportsteam_hawks

concept_sportsteam_edmonton_oilers

concept_sportsteam_chicago_blackhawks

concept_sportsteam_flyers_playoff_ticketsconcept_sportsteam_pittsburgh_penguins

concept_sportsteam_l_a__kings

concept_sportsteam_colorado_avalanche

concept_sportsteam_dallas_stars

concept_stateorprovince_new_york

concept_sportsteam_devils

concept_sportsteam_buffalo_sabres

concept_sportsgame_series

concept_sportsteam_bruins

concept_sportsteam_kings_college

concept_sport_hockey

concept_sportsleague_nhl

concept_sport_basketball

concept_sportsteam_leafsconcept_stadiumoreventvenue_joe_louis_arena

concept_sportsteam_blackhawks

concept_sportsteam_rangers

concept_sportsteam_capitals

concept_sportsteam_new_york_islanders

concept_sportsteam_oilers

concept_sportsteam_ottawa_senators

concept_sportsteam_spurs

concept_sportsteam_tampa_bay_lightning

concept_country___america

concept_country_republic_of_india

concept_country_canada_canada

concept_sportsequipment_ball

concept_awardtrophytournament_championshipsconcept_sportsequipment_clubs

concept_sportsequipment_hockey_sticks

concept_stadiumoreventvenue_td_banknorth_garden

concept_tool_accessories

concept_stadiumoreventvenue_pete_times_forum

concept_stadiumoreventvenue_staples_center

concept_sportsteamposition_forward

concept_hobby_hobbies

concept_sport_football

concept_sport_baseballconcept_sport_golf

concept_company_national

concept_sport_soccer

concept_country_usaconcept_sport_wrestling

concept_sport_team concept_sport_fitness

concept_sportsteam_packers

concept_hobby_hiking

concept_sport_sports

concept_hobby_fishing

concept_country_england

concept_country_iraq

concept_country_russia

concept_coach_billy_butler

concept_sportsteamposition_quarterback

concept_sportsteam_red_wings

concept_athlete_lidstrom

concept_sportsteam_blue_jackets

concept_sportsteam_montreal_canadiens

concept_sportsteam_anaheim_ducksconcept_sportsteam_columbus_blue_jackets

concept_sport_hockeyconcept_sportsleague_nhl

concept_sport_basketball

concept_sportsteam_leafs

concept_country___america

concept_sport_football

concept_sport_baseball

concept_sport_golf

Figure 13: TeamPlaysSport. The head is concept_sportsteam_red_wings, the query relation isconcept:teamplayssport, and the desired tail is concept_sport_hockey. The left is a full subgraphderived with max_attended_nodes_per_step = 20, and the right is a further extracted subgraphfrom the left based on attention. The big yellow node represents the head, and the big red noderepresents the tail. Colors indicate how important a node is attended to in a local subgraph. Greymeans less important, yellow means it is more attended during the early steps, and red means it ismore attended when getting close to the final step.

Selected key edges :concept_company_disney , concept : headquarteredin , concept_c i ty_burbankconcept_company_disney , concept : subpar to fo rgan i za t i on_ inv , concept_website_networkconcept_company_disney , concept : worksfor_ inv , concept_ceo_rober t_ igerconcept_company_disney , concept : p roxy fo r_ inv , concept_ceo_rober t_ igerconcept_company_disney , concept : person leadsorgan iza t ion_ inv , concept_ceo_rober t_ igerconcept_company_disney , concept : ceoof_inv , concept_ceo_rober t_ igerconcept_company_disney , concept : person leadsorgan iza t ion_ inv , concept_ceo_je f f rey_katzenbergconcept_company_disney , concept : o rgan iza t ionh i redperson , concept_ceo_je f f rey_katzenbergconcept_company_disney , concept : o rgan iza t ion terminatedperson , concept_ceo_je f f rey_katzenbergconcept_ci ty_burbank , concept : headquar teredin_inv , concept_company_disneyconcept_ci ty_burbank , concept : headquar teredin_inv , concept_biotechcompany_the_walt_disney_co_concept_website_network , concept : subpar to fo rgan iza t i on , concept_company_disneyconcept_ceo_rober t_ iger , concept : worksfor , concept_company_disneyconcept_ceo_rober t_ iger , concept : p roxy for , concept_company_disneyconcept_ceo_rober t_ iger , concept : person leadsorgan izat ion , concept_company_disneyconcept_ceo_rober t_ iger , concept : ceoof , concept_company_disneyconcept_ceo_rober t_ iger , concept : topmemberoforganizat ion , concept_biotechcompany_the_walt_disney_co_concept_ceo_rober t_ iger , concept : o rgan iza t ion te rmina tedperson_ inv , concept_biotechcompany_the_walt_disney_co_concept_ceo_jef f rey_katzenberg , concept : person leadsorgan iza t ion , concept_company_disneyconcept_ceo_jef f rey_katzenberg , concept : o rgan iza t ionh i redperson_ inv , concept_company_disneyconcept_ceo_jef f rey_katzenberg , concept : o rgan iza t ion te rmina tedperson_ inv , concept_company_disneyconcept_ceo_jef f rey_katzenberg , concept : worksfor , concept_recordlabel_dreamworks_skgconcept_ceo_jef f rey_katzenberg , concept : topmemberoforganizat ion , concept_recordlabel_dreamworks_skgconcept_ceo_jef f rey_katzenberg , concept : o rgan iza t ion te rmina tedperson_ inv , concept_recordlabel_dreamworks_skgconcept_ceo_jef f rey_katzenberg , concept : ceoof , concept_recordlabel_dreamworks_skgconcept_biotechcompany_the_walt_disney_co_ , concept : headquarteredin , concept_c i ty_burbankconcept_biotechcompany_the_walt_disney_co_ , concept : o rgan i za t i onheadquar te red inc i t y , concept_c i ty_burbankconcept_recordlabel_dreamworks_skg , concept : worksfor_ inv , concept_ceo_je f f rey_katzenbergconcept_recordlabel_dreamworks_skg , concept : topmemberoforganizat ion_inv , concept_ceo_je f f rey_katzenbergconcept_recordlabel_dreamworks_skg , concept : o rgan iza t ion terminatedperson , concept_ceo_je f f rey_katzenbergconcept_recordlabel_dreamworks_skg , concept : ceoof_inv , concept_ceo_je f f rey_katzenbergconcept_ci ty_burbank , concept : a i r p o r t i n c i t y _ i n v , concept_transportat ion_burbank_glendale_pasadenaconcept_transportat ion_burbank_glendale_pasadena , concept : a i r p o r t i n c i t y , concept_c i ty_burbank

For the WorksFor taskQuery : ( concept_sc ien t is t_ba lmer , concept : worksfor , concep t_un i ve rs i t y_m ic roso f t )

23

Page 24: Neural Consciousness Flow - GitHub Pages

concept_company_disney

concept_city_burbank

concept_ceo_robert_iger

concept_ceo_jeffrey_katzenberg

concept_city_abc

concept_ceo_michael_eisnerconcept_politicsblog_rights

concept_blog_espn_the_magazine

concept_website_network

concept_academicfield_media

concept_politicsissue_entertainment

concept_publication_espn

concept_company_abc_television_network

concept_personaustralia_jobs

concept_company_club_penguin

concept_televisionnetwork_abc

concept_website_infoseek

concept_company_pixar

concept_person_disney

concept_biotechcompany_the_walt_disney_co_

concept_city_new_york

concept_recordlabel_dreamworks_skg

concept_company_disney_feature_animation

concept_ceo_george_bodenheimerconcept_company_walt_disney

concept_city_emeryville

concept_personeurope_disney

concept_university_search

concept_personus_david_geffen

concept_sportsleague_espn

concept_personus_steven_spielberg

concept_musicartist_toy

concept_person_steven_spielberg

concept_city_lego

concept_personaustralia_david_geffen

concept_transportation_burbank_glendale_pasadena

concept_company_asylum

concept_politicianus_rudy_giuliani

concept_company_walt_disney_world_resort

concept_stateorprovince_illinois

concept_company_walt_disney_world

concept_company_disney

concept_city_burbank

concept_ceo_robert_iger

concept_ceo_jeffrey_katzenberg

concept_city_abc

concept_ceo_michael_eisner

concept_biotechcompany_the_walt_disney_co_

concept_city_new_york

concept_recordlabel_dreamworks_skg

concept_website_network

concept_ceo_george_bodenheimer

concept_transportation_burbank_glendale_pasadena

Figure 14: OrganizationHeadQuarteredInCity. The head is concept_company_disney, the queryrelation is concept:organizationheadquarteredincity, and the desired tail is concept_city_burbank.The left is a full subgraph derived with max_attended_nodes_per_step = 20, and the right is afurther extracted subgraph from the left based on attention. The big yellow node represents the head,and the big red node represents the tail. Colors indicate how important a node is attended to in a localsubgraph. Grey means less important, yellow means it is more attended during the early steps, andred means it is more attended when getting close to the final step.

Selected key edges :concept_sc ien t is t_ba lmer , concept : topmemberoforganizat ion , concept_company_microsoftconcept_sc ien t is t_ba lmer , concept : o rgan iza t ion te rmina tedperson_ inv , concep t_un i ve rs i t y_m ic roso f tconcept_company_microsoft , concept : topmemberoforganizat ion_inv , concept_personus_steve_bal lmerconcept_company_microsoft , concept : topmemberoforganizat ion_inv , concep t_sc ien t i s t_ba lmerconcep t_un ive rs i t y_mic roso f t , concept : agentco l labora teswi thagent , concept_personus_steve_bal lmerconcep t_un ive rs i t y_mic roso f t , concept : person leadsorgan iza t ion_ inv , concept_personus_steve_bal lmerconcep t_un ive rs i t y_mic roso f t , concept : person leadsorgan iza t ion_ inv , concep t_person_b i l lconcep t_un ive rs i t y_mic roso f t , concept : o rgan iza t ion terminatedperson , concep t_sc ien t i s t_ba lmerconcep t_un ive rs i t y_mic roso f t , concept : person leadsorgan iza t ion_ inv , concept_person_robbie_bachconcept_personus_steve_bal lmer , concept : topmemberoforganizat ion , concept_company_microsoftconcept_personus_steve_bal lmer , concept : agen tco l labora tesw i thagen t_ inv , concep t_un i ve rs i t y_m ic roso f tconcept_personus_steve_bal lmer , concept : person leadsorgan izat ion , concep t_un i ve rs i t y_m ic roso f tconcept_personus_steve_bal lmer , concept : worksfor , concep t_un i ve rs i t y_m ic roso f tconcept_personus_steve_bal lmer , concept : p roxy for , c o n c e p t _ r e t a i l s t o r e _ m i c r o s o f tconcept_personus_steve_bal lmer , concept : subpar to f , c o n c e p t _ r e t a i l s t o r e _ m i c r o s o f tconcept_personus_steve_bal lmer , concept : agentcont ro ls , c o n c e p t _ r e t a i l s t o r e _ m i c r o s o f tconcept_person_b i l l , concept : person leadsorgan izat ion , concep t_un i ve rs i t y_m ic roso f tconcept_person_b i l l , concept : worksfor , concep t_un i ve rs i t y_m ic roso f tconcept_person_robbie_bach , concept : person leadsorgan izat ion , concep t_un i ve rs i t y_m ic roso f tconcept_person_robbie_bach , concept : worksfor , concep t_un i ve rs i t y_m ic roso f tc o n c e p t _ r e t a i l s t o r e _ m i c r o s o f t , concept : p roxy fo r_ inv , concept_personus_steve_bal lmerc o n c e p t _ r e t a i l s t o r e _ m i c r o s o f t , concept : subpar to f_ inv , concept_personus_steve_bal lmerc o n c e p t _ r e t a i l s t o r e _ m i c r o s o f t , concept : agen tcon t ro ls_ inv , concept_personus_steve_bal lmer

For the PersonBornInLocation taskQuery : ( concept_person_mark001 , concept : personborn in loca t ion , concept_county_york_c i ty )

Selected key edges :concept_person_mark001 , concept : persongraduatedf romunivers i ty , concep t_un i ve rs i t y_co l l egeconcept_person_mark001 , concept : persongraduatedschool , concep t_un i ve rs i t y_co l l egeconcept_person_mark001 , concept : persongraduatedf romunivers i ty , c o n c e p t _ u n i v e r s i t y _ s t a t e _ u n i v e r s i t yconcept_person_mark001 , concept : persongraduatedschool , c o n c e p t _ u n i v e r s i t y _ s t a t e _ u n i v e r s i t yconcept_person_mark001 , concept : personborn inc i t y , concept_ci ty_hampshireconcept_person_mark001 , concept : hasspouse , concept_person_diane001concept_person_mark001 , concept : hasspouse_inv , concept_person_diane001

24

Page 25: Neural Consciousness Flow - GitHub Pages

concept_scientist_balmer

concept_university_microsoft

concept_company_microsoft

concept_personus_steve_ballmer

concept_person_robbie_bach

concept_politician_jobs

concept_person_bill

concept_personaustralia_paul_allen

concept_ceo_steve_ballmerconcept_company_gates

concept_buildingfeature_windows

concept_date_bill

concept_website_download

concept_product_powerpoint

concept_charactertrait_vista

concept_product_word_documents

concept_mlalgorithm_microsoft_word

concept_museum_steve

concept_city_outlook

concept_emotion_word

concept_consumerelectronicitem_ms_word

concept_hallwayitem_access

concept_personmexico_ryan_whitney

concept_retailstore_microsoft

concept_company_adobe

concept_beverage_new

concept_sportsteam_harvard_divinity_school

concept_coach_vulcan_inc

concept_sportsteam_state_universityconcept_biotechcompany_microsoft_corp

concept_person_edwards

concept_company_clinton

concept_politicianus_rodham_clinton

concept_geopoliticallocation_kerry

concept_person_mccain

concept_governmentorganization_representativesconcept_politicalparty_senate

concept_politicalparty_college

concept_governmentorganization_house

concept_personus_paul_allen

concept_coach_tim_murphy

concept_software_microsoft_excel

concept_ceo_paul_allen

concept_sportsteam_new_york_jets

concept_company_sunconcept_software_office_2003

concept_company_yahoo001

concept_female_hillary

concept_personafrica_george_bush

concept_stateorprovince_last_year

concept_university_yahoo

concept_geopoliticallocation_world

concept_university_google

concept_company_microsoft_corporation

concept_automobilemaker_jeff_bezos

concept_scientist_balmer

concept_university_microsoft

concept_company_microsoft

concept_personus_steve_ballmer

concept_person_robbie_bachconcept_politician_jobs

concept_person_bill

concept_personmexico_ryan_whitney

concept_retailstore_microsoft

concept_company_adobe

concept_personaustralia_paul_allen

concept_sportsteam_harvard_divinity_school

concept_sportsteam_state_university

Figure 15: WorksFor. The head is concept_scientist_balmer, the query relation is concept:worksfor,and the desired tail is concept_university_microsoft. The left is a full subgraph derived withmax_attended_nodes_per_step = 20, and the right is a further extracted subgraph from theleft based on attention. The big yellow node represents the head, and the big red node representsthe tail. Colors indicate how important a node is attended to in a local subgraph. Grey means lessimportant, yellow means it is more attended during the early steps, and red means it is more attendedwhen getting close to the final step.

concep t_un ive rs i t y_co l l ege , concept : persongraduatedf romunivers i ty_ inv , concept_person_mark001concep t_un ive rs i t y_co l l ege , concept : persongraduatedschool_inv , concept_person_mark001concep t_un ive rs i t y_co l l ege , concept : persongraduatedf romunivers i ty_ inv , concep t_person_b i l lconcep t_un ive rs i t y_co l l ege , concept : persongraduatedschool_inv , concep t_person_b i l lc o n c e p t _ u n i v e r s i t y _ s t a t e _ u n i v e r s i t y , concept : persongraduated f romunivers i ty_ inv , concept_person_mark001c o n c e p t _ u n i v e r s i t y _ s t a t e _ u n i v e r s i t y , concept : persongraduatedschool_inv , concept_person_mark001c o n c e p t _ u n i v e r s i t y _ s t a t e _ u n i v e r s i t y , concept : persongraduated f romunivers i ty_ inv , concep t_person_b i l lc o n c e p t _ u n i v e r s i t y _ s t a t e _ u n i v e r s i t y , concept : persongraduatedschool_inv , concep t_person_b i l lconcept_ci ty_hampshire , concept : pe rsonborn inc i t y_ inv , concept_person_mark001concept_person_diane001 , concept : persongraduatedf romunivers i ty , c o n c e p t _ u n i v e r s i t y _ s t a t e _ u n i v e r s i t yconcept_person_diane001 , concept : persongraduatedschool , c o n c e p t _ u n i v e r s i t y _ s t a t e _ u n i v e r s i t yconcept_person_diane001 , concept : hasspouse , concept_person_mark001concept_person_diane001 , concept : hasspouse_inv , concept_person_mark001concept_person_diane001 , concept : personborn in loca t ion , concept_county_york_c i tyc o n c e p t _ u n i v e r s i t y _ s t a t e _ u n i v e r s i t y , concept : persongraduated f romunivers i ty_ inv , concept_person_diane001c o n c e p t _ u n i v e r s i t y _ s t a t e _ u n i v e r s i t y , concept : persongraduatedschool_inv , concept_person_diane001concept_person_b i l l , concept : pe rsonborn inc i t y , concept_c i ty_yorkconcept_person_b i l l , concept : personborn in loca t ion , concept_c i ty_yorkconcept_person_b i l l , concept : persongraduatedf romunivers i ty , concep t_un i ve rs i t y_co l l egeconcept_person_b i l l , concept : persongraduatedschool , concep t_un i ve rs i t y_co l l egeconcept_person_b i l l , concept : persongraduatedf romunivers i ty , c o n c e p t _ u n i v e r s i t y _ s t a t e _ u n i v e r s i t yconcept_person_b i l l , concept : persongraduatedschool , c o n c e p t _ u n i v e r s i t y _ s t a t e _ u n i v e r s i t yconcept_c i ty_york , concept : pe rsonborn inc i t y_ inv , concep t_person_b i l lconcept_c i ty_york , concept : pe rsonborn inc i t y_ inv , concept_person_diane001concep t_un ive rs i t y_co l l ege , concept : persongraduatedf romunivers i ty_ inv , concept_person_diane001concept_person_diane001 , concept : pe rsonborn inc i t y , concept_c i ty_york

For the PersonLeadsOrganization taskQuery : ( c o n c e p t _ j o u r n a l i s t _ b i l l _ p l a n t e , concept : person leadsorgan izat ion , concept_company_cnn__pbs )

Selected key edges :c o n c e p t _ j o u r n a l i s t _ b i l l _ p l a n t e , concept : worksfor , concept_ te lev is ionnetwork_cbsc o n c e p t _ j o u r n a l i s t _ b i l l _ p l a n t e , concept : agen tco l labora tesw i thagent_ inv , concept_ te lev is ionnetwork_cbsconcept_ te lev is ionnetwork_cbs , concept : worksfor_ inv , c o n c e p t _ j o u r n a l i s t _ w a l t e r _ c r o n k i t econcept_ te lev is ionnetwork_cbs , concept : agentco l labora teswi thagent , c o n c e p t _ j o u r n a l i s t _ w a l t e r _ c r o n k i t econcept_ te lev is ionnetwork_cbs , concept : worksfor_ inv , concept_personus_scot t_pe l leyconcept_ te lev is ionnetwork_cbs , concept : worksfor_ inv , concept_ac tor_dan ie l_schor r

25

Page 26: Neural Consciousness Flow - GitHub Pages

concept_person_mark001

concept_university_collegeconcept_university_state_university

concept_person_diane001

concept_male_world

concept_sportsteam_state_university

concept_politicalparty_college

10

concept_university_syracuse_university

concept_stateorprovince_california

concept_university_high_school

concept_sportsgame_series

concept_city_hampshire

concept_stateorprovince_georgia

concept_geopoliticallocation_world

concept_stateorprovince_massachusettsconcept_stateorprovince_maine

concept_stateorprovince_illinoisconcept_jobposition_king

concept_charactertrait_world

concept_county_york_city

concept_person_bill

concept_person_david

concept_person_greg001

concept_person_james001concept_politician_jobs

concept_language_english

concept_person_michael002

concept_person_joe002concept_person_aaron_brooks

concept_journalist_dan

concept_person_john003

concept_person_andrew001

concept_person_kevin

concept_person_jim

concept_person_adam001

concept_academicfield_science

concept_city_york

concept_lake_new

concept_buildingfeature_american

concept_city_new_yorkconcept_building_the_metropolitan

concept_stateorprovince_new_york

concept_building_metropolitan

concept_room_contemporary

concept_river_arts

concept_country_orleans

concept_governmentorganization_federal

concept_geopoliticallocation_state

concept_personeurope_whitney

concept_person_sean002

concept_person_charles001

concept_person_princess

concept_river_state

concept_person_robert003

concept_female_mary

concept_person_prince

concept_sportsleague_newconcept_book_new

concept_country_monaco

concept_country_luxembourg

concept_writer_new

concept_musicinstrument_guitar

concept_country_brazil

concept_country_sweden

concept_geopoliticalorganization_wurttemberg

concept_country_mecklenburg

concept_country_romania

concept_person_mark001

concept_university_college

concept_university_state_university

concept_person_diane001

concept_male_world

concept_sportsteam_state_university

concept_county_york_city

concept_person_bill

concept_university_syracuse_university

concept_person_david

concept_city_york

concept_lake_new

concept_city_hampshire

concept_person_adam001

concept_country_orleans

Figure 16: PersonBornInLocation. The head is concept_person_mark001, the query relation isconcept:personborninlocation, and the desired tail is concept_county_york_city. The left is a fullsubgraph derived with max_attended_nodes_per_step = 20, and the right is a further extractedsubgraph from the left based on attention. The big yellow node represents the head, and the big rednode represents the tail. Colors indicate how important a node is attended to in a local subgraph.Grey means less important, yellow means it is more attended during the early steps, and red means itis more attended when getting close to the final step.

concept_ te lev is ionnetwork_cbs , concept : worksfor_ inv , concept_person_edward_r__murrowconcept_ te lev is ionnetwork_cbs , concept : agentco l labora teswi thagent , concept_person_edward_r__murrowconcept_ te lev is ionnetwork_cbs , concept : worksfor_ inv , c o n c e p t _ j o u r n a l i s t _ b i l l _ p l a n t econcept_ te lev is ionnetwork_cbs , concept : agentco l labora teswi thagent , c o n c e p t _ j o u r n a l i s t _ b i l l _ p l a n t ec o n c e p t _ j o u r n a l i s t _ w a l t e r _ c r o n k i t e , concept : worksfor , concept_ te lev is ionnetwork_cbsc o n c e p t _ j o u r n a l i s t _ w a l t e r _ c r o n k i t e , concept : agen tco l labora tesw i thagen t_ inv , concept_ te lev is ionnetwork_cbsc o n c e p t _ j o u r n a l i s t _ w a l t e r _ c r o n k i t e , concept : worksfor , concept_nonpro f i to rgan iza t ion_cbs_even ingconcept_personus_scot t_pel ley , concept : worksfor , concept_ te lev is ionnetwork_cbsconcept_personus_scot t_pel ley , concept : person leadsorgan izat ion , concept_ te lev is ionnetwork_cbsconcept_personus_scot t_pel ley , concept : person leadsorgan izat ion , concept_company_cnn__pbsconcept_actor_danie l_schor r , concept : worksfor , concept_ te lev is ionnetwork_cbsconcept_actor_danie l_schor r , concept : person leadsorgan iza t ion , concept_ te lev is ionnetwork_cbsconcept_actor_danie l_schor r , concept : person leadsorgan iza t ion , concept_company_cnn__pbsconcept_person_edward_r__murrow , concept : worksfor , concept_ te lev is ionnetwork_cbsconcept_person_edward_r__murrow , concept : agen tco l labora tesw i thagen t_ inv , concept_ te lev is ionnetwork_cbsconcept_person_edward_r__murrow , concept : person leadsorgan izat ion , concept_ te lev is ionnetwork_cbsconcept_person_edward_r__murrow , concept : person leadsorgan izat ion , concept_company_cnn__pbsconcept_ te lev is ionnetwork_cbs , concept : o rgan i za t i onheadquar te red inc i t y , concept_ci ty_new_yorkconcept_ te lev is ionnetwork_cbs , concept : headquarteredin , concept_ci ty_new_yorkconcept_ te lev is ionnetwork_cbs , concept : agentco l labora teswi thagent , concept_personeurope_wi l l iam_paleyconcept_ te lev is ionnetwork_cbs , concept : topmemberoforganizat ion_inv , concept_personeurope_wi l l iam_paleyconcept_company_cnn__pbs , concept : headquarteredin , concept_ci ty_new_yorkconcept_company_cnn__pbs , concept : personbe longstoorgan iza t ion_ inv , concept_personeurope_wi l l iam_paleyconcept_nonpro f i to rgan iza t ion_cbs_even ing , concept : worksfor_ inv , c o n c e p t _ j o u r n a l i s t _ w a l t e r _ c r o n k i t econcept_ci ty_new_york , concept : o rgan i za t i onheadquar te red inc i t y_ inv , concept_ te lev is ionnetwork_cbsconcept_ci ty_new_york , concept : headquarteredin_inv , concept_ te lev is ionnetwork_cbsconcept_ci ty_new_york , concept : headquarteredin_inv , concept_company_cnn__pbsconcept_personeurope_wi l l iam_paley , concept : agen tco l labora tesw i thagen t_ inv , concept_ te lev is ionnetwork_cbsconcept_personeurope_wi l l iam_paley , concept : topmemberoforganizat ion , concept_ te lev is ionnetwork_cbsconcept_personeurope_wi l l iam_paley , concept : personbelongstoorgan izat ion , concept_company_cnn__pbsconcept_personeurope_wi l l iam_paley , concept : person leadsorgan iza t ion , concept_company_cnn__pbs

For the OrganizationHiredPerson taskQuery : ( concept_s ta teorprov ince_af te rnoon , concept : o rgan iza t ionh i redperson , concept_personmexico_ryan_whitney )

Selected key edges :concept_s ta teorprov ince_af te rnoon , concept : atdate , concep t_da te l i t e ra l_n2007

26

Page 27: Neural Consciousness Flow - GitHub Pages

concept_journalist_bill_plante

concept_televisionnetwork_cbs

concept_person_edward_r__murrow

concept_personus_scott_pelley

concept_actor_daniel_schorr

concept_journalist_wyatt_andrews

concept_athlete_anthony_mason

concept_journalist_connie_chung

concept_journalist_andy_rooney

concept_journalist_ed_bradley

concept_writer_bill_geist

concept_person_byron_pitts

concept_personaustralia_harry_smith

concept_journalist_morley_safer

concept_journalist_gloria_borger

concept_writer_bernard_goldberg

concept_person_eric_sevareid

concept_journalist_lesley_stahl

concept_personmexico_david_martin

concept_journalist_walter_cronkite

concept_personaustralia_phil_jones concept_company_cnn__pbs

concept_nonprofitorganization_cbs_evening

concept_musicartist_television

concept_crustacean_tv

concept_website_cbs_evening_news

concept_city_new_york

concept_personeurope_william_paley

concept_coach_douglas_edwards

concept_academicfield_media

concept_person_sean_mcmanus

concept_televisionstation_wcscconcept_comedian_leslie_moonves

concept_person_nina_tasslerconcept_televisionstation_kion_tv

concept_televisionstation_wowk_tv

concept_televisionstation_kbci_tv

concept_televisionstation_wjtv

concept_politicsissue_entertainment

concept_person_les_moonves

concept_televisionstation_kfda

concept_televisionstation_wdtv

concept_televisionstation_khqa_tv

concept_televisionstation_wrbl

concept_televisionstation_wltx_tv

concept_website_cbs

concept_company_cbs_corp_

concept_recordlabel_universal

concept_blog_mtv

concept_journalist_bill_plante

concept_televisionnetwork_cbs

concept_person_edward_r__murrow

concept_personus_scott_pelley

concept_actor_daniel_schorr

concept_journalist_wyatt_andrews

concept_athlete_anthony_mason

concept_company_cnn__pbs

concept_nonprofitorganization_cbs_evening

concept_musicartist_television

concept_crustacean_tv

concept_journalist_walter_cronkite

concept_city_new_york

concept_personeurope_william_paley

concept_coach_douglas_edwards

concept_academicfield_media

concept_website_cbs

concept_televisionstation_wcsc

Figure 17: PersonLeadsOrganization. The head is concept_journalist_bill_plante, the query rela-tion is concept:organizationheadquarteredincity, and the desired tail is concept_company_cnn__pbs.The left is a full subgraph derived with max_attended_nodes_per_step = 20, and the right is afurther extracted subgraph from the left based on attention. The big yellow node represents the head,and the big red node represents the tail. Colors indicate how important a node is attended to in a localsubgraph. Grey means less important, yellow means it is more attended during the early steps, andred means it is more attended when getting close to the final step.

concept_s ta teorprov ince_af te rnoon , concept : atdate , concept_date_n2003concept_s ta teorprov ince_af te rnoon , concept : atdate , concep t_da te l i t e ra l_n2006concept_date l i te ra l_n2007 , concept : a tdate_ inv , concept_count ry_un i ted_sta tesconcept_date l i te ra l_n2007 , concept : a tdate_ inv , concept_city_homeconcept_date l i te ra l_n2007 , concept : a tdate_ inv , concep t_c i t y_se rv i ceconcept_date l i te ra l_n2007 , concept : a tdate_ inv , concep t_coun t r y_ l e f t _pa r t i esconcept_date_n2003 , concept : a tdate_ inv , concept_count ry_un i ted_s ta tesconcept_date_n2003 , concept : a tdate_ inv , concept_city_homeconcept_date_n2003 , concept : a tdate_ inv , concep t_c i t y_se rv i ceconcept_date_n2003 , concept : a tdate_ inv , concep t_coun t r y_ l e f t _pa r t i esconcept_date l i te ra l_n2006 , concept : a tdate_ inv , concept_count ry_un i ted_sta tesconcept_date l i te ra l_n2006 , concept : a tdate_ inv , concept_city_homeconcept_date l i te ra l_n2006 , concept : a tdate_ inv , concep t_c i t y_se rv i ceconcept_date l i te ra l_n2006 , concept : a tdate_ inv , concep t_coun t r y_ l e f t _pa r t i esconcept_count ry_un i ted_sta tes , concept : atdate , concept_year_n1992concept_count ry_un i ted_sta tes , concept : atdate , concept_year_n1997concept_count ry_un i ted_sta tes , concept : o rgan iza t ionh i redperson , concept_personmexico_ryan_whitneyconcept_city_home , concept : atdate , concept_year_n1992concept_city_home , concept : atdate , concept_year_n1997concept_city_home , concept : o rgan iza t ionh i redperson , concept_personmexico_ryan_whitneyconcep t_c i t y_serv ice , concept : atdate , concept_year_n1992concep t_c i t y_serv ice , concept : atdate , concept_year_n1997concep t_c i t y_serv ice , concept : o rgan iza t ionh i redperson , concept_personmexico_ryan_whitneyconcep t_coun t r y_ le f t _pa r t i es , concept : worksfor_ inv , concept_personmexico_ryan_whitneyconcep t_coun t r y_ le f t _pa r t i es , concept : o rgan iza t ionh i redperson , concept_personmexico_ryan_whitneyconcept_year_n1992 , concept : a tdate_ inv , concept_governmentorganizat ion_houseconcept_year_n1992 , concept : a tdate_ inv , concept_count ry_un i ted_sta tesconcept_year_n1992 , concept : a tdate_ inv , concept_city_homeconcept_year_n1992 , concept : a tdate_ inv , concept_tradeunion_congressconcept_year_n1997 , concept : a tdate_ inv , concept_governmentorganizat ion_houseconcept_year_n1997 , concept : a tdate_ inv , concept_count ry_un i ted_sta tesconcept_year_n1997 , concept : a tdate_ inv , concept_city_homeconcept_personmexico_ryan_whitney , concept : worksfor , concept_governmentorganizat ion_houseconcept_personmexico_ryan_whitney , concept : worksfor , concept_tradeunion_congressconcept_personmexico_ryan_whitney , concept : worksfor , concep t_coun t r y_ l e f t _pa r t i e sconcept_governmentorganizat ion_house , concept : personbe longstoorgan iza t ion_ inv , concept_personus_partyconcept_governmentorganizat ion_house , concept : worksfor_ inv , concept_personmexico_ryan_whitneyconcept_governmentorganizat ion_house , concept : o rgan iza t ionh i redperson , concept_personmexico_ryan_whitney

27

Page 28: Neural Consciousness Flow - GitHub Pages

concept_stateorprovince_afternoon

concept_dateliteral_n2007

concept_dateliteral_n2006concept_date_n2003

concept_dateliteral_n2002

concept_dateliteral_n2008

concept_dateliteral_n2005concept_date_n2004concept_date_n2001

concept_date_n1996

concept_date_n1999concept_year_n1991

concept_year_n1998concept_date_n2000

concept_book_new

concept_lake_new

concept_university_national

concept_city_home

concept_country_united_states

concept_city_service

concept_governmentorganization_law

concept_governmentorganization_program

concept_book_morning

concept_programminglanguage_project

concept_academicfield_directors

concept_musicsong_nightconcept_website_tour

concept_country_israel

concept_musicsong_end

concept_book_years

concept_country_left_parties

concept_website_visit

concept_governmentorganization_fire

concept_academicfield_trial

concept_website_trip

concept_governmentorganization_launch

concept_company_case

concept_personmexico_ryan_whitney

concept_year_n1997

concept_year_n1992

concept_year_n1994concept_year_n1995

concept_date_n1993

concept_date_n2009

concept_year_n1986concept_date_n1968

concept_year_n1989

concept_year_n1975

concept_year_n1982

concept_director_committee

concept_dayofweek_wednesday

concept_date_n1944

concept_dateliteral_n1990

concept_year_n1978

concept_year_n1967concept_dateliteral_n1987

concept_tradeunion_congress

concept_governmentorganization_house

concept_governmentorganization_epa

concept_governmentorganization_commission

concept_nongovorganization_council

concept_politicsblog_white_house

concept_country_party

concept_geopoliticallocation_iraqconcept_county_records

concept_city_capital

concept_city_team

concept_visualizablething_use

concept_biotechcompany_china

concept_governmentorganization_action

concept_governmentorganization_representatives

concept_city_members

concept_personus_party

concept_person_state

concept_buildingfeature_window

concept_politicianus_president_george_w__bush

concept_person_president

concept_person_mugabe

concept_personasia_number

concept_buildingfeature_windows

concept_astronaut_herbert_hoover

concept_stateorprovince_afternoon

concept_dateliteral_n2007concept_dateliteral_n2006

concept_date_n2003

concept_dateliteral_n2002

concept_dateliteral_n2008

concept_city_home

concept_country_united_states

concept_city_service

concept_governmentorganization_lawconcept_governmentorganization_program

concept_personmexico_ryan_whitneyconcept_year_n1997

concept_year_n1992

concept_year_n1994

concept_tradeunion_congressconcept_governmentorganization_house

concept_governmentorganization_epa

concept_governmentorganization_commission

concept_personus_party

concept_country_left_parties

concept_person_state

Figure 18: OrganizationHiredPerson. The head is concept_stateorprovince_afternoon,the query relation is concept:organizationhiredperson, and the desired tail isconcept_personmexico_ryan_whitney. The left is a full subgraph derived withmax_attended_nodes_per_step = 20, and the right is a further extracted subgraph fromthe left based on attention. The big yellow node represents the head, and the big red node representsthe tail. Colors indicate how important a node is attended to in a local subgraph. Grey means lessimportant, yellow means it is more attended during the early steps, and red means it is more attendedwhen getting close to the final step.

concept_tradeunion_congress , concept : o rgan iza t ionh i redperson , concept_personus_partyconcept_tradeunion_congress , concept : worksfor_ inv , concept_personmexico_ryan_whitneyconcept_tradeunion_congress , concept : o rgan iza t ionh i redperson , concept_personmexico_ryan_whitneyconcep t_coun t r y_ le f t _pa r t i es , concept : o rgan iza t ionh i redperson , concept_personus_party

For the AgentBelongsToOrganization taskQuery : ( concept_person_mark001 , concept : agentbe longstoorgan iza t ion , c o n c e p t _ g e o p o l i t i c a l l o c a t i o n _ w o r l d )

Selected key edges :concept_person_mark001 , concept : personbe longstoorgan izat ion , concep t_spor ts team_sta te_un ive rs i t yconcept_person_mark001 , concept : agentco l labora teswi thagent , concept_male_worldconcept_person_mark001 , concept : agen tco l labora tesw i thagen t_ inv , concept_male_worldconcept_person_mark001 , concept : personbe longstoorgan izat ion , c o n c e p t _ p o l i t i c a l p a r t y _ c o l l e g econcept_spor ts team_sta te_un ivers i ty , concept : personbe longstoorgan iza t ion_ inv , c o n c e p t _ p o l i t i c i a n _ j o b sconcept_spor ts team_sta te_un ivers i ty , concept : personbe longstoorgan iza t ion_ inv , concept_person_mark001concept_spor ts team_sta te_un ivers i ty , concept : personbe longstoorgan iza t ion_ inv , concept_person_greg001concept_spor ts team_sta te_un ivers i ty , concept : personbe longstoorgan iza t ion_ inv , concept_person_michael002concept_male_world , concept : agentco l labora teswi thagent , c o n c e p t _ p o l i t i c i a n _ j o b sconcept_male_world , concept : agen tco l labora tesw i thagent_ inv , c o n c e p t _ p o l i t i c i a n _ j o b sconcept_male_world , concept : agentco l labora teswi thagent , concept_person_mark001concept_male_world , concept : agen tco l labora tesw i thagent_ inv , concept_person_mark001concept_male_world , concept : agentco l labora teswi thagent , concept_person_greg001concept_male_world , concept : agen tco l labora tesw i thagent_ inv , concept_person_greg001concept_male_world , concept : agentcont ro ls , concept_person_greg001concept_male_world , concept : agentco l labora teswi thagent , concept_person_michael002concept_male_world , concept : agen tco l labora tesw i thagent_ inv , concept_person_michael002c o n c e p t _ p o l i t i c a l p a r t y _ c o l l e g e , concept : personbe longstoorgan iza t ion_ inv , concept_person_mark001c o n c e p t _ p o l i t i c a l p a r t y _ c o l l e g e , concept : personbe longstoorgan iza t ion_ inv , concept_person_greg001c o n c e p t _ p o l i t i c a l p a r t y _ c o l l e g e , concept : personbe longstoorgan iza t ion_ inv , concept_person_michael002concep t_po l i t i c i an_ jobs , concept : personbe longstoorgan izat ion , concep t_spor ts team_sta te_un ivers i t yconcep t_po l i t i c i an_ jobs , concept : agentco l labora teswi thagent , concept_male_worldconcep t_po l i t i c i an_ jobs , concept : agen tco l labora tesw i thagent_ inv , concept_male_worldconcep t_po l i t i c i an_ jobs , concept : worksfor , c o n c e p t _ g e o p o l i t i c a l l o c a t i o n _ w o r l dconcept_person_greg001 , concept : personbe longstoorgan izat ion , concep t_spor ts team_sta te_un ivers i t yconcept_person_greg001 , concept : agentco l labora teswi thagent , concept_male_world

28

Page 29: Neural Consciousness Flow - GitHub Pages

concept_person_mark001

concept_sportsteam_state_university

concept_male_world

concept_politicalparty_college

concept_sportsgame_series

concept_university_state_university

concept_charactertrait_world

concept_university_college

concept_person_diane001

concept_university_high_school

concept_televisionshow_passionconcept_musicsong_gospel

10

concept_university_syracuse_university

concept_city_hampshireconcept_person_louise

concept_jobposition_king

concept_county_york_city

concept_stateorprovince_california

concept_stateorprovince_maine

concept_stateorprovince_georgia

concept_person_greg001concept_person_michael002

concept_stateorprovince_author

concept_person_bill

concept_person_kevin

concept_person_john003

concept_politician_jobs

concept_person_mike

concept_person_stephen

concept_person_fred

concept_person_karen

concept_person_william001

concept_person_tom001

concept_journalist_dan

concept_person_joe002

concept_person_adam001

concept_politician_james

concept_eventoutcome_result

concept_geopoliticallocation_world

concept_recordlabel_friends

concept_politicalparty_house

concept_company_apple001

concept_televisionnetwork_cbs

concept_company_apple002

3

concept_automobilemaker_announcement

concept_personmexico_ryan_whitney

concept_museum_steve

concept_city_downtown_manhattanconcept_attraction_nycconcept_city_nyc_

concept_personasia_number

concept_scientist_no_

concept_country_united_statesconcept_geopoliticallocation_agencies

concept_company_nbc

concept_city_team

concept_terroristorganization_state

concept_automobilemaker_jeff_bezos

concept_ceo_richard

concept_lake_new

concept_city_service

concept_governmentorganization_u_s__department

concept_person_mark001

concept_sportsteam_state_university

concept_male_world

concept_politicalparty_college

concept_sportsgame_series concept_university_state_university

concept_person_greg001

concept_person_michael002

concept_stateorprovince_author

concept_geopoliticallocation_world

concept_recordlabel_friends

concept_personmexico_ryan_whitney

concept_politician_jobs

concept_eventoutcome_result

concept_museum_steve

Figure 19: AgentBelongsToOrganization. The head is concept_person_mark001, the query relationis concept:agentbelongstoorganization, and the desired tail is concept_geopoliticallocation_world.The left is a full subgraph derived with max_attended_nodes_per_step = 20, and the right is afurther extracted subgraph from the left based on attention. The big yellow node represents the head,and the big red node represents the tail. Colors indicate how important a node is attended to in a localsubgraph. Grey means less important, yellow means it is more attended during the early steps, andred means it is more attended when getting close to the final step.

concept_person_greg001 , concept : agen tco l labora tesw i thagent_ inv , concept_male_worldconcept_person_greg001 , concept : agen tcon t ro ls_ inv , concept_male_worldconcept_person_greg001 , concept : agentbe longstoorgan iza t ion , c o n c e p t _ g e o p o l i t i c a l l o c a t i o n _ w o r l dconcept_person_greg001 , concept : personbe longstoorgan izat ion , c o n c e p t _ p o l i t i c a l p a r t y _ c o l l e g econcept_person_greg001 , concept : agentbe longstoorgan iza t ion , concep t_ reco rd labe l_ f r i endsconcept_person_michael002 , concept : personbe longstoorgan izat ion , concep t_spor ts team_sta te_un ive rs i t yconcept_person_michael002 , concept : agentco l labora teswi thagent , concept_male_worldconcept_person_michael002 , concept : agen tco l labora tesw i thagen t_ inv , concept_male_worldconcept_person_michael002 , concept : agentbe longstoorgan iza t ion , c o n c e p t _ g e o p o l i t i c a l l o c a t i o n _ w o r l dconcept_person_michael002 , concept : personbe longstoorgan izat ion , c o n c e p t _ p o l i t i c a l p a r t y _ c o l l e g econcep t_geopo l i t i ca l l o ca t i on_wor l d , concept : worksfor_ inv , concept_personmexico_ryan_whitneyconcep t_geopo l i t i ca l l o ca t i on_wor l d , concept : o rgan iza t ionh i redperson , concept_personmexico_ryan_whitneyconcep t_geopo l i t i ca l l o ca t i on_wor l d , concept : worksfor_ inv , c o n c e p t _ p o l i t i c i a n _ j o b sconcep t_ record labe l_ f r iends , concept : o rgan iza t ionh i redperson , concept_personmexico_ryan_whitneyconcept_personmexico_ryan_whitney , concept : worksfor , c o n c e p t _ g e o p o l i t i c a l l o c a t i o n _ w o r l dconcept_personmexico_ryan_whitney , concept : o rgan iza t ionh i redperson_ inv , c o n c e p t _ g e o p o l i t i c a l l o c a t i o n _ w o r l dconcept_personmexico_ryan_whitney , concept : o rgan iza t ionh i redperson_ inv , concep t_ reco rd labe l_ f r i ends

For the TeamPlaysInLeague taskQuery : ( concept_sportsteam_mavericks , concept : teamplaysinleague , concept_sportsleague_nba )

Selected key edges :concept_sportsteam_mavericks , concept : teamplayssport , concept_spor t_baske tba l lconcept_sportsteam_mavericks , concept : teamplaysagainstteam , concept_spor ts team_boston_ce l t icsconcept_sportsteam_mavericks , concept : teamplaysagainstteam_inv , concept_spor ts team_boston_ce l t icsconcept_sportsteam_mavericks , concept : teamplaysagainstteam , concept_sportsteam_spursconcept_sportsteam_mavericks , concept : teamplaysagainstteam_inv , concept_sportsteam_spursconcept_spor t_basketba l l , concept : teamplaysspor t_ inv , concept_sports team_col legeconcept_spor t_basketba l l , concept : teamplaysspor t_ inv , concept_spor ts team_marsha l l_un ivers i tyconcept_spor ts team_boston_cel t ics , concept : teamplaysinleague , concept_sportsleague_nbaconcept_sportsteam_spurs , concept : teamplaysinleague , concept_sportsleague_nbaconcept_sportsleague_nba , concept : agentcompeteswithagent , concept_sportsleague_nbaconcept_sportsleague_nba , concept : agentcompeteswithagent_inv , concept_sportsleague_nbaconcept_sportsteam_col lege , concept : teamplaysinleague , concep t_spo r t s league_ in te rna t i ona l

29

Page 30: Neural Consciousness Flow - GitHub Pages

concept_sportsteam_mavericks

concept_sport_basketball

concept_sportsteam_boston_celtics

concept_sportsteam_spurs

concept_sportsteam_suns

concept_sportsteam_esu_hornets

concept_sportsteam_rockets

concept_sportsteam_knicks

concept_sportsteam_memphis_grizzlies

concept_sportsteam_kings_college

concept_sportsteam_san_antonio

concept_sportsgame_series

concept_sportsteam_chicago_bulls

concept_sportsteam_golden_state_warriors

concept_sportsteam_utah_jazz

concept_sportsgame_championship

concept_sportsteam_la_clippers

concept_convention_games

concept_athlete_josh_howard

concept_awardtrophytournament_nba_finals

concept_sportsteam_marshall_university

concept_sportsleague_nba

concept_sportsteam_college

concept_sportsteam_pacers

concept_athlete_o__j__simpson

concept_sportsteam_devils

concept_athlete_dikembe_mutombo

concept_sportsleague_nascar

concept_sportsteam_pistons

concept_sportsteam_los_angeles_lakers

concept_athlete_shane_battier

concept_sportsteam_hawks

concept_sportsteam_washington_wizards

concept_sportsteam_michigan_state_university

concept_sportsteam_new_jersey_nets

concept_sportsteam_astros

concept_sportsteam_trail_blazers

concept_athlete_cuttino_mobley

concept_sportsteam_george_mason_university

concept_athlete_colin_long

concept_sportsleague_international

concept_city_huntington

concept_sportsleague_nhl

concept_sportsleague_mlb

concept_stadiumoreventvenue_toyota_center

concept_sportsteam_mavericks

concept_sport_basketball

concept_sportsteam_boston_celtics

concept_sportsteam_spurs

concept_sportsteam_suns

concept_sportsteam_esu_hornets

concept_sportsteam_marshall_university

concept_sportsleague_nba

concept_sportsteam_college

concept_sportsteam_pacers

concept_athlete_o__j__simpson

concept_sportsleague_international

concept_city_huntington

concept_sportsleague_nhl

concept_sportsleague_nascar

Figure 20: TeamPlaysInLeague. The head is concept_sportsteam_mavericks, the query relationis concept:teamplaysinleague, and the desired tail is concept_sportsleague_nba. The left is a fullsubgraph derived with max_attended_nodes_per_step = 20, and the right is a further extractedsubgraph from the left based on attention. The big yellow node represents the head, and the big rednode represents the tail. Colors indicate how important a node is attended to in a local subgraph.Grey means less important, yellow means it is more attended during the early steps, and red means itis more attended when getting close to the final step.

30