21
Toward a General Theory of Embodied, Social General Intelligence Under Feasible Computational Resources Ben Goertzel Abstract An attempt is made to outline a general conceptual theory of real-world general intelligence, that avoids the twin pitfalls of excessive mathematical generality, and excessive specialization to the case of human intelligence. The foundation of the theory is a choice to measure the simplicity of an entity in terms of the ease of communicating it within a community of embodied agents (the so-called Embodied Communication Prior or ECP). Augmented by some further assumptions and a definition of general intelligence in terms of goal-achievement, this choice is seen to lead to a model of intelligence in terms of distinct memory and cognitive subsystems dealing with procedural, declarative, sensory/episodic, attentional and cybernetic knowledge. Furthermore, the need for subtle interactions among these subsystems is highlighted; and these ideas are applied to analyze the OpenCogPrime design for general intelligence, with the conclusion that if certain conjectures about OpenCogPrime’s subsystems and their interactions hold true, then the design may be capable of leading to a system displaying general intelligence relative to the ECP, including cognition at the level identified by Piaget as “formal.” Introduction What can one say, in general, about general intelligence? The answer seems to be: a fair bit of interesting mathematics which has worthwhile philosophical implications, but not much of direct practical value. The tradition of Solomonoff (1964) induction, pursued recently by Juergen Schmidhuber (2006), Marcus Hutter (2005) and others, seems to say what there is to be said about truly general intelligence. There are many more mathematical details to be unraveled, but the conceptual picture seems clear. Without some sort of special assumptions about the environment and goals relevant to an intelligent system, the best way to achieve general intelligence -- in the sense of generally effective goal-achieving behavior, for complex goal/environment combinations -- is essentially to carry out some form of brute-force search of the space of possible behavior- control programs, continually re-initiating the search as one's actions lead to new information. The problem with this approach is that it's incredibly computationally expensive -- which leads to the question of the best way to achieve reasonable levels of general intelligence given feasible computational resources. Here the picture is less clear mathematically at this point, but, intuitively, it seems fairly clear that achieving useful

Toward a General Theory of Embodied, Social General

  • Upload
    others

  • View
    0

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Toward a General Theory of Embodied, Social General

Toward a General Theory of Embodied, Social General Intelligence

Under Feasible Computational Resources

Ben Goertzel

Abstract An attempt is made to outline a general conceptual theory of real-world general intelligence, that avoids the twin pitfalls of excessive mathematical generality, and excessive specialization to the case of human intelligence. The foundation of the theory is a choice to measure the simplicity of an entity in terms of the ease of communicating it within a community of embodied agents (the so-called Embodied Communication Prior or ECP). Augmented by some further assumptions and a definition of general intelligence in terms of goal-achievement, this choice is seen to lead to a model of intelligence in terms of distinct memory and cognitive subsystems dealing with procedural, declarative, sensory/episodic, attentional and cybernetic knowledge. Furthermore, the need for subtle interactions among these subsystems is highlighted; and these ideas are applied to analyze the OpenCogPrime design for general intelligence, with the conclusion that if certain conjectures about OpenCogPrime’s subsystems and their interactions hold true, then the design may be capable of leading to a system displaying general intelligence relative to the ECP, including cognition at the level identified by Piaget as “formal.”

Introduction What can one say, in general, about general intelligence? The answer seems to be: a fair bit of interesting mathematics which has worthwhile philosophical implications, but not much of direct practical value. The tradition of Solomonoff (1964) induction, pursued recently by Juergen Schmidhuber (2006), Marcus Hutter (2005) and others, seems to say what there is to be said about truly general intelligence. There are many more mathematical details to be unraveled, but the conceptual picture seems clear. Without some sort of special assumptions about the environment and goals relevant to an intelligent system, the best way to achieve general intelligence -- in the sense of generally effective goal-achieving behavior, for complex goal/environment combinations -- is essentially to carry out some form of brute-force search of the space of possible behavior-control programs, continually re-initiating the search as one's actions lead to new information. The problem with this approach is that it's incredibly computationally expensive -- which leads to the question of the best way to achieve reasonable levels of general intelligence given feasible computational resources. Here the picture is less clear mathematically at this point, but, intuitively, it seems fairly clear that achieving useful

Page 2: Toward a General Theory of Embodied, Social General

levels of truly general intelligence (without special assumptions about the world) using feasible computational resources is just not possible. But what if one restricts the scope of generality, via making appropriate special assumptions about the goal/environment combinations of interest? In this case, one's claim for generality of intelligence is less, but if the restrictions are not too severe, one still has a case of interest. But the question is whether there is anything of elegance and scope to say about this case. Plausibly, this case might degenerate into a collection of highly specialized statements about particular classes of goals and environments. On the other hand, maybe there is something sensible one can say about "general intelligence in the everyday world using feasible computational resources," as a topic more restrictive than mathematically general intelligence, but less restrictive than task-specific intelligence like chess-playing, car-driving, biological data analysis, etc. My goal here is to sketch some ideas that I think may eventually lead to a reasonably general theory of everyday-world general intelligence using feasible computational resources. I describe a particular class of probability distributions over the space of goals and environments called the Embodied Communication Prior (ECP), and argue that according to this distribution, there are certain general properties that a system should possess in order to display a level of intelligence equivalent to Gregory Bateson's notion of third-order learning ("learning how to learn how to learn"), which is argued to be equivalent (under the assumption of highly limited computational resources) to Piaget's formal stage of cognitive development. The arguments presented are intuitive rather than rigorous, and are intended to serve not as definitive conclusions, but rather as tentative, hypothetical conclusions pointing the direction for future work to be conducted with greater rigor. After the general conclusions are outlined, arguments are given as to why the Novamente Cognition Engine and OpenCog AI system designs appear to fulfill the specified properties for general intelligence under feasible computational resources. Finally, some indications are given regarding how these ideas might be extended yet further, to explain the properties of systems displaying general intelligence relative to various assumptions other than the ECP.

Real-World General Intelligence

The concept of intelligence, like most folk-psychology concepts, is a slippery one with multiple overlapping meanings, and defies straightforward formalization. Legg and Hutter (2006a) have reviewed over 70 published definitions of intelligence, and have proposed a fully formalized definition of their own (Legg and Hutter, 2006). Here we will content ourselves with a semi-formalized definition that is largely in the spirit of Legg and Hutter, and also in the spirit of the earlier semi-formalized definition I gave in (Goertzel, 1993). The intuition underlying this definition is that “intelligence is the ability to achieve complex goals in complex environments.” To capture this more precisely, we may assess intelligence by averaging achievability (by the system) over the space of all (goal, environmental situation) pairs, where the achievability of the pair is weighted by a coefficient proportional to the simplicity of the pair according to the ECP. This begs the question of how to measure simplicity – but we will defer this to the following section of the paper, as it’s one of the key original themes presented here.

Page 3: Toward a General Theory of Embodied, Social General

So: given a system M, a goal G and a situation S, let ach(M,G,S) denote the quality with which M can achieve G in S; and let simp(G,S) denote the simplicity of the pair (G,S). Then, we may conceive the intelligence of M as the sum over (G,S) pairs of

ach(M,G,S) simp(G,S) Fully formalizing this along the lines of (Legg and Hutter, 2006) would be possible but would serve little function as this is a qualitative paper that won't be pursuing rigorous proofs.

One may also define the notion of “efficient intelligence” (Goertzel, 2006), obtained by normalizing general intelligence according to computational resources utilized. There is essentially no formal theory of efficient intelligence at this point. But efficient intelligence is what matters in the real world, since theoretical general intelligences utilizing excessive amounts of computational resources can never actually be constructed. According to this approach, different simplicity measures lead to different notions of intelligence, and hence may correspond to intelligent systems with very different properties. As noted above, it seems extremely likely to me that, if one measures simplicity as program length in some standard computational model (say, a classical Turing machine; or lambda calculus), then the best approach to general intelligence is going to be some sort of simplistic brute force search algorithm such as Hutter’s (2005) AIXI. Essentially, what AIXI does, at each time point, is to search the space of all computer programs to find the program P so that, if AIXI uses P to determine its next action, it will be maximally likely to achieve its goals. After taking a single action, it then re-evaluates and chooses a new program P to execute; etc. This is a great approach if one has the computational resources to perform a thorough search of program space at each time step. Unfortunately AIXI requires infinitely much computational resources, but there is a related approach called AIXItl that restricts its search to programs with specifically bounded length and runtime, whose resource requirements are merely infeasibly humongous rather than infinite. This kind of approach has very high general intelligence, but very poor efficient intelligence. Yet, my suspicion is that if one measures simplicity using standard computational models, there’s probably not any way to do significantly better. The trick to real-world general intelligence, I suggest, lies in the simplicity measure. If one assumes a simplicity measure that is biased to certain (goal, situation) pairs, then one can create intelligent systems displaying general intelligence with respect to this simplicity measure, via subtly constructing the systems in a way that embodies the patterns implicit in the simplicit measure. From this perspective, there are two interesting approaches to the “general theory of general intelligence under feasible computational resources”:

1. Discovering general principles that determine how, given a simplicity measure, to construct systems displaying general intelligence relative to that simplicity measure

Page 4: Toward a General Theory of Embodied, Social General

2. Exploring a particular class of simplicity measures that seems to have high relevance, and figuring out how to create systems displaying general intelligence relative to simplicity measures in this class

The former would be the more ambitious approach: potentially, one could envision a theory that would let one specify a simplicity measure, and would then produce an AI design displaying a near-maximal amount of efficient general intelligence relative to that simplicity measure. While we are currently conducting research aimed in this direction, our focus in this paper is on the second and less ambitious goal. In the following section we will articulate a specific class of simplicity measures relevant to broadly humanlike intelligence, and then explore the properties of systems displaying general intelligence relative to these simplicity measures. We will then discuss a specific AI design (the Novamente Cognition Engine / OpenCogPrime design) and argue that it is likely to possess these properties.

The Embodied Communication Prior In this section I will describe a simplicity measure on (goal, environment) pairs; or, equivalently, a probability distribution over such pairs. This distribution is what I call the Embodied Communication Prior (ECP); and I will describe it intuitively rather than formalizing it rigorously, though the latter would not be extremely difficult. In (Goertzel, 2009) a probability distribution called the Communication Prior is introduced, and used as part of a formalization of the scientific process. The ECP is an extension of the Communication Prior (which may also have applicability to the formalization of the scientific process, but that’s another story not to be entered into here).

Consider a community of embodied agents living in a shared world, and suppose that the agents can communicate with each other via a set of mechanisms including:

• Linguistic communication, in a language whose semantics is largely (not necessarily wholly) interpretable based on mapping linguistic utterances into finite combinations of entities drawn from a finite vocabulary

• Indicative communication, in which e.g. one agent points to some part of the world or delimits some interval of time, and another agent is able to interpret the meaning

• Demonstrative communication, in which an agent carries out a set of actions in the world, and the other agent is able to imitate these actions, or instruct another agent as to how to imitate these actions

• Depictive communication, in which an agent creates some sort of (visual, auditory, etc.) construction to show another agent, with a goal of causing the other agent to experience phenomena similar to what they would experience upon experiencing some particular entity in the shared environment

• Cybernetic communication, in which an agent explicitly communicates to another agent what its goal is in a certain situation (note interesting recent results showing that mirror neurons fire in response to some cases of cybernetic communication; Fogassi et al, 2005)

Page 5: Toward a General Theory of Embodied, Social General

It is clear that ordinary everyday communication between humans possesses all these aspects. The Embodied Communication Prior is defined as the probability distribution in which the probability of an entity (e.g. a goal or environment) is proportional to the difficulty of describing that entity, for a typical member of the community in question, using a particular set of communication mechanisms including the above five modes. We will sometimes refer to the prior probability of an entity under this distribution, as its "simplicity" under the distribution. Next, to further specialize the Embodied Communication Prior as I'll consider it here, I will assume that for each of these modes of communication, there are some aspects of the world that are much more easily communicable using that mode than the other modes. For instance, in the human everyday world:

• Abstract statements spanning large classes of situations are generally much easier to communicate linguistically

• Complex, multi-part procedures are much easier to communicate either demonstratively, or using a combination of demonstration with other modes

• Sensory or episodic data is often much easier to communicate demonstratively • The current value of attending to some portion of the shared environment is often

much easier to communicate indicatively • Information about what goals to follow in a certain situation is often much easier

to communicate cybernetically, i.e. via explicitly indicating what one’s own goal is

These simple observations have significant implications for the nature of the Embodied Communication Prior. For one thing they let us define multiple forms of knowledge:

• Isolatedly declarative knowledge is that which is much more easily communicable linguistically

• Isolatedly procedural knowledge is that which is much more easily communicable demonstratively

• Isolatedly sensory knowledge is that which is much more easily communicable depictively

• Isolatedly attentive knowledge is that which is much more easily communicable indicatively

• Isolatedly cybernetic knowledge is that which is much more easily communicable cybernetically

Of course there may be much knowledge, of relevance to systems seeking intelligence according to the ECP, that does not fall into any of these categories and constitutes "mixed knowledge." There are some very important specific subclasses of mixed knowledge. For instance, episodic knowledge (knowledge about specific real or hypothetical sets of events) will most easily be communicated via a combination of declarative, sensory and (in some cases) procedural communication. Scientific and

Page 6: Toward a General Theory of Embodied, Social General

mathematical knowledge are generally mixed knowledge, as is most everyday commonsense knowledge. Some cases of mixed knowledge are reasonably well decomposable, in the sense that they decompose into knowledge items that individually fall into some specific knowledge type. For instance, an experimental chemistry procedure may be much better communicable procedurally than any other way, whereas an allied piece of knowledge from theoretical chemistry may be much better communicable declaratively; but in order to fully communicate either the experimental procedure or the abstract piece of knowledge, one may ultimately need to communicate both aspects. Also, even when the best way to communicate something is mixed-mode, it may be possible to identify one mode that poses the most important part of the communication. An example would be a chemistry experiment that is best communicated via a practical demonstration together with a running narrative. It may be that the demonstration without the narrative would be vastly more valuable than the narrative without the demonstration. To cover this case we may make less restrictive definitions such as

• Interactively declarative knowledge is that which is much more easily communicable in a manner dominated by linguistic communication

and so forth. We call these “interactive knowledge categories,” by contrast to the “isolated knowledge categories” introduced earlier. Next we introduce an assumption we call NKC, for Naturalness of Knowledge Categories. The NKC assumption states that the knowledge in each of the above isolated and interactive communication-modality-focused categories forms a "natural category," in the sense that for each of these categories, there are many different properties shared by a large percentage of the knowledge in the category, but not by a large percentage of the knowledge in the other categories. This means that, for instance, procedural knowledge systematically (and statistically) has different characteristics than the other kinds of knowledge. The NKC assumption seems commonsensically to be the case for human everyday knowledge, and it has fairly dramatic implications for general intelligence. Suppose we conceive general intelligence as the ability to achieve goals in the environment shared by the communicating agents underlying the Embodied Communication Prior. Then, NKC suggests that the best way to achieve general intelligence according to the Embodied Communication Prior is going to involve

• specialized methods for handling declarative, procedural, sensory and attentional knowledge (due to the naturalness of the isolated knowledge categories)

• specialized methods for handling interactions between different types of knowledge, including methods focused on the case where one type of knowledge is primary and the others are supporting (the latter due to the naturalness of the interactive knowledge categories)

Page 7: Toward a General Theory of Embodied, Social General

Levels of General Intelligence According to the ECP In the context of the ECP in particular, one can refine the notion of general intelligence introduced above, obtaining something more reminiscent of everyday human notions of intelligence. In this section we move in this direction via the ideas of two previous theorists concerned with articulating hierarchical levels of intelligence in humans and other organisms, Piaget and Bateson. Piaget introduced a series of cognitive developmental stages, which may be approximately summarized as

1. Infantile: learns to carry out procedures oriented toward achieving goals in relevant contexts

2. Operational: learns abstractions allowing it to adapt its learning to different contexts and subgoals

3. Formal: learns abstractions allowing it to adapt its adaptation Piaget also introduced a “pre-operational” stage between the infantile and operational ones; and some more recent thinkers have introduced a “post-formal” stage involving the ability of the system to apply formal reasoning to its own basic structure and outlook. But these three phases will suffice for our current purposes. With cybernetic considerations in mind, Gregory Bateson proposed a different sort of hierarchy:

1. Learning 2. Learning how to learn 3. Learning how to learn how to learn

Of course, there is no reason the hierarchy of levels of learning needs to stop at that point; but Bateson suggests that in actual human practice it does. In (Goertzel and Bugaj, 2007) it is suggested that these two hierarchies, which look quite different on the surface, are actually closely aligned, so that in certain types of intelligent systems, the Piagetan and Batesonian stages correspond closely. Specifically it is argued there that this correpsondence holds both in humans and in AI systems whose operations are centrally based on uncertain inference. What we suggest here is that this correspondence is actually a conceptual consequence of the Embodied Communication Prior. In the context of the ECP, the operational stage may be viewed as requiring a thorough integration of declarative and procedural learning. Declarations about procedures must be learned, and then manipulated in order to lead to new procedures. This is precisely Batesonian “learning how to learn.” Of course sensory knowledge must also be drawn into the picture here, as these declarations will often be highly dependent on sensorially-defined contexts. On the other hand, the formal stage requires a yet deeper integration, in which procedures are learned for controlling the application of declarative knowledge to guiding procedure learning – and declarative knowledge is then learned regarding these higher-level procedures. This is a direct translation of the ideas of (Goertzel and Bugaj, 2007)

Page 8: Toward a General Theory of Embodied, Social General

into the language of the ECP; and clearly it is an instance of Batesonian “learning how to learn how to learn.” Thus, we suggest that in the context of the ECP, Bateson’s third level of learning and Piaget’s formal stage come together in a dynamic of “mixed declarative/procedural learning about declarative learning about procedural learning.” And this dynamic, we suggest, is critical to the achievement of a high level of efficient general intelligence according to the ECP.

Analysis of Goal-Directed Systems According to the ECP

Now we seek to explore a little more deeply what the above conclusions imply about the internal operations of AI systems displaying general intelligence with respect to the ECP (and adopting the NKC assumption as well). To simplify analysis we consider an AI system that possesses a specific set of goals, together with a predefined set of actuators and sensors. We further assume that the action of the system may be modeled in terms of the “cognitive schematic” equation

Context & Procedure Goal <p>

interpreted to mean “If the context C appears to hold currently, then if I enact the procedure P, I can expect to achieve the goal G with certainty p.” A procedure is defined as some systematic pattern of activity within the system (which may involve activation of the external actuators, or may in some cases involve purely internal activities). A context is defined as a fuzzy logical predicate that holds, to a certain extent, during each interval of time. A goal is simply some fuzzy logical predicate that has a certain value at each interval of time, as well. We will also us the shorthand

C & P G <p>

Note that we don’t assume the system explicitly uses the cognitive schematic to regulate its activities; rather, we are introducing this as an external model of the system. If the system explicitly uses some form of the cognitive schematic in its internal operations, that’s fine, but our analysis does not require this.

This formalization leads to a conceptualization of the internal action of an intelligent system as involving two “key learning processes”:

1. Estimating the probability p of a posited C & P G relationship 2. Filling in one or two of the variables in the cognitive schematic, given

assumptions regarding the remaining variables, and directed by the goal of maximizing the probability of the cognitive schematic

Again, we stress that we don’t assume the system’s internal dynamics are explicitly oriented around these two types of activity. What we assume is that the system can be modeled this way, basically as a combination of:

Page 9: Toward a General Theory of Embodied, Social General

1. Evaluating conjectured relationships between procedures, contexts and goals (“analysis”)

2. Conceiving novel possible relationships between procedures, contexts and goals (“synthesis”)

Note that this dichotomy is the same as the one posited in (Goertzel, 2006a), where is synthesis referred to as a “forward cognitive process”, and analysis is referred to as a “backward cognitive process.”

Given this conceptualization, we can see that, where synthesis is concerned,

• Isolatedly procedural knowledge can be useful for choosing P, given fixed C and G

• Isolatedly sensory knowledge, isolatedly declarative knowledge, or mixed sensory/declarative knowledge can be useful for choosing C, given fixed P and G

• Isolatedly declarative knowledge can be useful for choosing G, given fixed P and C

On the other hand, where analysis is concerned:

• Isolatedly declarative knowledge can be useful for estimating the probability of the implication in the schematic equation, given fixed C, P and G. Episodic knowledge can also be useful in this regard, via enabling estimation of the probability via simple similarity matching against past experience.

• Isolatedly procedural knowledge can be useful for estimating the probability of the implication C & P G, in cases where the probability of C & P1 G is known for some P1 related to P

• Isolatedly declarative or isolatedly sensory knowledge can be useful for estimating the probability of the implication C & P G, in cases where the probability of C1 & P G is known for some C1 related to C

• Isolatedly declarative knowledge can be useful for estimating the probability of the implication C & P G, in cases where the probability of C & P G1 is known for some G1 related to G

We can also see the role of mixed knowledge here, because sometimes the best way to handle the schematic equation will be to fix only one of the terms. For instance, if we fix G, sometimes the best approach will be to collectively learn C and P (for instance, using evolutionary learning methods, to allow them to “co-evolve”). Dominantly procedural knowledge, for example, corresponds to the case where one mainly wants to learn P, but accepts that one may also need to adapt C or G during the learning process, rather than leaving them fixed. The final fact we need to account for is that, in any real-world context, a system will be presented with a huge number of possibly relevant analysis and synthesis problems. Choosing which ones to explore is a difficult cognitive problem in itself – a problem that also takes the form of the cognitive schematic, but where the procedures are

Page 10: Toward a General Theory of Embodied, Social General

internal rather than external. Thus this problem may be addressed via the analysis and synthesis methods describe above. This is the role of attentional knowledge ; it gives the system some base knowledge regarding what to attend to (which in some cases will be the problem of using complex analysis and/or synthesis to figure out what to attend to). Attentional knowledge may be built up analytically or synthetically, and isolatedly or interactively, just like other types of knowledge. The NKC suggests that attentional knowledge forms a natural category just like the other types of knowledge. Suppose we conceive an AI system as consisting of a set of learning capabilities, each one characterized by three features:

• One or more knowledge types that it is competent to deal with, in the sense of the two key learning problems mentioned above

• At least one learning type: either analysis, or synthesis, or both • At least one interaction type, for each (knowledge type, learning type) pair it

handles: “isolated” (meaning it deals mainly with that knowledge type in isolation), or “interactive” (meaning it focuses on that knowledge type but in a way that explicitly incorporates other knowledge types into its process), or “fully mixed” (meaning that when it deals with the knowledge type in question, no particular knowledge type tends to dominate the learning process).

Then, it seems to follow from the ECP with NKC that systems with high efficient general intelligence should have the following properties, which collectively I’ll call “cognitive completeness”:

• For each (knowledge type, learning type, interaction type) triple, there should be a learning capability corresponding to that triple.

• Furthermore the capabilities corresponding to different (knowledge type, interaction type) pairs should have distinct characteristics (since according to the NKC the isolated knowledge corresponding to a knowledge type is a natural category, as is the dominant knowledge corresponding to a knowledge type)

• For each (knowledge type, learning type) pair (K,L), and each other knowledge type K1 distinct from K, there should be a capability with interaction type “interactive” and dealing with knowledge that is interactively K but also includes aspects of K1; and this capability should have distinct characteristics

Furthermore, it seems plain that according to the ECP with NKC, if the capabilities mentioned in the above points are reasonably able, then the system possessing the capabilities will display general intelligence relative to the ECP. Thus we arrive at the hypothesis that Under the assumption of the Embodied Communication Prior (with the Natural Knowledge Categories assumption), the property above called “cognitive completeness” is necessary and sufficient for efficient general intelligence at the Piagetan formal level. Of course, the above considerations are very far from a rigorous mathematical proof (or even precise formulation) of this hypothesis. But we are presenting this here as a

Page 11: Toward a General Theory of Embodied, Social General

conceptual hypothesis, in order to qualitatively guide R&D and also to motivate further, more rigorous theoretical work.

Analysis of the OpenCogPrime AGI Design Next, in this section we review some of the properties of one particular AGI design, OpenCogPrime (OCP), and argue that it is cognitively complete in the above sense. We will not review the OCP design in detail here, as it and the closely related Novamente Cognition Engine design have been described in fair detail in prior publications (Goertzel et al, 2003; Goertzel, 2006, 2008). Here we will restrict ourselves to presenting a high level analysis connecting OCP with the ideas presented here. Unlike the human brain or many other AI designs, OCP explicitly makes use of the cognitive schematic: it has an explicit list of goals (which it refines via a “goal refinement” process), and explicitly seeks to enact procedures according to the cognitive schematic, constructed relative to these goals. Figure 1 gives a general overview of the OCP architecture, oriented toward the ECP conceptual framework via indicating the cognitive algorithms that OCP uses to manage each type of knowledge identified in ECP. As noted there, the key cognitive algorithms of OCP are:

• Probabilistic Logic Networks (PLN), a logical inference framework capable of uncertain reasoning about abstract knowledge, everyday commonsense knowledge, and low-level perceptual and motor knowledge (Goertzel et al, 2008)

• MOSES, a probabilistic evolutionary learning algorithm, which learns procedures (represented as LISP-like program trees) based on specifications (Looks, 2006)

• Economic Attention Networks (ECAN), a framework for allocating (memory and processor) attention among items of knowledge and cognitive processes, utilizing a synthesis of ideas from neural networks and artificial economics

• Map formation, the process of scanning the knowledge base of the system for patterns and then embodying these patterns explicitly as new knowledge items

• Concept creation, the process of forming new concepts via blending and otherwise merging existing ones

• Simulation, the running of simulations of (remembered or imagined) external-world scenarios in an internal world-simulation engine

• Goal refinement, that transforms given goals into sets of subgoals and there are also other supporting cognitive algorithms.

Page 12: Toward a General Theory of Embodied, Social General

Figure 1. High-level overview of the cognitive dynamics in the OpenCogPrime design. For detailed explanations the reader is directed to the references. The overall operation of the system involved the pursuit of a set of goals provided by the programmer, which are then refined by inference. At each time the system chooses a set of procedures to execute, based on its judgments regarding which procedures will best help it achieve its goals in the current context. These procedures may involve external actions (e.g. involving conversation, or controlling an agent in a simulated world) and/or internal cognitive actions. In order to make these judgments it must effectively manage declarative, procedural, episodic, sensory and attentional memory, each of which is associated with specific algorithms and structures as depicted in the diagram. There are also global processes spanning all the forms of memory, including the allocation of attention to different memory items and cognitive processes, and the identification and reification of system-wide activity patterns. Tables 1 and 2 categorize OCP’s cognitive processes along the axes of analysis/synthesis and isolated/interactive, as articulated in the theory given above. Finally, Table 3 (which is a 16x16 table broken up into 4 parts so as to fit more easily on the page) explores the interactions between different OCP cognitive processes, indicating explicitly which processes can help which other ones, and how. Table 3 explicitly indicates why we believe that OCP is “cognitively complete” in the sense articulated above. If all the interactions given in Table 3 are actually effective (which is not yet known to be the case; so far some have been demonstrated experimentally and some have not, due to the incompleteness of the OCP software implementation), then the cognitive completeness of OCP is a consequence. Thus, if

Page 13: Toward a General Theory of Embodied, Social General

Table 3 holds true and the hypotheses given earlier in this paper are correct, it would follow that OCP can display efficient general intelligence at the Piagetan formal level, relative to the ECP with the NKC assumption. Clearly there is a significant amount of conjecture here, so these arguments are nowhere near constituting a proof of the inevitable success of OCP at real-world general intelligence. But, our goal here is somewhat more modest – it is merely to clarify, in a systematic and precise way, the core conceptual argument as to why we believe OCP can succeed.

Table 1. Categorization of OCP’s primary cognitive dynamics according to the categories of analysis vs. synthesis

Page 14: Toward a General Theory of Embodied, Social General

Table 2. Categorization of OCP’s primary cognitive dynamics according to the categories of isolated vs. interactive (isolated vs. interactive). Note that in many cases the same cognitive dynamics can be used both an isolated or interactive way.

Page 15: Toward a General Theory of Embodied, Social General
Page 16: Toward a General Theory of Embodied, Social General
Page 17: Toward a General Theory of Embodied, Social General
Page 18: Toward a General Theory of Embodied, Social General

Table 3. Systematic enumeration of key interactions between OCP cognitive processes. Some of these interactions have been demonstrated empirically to date, and some have not, due to the incomplete status of the OCP implementation. If all these interactions prove workable, then according to the theory presented in this paper, OCP will demonstrate efficient general intelligence at the Piagetan formal level, relative to the ECP prior with the NKC assumption.

Page 19: Toward a General Theory of Embodied, Social General

Beyond the Embodied Cognition Prior One interesting direction for further research would be to broaden the scope of the inquiry, in a manner suggested earlier: instead of just looking at the ECP, look at simplicity measures in general, and attack the question of how a mind must be structured in order to display efficient general intelligenc relative to a specified simplicity measure. This problem seems unapproachable in general, but some special cases may be more tractable. For instance, suppose one has

• a simplicity measure that (like the ECP) is approximately decomposable into a set of fairly distinct components, plus their interactions

• an assumption similar to NKC, which states that the entities displaying simplicity according to each of the distinct components, are roughly clustered together in entity-space

Then one should be able to say that, to achieve efficient general intelligence relative to this decomposable simplicity measure, a system should have

• distinct capabilities corresponding to each of the components of the simplicity measure

• interactions between these capabilities, corresponding to the interaction terms in the simplicity measure

With copious additional work, these simple observations could potentially serve as the seed for a novel sort of theory of general intelligence – a theory of how the structure of a system depends on the structure of the simplicity measure with which it achieves efficient general intelligence.

Conclusion We have given the broad outlines of a general theory of real-world general intelligence. While lacking true mathematical rigor, the ideas presented have been semi-formalized, in an attempt to maintain a high level of conceptual precision; but nevertheless, they must be considered at this stage to fall into the category of “philosophy of mind” rather than mathematics or science. Yet, we do not consider these ideas pragmatically irrelevant due to their predominantly philosophical nature. Given the current lack of a rigorous mathematical and scientific theory of real-world general intelligence, those of us concerned with constructing AGI systems or analyzing human intelligence require some sort of guidance. And, it seems better to draw guidance from conceptually clear, only partially rigorous thinking about the right issues, than from more rigorously mathematically or empirically grounded thinking about the wrong issues.

Page 20: Toward a General Theory of Embodied, Social General

As the discussion here has hopefully made clear, theorizing about “fully general intelligence” is not likely to be useful for understanding real-world general intelligence. And yet, one would like to be able to say something about general intelligence going beyond the description of particular systems like human brains or particular software systems. From an AGI point of view, studying the human brain and mind is valuable, but can also be confusing and misleading, if one’s goal is not to precisely emulate human intelligence but rather to make a different, perhaps in some respects better, sort of intelligence, operating in the same everyday world as humans. The approach taken here seeks to find a middle path, via qualitatively characterizing the class of systems that are generally intelligent with respect to the Embodied Communication Prior (with the Natural Knowledge Categories assumption); a class which, intuitively, appears to include both human brains, and the systems that would result from fully implementing and teaching certain contemporary AGI designs such as OpenCogPrime. Further research directions suggested are numerous. One would be to attempt to create a fully mathematically rigorous version of the ideas presented here – a quest that would doubtless involve a number of refinements and revisions of the ideas, though I suspect that the spirit would remain intact. Another, noted above, is to extend the scope of these ideas more thoroughly beyond the ECP to deal with more general classes of simplicity measures. One could also attempt to go further in the direction of directly deriving AGI designs from these theoretical considerations (rather than using the theory to analyze existing AGI designs, as done above), or using these considerations to analyze human brain function. All these avenues seem valuable, along with many others.

References

• Bateson, Gregory (1980). Mind and Nature. Bantam. • Fogassi, Leonardo, Pier Francesco Ferrari, Benno Gesierich, Stefano Rozzi,

Fabian Chersi, Giacomo Rizzolatti (2005). Parietal lobe: from action organization to intention understanding. Science 308: 662-667.

• Goertzel, Ben (2009). Might Humanlike Intelligence Require Hypercomputation?, submitted for publication

• Goertzel, Ben (1993). The Structure of Intelligence. Springer. • Goertzel, Ben and Stephan Vladimir Bugaj (2007). Stages of Cognitive

Development In Uncertain Logic Based AI Systems. Advances in Artificial General Intelligence, IOS Press.

• Goertzel, Ben, Matthew Ikle’, Izabela Freire Goertzel and Ari Heljakka (2008). Probabilistic Logic Networks. Springer Verlag

• Goertzel, Ben (2008). OpenCog Prime: Design for a Thinking Machine. Online wikibook, at http://opencog.org/wiki/OpenCogPrime

• Goertzel, Ben (2006). The Hidden Pattern: A Patternist Philosophy of Mind, Brown-Walker Press

• Goertzel, Ben (2006a). A System-Theoretic Analysis of Focused Cognition, and its Implications for the Emergence of Self and Attention. Dynamical Psychology.

Page 21: Toward a General Theory of Embodied, Social General

• Goertzel, Ben, Cassio Pennachin, Andre Senna, Thiago Maia and Guilherme Lamacie (2003). Novamente: An Integrative Architecture for Artificial General Intelligence. Proceedings of IJCAI-03 Workshop on Agents and Cognitive Modeling, Acapulco, August 2003

• Hutter, Marcus (2005). Universal Artificial Intelligence: Sequential Decisions based on Algorithmic Probability. Springer, 2005

• Looks, Moshe (2006). Competent Program Evolution. PhD Thesis, Computer Science Department, Washington University at St. Louis

• Piaget, Jean. The Construction of Reality in the Child. Routledge and Kegan Paul.

• Legg, S. and Marcus Hutter. (2006). A Formal Measure of Machine Intelligence. In Proc. Annual machine learning conference of Belgium and The Netherlands (Benelearn-2006). Ghent, 2006.

• Legg, S. and Marcus Hutter. (2006a). A Collection of Definitions of Intelligence. In B. Goertzel, editor, Advances in Artificial General Intelligence, IOS Press.

• Schmidhuber, J. (2006). Gödel machines: Fully Self-Referential Optimal Universal Self-Improvers. In B. Goertzel and C. Pennachin, eds.: Artificial General Intelligence, p. 119-226, 2006.

• Solomonoff, Ray. 1964. "A Formal Theory of Inductive Inference, Part I" Information and Control, Vol 7, No. 1 pp 1-22, March 1964.

• Solomonoff, Ray. 1964. "A Formal Theory of Inductive Inference, Part II" Information and Control, Vol 7, No. 2 pp 224-254, June 1964