Symbol grounding
Allikas: Lambda
Harnad, S. (1990) The Symbol Grounding Problem. Physica D 42: 335-346. --------------------------------------------------------------- THE SYMBOL GROUNDING PROBLEM Stevan Harnad Department of Psychology Princeton University Princeton NJ 08544 ABSTRACT: There has been much discussion recently about the scope and limits of purely symbolic models of the mind and about the proper role of connectionism in cognitive modeling. This paper describes the "symbol grounding problem": How can the semantic interpretation of a formal symbol system be made intrinsic to the system, rather than just parasitic on the meanings in our heads? How can the meanings of the meaningless symbol tokens, manipulated solely on the basis of their (arbitrary) shapes, be grounded in anything but other meaningless symbols? The problem is analogous to trying to learn Chinese from a Chinese/Chinese dictionary alone. A candidate solution is sketched: Symbolic representations must be grounded bottom-up in nonsymbolic representations of two kinds: (1) "iconic representations" , which are analogs of the proximal sensory projections of distal objects and events, and (2) "categorical representations" , which are learned and innate feature-detectors that pick out the invariant features of object and event categories from their sensory projections. Elementary symbols are the names of these object and event categories, assigned on the basis of their (nonsymbolic) categorical representations. Higher-order (3) "symbolic representations" , grounded in these elementary symbols, consist of symbol strings describing category membership relations (e.g., "An X is a Y that is Z"). Connectionism is one natural candidate for the mechanism that learns the invariant features underlying categorical representations, thereby connecting names to the proximal projections of the distal objects they stand for. In this way connectionism can be seen as a complementary component in a hybrid nonsymbolic/symbolic model of the mind, rather than a rival to purely symbolic modeling. Such a hybrid model would not have an autonomous symbolic "module," however; the symbolic functions would emerge as an intrinsically "dedicated" symbol system as a consequence of the bottom-up grounding of categories' names in their sensory representations. Symbol manipulation would be governed not just by the arbitrary shapes of the symbol tokens, but by the nonarbitrary shapes of the icons and category invariants in which they are grounded. KEYWORDS: symbol systems, connectionism, category learning, cognitive models, neural models 1. Modeling the Mind "1.1 From Behaviorism to Cognitivism." For many years the only empirical approach in psychology was behaviorism, its only explanatory tools input/input and input/output associations (in the case of classical conditioning; Turkkan 1989) and the reward/punishment history that "shaped" behavior (in the case of operant conditioning; Catania & Harnad 1988). In a reaction against the subjectivity of armchair introspectionism, behaviorism had declared that it was just as illicit to theorize about what went on in the head of the organism to generate its behavior as to theorize about what went on in its mind. Only observables were to be the subject matter of psychology; and, apparently, these were expected to explain themselves. Psychology became more like an empirical science when, with the gradual advent of cognitivism (Miller 1956, Neisser 1967, Haugeland 1978), it became acceptable to make inferences about the unobservable processes underlying behavior. Unfortunately, cognitivism let mentalism in again by the back door too, for the hypothetical internal processes came embellished with subjective interpretations. In fact, semantic interpretability (meaningfulness), as we shall see, was one of the defining features of the most prominent contender vying to become the theoretical vocabulary of cognitivism, the "language of thought" (Fodor 1975), which became the prevailing view in cognitive theory for several decades in the form of the "symbolic" model of the mind: The mind is a symbol system and cognition is symbol manipulation. The possibility of generating complex behavior through symbol manipulation was empirically demonstrated by successes in the field of artificial intelligence (AI). "1.2 Symbol Systems." What is a symbol system? From Newell (1980) Pylyshyn (1984), Fodor (1987) and the classical work of Von Neumann, Turing, Goedel, Church, etc. (see Kleene 1969) on the foundations of computation, we can reconstruct the following definition: A symbol system is: (1) a set of arbitrary "physical tokens" (scratches on paper, holes on a tape, events in a digital computer, etc.) that are (2) manipulated on the basis of "explicit rules" that are (3) likewise physical tokens and strings of tokens. The rule-governed symbol-token manipulation is based (4) purely on the shape of the symbol tokens (not their "meaning"), i.e., it is purely syntactic, and consists of (5) "rulefully combining" and recombining symbol tokens. There are (6) primitive atomic symbol tokens and (7) composite symbol-token strings. The entire system and all its parts -- the atomic tokens, the composite tokens, the syntactic manipulations (both actual and possible) and the rules -- are all (8) "semantically interpretable:" The syntax can be systematically assigned a meaning (e.g., as standing for objects, as describing states of affairs). According to proponents of the symbolic model of mind such as Fodor (1980) and Pylyshyn (1980, 1984), symbol-strings of this sort capture what mental phenomena such as thoughts and beliefs are. Symbolists emphasize that the symbolic level (for them, the mental level) is a natural functional level of its own, with ruleful regularities that are independent of their specific physical realizations. For symbolists, this implementation-independence is the critical difference between cognitive phenomena and ordinary physical phenomena and their respective explanations. This concept of an autonomous symbolic level also conforms to general foundational principles in the theory of computation and applies to all the work being done in symbolic AI, the branch of science that has so far been the most successful in generating (hence explaining) intelligent behavior. All eight of the properties listed above seem to be critical to this definition of symbolic.\** [footnote start] Paul Kube (personal communication) has suggested that (2) and (3) may be too strong, excluding some kinds of Turing Machine and perhaps even leading to an infinite regress on levels of explicitness and systematicity. [footnote end] Many phenomena have some of the properties, but that does not entail that they are symbolic in this explicit, technical sense. It is not enough, for example, for a phenomenon to be interpretable as rule-governed, for just about anything can be interpreted as rule-governed. A thermostat may be interpreted as following the rule: Turn on the furnace if the temperature goes below 70 degrees and turn it off if it goes above 70 degrees, yet nowhere in the thermostat is that rule explicitly represented. Wittgenstein (1953) emphasized the difference between explicit and implicit rules: It is not the same thing to "follow" a rule (explicitly) and merely to behave "in accordance with" a rule (implicitly).\** [footnote start] Similar considerations apply to Chomsky's (1980) concept of "psychological reality" (i. e., whether Chomskian rules are really physically represented in the brain or whether they merely "fit" our performance regularities, without being what actually governs them). Another version of the distinction concerns explicitly represented rules versus hard-wired physical constraints (Stabler 1985). In each case, an explicit representation consisting of elements that can be recombined in systematic ways would be symbolic whereas an implicit physical constraint would not, although both would be semantically "intepretable" as a "rule" if construed in isolation rather than as part of a system. [footnote end] The critical difference is in the compositeness (7) and systematicity (8) criteria. The explicitly represented symbolic rule is part of a formal system, it is decomposable (unless primitive), its application and manipulation is purely formal (syntactic, shape-dependent), and the entire system must be semantically interpretable, not just the chunk in question. An isolated ("modular") chunk cannot be symbolic; being symbolic is a systematic property. So the mere fact that a behavior is "interpretable" as ruleful does not mean that it is really governed by a symbolic rule.\** [footnote start] Analogously, the mere fact that a behavior is interpretable as purposeful or conscious or meaningful does not mean that it really is purposeful or conscious. (For arguments to the contrary, see Dennett 1983). [footnote end] Semantic interpretability must be coupled with explicit representation (2), syntactic manipulability (4), and systematicity (8) in order to be symbolic. None of these criteria is arbitrary, and, as far as I can tell, if you weaken them, you lose the grip on what looks like a natural category and you sever the links with the formal theory of computation, leaving a sense of "symbolic" that is merely unexplicated metaphor (and probably differs from speaker to speaker). Hence it is only this formal sense of "symbolic" and "symbol system" that will be considered in this discussion of the grounding of symbol systems. "1.3 Connectionist systems." An early rival to the symbolic model of mind appeared (Rosenblatt 1962), was overcome by symbolic AI (Minsky & Papert 1969) and has recently re-appeared in a stronger form that is currently vying with AI to be the general theory of cognition and behavior (McClelland, Rumelhart et al. 1986, Smolensky 1988). Variously described as "neural networks," "parallel distributed processing" and "connectionism," this approach has a multiple agenda, which includes providing a theory of brain function. Now, much can be said for and against studying behavioral and brain function independently, but in this paper it will be assumed that, first and foremost, a cognitive theory must stand on its own merits, which depend on how well it explains our observable behavioral capacity. Whether or not it does so in a sufficiently brainlike way is another matter, and a downstream one, in the course of theory development. Very little is known of the brain's structure and its "lower" (vegetative) functions so far; and the nature of "higher" brain function is itself a theoretical matter. To "constrain" a cognitive theory to account for behavior in a brainlike way is hence premature in two respects: (1) It is far from clear yet what "brainlike" means, and (2) we are far from having accounted for a lifesize chunk of behavior yet, even without added constraints. Moreover, the formal principles underlying connectionism seem to be based on the associative and statistical structure of the causal interactions in certain dynamical systems; a neural network is merely one possible implementation of such a dynamical system.\** [footnote start] It is not even clear yet that a "neural network" needs to be implemented as a net (i.e., a parallel system of interconnected units) in order to do what it can do; if symbolic simulations of nets have the same functional capacity as real nets, then a connectionist model is just a special kind of symbolic model, and connectionism is just a special family of symbolic algorithms. [footnote end] Connectionism will accordingly only be considered here as a cognitive theory. As such, it has lately challenged the symbolic approach to modeling the mind. According to connectionism, cognition is not symbol manipulation but dynamic patterns of activity in a multilayered network of nodes or units with weighted positive and negative interconnections. The patterns change according to internal network constraints governing how the activations and connection strengths are adjusted on the basis of new inputs (e.g., the generalized "delta rule," or "backpropagation," McClelland, Rumelhart et al. 1986). The result is a system that learns, recognizes patterns, solves problems, and can even exhibit motor skills. "1.4 Scope and Limits of Symbols and Nets." It is far from clear what the actual capabilities and limitations of either symbolic AI or connectionism are. The former seems better at formal and language-like tasks, the latter at sensory, motor and learning tasks, but there is considerable overlap and neither has gone much beyond the stage of "toy" tasks toward lifesize behavioral capacity. Moreover, there has been some disagreement as to whether or not connectionism itself is symbolic. We will adopt the position here that it is not, because connectionist networks fail to meet several of the criteria for being symbol systems, as Fodor & Pylyshyn (1988) have argued recently. In particular, although, like everything else, their behavior and internal states can be given isolated semantic interpretations, nets fail to meet the compositeness (7) and systematicity (8) criteria listed earlier: The patterns of interconnections do not decompose, combine and recombine according to a formal syntax that can be given a systematic semantic interpretation.\** [footnote start] There is some misunderstanding of this point because it is often conflated with a mere implementational issue: Connectionist networks can be simulated using symbol systems, and symbol systems can be implemented using a connectionist architecture, but that is independent of the question of what each can do qua symbol system or connectionist network, respectively. By way of analogy, silicon can be used to build a computer, and a computer can simulate the properties of silicon, but the functional properties of silicon are not those of computation, and the functional properties of computation are not those of silicon. [footnote end] Instead, nets seem to do what they do non symbolically. According to Fodor & Pylyshyn, this is a severe limitation, because many of our behavioral capacities appear to be symbolic, and hence the most natural hypothesis about the underlying cognitive processes that generate them would be that they too must be symbolic. Our linguistic capacities are the primary examples here, but many of the other skills we have -- logical reasoning, mathematics, chess-playing, perhaps even our higher-level perceptual and motor skills -- also seem to be symbolic. In any case, when we interpret our sentences, mathematical formulas, and chess moves (and perhaps some of our perceptual judgments and motor strategies) as having a systematic meaning or content, we know at first hand that that's literally true, and not just a figure of speech. Connectionism hence seems to be at a disadvantage in attempting to model these cognitive capacities. Yet it is not clear whether connectionism should for this reason aspire to be symbolic, for the symbolic approach turns out to suffer from a severe handicap, one that may be responsible for the limited extent of its success to date (especially in modeling human-scale capacities) as well as the uninteresting and ad hoc nature of the symbolic "knowledge" it attributes to the "mind" of the symbol system. The handicap has been noticed in various forms since the advent of computing; I have dubbed a recent manifestation of it the "symbol grounding problem" (Harnad 1987b). "2. The Symbol Grounding Problem" "2.1 The Chinese Room." Before defining the symbol grounding problem I will give two examples of it. The first comes from Searle's (1980) celebrated "Chinese Room Argument," in which the symbol grounding problem is referred to as the problem of intrinsic meaning (or "intentionality"): Searle challenges the core assumption of symbolic AI that a symbol system able to generate behavior indistinguishable from that of a person must have a mind. More specifically, according to the symbolic theory of mind, if a computer could pass the Turing Test (Turing 1964) in Chinese -- i.e., if it could respond to all Chinese symbol strings it receives as input with Chinese symbol strings that are indistinguishable from the replies a real Chinese speaker would make (even if we keep testing for a lifetime) -- then the computer would understand the meaning of Chinese symbols in the same sense that I understand the meaning of English symbols. Searle's simple demonstration that this cannot be so consists of imagining himself doing everything the computer does -- receiving the Chinese input symbols, manipulating them purely on the basis of their shape (in accordance with (1) to (8) above), and finally returning the Chinese output symbols. It is evident that Searle (who knows no Chinese) would not be understanding Chinese under those conditions -- hence neither could the computer. The symbols and the symbol manipulation, being all based on shape rather than meaning, are systematically interpretable as having meaning -- that, after all, is what it is to be a symbol system, according to our definition. But the interpretation will not be intrinsic to the symbol system itself: It will be parasitic on the fact that the symbols have meaning for us, in exactly the same way that the meanings of the symbols in a book are not intrinsic, but derive from the meanings in our heads. Hence, if the meanings of symbols in a symbol system are extrinsic, rather than intrinsic like the meanings in our heads, then they are not a viable model for the meanings in our heads: Cognition cannot be just symbol manipulation. "2.2 The Chinese/Chinese Dictionary-Go-Round." My own example of the symbol grounding problem has two versions, one difficult, and one, I think, impossible. The difficult version is: Suppose you had to learn Chinese as a second language and the only source of information you had was a Chinese/Chinese dictionary. The trip through the dictionary would amount to a merry-go-round, passing endlessly from one meaningless symbol or symbol-string (the definientes) to another (the definienda), never coming to a halt on what anything meant.\** [footnote start] Symbolic AI abounds with symptoms of the symbol grounding problem. One well-known (though misdiagnosed) manifestation of it is the so-called "frame" problem (McCarthy & Hayes 1969; Minsky 1974; NcDermott 1976; Pylyshyn 1987): It is a frustrating but familiar experience in writing "knowledge-based" programs that a system apparently behaving perfectly intelligently for a while can be foiled by an unexpected case that demonstrates its utter stupidity: A "scene-understanding" program will blithely describe the goings-on in a visual scene and answer questions demonstrating its comprehension (who did what, where, why?) and then suddenly reveal that it doesn't "know" that hanging up the phone and leaving the room does not make the phone disappear, or something like that. (It is important to note that these are not the kinds of lapses and gaps in knowledge that people are prone to; rather, they are such howlers as to cast serious doubt on whether the system has anything like "knowledge" at all.) The "frame" problem has been optimistically defined as the problem of formally specifying ("framing") what varies and what stays constant in a particular "knowledge domain," but in reality it's the problem of second-guessing all the contingencies the programmer has not anticipated in symbolizing the knowledge he is attempting to symbolize. These contingencies are probably unbounded, for practical purposes, because purely symbolic "knowledge" is ungrounded. Merely adding on more symbolic contingencies is like taking a few more turns in the Chinese/Chinese Dictionary-Go-Round. There is in reality no ground in sight: merely enough "intelligent" symbol-manipulation to lull the programmer into losing sight of the fact that its meaningfulness is just parasitic on the meanings he is projecting onto it from the grounded meanings in his own head. (I've called this effect the "hermeneutic hall of mirrors" [Harnad 1990]; it's the reverse side of the symbol grounding problem). Yet parasitism it is, as the next "frame problem" lurking around the corner is ready to confirm. (A similar form of over-interpretation has occurred in the ape "language" experiments [Terrace 1979]. Perhaps both apes and computers should be trained using Chinese code, to immunize their experimenters and programmers against spurious over-interpretations. But since the actual behavioral tasks in both domains are still so trivial, there's probably no way to prevent their being decrypted. In fact, there seems to be an irresistible tendency to overinterpret toy task performance itself, preemptively extrapolating and "scaling it up" conceptually to lifesize without any justification in practice.) [footnote end] -- Figure 1 (Chinese Dictionary Entry) about here. -- The only reason cryptologists of ancient languages and secret codes seem to be able to successfully accomplish something very like this is that their efforts are grounded in a first language and in real world experience and knowledge.\** [footnote start] Cryptologists also use statistical information about word frequencies, inferences about what an ancient culture or an enemy government are likely to be writing about, decryption algorithms, etc. [footnote end] The second variant of the Dictionary-Go-Round, however, goes far beyond the conceivable resources of cryptology: Suppose you had to learn Chinese as a first language and the only source of information you had was a Chinese/Chinese dictionary!\** [footnote start] There is of course no need to restrict the symbolic resources to a dictionary; the task would be just as impossible if one had access to the entire body of Chinese-language literature, including all of its computer programs and anything else that can be codified in symbols. [footnote end] This is more like the actual task faced by a purely symbolic model of the mind: How can you ever get off the symbol/symbol merry-go-round? How is symbol meaning to be grounded in something other than just more meaningless symbols?\** [footnote start] Even mathematicians, whether Platonist or formalist, point out that symbol manipulation (computation) itself cannot capture the notion of the intended interpretation of the symbols (Penrose 1989). The fact that formal symbol systems and their interpretations are not the same thing is hence evident independently of the Church-Turing thesis (Kleene 1969) or the Goedel results (Davis 1958, 1965), which have been zealously misapplied to the problem of mind-modeling (e.g., by Lucas 1964) -- to which they are largely irrelevant, in my view. [footnote end] This is the symbol grounding problem.\** [footnote start] Note that, strictly speaking, symbol grounding is a problem only for cognitive modeling, not for AI in general. If symbol systems alone succeed in generating all the intelligent machine performance pure AI is interested in -- e.g., an automated dictionary -- then there is no reason whatsoever to demand that their symbols have intrinsic meaning. On the other hand, the fact that our own symbols do have intrinsic meaning whereas the computer's do not, and the fact that we can do things that the computer so far cannot, may be indications that even in AI there are performance gains to be made (especially in robotics and machine vision) from endeavouring to ground symbol systems. [footnote end] "2.3 Connecting to the World." The standard reply of the symbolist (e.g., Fodor 1980, 1985) is that the meaning of the symbols comes from connecting the symbol system to the world "in the right way." But it seems apparent that the problem of connecting up with the world in the right way is virtually coextensive with the problem of cognition itself. If each definiens in a Chinese/Chinese dictionary were somehow connected to the world in the right way, we'd hardly need the definienda! Many symbolists believe that cognition, being symbol-manipulation, is an autonomous functional module that need only be hooked up to peripheral devices in order to "see" the world of objects to which its symbols refer (or, rather, to which they can be systematically interpreted as referring).\** [footnote start] The homuncular viewpoint inherent in this belief is quite apparent, as is the effect of the "hermeneutic hall of mirrors" (Harnad 1990). [footnote end] Unfortunately, this radically underestimates the difficulty of picking out the objects, events and states of affairs in the world that symbols refer to, i.e., it trivializes the symbol grounding problem. It is one possible candidate for a solution to this problem, confronted directly, that will now be sketched: What will be proposed is a hybrid nonsymbolic/symbolic system, a "dedicated" one, in which the elementary symbols are grounded in two kinds of nonsymbolic representations that pick out, from their proximal sensory projections, the distal object categories to which the elementary symbols refer. Most of the components of which the model is made up (analog projections and transformations, discretization, invariance detection, connectionism, symbol manipulation) have also been proposed in various configurations by others, but they will be put together in a specific bottom-up way here that has not, to my knowledge, been previously suggested, and it is on this specific configuration that the potential success of the grounding scheme critically depends. Table 1 summarizes the relative strengths and weaknesses of connectionism and symbolism, the two current rival candidates for explaining all of cognition single-handedly. Their respective strengths will be put to cooperative rather than competing use in our hybrid model, thereby also remedying some of their respective weaknesses. Let us now look more closely at the behavioral capacities such a cognitive model must generate. -- Table 1 about here -- 3. Human Behavioral Capacity Since the advent of cognitivism, psychologists have continued to gather behavioral data, although to a large extent the relevant evidence is already in: We already know what human beings are able to do. They can (1) discriminate, (2) manipulate,\** [footnote start] Although they are no doubt as important as perceptual skills, motor skills will not be explicitly considered here. It is assumed that the relevant features of the sensory story (e.g., iconicity) will generalize to the motor story (e.g., in motor analogs; Liberman 1982). In addition, large parts of the motor story may not be cognitive, drawing instead upon innate motor patterns and sensorimotor feedback. Gibson's (1979) concept of "affordances" -- the invariant stimulus features that are detected by the motor possibilities they "afford" -- is relevant here too, though Gibson underestimates the processing problems involved in finding such invariants (Ullman 1980). In any case, motor and sensory-motor grounding will no doubt be as important as the sensory grounding that is being focused on here. [footnote end] (3) identify and (4) describe the objects, events and states of affairs in the world they live in, and they can also (5) "produce descriptions" and (6) "respond to descriptions" of those objects, events and states of affairs. Cognitive theory's burden is now to explain how human beings (or any other devices) do all this.\** [footnote start] If a candidate model were to exhibit all these behavioral capacities, both linguistic (5-6) and robotic (i.e., sensorimotor), (1-3) it would pass the "Total Turing Test" (Harnad 1989). The standard Turing Test (Turing 1964) calls for linguistic performance capacity only: symbols in and symbols out. This makes it equivocal about the status, scope and limits of pure symbol manipulation, and hence subject to the symbol grounding problem. A model that could pass the Total Turing Test, however, would be grounded in the world. [footnote end] "3.1 Discrimination and Identification." Let us first look more closely at discrimination and identification. To be able to discriminate is to able to judge whether two inputs are the same or different, and, if different, how different they are. Discrimination is a relative judgment, based on our capacity to tell things apart and discern their degree of similarity. To be able to identify is to be able to assign a unique (usually arbitrary) response -- a "name" -- to a class of inputs, treating them all as equivalent or invariant in some respect. Identification is an absolute judgment, based on our capacity to tell whether or not a given input is a member of a particular category. Consider the symbol "horse." We are able, in viewing different horses (or the same horse in different positions, or at different times) to tell them apart and to judge which of them are more alike, and even how alike they are. This is discrimination. In addition, in viewing a horse, we can reliably call it a horse, rather than, say, a mule or a donkey (or a giraffe, or a stone). This is identification. What sort of internal representation would be needed in order to generate these two kinds of performance? "3.2 Iconic and categorical representations." According to the model being proposed here, our ability to discriminate inputs depends on our forming "iconic representations" of them (Harnad 1987b). These are internal analog transforms of the projections of distal objects on our sensory surfaces (Shepard & Cooper 1982). In the case of horses (and vision), they would be analogs of the many shapes that horses cast on our retinas.\** [footnote start] There are many problems having to do with figure/ground discrimination, smoothing, size constancy, shape constancy, stereopsis, etc., that make the problem of discrimination much more complicated than what is described here, but these do not change the basic fact that iconic representations are a natural candidate substrate for our capacity to discriminate. [footnote end] Same/different judgments would be based on the sameness or difference of these iconic representations, and similarity judgments would be based on their degree of congruity. No homunculus is involved here; simply a process of superimposing icons and registering their degree of disparity. Nor are there memory problems, since the inputs are either simultaneously present or available in rapid enough succession to draw upon their persisting sensory icons. So we need horse icons to discriminate horses. But what about identifying them? Discrimination is independent of identification. I could be discriminating things without knowing what they were. Will the icon allow me to identify horses? Although there are theorists who believe it would (Paivio 1986), I have tried to show why it could not (Harnad 1982, 1987b). In a world where there were bold, easily detected natural discontinuities between all the categories we would ever have to (or choose to) sort and identify -- a world in which the members of one category couldn't be confused with the members of any another category -- icons might be sufficient for identification. But in our underdetermined world, with its infinity of confusable potential categories, icons are useless for identification because there are too many of them and because they blend continuously\** [footnote start] Elsewhere (Harnad 1987a,b) I have tried to show how the phenomenon of "categorical perception" could generate internal discontinuities where there is external continuity. There is evidence that our perceptual system is able to segment a continuum, such as the color spectrum, into relatively discrete, bounded regions or categories. Physical differences of equal magnitude are more discriminable across the boundaries between these categories than within them. This boundary effect, both innate and learned, may play an important role in the representation of the elementary perceptual categories out of which the higher-order ones are built. [footnote end] into one another, making it an independent problem to identify which of them are icons of members of the category and which are not! Icons of sensory projections are too unselective. For identification, icons must be selectively reduced to those "invariant features" of the sensory projection that will reliably distinguish a member of a category from any nonmembers with which it could be confused. Let us call the output of this category-specific feature detector the "categorical representation" . In some cases these representations may be innate, but since evolution could hardly anticipate all of the categories we may ever need or choose to identify, most of these features must be learned from experience. In particular, our categorical representation of a horse is probably a learned one. (I will defer till section 4 the problem of how the invariant features underlying identification might be learned.) Note that both iconic and categorical representations are nonsymbolic. The former are analog copies of the sensory projection, preserving its "shape" faithfully; the latter are icons that have been selectively filtered to preserve only some of the features of the shape of the sensory projection: those that reliably distinguish members from nonmembers of a category. But both representations are still sensory and nonsymbolic. There is no problem about their connection to the objects they pick out: It is a purely causal connection, based on the relation between distal objects, proximal sensory projections and the acquired internal changes that result from a history of behavioral interactions with them. Nor is there any problem of semantic interpretation, or whether the semantic interpretation is justified. Iconic representations no more "mean" the objects of which they are the projections than the image in a camera does. Both icons and camera-images can of course be interpreted as meaning or standing for something, but the interpretation would clearly be derivative rather than intrinsic.\** [footnote start] On the other hand, the resemblance on which discrimination performance is based -- the degree of isomorphism between the icon and the sensory projection, and between the sensory projection and the distal object -- seems to be intrinsic, rather than just a matter of interpretation. The resemblance can be objectively characterized as the degree of invertibility of the physical transformation from object to icon (Harnad 1987b). [footnote end] "3.3 Symbolic Representations." Nor can categorical representations yet be interpreted as "meaning" anything. It is true that they pick out the class of objects they "name," but the names do not have all the systematic properties of symbols and symbol systems described earlier. They are just an inert taxonomy. For systematicity it must be possible to combine and recombine them rulefully into propositions that can be semantically interpreted. "Horse" is so far just an arbitrary response that is reliably made in the presence of a certain category of objects. There is no justification for interpreting it holophrastically as meaning "This is a [member of the category] horse" when produced in the presence of a horse, because the other expected systematic properties of "this" and "a" and the all-important "is" of predication are not exhibited by mere passive taxonomizing. What would be required to generate these other systematic properties? Merely that the grounded names in the category taxonomy be strung together into propositions about further category membership relations. For example: (1) Suppose the name "horse" is grounded by iconic and categorical representations, learned from experience, that reliably discriminate and identify horses on the basis of their sensory projections. (2) Suppose "stripes" is similarly grounded. Now consider that the following category can be constituted out of these elementary categories by a symbolic description of category membership alone: (3) "Zebra" = "horse" & "stripes"\** [footnote start] Figure 1 is actually the Chinese dictionary entry for "zebra," which is "striped horse." Note that the character for "zebra" actually happens to be the character for "horse" plus the character for "striped." Although Chinese characters are iconic in structure, they function just like arbitrary alphabetic lexigrams at the level of syntax and semantics. [footnote end] What is the representation of a zebra? It is just the symbol string "horse & stripes." But because "horse" and "stripes" are grounded in their respective iconic and categorical representations, "zebra" inherits the grounding, through its grounded symbolic representation. In principle, someone who had never seen a zebra (but had seen and learned to identify horses and stripes) could identify a zebra on first acquaintance armed with this symbolic representation alone (plus the nonsymbolic -- iconic and categorical -- representations of horses and stripes that ground it). Once one has the grounded set of elementary symbols provided by a taxonomy of names (and the iconic and categorical representations that give content to the names and allow them to pick out the objects they identify), the rest of the symbol strings of a natural language can be generated by symbol composition alone,\** [footnote start] Some standard logical connectives and quantifiers are needed too, such as not, and, all, etc. [footnote end] and they will all inherit the intrinsic grounding of the elementary set.\** [footnote start] Note that it is not being claimed that "horse," "stripes," etc. are actually elementary symbols, with direct sensory grounding; the claim is only that some set of symbols must be directly grounded. Most sensory category representations are no doubt hybrid sensory/symbolic; and their features can change by bootstrapping: "Horse" can always be revised, both sensorily and symbolically, even if it was previously elementary. Kripke (1980) gives a good example of how "gold" might be baptized on the shiny yellow metal in question, used for trade, decoration and discourse, and then we might discover "fool's gold," which would make all the sensory features we had used until then inadequate, forcing us to find new ones. He points out that it is even possible in principle for "gold" to have been inadvertently baptized on "fool's gold"! Of interest here are not the ontological aspects of this possibility, but the epistemic ones: We could bootstrap successfully to real gold even if every prior case had been fool's gold. "Gold" would still be the right word for what we had been trying to pick out all along, and its original provisional features would still have provided a close enough approximation to ground it, even if later information were to pull the ground out from under it, so to speak. [footnote end] Hence, the ability to discriminate and categorize (and its underlying nonsymbolic representations) has led naturally to the ability to describe and to produce and respond to descriptions through symbolic representations. 4. A Complementary Role for Connectionism The symbol grounding scheme just described has one prominent gap: No mechanism has been suggested to explain how the all-important categorical representations could be formed: How does the hybrid system find the invariant features of the sensory projection that make it possible to categorize and identify objects correctly?\** [footnote start] Although it is beyond the scope of this paper to discuss it at length, it must be mentioned that this question has often been begged in the past, mainly on the grounds of "vanishing intersections." It has been claimed that one cannot find invariant features in the sensory projection because they simply do not exist: The intersection of all the projections of the members of a category such as "horse" is empty. The British empiricists have been criticized for thinking otherwise; for example, Wittgenstein's (1953) discussion of "games" and "family resemblances" has been taken to have discredited their view. And current research on human categorization (Rosch & Lloyd 1978) has been interpreted as confirming that intersections vanish and that hence categories are not represented in terms of invariant features. The problem of vanishing intersections (together with Chomsky's [1980] "poverty of the stimulus argument") has even been cited by thinkers such as Fodor (1985, 1987) as a justification for extreme nativism. The present paper is frankly empiricist. In my view, the reason intersections have not been found is that no one has yet looked for them properly. Introspection certainly isn't the way to look. And general pattern learning algorithms such as connectionism are relatively new; their inductive power remains to be tested. In addition, a careful distinction has not been made between pure sensory categories (which, I claim, must have invariants, otherwise we could not successfully identify them as we do) and higher-order categories that are grounded in sensory categories; these abstract representations may be symbolic rather than sensory, and hence not based directly on sensory invariants. For further discussion of this problem, see Harnad 1987b). [footnote end] Connectionism, with its general pattern learning capability, seems to be one natural candidate (though there may well be others): Icons, paired with feedback indicating their names, could be processed by a connectionist network that learns to identify icons correctly from the sample of confusable alternatives it has encountered by dynamically adjusting the weights of the features and feature combinations that are reliably associated with the names in a way that (provisionally) resolves the confusion, thereby reducing the icons to the invariant (confusion-resolving) features of the category to which they are assigned. In effect, the "connection" between the names and the objects that give rise to their sensory projections and their icons would be provided by connectionist networks. This circumscribed complementary role for connectionism in a hybrid system seems to remedy the weaknesses of the two current competitors in their attempts to model the mind independently. In a pure symbolic model the crucial connection between the symbols and their referents is missing; an autonomous symbol system, though amenable to a systematic semantic interpretation, is ungrounded. In a pure connectionist model, names are connected to objects through invariant patterns in their sensory projections, learned through exposure and feedback, but the crucial compositional property is missing; a network of names, though grounded, is not yet amenable to a full systematic semantic interpretation. In the hybrid system proposed here, there is no longer any autonomous symbolic level at all; instead, there is an intrinsically dedicated symbol system, its elementary symbols (names) connected to nonsymbolic representations that can pick out the objects to which they refer, via connectionist networks that extract the invariant features of their analog sensory projections. 5. Conclusions The expectation has often been voiced that "top-down" (symbolic) approaches to modeling cognition will somehow meet "bottom-up" (sensory) approaches somewhere in between. If the grounding considerations in this paper are valid, then this expectation is hopelessly modular and there is really only one viable route from sense to symbols: from the ground up. A free-floating symbolic level like the software level of a computer will never be reached by this route (or vice versa) -- nor is it clear why we should even try to reach such a level, since it looks as if getting there would just amount to uprooting our symbols from their intrinsic meanings (thereby merely reducing ourselves to the functional equivalent of a programmable computer). In an intrinsically dedicated symbol system there are more constraints on the symbol tokens than merely syntactic ones. Symbols are manipulated not only on the basis of the arbitrary shape of their tokens, but also on the basis of the decidedly nonarbitrary "shape" of the iconic and categorical representations connected to the grounded elementary symbols out of which the higher-order symbols are composed. Of these two kinds of constraints, the iconic/categorical ones are primary. I am not aware of any formal analysis of such dedicated symbol systems,\** [footnote start] Although mathematicians investigate the formal properties of uninterpreted symbol systems, all of their motivations and intuitions clearly come from the intended interpretations of those systems (see Penrose 1989). Perhaps these too are grounded in the iconic and categorical representations in their heads. [footnote end] but this may be because they are unique to cognitive and robotic modeling and their properties will depend on the specific kinds of robotic (i.e., behavioral) capacities they are designed to exhibit. It is appropriate that the properties of dedicated symbol systems should turn out to depend on behavioral considerations. The present grounding scheme is still in the spirit of behaviorism in that the only tests proposed for whether a semantic interpretation will bear the semantic weight placed on it consist of one formal test (does it meet the eight criteria for being a symbol system?) and one behavioral test (can it discriminate, identify and describe all the objects and states of affairs to which its symbols refer?). If both tests are passed, then the semantic interpretation of its symbols is "fixed" by the behavioral capacity of the dedicated symbol system, as exercised on the objects and states of affairs in the world to which its symbols refer; the symbol meanings are accordingly not just parasitic on the meanings in the head of the interpreter, but intrinsic to the dedicated symbol system itself. This is still no guarantee that our model has captured subjective meaning, of course. But if the system's behavioral capacities are lifesize, it's as close as we can ever hope to get. References Catania, A. C. & Harnad, S. (eds.) (1988) The Selection of Behavior. The Operant Behaviorism of B. F. Skinner: Comments and Consequences. New York: Cambridge University Press. Chomsky, N. (1980) Rules and representations. "Behavioral and Brain Sciences" 3: 1-61. Davis, M. (1958) "Computability and unsolvability." Manchester: McGraw-Hill. Davis, M. (1965) "The undecidable." New York: Raven. Dennett, D. C. (1983) Intentional systems in cognitive ethology. Behavioral and Brain Sciences 6: 343 - 90. Fodor, J. A. (1975) "The language of thought" New York: Thomas Y. Crowell Fodor, J. A. (1980) Methodological solipsism considered as a research strategy in cognitive psychology. "Behavioral and Brain Sciences 3: 63 - 109. Fodor, J. A. (1985) Pr\*'ecis of "The Modularity of Mind." "Behavioral and Brain Sciences" 8: 1 - 42. Fodor, J. A. (1987) Psychosemantics Cambridge MA: MIT/Bradford. Fodor, J. A. & Pylyshyn, Z. W. (1988) Connectionism and cognitive architecture: A critical appraisal. Cognition 28: 3 - 71. Gibson, J. J. (1979) "An ecological approach to visual perception." Boston: Houghton Mifflin Harnad, S. (1982) Metaphor and mental duality. In T. Simon & R. Scholes, R. (Eds.) "Language, mind and brain." Hillsdale, N. J.: Lawrence Erlbaum Associates Harnad, S. (1987a) Categorical perception: A critical overview. In S. Harnad (Ed.) "Categorical perception: The groundwork of Cognition." New York: Cambridge University Press Harnad, S. (1987b) Category induction and representation. In S. Harnad (Ed.) "Categorical perception: The groundwork of Cognition." New York: Cambridge University Press Harnad, S. (1989) Minds, Machines and Searle. "Journal of Theoretical and Experimental Artificial Intelligence" 1: 5-25. Harnad, S. (1990) Computational Hermeneutics. "Social Epistemology" in press. Haugeland, J. (1978) The nature and plausibility of cognitivism. "Behavioral and Brain Sciences" 1: 215-260. Kleene, S. C. (1969) "Formalized recursive functionals and formalized realizability." Providence, R.: American Mathematical Society. Kripke, S.A. (1980) "Naming and Necessity." Cambridge MA: Harvard University Press Liberman, A. M. (1982) On the finding that speech is special. "American Psychologist" 37: 148-167. Lucas, J. R. (1961) Minds, machines and G\*"odel. Philosophy 36: 112-117. McCarthy, J. & Hayes, P. (1969) Some philosophical problems from the standpoint of artificial intelligence. In: Meltzer B. & Michie, P. "Machine Intelligence" Volume 4. Edinburgh: Edinburgh University Press. McDermott, D. (1976) Artificial intelligence meets natural stupidity. "SIGART Newsletter" 57: 4 - 9. McClelland, J. L., Rumelhart, D. E., and the PDP Research Group (1986) "Parallel distributed processing: Explorations in the microstructure of cognition," Volume 1. Cambridge MA: MIT/Bradford. Miller, G. A. (1956) The magical number seven, plus or minus two: Some limits on our capacity for processing information. "Psychological Review" 63: 81 - 97. Minsky, M. (1974) A framework for Representing knowledge. "MIT Lab Memo" # 306. Minsky, M. & Papert, S. (1969) "Perceptrons: An introduction to computational geometry." Cambridge MA: MIT Press (Reissued in an Expanded Edition, 1988). Newell, A. (1980) Physical Symbol Systems. "Cognitive Science 4:" 135 - 83. Neisser, U. (1967) "Cognitive Psychology" NY: Appleton-Century-Crofts. Cognitive Psychology Paivio, A. (1986) "Mental representation: A dual coding approach." New York: Oxford Penrose, R. (1989) "The emperor's new mind." Oxford: Oxford University Press Pylyshyn, Z. W. (1980) Computation and cognition: Issues in the foundations of cognitive science. "Behavioral and Brain Sciences" 3: 111-169. Pylyshyn, Z. W. (1984) "Computation and cognition." Cambridge MA: MIT/Bradford Pylyshyn, Z. W. (Ed.) (1987) "The robot's dilemma: The frame problem in artificial intelligence." Norwood NJ: Ablex Rosch, E. & Lloyd, B. B. (1978) "Cognition and categorization." Hillsdale NJ: Erlbaum Associates Rosenblatt, F. (1962) Principles of neurodynamics. NY: Spartan Searle, J. R. (1980) Minds, brains and programs. "Behavioral and Brain Sciences" 3: 417-457. Shepard, R. N. & Cooper, L. A. (1982) "Mental images and their transformations." Cambridge: MIT Press/Bradford. Smolensky, P. (1988) On the proper treatment of connectionism. "Behavioral and Brain Sciences" 11: 1 - 74. Stabler, E. P. (1985) How are grammars represented? "Behavioral and Brain Sciences" 6: 391-421. Terrace, H. (1979) Nim. NY: Random House. Turkkan, J. (1989) Classical conditioning: The new hegemony. "Behavioral and Brain Sciences 12:" 121 - 79. Turing, A. M. (1964) Computing machinery and intelligence. In: "Minds and machines, A. R. Anderson (ed.), Engelwood Cliffs NJ: Prentice Hall. Ullman, S. (1980) Against direct perception. "Behavioral and Brain Sciences" 3: 373 - 415. Wittgenstein, L. (1953) "Philosophical investigations." New York: Macmillan This figure should consist of the Chinese characters for "zebra," "horse" and "stripes," formatted as a dictionary entry, thus: "ZEBRA": "HORSE" with "STRIPES" Table 1. Connectionism Vs. Symbol Systems Strengths of Connectionism: (1) Nonsymbolic Function: As long as it does not aspire to be a symbol system, a connectionist network has the advantage of not being subject to the symbol grounding problem. (2) Generality: Connectionism applies the same small family of algorithms to many problems, whereas symbolism, being a methodology rather than an algorithm, relies on endless problem-specific symbolic rules. (3) "Neurosimilitude": Connectionist architecture seems more brain-like than a Turing machine or a digital computer. (4) Pattern Learning: Connectionist networks are especially suited to the learning of patterns from data. Weaknesses of Connectionism: (1) Nonsymbolic Function: Connectionist networks, because they are not symbol systems, do not have the systematic semantic properties that many cognitive phenomena appear to have. (2) Generality: Not every problem amounts to pattern learning. Some cognitive tasks may call for problem-specific rules, symbol manipulation, and standard computation. (3) "Neurosimilitude" : Connectionism's brain-likeness may be superficial and may (like toy models) camoflauge deeper performance limitations. Strengths of Symbol Systems: (1) Symbolic Function: Symbols have the computing power of Turing Machines and the systematic properties of a formal syntax that is semantically interpretable. (2) Generality: All computable functions (including all cognitive functions) are equivalent to a computational state in a Turing Machine. (3) Practical Successes: Symbol systems' ability to generate intelligent behavior is demonstrated by the successes of Artificial Intelligence. Weaknesses of Symbol Systems: (1) Symbolic Function: Symbol systems are subject to the symbol grounding problem. (2) Generality: Turing power is too general. The solutions to AI's many toy problems do not give rise to common principles of cognition but to a vast variety of ad hoc symbolic strategies.