From Neural Circuitry to Mechanistic Model-based Reasoning
-
Upload
independent -
Category
Documents
-
view
0 -
download
0
Transcript of From Neural Circuitry to Mechanistic Model-based Reasoning
1 From Neural Circuitry to Mechanistic Model-based Reasoning
Jonathan Waskan
To appear in Springer Handbook of Model-based Science
Model-based reasoning in science is often carried out in an attempt to understand the kinds of
mechanical interactions that might give rise to particular occurrences. One hypothesis regarding
in-the-head reasoning about mechanisms is that scientist rely upon mental models that are like
scale models in crucial respects. Behavioral evidence points to the existence of these mental
models, but questions remain about the neural plausibility of this hypothesis.
This chapter will provide an overview of the psychological literature on mental models of
mechanisms with a specific focus on the question of how representations that share the
distinctive features of scale models might be realized by neural machinations. It is shown how
lessons gleaned from the computational simulation of mechanisms and from neurological
research on mental maps in rats can be applied to make sense of how neurophysiological
processes might realize mental models.
The goal of this chapter is to provide readers with a general introduction to the central
challenge facing those would maintain that in-the-head model based reasoning about
mechanisms in science is achieved through the use of scale-model-like mental representations.
1.1 Overview
A central form of model-based reasoning in science, particularly in the special sciences, is
model-based reasoning about mechanisms. This form of reasoning can be effected with the aid of
external representational aids (e.g., formalisms, diagrams, and computer simulations) and
through the in-the-head manipulation of representations. Philosophers of science have devoted
most of their attention to the former, but the latter is arguably at the heart of most of what passes
for explanatory understanding in science (Sec. 1.2). Psychologists have long theorized that the
humans and other creatures (e.g., rats) reason about spatial, kinematic, and dynamic relationships
through the use of mental representations, often termed mental models, that are structurally
similar to scale models, though clearly the brain does not instantiate the very properties of a
modeled system in the way that scale models do (1.3). A key challenge facing this view is thus to
show that brains are capable of realizing representations that are like scale models in crucial
respects. There have been several failed attempts to show precisely this, but a look at how
computers are utilized to model mechanical interactions offers a useful way of understanding
how brains might realize mental representations of the relevant sort (Sec. 1.4). This approach
meshes well with current research on mental maps in rats. In addition, it has useful ramifications
for research in A.I. and logic (Sec. 1.5), and it offers a promising account of the generative
knowledge that scientists bring to bear when testing mechanistic theories while also shedding
light on the role that external representations of mechanisms play in scientific reasoning (1.6).
1.2 Mechanistic Reasoning in Science
A common reason that scientists engage in model-based reasoning is to derive information
that will enable them to explain or predict the behavior of some target system. Model-based
explanations provide scientists with a way of understanding how or why one or more
explanandum occurrences came about. A good model-based explanation will typically provide
the means for determining what else one ought to expect if that explanation is accurate1 – that is,
it will enable one to formulate predictions so that the explanation may (within widely known
limits) be tested.
1 One must bear in mind, however, that models are often accurate only in certain respects and to certain degrees [2].
Model-based reasoning can, corresponding to the diversity of representational structures
that count as models – including external scale models, biological models, mathematical
formalisms, and computer simulations – take many forms in science. As for what models
represent, it is now widely accepted that mechanisms are one of the principal targets of model-
based reasoning. This is most obviously true in the non-basic sciences (e.g., biology, medicine,
cognitive science, economics, and geology).
In philosophy of science, much of the focus on mechanisms has thus far been on the role
they play in scientific explanation. The idea that all genuine scientific explanation is mechanistic
began to gain traction in contemporary philosophy of science with the work of Peter Railton,
who claimed that “if the world is a machine – a vast arrangement of nomic connections – then
our theory ought to give us some insight into the structure and workings of the mechanism,
above and beyond the capability of predicting and controlling its outcomes…” [1]. Inspired by
Railton, Wesley Salmon abandoned his statistical-relevance model of explanation in favor of the
view that “the underlying causal mechanisms hold the key to our understanding of the world”
[3]. On this view, an “explanation of an event involves exhibiting that event as it is embedded in
its causal network and/or displaying its internal causal structure"[4]. Salmon was working in the
shadow of Carl Hempel’s covering law model of explanation, according to which explanations
involve inferences from statements describing laws and, in some cases, particular conditions.
Salmon tended, in contrast, to favor an ontic account, according to which explanations are out in
the world. He thought that progress in understanding those explanations requires ‘exhibiting’ the
relevant mechanisms. However, even though he rejected representational and inferential
accounts of explanation, he naturally recognized that reasoning about mechanisms, which
requires representations (models), plays a big part in the process of exhibiting those mechanisms.
A more recent formulation of the mechanistic account of explanation is supplied by
Machamer, Darden, & Craver, who claim that “Mechanisms are entities and activities organized
such that they are productive of regular changes from start or set-up to finish or termination
conditions” [5]. A central goal of science, on their view, is to formulate models, which take the
form of descriptions of mechanisms that render target occurrences intelligible:
Mechanism descriptions show how possibly, how plausibly, or how actually things work.
Intelligibility arises...from an elucidative relation between the explanans (the set-up
conditions and intermediate entities and activities) and the explanandum (the termination
condition or the phenomenon to be explained)....
As with ‘exhibiting’ for Salmon, the process of ‘elucidating’ how set-up conditions lead to
termination conditions requires a significant contribution from model-based reasoning.
Bechtel offers a related account of mechanisms. He claims: “A mechanisms is a structure
performing a function in virtue of its component parts. The orchestrated functioning of the
mechanism is responsible for one or more phenomena” [6]. As compared with other mechanists,
Bechtel is much more explicit about the role that model-based reasoning plays in science and
about the diverse forms of representation that may be involved (e.g., descriptions, diagrams,
scale models, and animal models). He is, moreover, among the few to acknowledge the
importance of ‘in-the-head’ model-based reasoning. He suggests that its central form may
involve a kind of mental animation. As Bechtel and Wright put it, “One strategy is to use
imagination to put one’s representation of the mechanism into motion so as to visualize how that
phenomenon is generated” [7]. Bechtel claims that the representations underlying this mental
animation process may have a structure similar to that of the diagrams scientist use in their
thinking and to the animated renderings of computer simulations scientists construct to represent
proposed mechanisms in action [6]. As for prediction, he notes:
what the scientist advances is a representation of a mechanism… She or he then evaluates
the representation by using it to reason about how such a mechanism would be expected to
behave under a variety of circumstances and testing these expectations against the behavior
of the actual mechanism [6].
In other words, once the scientist possesses a model of the mechanisms that may be responsible
for an occurrence, which may take the form of a mental model, he or she may then use it to
formulate predictions in order to test that model.
While external representational artifacts may sometimes be required in order to achieve
explanatory understanding of how a mechanism could produce a given phenomenon, plausibly
those artifacts are not themselves sufficient for explanatory understanding.2 Instead,
representational artifacts may have the important function of facilitating understanding by
enhancing the scientist’s ability to mentally simulate the process by which the proposed
mechanism would produce the target phenomenon.3 Through manipulation of those mental
simulations, scientists may also discover novel predictions of a given model.
1.3 The Psychology of Model-based Reasoning
Given the potentially crucial role that mental models play in the process of mechanistic
explanation and prediction, it may be that we cannot hope to attain a truly adequate, deep
understanding of science without first understanding how the mental modeling process works.
An obvious way of going about making sense of the role mental models play in science is to
inquire into the nature of those models themselves. A good question to ask here is: What form
2 For evidence that there is a crucial psychological component to explanatory understanding, see [8]. 3 As shown in Section 5.2, external representational aids may also enable forms of reasoning that would otherwise
(e.g., due to the complexity of the mechanism) be impossible.
must our mental models take if they are to play the role that they do in science? One increasingly
popular answer has its origins in Craik’s landmark monograph, The Nature of Explanation.
Regarding everyday reasoning, Craik suggests:
If the organism carries a 'small-scale model' of external reality and of its own possible
actions within its head, it is able to try out various alternatives, conclude which is the best
of them, react to future situations before they arise ... and in every way to react in a much
fuller, safer, and more competent manner to the emergencies which face it [9].
On Craik’s view, scientific explanation is just an extension of this everyday reasoning process –
that is, it involves the construction of internal world models that are akin to scale models.4
What may be considered the first attempt to put this view to experimental scrutiny came in
the prelude to the cognitive revolution with Edward Tolman’s seminal studies of spatial
navigation in rats [9]. In his most famous experiment, Tolman’s team placed rats in a simple
alley maze, similar to the one depicted in Figure 1a, and rewarded the animals with food when
they reached the end. After learning to perform the task without hesitation, the maze was
replaced with a radial maze similar to the one in Figure 1b, where the alley that the rats had
previously learned to traverse was blocked. Upon discovering this, the vast preponderance of rats
then chose the alley that led most directly to where the food source had been in previous trials.
On the basis such experiments, Tolman concluded that rats navigate with the aid of cognitive
maps of the relative spatial locations of objects in their environment.
[Insert Figure 1 About Here]
Figure 1. Alley maze (a) and radial maze (b), based on Tolman (1948).
4 Bechtel is explicit in crediting Craik, when he maintains that the use of mental models in scientific reasoning about
mechanisms is to be understood by analogy with the use of external images and scale models [6]. Fellow mechanists
Nancy Nersessian [11] and Paul Thagard [12] also credit Craik.
Later, Shepard & Metzler would show that the time it takes for people to determine if two
3D structures have the same shape is proportional to the relative degree of rotational
displacement of those structures [13]. One neat explanation for this finding is that people engage
in the mental rotation of 3D models of the two structures until they are aligned in such a fashion
as to enable easier comparison. In another landmark study of mental imagery, Kosslyn showed
that reaction times for scanning across mental images of a map was proportional to distance, but
not to the number of intervening objects, suggesting that spatial reasoning is better explained by
a process akin to scanning across a real map than to a process of sentence-based reasoning (e.g.,
working through a list structure) [14].
All of this research points to the existence of mental models of 2D and 3D spatial
relationships, but to support the full range of inferences implicated in mechanistic model-based
scientific reasoning, mental models would need to capture kinematic and dynamic relations as
well. There is some support for the existence of these models as well. For instance, Schwartz &
Black observed similar, proportional reaction times when subjects were asked to determine
whether or not a knob on one gear would, when that gear is rotated, fit into a grove on a
connecting gear (Figure 2a) [15]. Schwartz & Black found, moreover, that subjects were able to
“induce patterns of behavior from the results depicted in their imaginations” [16]. Subjects
might, for instance, infer and remember that the second in a series of gears will, along with every
other even-numbered gear, turn in the opposite direction of the drive gear (Figure 2b). Having
inferred this through simulation, the information becomes stored as explicit knowledge, thereby
eliminating the need to generate the knowledge anew for each new application.
[Insert Figure 2a & 2b About Here]
Figure 2. Knob and groove on connecting gears (a), based on Schwartz & Black [16]. Gears in
series (b), based on Schwartz & Black [17].
In addition, Hegarty [17] and Hegarty, Kriz, & Cate [16] have shown that mental modeling
of dynamic relationships is often effected in piecemeal fashion, a process that is much better
suited for tracing a sequence of interactions through a system than for simulating collections of
dynamic effects all at once. All of this research fits well with Norman’s early assessment of
mental models. He notes:
1. Mental models are incomplete.
2. People’s abilities to “run” their models are severely limited.
3. Mental models are unstable: People forget the details of the system they are using…
4. Mental models do not have firm boundaries: similar devices and operations get confused
with one another [19].
These limitations on the human ability to construct and manipulate mental models surely have a
great deal to do with more general limitations on the capacity of human working memory and
with the high cognitive load associated with creating, maintaining, and manipulating mental
models.
In everyday reasoning with mental models, the behaviors of the components structures in
our models will not typically be tied in any direct way to fundamental physical laws (e.g.,
Newtonian, quantum mechanical, or relativistic). Rather, many of the kinematic and dynamic
principles governing object behavior in our mental simulations will be rooted in early
experiences of collisions, impenetrability, balance and support, projectiles, blocking, and so forth
[20-2]. In addition, in everyday reasoning, and even more so in scientific reasoning about
mechanisms, many of the behaviors of the components of our models will not be the result of
early learning. Some of these will be one-off ‘brute’ events – such a meteor striking the earth, a
gene mutating, or a latch coming undone – for which one does not have or require (i.e., in order
to formulate a satisfactory answer to the question of why the explanandum occurred) any deeper
explanation. Such occurrences might be imposed upon a mental model in much the same way
that one would impose them – that is, through direct intervention – on a scale model. In the same
way, one could also impose newly learned or hypothesized regularities on a model. Some of
these might be discovered through simple induction (e.g., one might notice that one’s car engine
becomes louder in cold weather) or through prior model-based reasoning (e.g., as in Schwartz’
study with gears). However, when formulating mechanical explanations, particularly in science,
one sometimes simply hypothesizes, as a way of making sense of the available data, that a
particular regularity obtains. A good example of this is the way that the hypothesis of periodic
geomagnetic pole flipping was used to make sense of the patterns of magnetization in rocks
found lateral to mid-ocean rifts [2]. Such ideas accord well with recent work regarding
mechanistic explanation in the philosophy of science, where it is generally recognized that our
models of mechanisms typically bottom out at brute ‘activities’ [5] or ‘functions’ [6].
The above empirical research sheds light on the properties of the models we use to reason
about mechanisms in everyday life and in science. There is, in addition, a great deal of research
that simply hypothesizes that we do utilize such models to understand other cognitive processes
such as language comprehension [23-8], concepts [29], or learning [30].
The hypothesis of mental models has also been invoked by Johnson-Laird to explain
deductive reasoning, though here the term ‘mental model’ is used somewhat differently than it is
in the research cited above [31].5 Like many proponents of mental models, Johnson-Laird does
claim to be directly inspired by Craik, an inspiration that shows up in his suggestion that mental
5 Below, I explain, in greater depth, the contrast between deductive reasoning more generally and the mental models
approach to mechanistic reasoning espoused here.
models have “a structure that is remote from verbal assertions, but close to the structure of the
world as humans conceive it” [32]. However, if we look more closely at the way in which
Johnson-Laird employs the mental models hypothesis in accounting for reasoning processes
(viz., deductive, inductive, and abductive), it begins to look as though he has something very
different in mind. For instance, with regard to deductive reasoning – that is, reasoning that
mainly involves the semantic properties of top-neutral logical operators such as ‘if…then…,’
‘and,’ ‘all,’ and ‘some’ – Johnson-Laird proposes that we reason internally through a process not
unlike the formal method of truth table analysis. For instance, on Johnson-Laird’s view, the
conditional, “If the door is pushed, then the bucket will fall,” would be mentally represented as
something like the following spatial array, which lists those scenarios (models) that would be
consistent with the truth of the statement:6
door pushed bucket falls
¬ door pushed bucket falls
¬ door pushed ¬ bucket falls
If presented with the additional premise, “The bucket did not fall,” one could then eliminate all
but the last of these models, enabling a valid deduction to “The door was not pushed.” The
formal, topic-neutral nature of this strategy means that it works in exactly the same way
regardless of what items (e.g., balloons, satellites, or mice) we are reasoning about. To say
nothing of the viability of the approach, Johnson-Laird’s proposals regarding deductive (as well
as inductive and abductive) reasoning thus seem, except insofar as they appeal to such structures
as spatial arrays, at odds with his avowed view that mental models have a structure closer to the
world than to our descriptions of it.
6 ‘¬’ here signals negation.
1.4 Mental Models in the Brain: Attempts at Psycho-neural Reduction
While there has been considerable research on mental models in recent years, what has
been very slow to materialize is a demonstration that brains do or, what is even more worrisome,
that they could harbor mental models that are like scale models in crucial respects. One can see
how this might raise concerns about the mental models hypothesis. After all, if there can be no
such models, then the above explanatory appeals to mental models come out looking misguided
from the outset. At the same time, there is a competing hypothesis which faces no such
difficulties. In its most audacious form, it is the proposal that all of cognition is effected through
formal computational operations – that is, operations that involve the application of syntax-
sensitive inference rules to syntactically structured (viz., sentential) representations.
Proponents of the computational theory of cognition know that they have nothing to fear at
least with regard to the matter of whether or not brains are capable of realizing the relevant kinds
of syntax-crunching operations. McCulloch & Pitts showed, decades ago, that collections of
neuron-like processing units can implement logic gates and, in principle, a universal Turing
machine [33]. Indeed, it was in no small part because von Neumann recognized the functional
similarities between McCulloch-Pitts neurons and electronic switches (e.g., transistors) that he
was inspired to create the first fully programmable computers, ENIAC and EDVAC. More
recently, it has been shown that recurrent neural networks are, memory limitations
notwithstanding, capable of implementing computers that are Turing complete [34]. There is,
then, no longer any doubt that it is possible to bridge the divide between neural machinations and
syntax-crunching operations.
In contrast, a satisfactory demonstration that neural machinations might realize mental
models – that is, non-sentential mental representations that are like scale models in crucial
respects – has proven far more elusive. Indeed, difficulties arise the moment one tries to specify
what the ‘crucial respects’ might be, as is evidenced by the fact that each past attempt at doing
this has been argued, not without justification, to run afoul of one or the other of the following
two desiderata:
(i) an adequate account of mental models must be compatible with basic facts about the
brain.
(ii) an adequate account of mental models must be specific enough to distinguish mental
models from other kinds of representation (viz., sentential representations).
Again, this is no small matter, for given that brains are known to be capable of formal
computational operations, if it cannot be shown that they are also capable of realizing mental
models, this will cast doubt on all those psychological theories mentioned above that advert to
mental models. This is a concern made all-the-more pressing by the fact that proponents of the
computational theory of cognition have no shortage of alternative explanations for the behavioral
data cited in support of mental models. For instance, to the extent that people report having
model-like phenomenology, this might be dismissed as a mere epiphenomenon of the actual,
underlying computational operations. Similarly, to the extent that behavioral data, such as
reaction times, suggests reliance upon model-like mental representations that undergo continuous
transformations, this might be chalked up to demand characteristics (e.g., subjects may feel
compelled to pretend that they are scanning a map). Some of these specific objections could be
vulnerable in that they give rise to their own testable predictions [35], but, as explained below,
proponents of the computational theory have an ace up their sleeve, for computational accounts
are flexible enough to handle virtually any behavioral data. All of this is quite general, so let us
turn to some of the specific attempts to spell out the distinctive features of mental models.
From Structural to Functional Isomorphism
As we have seen, from the outset, the claim made on behalf of putative mental models is
that they are like scale models in one or more crucial respects. Of course, scale models are
themselves like the actual systems they represent in a very obvious respect: They instantiate the
very same properties as what they represent. It is thus no surprise that the dominant theme in
attempts to specify what makes mental models models is the invocation of one form or another of
isomorphism between mental models, scale models, and the modeled world.
Mere Isomorphism
The most straightforward form of isomorphism invoked in this literature is what might be
termed bare isomorphism, or isomorphism simpliciter, which is a purported relationship between
mental models and what they represent. Despite initial appearances, this is the form of
isomorphism that Craik seems to have had in mind. He claims, for instance: “By a model we thus
mean any physical or chemical system which has a similar relation-structure to that of the
process it imitates” [9]. Latter-day proponents of this proposal include Cummins [36] and
Hegarty, who, in an attempt to summarize the dominant view of mental models in psychology,
notes: “a mental model (or situation model) is a representation that is isomorphic to the physical
situation that it represents and the inference processes simulate the physical processes being
reasoned about” [18].
[Insert Figure 3 About Here]
Figure 3. Kelvin’s first tide predicting device. From:
http://www.sciencemuseum.org.uk/images/ManualSSPL/10300041.aspx
One serious concern about this approach is that it is too liberal, which is to say that it leads
one to classify too wide a range of representations as models. Consider, for instance, that one of
Craik's favored examples of a representation with a ‘similar relation structure’ to what it
represents is Kelvin’s Tide Predictor, a device that consists of an ingenious system of gears and
pulleys arranged so as to support truth-preserving inferences regarding the tides (Figure 3). Says
Craik, “My hypothesis then is that thought models, or parallels, reality—that its essential feature
is...symbolism, and that this symbolism is largely of the same kind as that which is familiar to us
in mechanical devices which aid thought and calculation" [9]. This, of course, is no different
from what proponents of the computational theory of cognition currently maintain. After all, any
syntax-crunching system capable of supporting truth-preserving inferences with respect to a
given physical system will have to be isomorphic with it – that is, there will have to be
correspondences between the parts and relations in the system and the components of the
representation – in ways that get preserved over the course of computation. To that extent, one
might even say that the inference process ‘simulates,’ or even ‘pictures’ [37], the process being
reasoned about. In short, then, the proposal that mental models are merely isomorphic with what
they represent is thus far too vague to satisfy desideratum (ii) above. Indeed, it is for this very
reason that researchers have tried to find a more restrictive notion of isomorphism, one that can
distinguish models from sentential representations.
Physical isomorphism
Perhaps the most restrictive such notion is that of structural [38] or physical [39]
isomorphism, which involves instantiating the very same properties, and arrangements thereof,
as the represented system. This appears to be the kind of isomorphism that Thagard has in mind
when he claims, “Demonstrating that neural representation can constitute mental models requires
showing how they can have the same relational structure as what they represent, both statically
and dynamically” [12; also see 40]. Thagard cites Kosslyn’s research as indicative of how this
demand might be met, and in Kosslyn too, we do find frequent appeals to structural
isomorphisms. For instance, noting the retinotopic organization of areas of visual cortex that are
implicated in mental imagery, Kosslyn claims, "[t]hese areas represent depictively in the most
literal sense . . ." [41].
Unfortunately, the postulation of physically isomorphic mental representations is highly
suspect for several reasons. To start with, the kind of retinotopy that one finds in areas such as
V1 is highly distorted relative to the world due to the disproportionate amount of cortex devoted
to the central portion of the retina (i.e., the fovea). A square in the visual field is thus not
represented in the cortex by sets of neurons that lie in straight, let alone in parallel, lines.
Moreover, visual representation seems not to be carried out through the activity of any single
retinotopically organized neural ensemble. Rather, vision involves the combined activity of a
variety of systems that are, to a considerable extent, anatomically and functionally distinct [42-
4]. Lastly, the kind of retinotopy pointed out by Kosslyn is restricted to two spatial dimensions,
and a 2D representational medium cannot realize representations that are physically isomorphic
with what they represent in three dimensions.7 Nor, a fortiori, can such a medium realize
representations that are physically isomorphic in both 3D and causal respects. Crudely put, there
are no literal buckets, balls, or doors in the brain.
Functional Isomorphism
The main problem with the appeal to physical isomorphism, one that has long been
appreciated, is that it fails to satisfy desideratum (i). As Shepard and Chipman note, "With about
7 Perhaps it is worth noting, as well, how inessential structural isomorphism is to information processing in neural
networks, even in the case of 2D retinotopic maps. The relative physical locations of neural cell bodies seems
irrelevant when compared to the patterns of connectivity between neurons, the strengths and valences of
connections, and the schemes of temporal coding the neurons employ. One would expect then that, so long as all of
this is preserved, cell bodies might be tangled up in arbitrary ways without affecting processing.
as much logic, one might as well argue that the neurons that signal that the square is green
should themselves be green!" [38]. Recognizing this, and recognizing the weakness of appeals to
mere isomorphism, Shepard and Chipman push for the following moderate notion of
isomorphism:
isomorphism should be sought-not in the first-order relation between (a) an individual
object, and (b) its corresponding internal representation-but in the second-order relation
between (a) the relations among alternative external objects, and (b) the relations among
their corresponding internal representations. Thus, although the internal representation for a
square need not itself be square, it should . . . at least have a closer functional relation to
the internal representation for a rectangle than to that, say, for a green flash or the taste of
persimmon [38, italics mine].
The appeal to second-order isomorphism would, they hoped, provide an alternative to physical
isomorphism that is both consistent with basic brain facts (desideratum (i)) and distinct from
sentential accounts (desideratum (ii)).
Another moderate account of isomorphism was put forward at the same time by
Huttenlocher, Higgins, & Clark [45]. They had a particular interest in how subjects make
ordering inferences (viz., those involving the ordering of three items along such dimensions as
size, weight, and height) like this one:
Linus is taller than Prior.
Prior is taller than Mabel.
Linus is taller than Mabel.
Huttenlocher, Higgins, & Clark suggested that subjects might use representations that “are
isomorphic with the physically realized representations they use in solving analogous problems
(graphs, maps, etc.). ...” [45]. The essence of their proposal was that the mental representations
that subjects form in order to solve such problems might function like spatial arrays rather than
like sentences. For instance, what seems distinctive about external sentential representations of
three-term ordering syllogisms like the one above is that, because each premise is represented in
terms of a distinct expression, terms that denote particular individuals must be repeated. On the
other hand, when such inferences are made with the aid of external spatial arrays, the terms need
not be repeated. For instance, one can make inferences about the taller-than relation on the basis
of the left-of relation with the help of marks on a paper like these:
L P M
In fact, the introspective reports obtained by Huttenlocher, Higgins, & Clark did support the idea
that subjects were constructing the functional equivalents of spatial arrays – for instance, subjects
reported that symbols representing individuals were not repeated – and on this basis they claimed
that subjects might be carrying out three-term ordering inferences using mental representations
that function like actual spatial arrays and unlike lists of sentences (also see [40]). This kind of
isomorphism is thus sometimes termed ‘functional’ isomorphism [39].
Shepard & Chipman [38] and Huttenlocher, Higgins, & Clark [45] were clearly after a
notion of isomorphism that satisfies desideratum (i). Unfortunately, the solutions they offer
appears, at least at first glance, to run afoul of desideratum (ii) – that is, appeals to functional
isomorphism, of either the first or second-order variety, seem not to distinguish between
computational representations and model-like representations. Huttonlocher et al. were among
the first to suspect this. They note, “It is not obvious at present whether any theory which
postulates imagery as a mechanism for solving problems can or cannot, in general, be
reformulated in an abstract logical fashion that, nevertheless makes the same behavioral
predictions” [45]. Anderson is generally credited with confirming this suspicion by pointing out
the possible tradeoffs that can be made between assumptions about representational structure and
those concerning the processes that operate over the representations [46]. He showed that the
possible structure-process tradeoffs render computational accounts flexible enough to handle
virtually any behavioral finding. Most have since endorsed his thesis that it is, at least after the
fact, always possible to “generate a propositional [i.e., sentential] model to mimic an imaginal
model" [46]. Alternatively, as Palmer puts it, if you create the right sentential model it will be
functionally isomorphic to what it represents in just the sense that a non-sentential model is
supposed to be [40].
Imagery and Perception
One last way in which one might try to satisfy the above desiderata, at least with regard to
spatial models, is to point out that visual mental imagery involves the utilization of visual
processing resources. Brooks [47] and Segal & Fusella [48], for instance, discovered that
performance on visual imagery tasks is diminished when subjects must perform a concurrent
visual processing task but not when they perform an auditory task – that is, they found that there
is interference between mental imagery and auditory perception but not between mental imagery
and visual perception (also see [36]). However, if these findings are meant to provide a model-
based alternative to computational theories, the attempt would appear to have the same
fundamental flaw as the appeal to functional isomorphism. As Block notes, because perceptual
processing can, in principle, also be explained in terms of computational processes, “the claim
that the representations of imagery and perception are of the same kind is irrelevant to the
controversy over pictorialist vs. descriptionalist interpretation of experiments like the image
scanning and rotation experiments ...” [49] (also see [46, 50]). That is, the claim that imagery
utilizes visual processing resources fails to satisfy desideratum (ii).
Distinctive Features of Scale Models
The overall realization problem facing putative mental models, then, is just that it has
proven exceedingly difficult to specify what sorts of representational structures mental models
are in a way that is consistent with basic brain facts but that also distinguishes models from
sentential representations. In order to finally see our way past these concerns, it will helpful if we
first take stock of a handful of features that are widely taken, even by proponents of the
computational theory of mind, to distinguish external images and scale models from sentential
representations. Three such features concern the sorts of entities, properties, and processes that
each form of representation is naturally suited for representing:
(1) Images and scale models are not naturally suited for representing abstract entities,
properties, and processes (e.g., war criminal, ownership, or economic inflation). They
are much better suited for representing concrete entities, properties, and processes
(e.g., a bucket, one object being over another, or compression).
(2) Images and scale models are not naturally suited for representing general categories
(e.g., triangles or automobiles). They are better suited for representing specific
instances of categories.8
(3) Images and scale models are not naturally suited for singling out specific properties
of specific objects [37, 50]. For instance, if would be difficult, using a scale model, to
represent just the fact that Fred's car is green, for any such model will simultaneously
represent many other properties, such as the number of doors and wheels, the body
type, and so on.
8 Note: Genera differ from abstracta in that the former can be concrete (e.g., rocks) and the latter can be specific
(e.g., the Enlightenment).
In contrast, sentential representations (e.g., those constructed using natural and artificial
languages) have little trouble representing abstracta (e.g., ‘war criminal’), genera (‘triangle’), and
specific properties of specific objects (e.g., ‘Fred’s car is green’).
While images and scale models are relatively disadvantaged in the above respects, they
are much better suited for supporting inferences regarding the consequences of alterations to
specific, concrete systems. The fact that syntax-crunching systems are quite limited in this regard
first came to light as a consequence of early work in formal-logic-inspired, sentence-and-rule-
based artificial intelligence (AI). The general problem confronting syntax-crunching approaches
came to be known as the frame problem [51].
In its original formulation, the frame problem had much to do with the challenge of
endowing a sentence-and-rule-based representational system with the ability to anticipate what
will not change following an alteration to the world (e.g., tipping over a bottle changes its
orientation but not its color). Today, however, the frame problem is regarded as something more
general – namely, the problem of endowing computational systems (and other artifacts) with the
kind of commonsense knowledge that the average human possesses about what will change and
what will stay the same following alterations to the objects in the world. As Hayes puts it:
The frame problem arises in attempts to formalise problem-solving processes involving
interactions with a complex world. It concerns the difficulty of keeping track of the
consequences of the performance of an action in, or more generally of the making of some
alteration to, a representation of the world [52].
The frame problem can actually be broken down into at least two component problems, the
prediction problem [53] and the qualification problem [54].
[Insert Figure 4 About Here]
Figure 4. A ‘toy’ world: A doorway, a bucket, and a ball. From [55].
As it confronts computational devices, the prediction problem can be summed up as
follows: In order to support inferences about the consequences of alterations to even simple
physical systems, a sentence-and-rule system would have to contain innumerable rules that
explicitly specify how objects will behave relative to one another following each of innumerable
possible alterations. For a simple illustration, consider what we all know about the consequences
of different ways of altering the items in Figure 4. We know, for example, what would happen
were we to use the bucket to throw the ball through the open doorway, were we to place the
bucket over the ball and slide the bucket through the doorway, were we to set the bucket
containing the ball atop the slightly ajar door and then shove the door open, and so on
indefinitely. To endow a sentence-and-rule system with the ability to predict the consequences of
these various alterations, one would have to build in, corresponding to each one, a separate data
structure specifying the starting conditions, the alteration, and the consequences of that
alteration. If these take the form of conditional statements, the system could then make
inferences utilizing domain-general (e.g., topic-neutral, deductive) machinery. Alternatively, the
information could be encoded directly as domain-specific inference rules (e.g., production-
system operators). Either way, from an engineering standpoint, the problem that quickly arises is
that no matter how many of these statements or rules one builds into the knowledge base of the
system, there will generally be countless other bits of commonsense knowledge that one has
overlooked. Notice, moreover, that scaling the scenario up even slightly (e.g., such that it now
includes a board) has an exponential effect on the number of potential alterations and, as such, on
the number of new data structures that one would have to incorporate into one's model [53]. As
Hayes says, “One does not want to be obliged to give a law of motion for every aspect of the
new situation…especially as the number of frame axioms increases rapidly with the complexity
of the problem” [52]. Moreover, as explained in the manual for a past incarnation of the
production system Soar, “when working on large (realistic) problems, the number of operators
[i.e., domain-specific rules] that may be used in problem solving and the number of possible state
descriptions will be very large and probably infinite” [56].
As if the prediction problem were not problem enough, it is actually compounded by the
other facet of the frame problem, the qualification problem [54]. This is because, in order to
capture what the average human knows about the consequences of alterations to a physical
system, not only would innumerable distinct conditionals or inference rules be required, but each
would have to be qualified in an indefinite number of ways. Notice, for instance, that placing the
bucket over the ball and sliding it through the doorway will result in the ball being transferred to
the other side of the wall, but only if it is not the case that there is a hole in the floor into which
the ball might fall, there is a hole in the bucket through which it might escape, the ball is fastened
securely to the floor, and so on indefinitely. To once again quote Hayes, “Almost any general
belief about the result of his own actions may be contradicted by the robot’s observations….there
are no end to the different things that can go wrong, and he cannot be expected to hedge his
conclusions round with thousands of qualifications” [52]. Thus, to capture what the average
human knows, if only implicitly, about the consequences of this one alteration, all of the relevant
qualifications would have to be added to the relevant sentence or rule. Once again, in realistic
situations, the challenge of specifying all of the qualifications is magnified exponentially.
The general failing of sentence-and-rule-based representations that the frame problem
brings to light is that they only support predictions concerning the consequences of alterations
and the defeaters of those consequences if those alterations, consequences, and defeaters have
been spelled out, antecedently and explicitly, as distinct data structures. Representations of this
sort – that is, representations that require distinct structures to support predictions regarding the
consequences of each type of alteration to the represented system – are sometimes termed
extrinsic representations.9
It is worth a quick digression to note that, while the terminology has changed, these general
concerns about the limitations of extrinsic representations antedate work in contemporary AI by
over three-hundred years. They show up, for instance, in Descartes’ best-explanation arguments
for dualism in his Discourse on the Method. Descartes there despairs of there ever being a
mechanical explanation for, or an artifact that can duplicate, the average human’s boundless
knowledge of the consequences of interventions on the world:
If there were machines which bore a resemblance to our bodies and imitated our actions…
we should still have two very certain means of recognizing that they were not real men…
[Firstly, humans have the ability to converse.] Secondly, even though some machines might
do some things as well as we do them…they would inevitably fail in others, which would
reveal that they are acting not from understanding, but only from the disposition of their
organs. For whereas reason is a universal instrument, which can be used in all kinds of
situations, these organs need some particular action; hence it is for all practical purposes
impossible for a machine to have enough different organs to make it act in all the
contingencies of life in the way in which our reason makes us act.
Descartes thought that to match wits with even a ‘dull-witted’ human, any natural or artificial
device would need, per impossibile, to rely upon an infinite number of specific sensory-motor
routines – which bear a striking resemblance to production-system operators – for each new
9 The intrinsic-extrinsic distinction discussed here was introduced by Palmer [39] but modified by Waskan [57-8].
situation the device might confront. What Descartes could not imagine, because he thought that
all such knowledge had to be represented explicitly, was the possibility of (to use Chomsky’s
term) a generative inference mechanism – that is, one that embodies boundless knowledge of
implications through finite means.
What Descartes failed to notice was that there were already artifacts (i.e., scale models) that
exhibited the requisite generativity. Indeed, in contemporary AI, the benefits of an appeal to
scale-model-like representations are now well known. Starting with the prediction problem, one
can use a reasonably faithful scale model of the setup depicted in Figure 4 in order to predict
what would happen were one to use the bucket to throw the ball through the open doorway, were
one to place the bucket over the ball and slide the bucket through the doorway, were one to set
the bucket containing the ball atop the slightly ajar door and then shove the door open, and so on
indefinitely. To use Haugeland's terms, the side effects of alterations to such representations
mirror the side effects of alterations to the represented system automatically [59] – which is to
say, without requiring their explicit specification.10 Notice also that incremental additions to the
represented system will only have an incremental effect on what needs to be built into the
representation. The addition of a board to the system above, for instance, can be handled by the
simple addition of a scale model of the board to the representation.
Nor do scale models suffer from the qualification problem. To see why, notice that much of
what is true of a modeled domain will be true of a scale model of that domain. For instance, with
regard to a scale model of the setup in Figure 4, it is true that the scale model of the ball will fall
out of the scale model of the bucket when it is tipped over, but only if the ball is not wedged into
10 This only holds, of course, to the extent that the model is a faithful reproduction. Unless the model is a perfect
replica, which includes being to scale, there will be some limits on inferential fidelity, though this does not
undermine the claim that scale models are generative.
the bucket, there is no glue in the bucket, and so on indefinitely. Just like our own predictions,
the predictions generated using scale models are implicitly qualified in an open-ended number of
ways. With scale models, all of the relevant information is implicit in the models and so there is
no need to represent it all explicitly using innumerable distinct data structures. Representations
of this sort are termed intrinsic representations. Summing up, scale models are immune to the
frame problem, for one can use them to determine, on an as-needed basis, both the consequences
of countless alterations to the modeled system and the countless possible defeaters of those
consequences – that is, one simply manipulates the model in the relevant ways and reads off the
consequences.
Does Computational Realization Entail Sentential Representation?
The above distinguishing features can help us to know better whether we are dealing with
model-like or sentence-like representations and, ultimately, to appreciate how one might bridge
the gap from neurophysiology to mental models. As noted above, a similar bridge was
constructed from neurophysiology to computational processes by showing that artifacts (e.g.,
collections of McCulloch-Pitts neurons or wires and transistors) characterized by a complex
circuitry not unlike that of real brains can be configured so as to implement, at a higher level of
abstraction, processes that exhibit the hallmarks of traditional syntax-crunching. Because
neurons have similar information-processing capabilities as these artifacts, implementing a set of
formal operations on an electronic computer is already very nearly an existence proof that brain-
like systems can realize the same set of operations.
Might this strategy offer a template for constructing a similar bridge to high-level models?
There is surely no shortage of computer simulations of mechanical systems, and at least as they
are depicted on a computer’s display, these simulations look for all the world like images and
scale models. Many would argue, however, that this approach to bridging the neuron-model
divide is a non-starter. The worry, in short, is that it fails to satisfy desideratum (ii) above. To see
why, it will be helpful to look at the kinds of computational models of mental imagery offered up
by researchers such as Kosslyn [14] and Glasgow and Papadias [60].
Kosslyn's model of mental imagery has several components [14]. One is a long-term store
that contains sentential representations of the shape and orientation of objects. These descriptive
representations are utilized for the construction of representations in another component, the
visual buffer, which encodes the same information in terms of the filled and empty cells of a
computation matrix. The cells of the matrix are indexed by x, y coordinates, and the descriptions
in long-term memory take the form of polar coordinate specifications (i.e., specifications of the
angle and distance from a point of origin) of the locations of filled cells. Control processes
operate over the co-ordinate specifications in order to perform such functions as panning in and
out, scanning across, and mental rotation.
One distinctive feature of actual (e.g., paper-and-ink) spatial matrix representations is that
they embody some of the very same properties and relationships (viz., spatial ones) as – which is
just to say that they are physically isomorphic with – the things they represent. But Kosslyn's
computational matrix representations (CMRs) are clearly not physically isomorphic with what
they represent. After all, Kosslyn’s visual buffer representations are not 'real' matrix
representations that utilize cells arranged in Euclidean space; they are computational matrix
representations. To be sure, modelers may sometimes see literal pictures on the output displays
of their computers, but the representations of interest are located in the CPU (viz., in RAM) of
the computer running the model. Accordingly, the control operations responsible for executing
representational transformations like rotation do not make use of inherent spatial constraints, but
rather they operate over the coordinate specifications that are stored in the computer's memory.
Details aside, at a certain level of description, there can be no doubt that the computer is
implementing a set of syntax-sensitive rules for manipulating syntactically structured
representations; this is what computers do. As Block puts it, “Once we see what the computer
does, we realize that the representation of the line is descriptional” [49]. The received view, then,
a view that has gone nearly unchallenged, is that if a representation of spatial, kinematic, or
dynamic properties is implemented using a high-level computer program, then the resulting
representations must be sentential in character [49, 61-2].11
It would thus seem that the strongest claim that can possibly be supported with regard to
CMRs is that they function like images. Yet, as Anderson notes, it is always possible, through
clever structure-process trade-offs, to create a sentential system that mimics an imagistic one
[46]. Indeed, rather than supporting the mental models framework, one might well take computer
simulations of mental modeling as concrete evidence for Anderson’s claim. Likewise, there is a
case to be made that CMRs and their brethren are, unlike scale models, extrinsic representations
[62]. After all, the computers that run them implement syntax-sensitive rules that provide explicit
specifications of the consequences of alterations. This is no small matter. From the standpoint of
cognitive science, one of the most important virtues of the hypothesis that we utilize mental
representations akin to scale models was that scale models constitute intrinsic representations of
interacting worldly constraints and are thus immune to the frame problem. One could, then, be
forgiven for thinking that any attempt to build a bridge from neurons to models by following the
template set by computational theories – that is, by noting that certain computational artifacts
11 That Fodor shares this sentiment is suggested by his claim that "if ... you propose to co-opt Turing's account of the
nature of computation for use in a cognitive psychology of thought, you will have to assume that thoughts
themselves have syntactic structure" [63].
instantiate the relevant kind of processing – will be doomed to failure from the outset.
What About POPI?
Consider, however, that upon gazing directly at a vast collection of electrical or
electrochemical circuits, one will see no evidence of the harboring or manipulation of sentential
representations. In Monadology, Leibniz turned an analogous observation about perceptual
experience into an objection to materialism:
It must be confessed, moreover, that perception, and that which depends on it, are
inexplicable by mechanical causes, that is, by figures and motions, And, supposing that
there were a mechanism so constructed as to think, feel and have perception, we might
enter it as into a mill. And this granted, we should only find on visiting it, pieces which
push one against another, but never anything by which to explain a perception.
A similar objection might be leveled regarding computational processes. Again, one sees no
evidence of this kind of processing when one looks at electronic or electrochemical circuitry.
Clearly something has gone wrong.
What Leibniz overlooked – and this may be because he lacked the conceptual tools made
available by the information age – was a grasp of the principle of property independence (POPI).
The basic idea of POPI is that properties characterizing a system when it is studied at a relatively
low level of abstraction are often absent when it is studied at a higher level, and vice versa. It is
POPI that allows computer scientists to say that a system which is characterized by electronic
switches and relays at level n may nevertheless be best described in terms of the storing of bits of
information in numerically addressable memory registers at level n+1 and in terms of the
application of syntax-sensitive rules to syntactically structured representations at level n+2. It is
also the very thing that enables proponents of computational theories of cognition to say that
brains and computational artifacts are, despite superficial appearances, capable of implementing
the application of syntax-sensitive rules to syntactically structured representations.
However, when proponents of computational theories of cognition insist that computational
implementation (e.g., of CMRs) entails sentential representation, they are turning their backs on
the very principle that enabled them to bridge divide between low-level circuitry and high-level
computational operations; they are turning their back on POPI. Indeed, nothing about POPI
entails that all syntax-crunching systems must be characterized in terms of sentences and
inference rules at the highest level of abstraction. POPI thus opens up at least logical space for
systems that engage in syntax-crunching operations at one level but that harbor and manipulate
non-sentential models at a higher level.
In point of fact, in this logical space reside actual systems, including finite element models
(FEMs). These were first developed in the physical (e.g., civil and mechanical) engineering
disciplines for testing designs, but they have since become a staple tool in the sciences for
exploring the ramifications of theories, generating novel predictions, and facilitating
understanding. For our current purposes, what matters most about FEMs is that they provide an
existence proof that computational processes can realize non-sentential representations that are
like scale models and unlike sentential representations in all of the crucial respects listed above.
To see why, notice first that there are (among others) two important levels of abstraction at
which a given FEM may be understood. As with scale models, one may understand FEMs at the
relatively low level of the principles that govern their implementing medium. What one finds at
this level are sentential specifications of coordinates (e.g., for polygon vertices) along with rules,
akin to the fundamental laws of nature12, which constrain how those coordinates may change
12 For a close analogy, think the basic rules of Conway’s Game of Life.
(e.g., due to collisions and loads) (see Figure 5). When a given model is ‘run,’ at this low level
one finds a massive number of iterative number crunching operations. Not unlike Leibniz,
enemies of the idea of computationally realized non-sentential models have seized upon this low
level with their suggestion that computational systems harbor only sentential representations. At
this level, however, it is not even obvious that we are dealing with representations (i.e., of
worldly objects and properties) at all, any more than we are, for instance, when we fixate upon
the constraints governing the behaviors of individual Lego blocks.
One only finds representations of objects when one turns to the higher level of the models
that are realized, and multiply realizable, by the aforementioned modeling media. And when we
take a close look at the properties of these high-level FEMs, we find that they share several
characteristics that have long been taken, including by those who suggest that computational
implementation entails sentential representation, to distinguish sentential representations from
scale models.
[Insert Figure 5 About Here.]
Figure 5. Polymesh representation of a blunt impact to a semi-rigid sheet of material. From [57].
To start with, like scale models and unlike sentential representations, FEMs are not (by
themselves) naturally suited to representing abstract entities, properties, and processes (e.g., war
criminal, ownership, economic inflation). They are much better suited for representing concrete
entities, properties, and processes (e.g., a bucket, one object being over another, and
compression). Nor are FEMs naturally suited to representing general categories (e.g., triangles or
automobiles). They are far better suited for representing specific instances of those categories.
Lastly, FEMs are not naturally suited to singling out specific properties of specific objects. For
instance, using an FEM, if would be difficult to represent just the fact that Fred's car is green, for
any such model will simultaneously represent many other properties, such as the number of
doors and wheels, the body type, and so on. In short, just like scale models, FEMs are always
representations of specific, concrete systems. By these reasonable standards, FEMs ought to be
considered computationally-realized non-sentential models that are the close kin of scale models.
The case for this claim looks even stronger once we consider whether or not FEMs
constitute intrinsic representations. As we have seen, the received view is that FEMs and their
brethren (e.g., CMRs) are extrinsic representations, for the constraints governing how the
coordinates of primitive modeling elements may change must be encoded antecedently and
explicitly. Indeed, at the level of coordinates and transformation rules, one gets nothing ‘for
free.’ However, once a modeling medium has been used to construct a suitable FEM of a
collection of objects, the model can then be altered in any of countless ways in order to
determine the possible consequences of the corresponding alterations to the represented objects.
One can, for instance, use a high-fidelity FEM of the door, bucket, ball system to infer, among
other things, what would happen were we to place the bucket over the ball and slide the bucket
through the doorway, what would happen were the bucket used to throw the ball at the open
doorway, what would happen were the air pressure dramatically decreased, and so on
indefinitely [57]. The consequences of these alterations need not be anticipated or explicitly
incorporated into the system. Indeed, as with scale models, much of the point of constructing
FEMs is to find out how a system will behave in light of whichever alterations an engineer or
scientist can dream up.
It bears repeating that it is not at the level of the primitive operations of an implementation
base that we find intrinsic representations, but at the level of the representations realized by a
given, primitively constrained implementation base. Part of what justifies this claim is the fact
that certain constraints will be inviolable at the level of the model, and thus a great deal of
information will be implicit in the model, because it has been implemented using a particular
kind of medium. As Pylyshyn notes:
the greater number of formal properties built into a notation in advance, the weaker the
notational system’s expressive power (though the system may be more efficient for cases to
which it is applicable). This follows from the possibility that the system may no longer be
capable of expressing certain states of affairs that violate assumptions built into the
notation. For example, if Euclidean assumptions are built into a notation, the notation
cannot be used to describe non-Euclidean properties. . . . [61].
This, in fact, is very close to an apt characterization of what is going on in the case of FEMs.
Given that a particular model has been realized through the use of a primitively constrained
medium, certain constraints will be inviolable at the representational level and a great deal of
information will be implicit [57]. As Mark Bickhard (in correspondence) summarizes the point:
“Properties and regularities are only going to be 'intrinsic' at one level of description if they are
built-in in the realizing level – or else they are ontologically 'built-in' as in the case of strictly
spatial relationships in physical scale models.” While scale models are intrinsic for the latter
reason, FEMs are intrinsic for the former. This shows up in the fact that FEMs exhibit a
comparable degree of generativity to scale models and a correlative immunity to the frame
problem. Like scale models, FEMs provide a finite embodiment of boundless tacit knowledge,
which can be made explicit at any time, of the consequences of innumerable alterations to the
systems they represent.
So how does all of this square with Anderson’s [46] contention that it is always possible to
construct a sentential system to mimic an imagistic or model-based system or Palmer’s [39]
claim that if you create the right sentential model it will be functionally isomorphic to what it
represents in just the sense that a non-sentential model is supposed to be? Anderson and Palmer
are surely right that, post hoc, one can always constrain a sentential representational system so
that it mimics the output of a model-based system, but the post hoc character of the strategy is
precisely what gets sentential approaches into trouble vis-à-vis the frame problem. Consider, for
instance, that the traditional AI approach is to take any physical implication of which humans
express knowledge and, after the fact, to build it into the knowledge base of one’s system as a
sentence or inference rule.13 But to solve, or rather to avoid, the frame problem, one must rely
upon representations that embody all of this boundless information as tacit knowledge – that is,
the information cannot be explicitly encoded at the outset, but it can later be generated, and
thereby become explicit knowledge, on an as-needed basis. Put simply, to exhibit anything
approaching true functional isomorphism with scale models, what is needed are high-level,
intrinsic, non-sentential models.
To sum up, those who would contend that FEMs (or even CMRs of 2D spatial properties)
are, qua computational, necessarily extrinsic and sentential have overlooked the fact that there
are multiple levels of abstraction at which a given computational model can be understood. At
the relatively low level of the modeling medium, there are unquestionably extrinsic
representations of the principles governing the permissible transformation of primitive modeling
elements. At a higher level, one finds models that share many distinguishing features, including
immunity to the frame problem, with the scale models they were in large part invented to
replace. Thus, we find once again that FEMs are like scale models and unlike paradigmatic
sentential representations.
13 Despite its shortcomings, this strategy is alive and well, as is evidenced by Lenat’s massive ongoing Cyc project.
Bridging the Divide
All of this is bears directly on the longstanding concern that there is no way to bridge the
divide between neural machinations and the non-sentential models hypothesized by proponents
of mental models. What the foregoing makes clear is that computational processes can realize
non-sentential models that share with scale models the main characteristics that distinguish non-
sentential models from sentential representations. Given that the brain is capable, at least in
principle, of realizing any such computational processes, then one must also agree that brains can
realize non-sentential models. Thus, by appealing to the above distinguishing features of scale
models, we see that there is an account of mental models that (i) distinguishes them (on multiple
grounds) from sentential representations and (ii) is compatible with basic facts about how brains
operate. All of this provides a much-needed foundation for all of that psychological work cited
above that adverts to mental models.
One advantage of showing that low-level computations can realize higher-level mental
models is that it renders the mental models hypothesis robust enough to withstand the discovery
that the brain is a computational system at some level of description. Even if the brain is not a
computational system (i.e., in the syntax-crunching sense), the manner in which computational
systems realize intrinsic, non-sentential models will nevertheless remain quite instructive. It
suggests a general recipe for the creation of intrinsic models that can be followed even without
the computational intermediary: Start by creating a representational medium such that a large
number of primitive elements a constrained to obey a handful of simple behavioral principles.
Next constructs models from this highly productive medium.14 What emerges are generative
structures capable of supporting an open-ended number of mechanical inferences. At the level of
14 ‘Productive’ is here used in Fodor’s sense – that is, to denote a medium capable of representing an open-ended
number of distinct states of affairs [64].
the medium, ‘running’ such a model involves the recursive application of the basic constraints on
the modeling-element behaviors. This will typically be a massive, parallel, constraint-satisfaction
process. Given that this form of processing is the forte of neural networks, there should be little
doubt that neural machinations are up to the task.15
Bottom-up Approaches
Thus far, we have largely approached the question of the neural realizability of mental
models in the abstract, and from the top down. This is partly because there has been a relative
dearth of work that moves in the opposite direction, from the bottom up. One exception is
Thagard’s [12] recent work on the topic, which appeals to such biologically plausible simulations
of neural networks as those of Eliasmith and Anderson [65]. Unfortunately, Thagard has yet to
offer evidence that the neural encoding strategies he discusses exhibit any of the central features,
discussed here, that distinguish modeling from syntax-crunching. Most notably, the neural
representations he cites have not yet been shown to exhibit a significant degree of spatial,
kinematic, of causal generativity. The proof of the pudding here is in the eating.
To the extent that there have been significant advances in the bottom-up endeavor, they
mostly issue from research – such as that of Nobel laureates John O’Keefe, May-Britt Moser,
and Edward Moser – on the biological neural networks that underwrite spatial reasoning abilities
in rats. As you will recall, Tolman’s pioneering work on maze navigation suggested that rats
have an onboard medium for the construction of generative spatial maps of their location relative
to barriers and important items such as food and drink. O’Keefe is famous for showing that the
rat’s hippocampus contains ‘place’ cells which fire preferentially when an animal reaches a
15 At the same time, one should not overestimate inherent immunity to the frame problem of neural networks [58]. It
is only by implementing a primitively constrained modeling medium that neural networks can be expected to realize
intrinsic representations of complex, interacting worldly constraints.
particular location in its environment, cells that fire in sequence as a rat moves from one location
to another [66]. Moser & Moser subsequently showed that the rat’s uncanny spatial navigation
abilities also depend upon ‘grid’ cells in the nearby entorhinal cortex [67]. Individual grid cells
fire when an animal is in one any of several, roughly evenly spaced locations. When lines are
drawn to connect these points, they collectively form what (purely by coincidence) looks a great
deal like the kind of 2D polymesh shown in Figure 5. While each grid cell is tuned to a collection
of locations, different grid cells have sparser or denser coverage of the same region of space.
Collectively they provide effective coverage of the entire region of space in which the animal
finds itself.
Importantly, O’Keefe et al. note regarding place cells that “there does not appear to be any
obvious topographical relation between the field locations [i.e., the places to which cells become
temporarily tuned] and the anatomical locations of the cells relative to each other within the
hippocampus” [68]. Nor do grid cells in the entorhinal cortex exploit any obvious structural
isomorphisms between their respective anatomical locations and the spatial layout of the
environment. However, acting in concert, the two types of cells enable effective navigation, as if
the organism had an internal map that preserves relative locations (place cells) and distances
(grid cells). In other words, the two systems encode maps that are functionally isomorphic with
real maps of the environment. Moreover, they provide a productive modeling medium, one
which, not unlike a collection of Lego blocks, can be used and reused, through a process called
‘re-mapping,’ to encode information about an open-ended number of new environments [69].
The maps constructed in this medium are generative with regard to 2D spatial properties in the
aforementioned sense, as is shown by their role in enabling rats to find efficient new ways to a
destination when familiar routes are blocked. More recent research suggests that the rat’s place
cells are also somewhat sensitive to vertical displacement from a reference plane, perhaps
enabling 3D mapping capabilities [70]. Nor are the lessons learned here applicable only to rats,
for a large body of research suggests that the same anatomical systems may be implicated in
human spatial navigation [71].
Our deepest understanding of how real neural networks create spatial mental models thus
suggests that brains implement a reusable modeling medium and, by exploiting the kinds
functional, rather than physical, isomorphisms that make neural realizability feasible, nothing is
lost in the way of generativity. It also bears mentioning that this modeling medium is well suited
for producing models of the organism’s location relative to its specific, concrete environment. As
such, it may (taken in isolation) be ill suited for representing abstracta or genera. As for the
singling out of specific properties of specific objects, it may be that models that are realized by
neurophysiological processes have a natural advantage over scale models in that populations
representing specific properties may ‘cry out’ for attention (e.g., by oscillating at the appropriate
frequency). There is, moreover, no reason why these lessons could not scale up, so to speak, to
account for the human ability to run off-line models of spatial, kinematic, and dynamic
relationships. Of course, in humans, the neocortex is likely to play a much more prominent role.
As of yet, however, there is little understanding of the precise manner in which the neocortex
does, or might, realize mental models.
1.5 Realization Story Applied
Though we clearly have a long way to go, the above hypothesis about what mental models
are such that neural systems might realize them looks to have important ramifications for work in
several fields, ranging from AI to the philosophy of science.
AI and Psychology: Towards and Intuitive Physics Engine
One obvious ramification of the above is what it suggests about how one might go about
endowing computational artifacts with the kind of boundless commonsense knowledge of the
consequences of alterations to world that humans seem to possess. FEMs prove that there is a
determinate computational solution to the prediction and qualification problems. FEMs are
generative in that they can be manipulated in any of countless ways in order to make inferences
about how alterations to the environment might play out and, by the same token, about the ways
in which those consequences might be defeated. It would thus behoove AI researchers to
incorporate media for the construction of intrinsic models within the core inference machinery of
their devices. Indeed, there has been some movement in this direction in recent years. For
instance, though past manuals for the Soar production-system architecture evidence a certain
degree of exasperation when it comes to the frame problem, more recent manuals indicate that
Soar’s designers have begun to offload mechanical reasoning to non-sentential models. Laird
notes, for instance:
With the addition of visual imagery, we have demonstrated that it is possible to solve
spatial reasoning problems orders of magnitude faster than without it, and using
significantly less procedural knowledge. Visual imagery also enables processing that is not
possible with only symbolic reasoning, such as determining which letters in the
alphabet are symmetric along the vertical axis (A, H, I, M, O, T, U, V, W, X, Y) [72].
While Soar’s imagery module still only supports simple spatial reasoning, it is clearly a step in
the direction of richer, intrinsic models of 3D kinematics and dynamics.
There has also been some movement in the direction of using computationally realized
intrinsic models as a way of making sense of behavioral findings regarding how humans engage
in commonsense reasoning about the world. For instance, after paying homage to Craik, MIT
researchers Battaglia, Hamrick, & Tenenbaum describe their innovative approach to
commonsense reasoning as follows:
We posit that human judgments are driven by an “ intuitive physics engine ” (IPE), akin to
the computer physics engines used for quantitative but approximate simulation of rigid
body dynamics and collisions, soft body and fluid dynamics in computer graphics, and
interactive video games [73].
They simulate the IPE with FEMs of full-blown 3D kinematic and dynamic relationships. They
note that a similar IPE in humans might allow us to read off from our simulations the answers to
questions of ‘What will happen?’ regarding to innumerable novel scenarios. Their pioneering
work also breaks new ground in that it begins to account for probabilistic reasoning by building a
bit of uncertainty into models and treating multiple runs of a model as a statistical sample.
All of this work is very much in the spirit of Schwartz’ claim that “inferences can emerge
through imagined actions even though people may not know the answer explicitly” [74, italics
mine]. It also fits with the following suggestion of Moulton & Kosslyn:
the primary function of mental imagery is to allow us to generate specific predictions based
upon past experience. Imagery allows us to answer ‘what if’ questions by making explicit
and accessible the likely consequences of being in a specific situation or performing a
specific action [35, italics mine].
Exduction
Another important lesson to be learned from computationally realized intrinsic models is
that they support a form of mechanistic reasoning that has found its way into few, if any,
standard reasoning taxonomies. As Glasgow & Papadias claim: “The spatial structure of images
has properties not possessed by deductive sentential representations ... spatial image
representations ... support nondeductive inference using built-in constraints on the processes that
construct and access them” [60]. Of course, there is more to be said about the process of model-
based mechanistic reasoning than that it is not deductive. In fact, the process shares with (valid)
deductive reasoning the property of being monotonic. What makes deduction a monotonic (i.e.,
indefeasible) reasoning process is that the conclusion of a valid argument cannot be overturned
simply by adding premises; it can only be overturned by rejecting one or more of the premises
from which the conclusion was deduced. Other forms of reasoning (e.g., inductive
generalization, analogical reasoning, abduction) are defeasible in that one can overturn their
conclusions simply by adding relevant premises. For instance, if I hear a meowing noise
emanating from my daughter’s closet door, I may infer that the cat is trapped inside. But if I then
see the cat walking through the kitchen and am told that my daughter was given a new electronic
cat toy, my conclusion would be undermined while at the same time leaving the original premise
(i.e., that there is meowing coming from the closet) intact.
One diagnosis for why deduction is monotonic is that, in a certain sense, the premises of a
valid deduction already ‘contain’ the information stated in the conclusion, so adding information
takes nothing away from the support that those premises lend to the conclusion. That means that
insofar as the original premises are true, the conclusion must be as well, and insofar as the
conclusion is false, there must be something wrong with the premises used to derive it. But
deduction is formal, in that topic-neutral logical particles are what bear the entirety of the
inferential load – that is, the specific contents (consistently) connected and quantified over drop
out as irrelevant.
The use of scale models and FEMs makes evident that there is another form of monotonic
reasoning in addition to deduction. As explained above, information derived regarding the
consequences of interventions on a modeled system are to a significant extent already
‘contained’ (i.e., they are implicit) in the models themselves. The only way to overturn a model-
based inference is to call into question some aspect or other of the model from which it was
derived. By the same token, if the conclusion is incorrect, there must be something wrong with
the model. But unlike deduction, model-based reasoning is not effected by abstracting away from
specific contents and allowing logical particles to bear the inferential load. Instead, it is the
specific, concrete contents of the models that do all of the work. As yet, this form of monotonic
reasoning lacks a name. Let us thus call it exduction (ex- out + duce- lead). Like deduction,
exduction may be implemented externally through the use of representational artifacts, but the
hypothesis being explored here is just that we also sometimes engage in internal exductive
reasoning through the use of mental models. If this hypothesis is correct, then exduction must be
added to our standard taxonomy of internal reasoning processes and placed alongside deduction
under the broader heading of monotonic reasoning.
1.6 Mechanistic Explanation Revisited
It was noted at the end of Section 1 that mental models may well play a crucial role in the
process of mechanistic explanation and prediction. If that is so, then we can only hope to attain a
deep understanding of science if we first account for how the mental modeling process works.
Now that we have a clearer conception of the distinctive features of mental models and of the
way in which they might be realized by neurophysiological processes, we can begin to see what
the payoff might be in terms of our understanding of model-based reasoning about mechanisms
in science.
The Prediction and Ceteris Paribus Problems
To give some of the flavor of where this might lead, consider that one benefit of the
foregoing realization story regarding mental models for the philosophy of science is that it offers
a solution to two longstanding problems: the surplus-meaning problem and the ceteris paribus
problem (a.k.a., the problem of provisos). Both problems arose as a consequence of attempts to
apply the methods of formal, mostly deductive methods in an attempt to provide a logical
reconstruction of scientific reasoning.
The surplus-meaning problem has to do with the fact that explanatory hypotheses have, and
are known to have, countless implications beyond the happenings they explain. To keep things
simple, consider the non-scientific case of what a mechanic knows about the operation of an
automobile engine. Imagine, in particular, that someone has brought an automobile into the
mechanic’s shop complaining that the engine has suffered a drop-off in power. Listening to the
engine, the mechanic might decide that the engine has blown a ring. Setting aside the question of
creativity, one might provide the following formal, deductive reconstruction of his explanatory
model:
If one of the cylinders has lost a ring, then the engine will have lost power.
One of the cylinders has lost a ring.
The engine has lost power.
Consider, however, that the mechanic knows not only that the faulty ring will result in a loss of
power (the explananandum occurrence); he also knows many other implications of this
explanatory hypothesis, such as that oil will leak into the combustion chamber, the exhaust will
look smoky, the end of the tailpipe will become oily, and the sparkplugs will turn dark. He also
knows what interventions will and will not alleviate the problem – for instance, he knows that
replacing the rings will restore power but replacing the air filter will not, and so on indefinitely.
Any suitable deductive reconstruction of the mechanic’s mental model of the source of the
problem must thus account not just for the fact that it implies the explanandum; it must account
for its countless further implications. The general problem with deductive reconstructions of
explanations – a problem that will surely beset any attempt to reconstruct mechanistic
explanations using extrinsic representations – is that they fail to capture the full complexity of
what anyone who possesses that explanation must know, if only implicitly. The problem is that
there is too much surplus meaning to express it all explicitly [75], and these added implications
are essential to how we assess the adequacy of explanations, whether in everyday life of in
science. As Greenwood explains:
Where this “surplus meaning” comes from…is a matter of some dispute, but that genuine
theories poses [sic.] such surplus meaning is not—for this is precisely what accounts for
their explanatory power and creative predictive potential [76].
Notice also that the mechanic’s model of why the car lost power not only contains
information about the various other things he should expect to find if that model is correct; it also
contains information about the countless ways in which each of these expectations might,
consistent with the truth of the explanation, be defeated. The mechanic knows, for instance, that
replacing the rings will restore power, but only if it is not the case that one of the spark plug
wires was damaged in the process, the air filter has become clogged with dust from a nearby
construction project, and so on indefinitely.
Whether we are dealing with commonsense or scientific reasoning about mechanisms, the
problem with attempts at formalizing our knowledge of the ways in which a given implication is
qualified is that what we know far outstrips what can be expressed explicitly in the form of, say,
a conditional generalization. In philosophy of science, the qualifications that would have to be
added are termed ceteris paribus clauses and provisos. As Fodor claims, the general problem is
that “as soon as you try to make these generalizations explicit, you see that they have to be
hedged about with ceteris paribus clauses” [77]. And as Giere claims, “the number of provisos
implicit in any law is indefinitely large” [78, italics mine]. Deductive models, and others that rely
upon extrinsic representations, are thus unable to capture the full breadth of what we know when
we possess a mechanistic explanation for an occurrence. And all of this dark information plays a
crucial role in the testing and retention of mechanistic hypotheses [79]. Accounting for it must
thus be viewed as a central goal for the philosophy of science.
If we abandon deductive (or other extrinsic) schemes for making sense of scientific
reasoning about mechanisms and instead adopt an exductive (and intrinsic) model-based account
of reasoning, the surplus-meaning and ceteris paribus problems dissolve, and the source of the
dark information comes into focus. After all, these two problems are just variants on the
prediction and qualifications problems of AI. This is not surprising given that both sets of
problems were discovered through early attempts to deductively reconstruct everyday and
scientific reasoning about mechanisms. Naturally, the same solution applies in both cases:
Eschew the appeal to extrinsic representations and formal inferences in favor of an appeal to
intrinisic models and exductive inferences. The promising idea that emerges is that scientists
may be utilizing intrinsic mental models to understand the mechanisms that are (or might be)
responsible for particular phenomena. Such models would endow the scientist with boundless
tacit knowledge of the further implications of a given mechanistic hypothesis and of the
countless ways in which those implications are qualified.
Beyond Mental Models
We must not forget, however, that our mental models are limited by working memory
capacity and by the heavy cognitive load associated with mental modeling. Even so, scientists
are somehow able to formulate and comprehend some remarkably complex mechanical
explanations. Seen in this light, it is no surprise that, in their reasoning about mechanisms,
humans rely heavily upon external representational artifacts such as diagrams. These can act as
aids to memory, both short- and long-term, enabling us to off-load some of the cognitive burden
to the world and thereby compensating for our otherwise limited ability to keep track of the
simultaneous influences of many mechanical components (see [80]). Indeed, when aided by
external diagrams, Hegarty, Kriz, & Cate found that high- and low-imagery subjects performed
about equally well on a task that required model-based reasoning about mechanisms [19]. The
compensatory influence of external representations is strong indeed (also see Bechtel, this
volume).
Of course we do not just utilize static pictures to make sense of natural phenomena; we also
sometimes use scale models [81]. On the present view, the reason so many have come to view
these models as an apt metaphor for in-the-head reasoning may be that scale models recapitulate,
albeit in a way that overcomes many of our cognitive frailties, the structure of our internal
models of mechanisms. However, with the advent of sophisticated computer models, we now
have an even better tool for investigating the implications of mechanical hypotheses. As we have
seen, certain computer models (e.g., FEMs) are like scale models in that they constitute intrinsic
non-sentential representations of actual or hypothetical mechanisms. However, these models
have the added virtue that one can easily freeze the action, zoom in or out, slow things down, and
even watch things play out in reverse. These models thus constitute what Churchland &
Sejnowski term a “fortunate preparation” [82]. Such models provide an even more apt analogy
for understanding our own native mental models, for both sorts of models are realized at a low
level by complicated circuitry, and both tend to bottom-out well above the level of nature’s
fundamental laws.16
As noted in Section 3.2, over and above the quantitative limits imposed by working
memory capacity, scale models and FEMs are, and our own mental models may well be, limited
in certain qualitative respects, such as their ability to represent abstracta. But surely thoughts
about abstracta play a big role in scientific reasoning. One way of accounting for this is to say
that our deficiencies in regards to modeling abstracta are the precise problem that analogy and
metaphor were created to solve. This would make sense of why the language we use to represent
abstracta (e.g., ‘economic inflation’) is so shot through with analogies and metaphors rooted in
concrete domains [84-6].
[Insert Figure 6 About Here.]
Figure 6. Various configurations of circuitry, batteries, and resistors.
Gentner and Gentner’s study of human reasoning about electricity lends some credence to
this view [87]. Gentner and Gentner found that in conversation, non-expert subjects commonly
likened the flow of electricity to water moving through pipes or crowds moving through
corridors. Each analogy happens to yield its own unique set of inaccurate predictions about how
electrical current moves through particular configurations of electrical components (Figure 6).
Gentner and Gentner found that subjects’ errors in reasoning about these components tracked the
analogies they invoked when discussing electricity. This suggests that these analogies run deeper
16 One way of describing the interplay between external models (in their various forms) and internal mental models
would be to say that the latter are part and parcel of scientific cognition, whereas the former are representational
artifacts created to aid cognition. An alternative, somewhat speculative, proposal is that our external models are no
less part of fabric of scientific cognition than are our internal mental models [83].
than the surface features of language and penetrate right into subjects’ mental models of the flow
of electricity. Analogies and metaphors are perhaps not the whole story of how we think about
abstracta, but they may well be an important part of it.
1.7 Conclusion
It was noted at the end of Section 1 that mental models might play a crucial role in the
process of mechanistic explanation and prediction in science. As such, we can only attain a deep
understanding of science itself if we first understand that nature of this mental, model-based,
reasoning process. We then saw that experimental psychologists have long maintained that
mental models are distinct from sentential representations in much the way that scale models are.
Insofar as this hypothesis is viable, we can expect that experimental psychology will provide
crucial insight into both the nature and limits of our onboard mental models. At the same time, it
is important to recognize that the many appeals to distinctively model-like mental representations
in psychology will be considered suspect so long as we lack a reasonable way of spelling out
what sorts of representational structures mental models are supposed to be in a way that (i)
shows these models to be distinct from sentential representations while (ii) allowing for their
realization by neurophysiological processes. We can see the way forward, however, if we first
pay attention to some of the distinctive features of scale models that distinguish them from
sentential representations. If we then turn to the computational realm, we see that these very
features (including immunity to the notorious frame problem) are exhibited by certain
computational models of mechanisms such as FEMs. An appeal to the principle that sustains the
computational theory of cognition (i.e., POPI) enables us to understand how this could be so and
how high-level, non-sentential, intrinsic models of mechanisms could in principle be realized by
neurophysiological processes. The broader viability of this realization story for mental models is
suggested by recent work in both AI and experimental psychology and by the elegant solution it
offers to the surplus-meaning and ceteris paribus problems in the philosophy of science. Going
forward, the idea that our scientific reasoning about mechanisms might, to a large extent, involve
the manipulation of representations that are like scale models in crucial respects can be regarded
as at least one, sturdy pillar of a promising hypothesis regarding the nature of model-based
reasoning in science.
References
1. Railton, P.: A deductive-nomological model of probabilistic explanation. Philosophy of
Science 45 (1978) 206-226
2. Giere, R.: Explaining Science: A Cognitive Approach. University of Chicago Press,
Chicago (1988a)
3. Salmon, W.: Scientific Explanation and the Causal Structure of the World. Princeton
University Press, Princeton, NJ (1984)
4. Salmon, W.: Causality and Explanation. Oxford University Press, New York (1998)
5. Machamer, P., Darden, L., Craver, C.: Thinking about mechanisms. Philosophy of
Science 67 (2000) 1-25
6. Bechtel, W. Mental Mechanisms: Philosophical Perspectives on Cognitive Neuroscience.
Psychology Press, New York (2009)
7. Wright, C., Bechtel, W.: Mechanisms and psychological explanation. In Thagard, P., ed.:
Philosophy of Psychology and Cognitive Science. Elsevier, New York (2007) 31-79
8. Waskan, J., Harmon, I., Higgins, A., Spino, J.: Investigating lay and scientific norms for
using ‘explanation.’ In Lissack, M. Graber, A., eds.: Modes of Explanation: Affordances
for Action and Prediction. Palgrave Macmillan (2014) 198-205
9. Craik, K.: The Nature of Explanation. Cambridge University Press, Cambridge, UK
(1943)
10. Tolman, E.: Cognitive maps in rats and men. The Psychological Review 55 (1948) 189-
208
11. Nersessian, N.: Mental modeling in conceptual change. In Vosniadou, S., ed.:
Internatioinal Handbook of Conceptual Change. Routledge, London (2007) 391-416
12. Thagard, P.: The Cognitive Science of Science. The MIT Press, Cambridge (2012)
13. Shepard, R., Metzler, J.: Mental rotation of three-dimensional objects. Science 171
(1971) 701-703
14. Kosslyn, S. Image and Mind. Harvard University Press, Cambridge, MA (1980)
15. Schwartz, D., Black, J.: Analog imagery in mental model reasoning: Depictive models.
Cognitive Psychology 30 (1996a) 154–219
16. Schwartz, D., Black, J.: Shuttling between depictive models and abstract rules: Induction
and fall-back. Cognitive Science 20 (1996b) 457-497
17. Hegarty, M.: Mechanical reasoning by mental simulation. Trends in Cognitive Sciences 8
(2004) 280-285
18. Hegarty, M., Kriz, S. Cate, C.: The roles of mental animations and external animations in
understanding mechanical systems. Cognition and Instruction 21 (2003) 325–360
19. Norman, D.: Some observations on mental models. In Gentner, D., Stevens, A., eds.:
Mental Models. Lawrence Erlbaum Associates, Hillsdale, NJ (1983) 7-14
20. Leslie, A., Keeble, S.: Do six-month old infants perceive causality? Cognition 25 (1987)
265–88
21. Schlottman, A.: Seeing it happen and knowing how it works: How children understand
the relation between perceptual causality and underlying mechanism. Developmental
Psychology 35 (1999) 303-317
22. Baillargeon, R., Li, J., Gertner, Y., Wu, D.: How do infants reason about physical events?
In Goswami, U., ed.: The Wiley-Blackwell handbook of childhood cognitive development,
second edition. Blackwell, Oxford, UK (2011): 11-48
23. Langacker, R.: Concept, Image, and Symbol: The Cognitive Basis of Grammar. Mouton
de Gruyter, New York (1991)
24. Goldberg, A.: Constructions: A Construction Grammar Approach to Argument Structure.
The University of Chicago Press, Chicago (1995)
25. Fauconnier, G.: Mental Spaces. The MIT Press, Cambridge, MA (1985)
26. Garnham, A. Mental Models as Representations of Discourse and Text. John Wiley, New
York (1987)
27. Talmy, L.: Force dynamics in language and cognition. Cognitive Science 12 (1988) 49-
100
28. Johnson-Laird, P.: How is meaning mentally represented? In Eco, E.,Santambrogio, M,
Violi, P., eds.: Meaning and Mental Representation. Indiana University Press,
Bloomington, IN (1988) 99-118
29. Barsalou, L., Solomon, K., Wu, L.: Perceptual Simulation in Conceptual Tasks. In M.
Hiraga, C. Sinha, & S. Wilcox, eds.: Cultural, Typological, and Psychological
Perspectives in Cognitive Linguistics: The Proceedings of the 4th Conference of the
International Cognitive Linguistics Association, Vol. 2. John Benjamins, Amsterdam
(1999): 209-228
30. Brewer, W.: Scientific theories and naive theories as forms of mental representation:
psychologism revived. Science & Education 8 (1999) 489-505
31. Johnson-Laird, P.: Mental Models: Towards a Cognitive Science of Language, Inference,
and Consciousness. Harvard University Press, Cambridge, MA (1983)
32. Johnson-Laird, P., Byrne, R.: Deduction. Lawrence Erlbaum Associates, Hillsdale, New
Jersey (1991)
33. McCulloch, W., Pitts, W.: A logical calculus of the ideas immanent in nervous activity
Bulletin of Mathematical Biophysics 5 (1943)115-13
34. Franklin, S., Garzon, M.: Computation by discrete neural nets. In Smolensky, P., Mozer,
M., Rumelhart, D., eds.: Mathematical Perspectives on Neural Networks. Lawrence
Earlbaum Associates, Mahwah, New Jersey (1996) 41-84
35. Moulton, S., Kosslyn, S.: Imagining predictions: Mental imagery as mental emulation.
Philosophical Transactions of the Royal Society B 364 (2009) 1273–1280
36. Cummins, R.: Representations, Targets, and Attitudes. The MIT Press, Cambridge, MA
(1996)
37. Wittgenstein, L.: Philosophical Investigations. Macmillan, New York (1953)
38. Shepard, R., Chipman, S.: Second-order isomorphism of internal representations: Shapes
of states. Cognitive Psychology 1 (1970) 1-17
39. Palmer, S.: Fundamental aspects of cognitive representation. In Rosch, E., Lloyd, B.,
eds.: Cognition and Categorization. Lawrence Erlbaum Associates, Hillsdale, New
Jersey (1978) 259–303
40. Nersessian, N.: The cognitive basis of model-based reasoning in science. In Carruthers,
P, Stich, S., Siegal, M., eds.: The Cognitive Basis of Science. Cambridge University
Press, Cambridge, UK (2002) 133-153
41. Kosslyn, S.: Image and Brain: The Resolution of the Imagery Debate. The MIT Press,
Cambridge, Massachusetts (1994)
42. Zeki, S. The functional organization of projections from striate to prestriate visual cortex
in the rhesus monkey. Cold Spring Harbor Symposium on Quantitative Biology 40 (1976)
591-600
43. Mishkin, M., Ungerleider, L., & Macko, K. Object vision and spatial vision: Two cortical
pathways. Trends in Neurosciences 6 (1983) 414-417
44. DeYoe, E., Van Essen, D.: Concurrent processing streams in monkey visual cortex.
Trends in Neurosciences 11 (1988) 219-226
45. Huttenlocher, J., Higgins, E., Clark, H.: Adjectives, comparatives, and syllogisms.
Psychological Review 78 (1971) 487-514
46. Anderson, J.R.: Arguments concerning representations for mental imagery. Psychological
Review 85 (1978) 249-277
47. Brooks, L.: Spatial and verbal components in the act of recall. Canadian Journal of
Psychology 22 (1968) 349-368
48. Segal, S., Fusella, V.: Influence of imaged pictures and sounds on detection of visual and
auditory signals. Journal of Experimental Psychology 83 (1970) 458-464
49. Block, N.: Mental pictures and cognitive science. In Lycan, W.G., ed.: Mind and
Cognition. Basil Blackwell, Cambridge, Massachusetts (1990) 577-606
50. Fodor, J.A.: Imagistic Representation. In N. Block, ed.: Imagery. The MIT Press,
Cambridge, MA (1981), 63-86
51. McCarthy, J., Hayes, P.: Some philosophical problems from the standpoint of artificial
intelligence. In Meltzer, B., Michie, D., eds.: Machine Intelligence. Edinburgh University
Press, Edinburgh, UK (1969) 463-502
52. Hayes, P.: The frame problem and related problems in artificial intelligence. In Webber,
B., Nilsson, N., eds.: Readings in Artificial Intelligence. Morgan Kaufman, Los Altos,
CA (1981), 223-230
53. Janlert, L.: The frame problem: Freedom or stability? With pictures we can have both. In
Ford, K.M., Pylyshyn, Z., eds.: The Robot's Dilemma Revisited: The Frame Problem in
Artificial Intelligence. Ablex Publishing, Norwood, New Jersey (1996) 35-48
54. McCarthy, J. Applications of circumscription to formalizing common-sense knowledge.
Artificial Intelligence 28 (1986) 86-116
55. Waskan, J.: Applications of an implementation story for non-sentential models. In
Magnani, L, Carnielli, W., Pizzi, C.: Model-Based Reasoning in Science and Technology.
Springer, Berlin (2010) 463-476
56. Congdon, C., Laird, J.: The Soar User's Manual: Version 7.0.4. University of Michigan,
Ann Arbor (1997)
57. Waskan, J.: Intrinsic cognitive models. Cognitive Science 27 (2003) 259-283
58. Waskan, J.: Models and Cognition. The MIT Press, Cambridge, MA (2006)
59. Haugeland, J. An overview of the frame problem. In Pylyshyn, Z., ed.: Robot's Dilemma.
Ablex Publishing Corp, Norwood, NJ (1987) 77-93
60. Glasgow and Papadias (1992).
61. Pylyshyn, Z.: Computation and Cognition: Toward a Foundation for Cognitive Science.
The MIT Press, Cambridge, Massachusetts (1984)
62. Sterelny, K.: The imagery debate. In Lycan, W., ed.: Mind and Cognition. Blackwell,
Cambridge, MA (1990) 607-626
63. Fodor, J.A.: The Mind Doesn't Work That Way. The MIT Press, Cambridge, MA (2000)
64. Fodor, J.A.: The Language of Thought. Thomas Y. Crowell, New York (1975)
65. Eliasmith, C., Anderson, C.H.: Neural Engineering: Computation, Representation and
Dynamics in Neurobiological Systems. The MIT Press, Cambridge, MA (2003)
66. O'Keefe, J., Nadel, L.: The hippocampus as a cognitive map. Oxford Univeristy Press,
New York (1978)
67. Brun, V., Otnass, M., Molden, S., Steffenach, H., Witter, M., Moser, M., Moser, E.: Place
cells and place recognition maintained by direct entorhinal-hippocampal circuitry.
Science 296 (2002) 2243-2246
68. O’Keefe, J., Burgess, N., Donnett, J., Jeffery, K., Maguire, E.: Place cells, navigational
accuracy, and the human hippocampus. Philosophical Transactions of the Royal Society
B 353 (1998) 1333-1340
69. Fyhn, M., Hafting, T., Treves, A., Moser, M., Moser, E.: Hippocampal remapping and
grid realignment in entorhinal cortex. Nature 446 (2007) 190-194
70. Savelli, F. Knierim, J.: Coming up: In search of the vertical dimension in the brain.
Nature Neuroscience 14 (2011) 1102-1103
71. Woollett, K., Maguire, E.: Acquiring "the knowledge" of London's layout drives
structural brain changes. Current Biology 21 (2011) 2109-2114
72. Laird, J.: Extending the Soar cognitive architecture. In Wang, P., Goertzel, B., Franklin,
S., eds., Proceedings of the 2008 conference on Artificial General Intelligence:
Proceedings of the First AGI Conference. IOS Press, Amsterdam (2008) 224-235
73. Battaglia, P., Hamrick, J., Tenenbaum, J.: Simulation as an engine of physical scene
understanding. Proceedings of the National Academy of Science, 110 (2013), 18327-
18332
74. Schwartz, D.: Physical imagery: Kinematic versus dynamic models. Cognitive
Psychology 38 (1999) 433-464
75. Reichenbach, H.: Experience and Prediction: An Analysis of the Foundations and the
Structure of Knowledge. University of Chicago Press, Chicago (1938)
76. Greenwood, J. Folk psychology and scientific psychology. In Greenwood, J., ed.: The
Future of Folk Psychology. Cambridge University Press, Cambridge, UK (1991) 1-21
77. Fodor, J.A.: Psychosemantics: The Problem of Meaning in the Philosophy of Mind. The
MIT Press, Cambridge, MA (1984)
78. Giere, R.: Laws, theories, and generalizations. In Grünbaum, A., Salmon, W., eds.: The
Limits of Deductivism. University of California Press, Berkeley, CA (1988b), 37-46
79. Waskan, J.: Knowledge of counterfactual interventions through cognitive models of
mechanisms. International Studies in Philosophy of Science 22 (2008) 259-275
80. Clark, A.: Mindware: An Introduction to the Philosophy of Cognitive Science. Oxford
University Press, New York (2013)
81. Weisberg, M.: Who is a modeler? British Journal for Philosophy of Science 58 (2007)
207–233
82. Churchland, P., Sejnowski, T.: Neural representation and neural computation.
Philosophical Perspectives: Action Theory and Philosophy of Mind 4 (1988) 343-382
83. Giere, R.: Models as parts of distributed cognitive systems. In Magnani, L, Nersessian,
N., eds.: Model Based Reasoning: Science, Technology, Values. Kluwer Academic
Publishers, Amsterdam: (2002) 227-41
84. Lakoff, G., Johnson, M.: Metaphors We Live By. University of Chicago Press, Chicago,
IL (1980)
85. Lakoff, G.: Women, Fire, and Dangerous Things. University of Chicago Press, Chicago,
IL (1987)
86. Fernandez-Duque, D., Johnson, M.: Attention metaphors: how metaphors guide the
cognitive psychology of attention. Cognitive Science 23 (1999) 83-116
87. Gentner, D., Gentner, D.R.: Flowing waters or teeming crowds: Mental models of
electricity. In Gentner, D., Stevens, A., eds.: Mental Models. Lawrence Erlbaum
Associates, Hillsdale, NJ (1983) 99-129
Figures: