Naturalism and semantics - Carleton
Transcript of Naturalism and semantics - Carleton
Naturalism and semantics∗
Luke Jerzykiewicz ([email protected])
draft: October 2004
Abstract
Three attempts to furnish a scientific semantic theory are reviewed:
the natural semantic metalanguage (NSM); formal semantics (FS);
and conceptualist semantics (CS). It is argued that each can be made
compatible with naturalism, although the costs in each case are high.
NSM urges practitioners to recognize ‘meaning’ as a basic feature of
reality. FS requires that we countenance unlikely abstract entities as
basic explanatory posits. SC, while consonant with cognitive science’s
basic ontology, appears committed to the fallacy of psychologism. All
three approaches are rejected here as fully satisfactory approaches to
the study of meaning.
∗Carleton University Cognitive Science Technical Report 2004-09 (www.carleton.ca/iis/TechReports).
1
Contents
Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
Mise-en-scene . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
Natural Semantic Metalanguage . . . . . . . . . . . . . . . . . . . . 4
Characterization . . . . . . . . . . . . . . . . . . . . . . . . . 4
Discussion and Criticism . . . . . . . . . . . . . . . . . . . . . 9
Formal Semantics . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
Characterization . . . . . . . . . . . . . . . . . . . . . . . . . 12
Discussion and Criticism . . . . . . . . . . . . . . . . . . . . . 23
Conceptualist Semantics . . . . . . . . . . . . . . . . . . . . . . . . 26
Characterization . . . . . . . . . . . . . . . . . . . . . . . . . 26
Discussion and Criticism . . . . . . . . . . . . . . . . . . . . . 35
Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
Introduction
Semantics is the study of ‘meaning,’ whatever that turns out to be. The
nature of meaning is of interest to cognitive scientists because mental states,
as well as linguistic utterances, are paradigm cases of the meaningful. To
date, the issue of precisely how semantics ought to be practised has not
satisfactorily been resolved. Part of the difficulty involves competing visions
of how (and whether) semantics is to be integrated into the broader scientific
enterprise. In what follows, I review three current approaches to the study of
meaning. My aim is partly descriptive and partly critical. I’d like to clarify
(for my own peace of mind, as much as anything else) the extent to which
each constitutes a naturalist theory of meaning. The three projects discussed
are these:
• the natural semantic metalanguage (NSM);
• formal semantics (FS); and
• conceptualist semantics (CS).
Let me say at the outset that the conclusion of the paper is largely negative:
In my view, none of these projects (in their current form, at least) are ade-
quate to the task of characterizing meaning in a manner at once true to the
phenomena under scrutiny and compatible with naturalism. A way forward
that I find promising is briefly sketched toward the end of the paper.
Mise-en-scene
Outside of logic and mathematics, fully satisfactory definitions are hard to
come by. Still, we need a first-pass characterization of some key notions
before moving on to the particulars of the three semantic projects at issue.
2
Let me start with naturalism. I take naturalistic theories to be those
which do not posit basic entities, properties or events beyond those given
us by our best science. Furthermore, within science itself, broad ontological
continuity is recognized by the naturalist as a regulative ideal. The vari-
ous branches of science, whatever their domain of enquiry, aim ultimately to
arrive at a single, continuous understanding of the workings of that which
exists. In keeping with this ideal, new entities, properties or events are in-
troduced only sparingly. Obviously, one wants to recognize that non-basic
‘things’ exist: there are hydronium ions, ecosystems, GDPs, maximal pro-
jections, but also haircuts, kitsch, and Tuesdays. The non-basic posits are
understood by the naturalist to exist in virtue of being realized ultimately
by more basic entities and processes. Often, among a researchers’s goals is
to show precisely how higher level regularities—those regarding, for exam-
ple, locative alternation or strong-acid titration—fall out as a consequence
of lower level facts about lexical and semantic representations or hydrogen
bonding. Science often comes short of that goal. The discovery and de-
scription tight causal links does not come easily. Still, the puzzle is more
complete than it was a hundred years ago. And a hundred years ago, it was
more complete than the century before that.
Naturalism, as I have characterized it, comes in three flavours: there is
an eliminativist, a reductive and an emergentist variety. The eliminativist
proposes that all explanations, if they are to count as really explanatory,
must in the long run be couched in the terms of the most basic physics (this
is Dennett’s (1995) ‘greedy reductionism’). Hardly anyone favours this sort
of line these days. And with good reason: we all frequently rely on explana-
tions of all sorts that do not advert to photons or quarks. In contrast, the
emergentist holds that there are entities, processes or events which ought to
figure in our most basic descriptions of the world, yet which are not them-
selves posits of fundamental physics. This view finds many supporters. David
3
Chalmers (1996), has for instance argued that consciousness is probably just
such a basic phenomenon. Reductionists chart a middle course: they require
that all phenomena which are not themselves posits of fundamental physics
need to be reducible to some more basic phenomena or other. They do not
however require that all legitimate explanations be stated in physics’ terms.
With regard to semantics, the issue plays itself out as follows: Elimina-
tivists doubt that there is such a phenomenon as meaning to explain (here
we can recall the behaviourists). They tend therefore not to develop theo-
ries of meaning, and are not discussed here. Emergentists propose that we
should take meaning—or perhaps the terms in which meaning is to be expli-
cated—to be ontologically basic. I will argue that, for better or for worse,
both NSM and (perhaps surprisingly) FS come under this rubric. Finally,
the reductivist holds that both meaning and the posits in terms of which
meaning is cashed out are themselves explicable in more basic terms (and so
on). CS, the final semantic theory considered here, is a reductive naturalist
theory. Part of what needs to be settled with regard to semantics is the
variety of naturalism that we—and cognitive science as a whole—ought to
opt for.
Natural Semantic Metalanguage
Characterization
The NSM approach has been developed by Anna Wierzbicka, Cliff Goddard
and their associates. Its immediate intellectual motivations, according to
the proponents themselves, reach to the early days of the cognitive revolu-
tion (Wierzbicka (1996)). The decline of behaviourism and the advent of the
new science of mind was a tremendous opportunity to rehabilitate mentalis-
tic concepts—including meaning itself—and to once again place them at the
forefront of one’s research agenda. Unfortunately, according to Wierzbicka,
4
the opportunity is now on the verge of being squandered; linguistics and cog-
nitive science have been victims of their own success—the formal machinery
in computer science and in Chomsky-inspired syntax-studies have led to the
replacement of a science of meaning with a science of computation and infor-
mation processing. For their part, NSM proponents question the utility of a
rigid syntax-semantics distinction.1 To study syntax at a remove from work
on meaning is simply to miss the point of the linguistic enterprise. Cog-
nitive science, by their lights, succeeded in moving beyond the theoretical
commitments of the behaviourists precisely by taking the question meaning
seriously. NSM starts with a very reasonable observation: Hardly anyone,
with the exception perhaps of a few skeptical philosophers, can deny that
normal humans seem to possess a commonsensical, pretheoretic understand-
ing of meaning. We ‘just know’ that our words and signs have significance.
Admittedly, we can’t always define them precisely and we are apt to misuse
terms on occasion. But normative issues such as these are really not to the
point. Our words are not mere noises or inert marks; they are imbued with
significance. The study of meaning, according to NSM proponents, starts
with that fact.
Not all meanings are on a par. A long philosophical tradition, stretching
via German Idealism at least to Arnaud, Leibniz and Descartes, teaches that
the content of some concepts is constitutive of the content of others. How
might this work? Consider the concept2 [triangle] (though this is not,
in fact, an example that Wierzbicka uses). It seems impossible to grasp
the content of this notion without a prior understanding of [side] and of
[angle]. In this sense then, [side] and [angle] are more basic concepts
1This distinction, as we shall see, plays an important role in the other two theories
discussed below.2I follow standard practise by referring to concepts (as opposed to things or to words)
by using small-caps within square brackets.
5
than [triangle]. Indeed, the story goes, they are building blocks of which
[triangle] is made up. It seems reasonable to suppose then that some
concepts, at least, have constituent parts.3
A natural question to ask next is whether perhaps all concepts might
have further conceptual constituents. In other words, is it possible that all
concepts might be explicated fully in terms of other concepts? Here, NSM
answers in the negative. Two reasons are adduced. To see the first, consider
the meaning of the word ‘after.’ We can explicate by using synonymous
terms and phrases such as ‘subsequently’, ‘at a later time’, or even ‘afterward’
(OED). But it is fairly transparent that the defining terms are themselves just
as semantically involved, and perhaps more so, as the word being defined.
In fact, the definitions themselves presuppose a tacit understanding of the
very notion being explicated. This serves as evidence, Wierzbicka (1996)
and Goddard (1998) argue, for supposing that some lexical items, including
‘after’, do not decompose further. There is another reason also—this time one
with a more aprioristic flavour. Following Arnauld and Nicole (1996), NSM
proponents ask us to consider carefully what it would entail for all concepts
to be definable without remainder in terms of other concepts. Would this not
result in messy tangle of circular definitions, none of which would point to
anything beyond the web of concepts themselves? The concept-user would,
it would appear, be trapped in a sort of conceptual solipsism, unable to make
sense of any one notion except in terms of all others. But since, happily, we
are not stuck in that lamentable predicament, at least some concepts must
act as a sort of ground. We are invited to conclude that some concepts must
not be definable in terms of other concepts at all. Let us call such basic
concepts the primes.4
3Even Jerry Fodor (1987), in an apparent moment of weakness, makes this sort of claim
with regard to [triangle]. For widely accepted counter-arguments, cf. Fodor and Lepore
(1992).4In fact, it is far from obvious whether either of the two arguments really work. I return
6
Let us, for the sake of argument, grant that the picture just presented
is, broadly speaking, correct. What, on the NSM account, does research in
semantics consist in? The task is two-fold. On the one hand, semanticists
work to identify those linguistic items which truly do not admit of defini-
tion in other terms—they dig up the bedrock of meaning, as it were. In
one sense, this sort of digging about the foundations is a familiar endeav-
our. Philosophers once spent a great deal of time trying to discern precisely
which concepts were foundational and which were derivative. In Hegel’s
(1812) work especially, this enterprise reached a high art-form. Of course,
Wierzbicka is not interested in a priori speculation of this sort, even if the in-
tellectual motivation seems superficially similar. Rather, NSM linguists rely
on extensive empirical investigation to arrive at a hypothetical list of primes.
They reason that if primes are the primitive carriers of content, they ought
to be lexicalized—or, at the very least, they ought to appear as a morpho-
logical unit—in every language. (Though it does not follow, of course that
any universally lexicalized item is necessarily a prime.) As one can imagine,
to discover whether some particular concept—such as [after]—is in fact a
prime requires a tremendous amount of painstaking research in comparative
linguistics and anthropology (work for which, NSM practitioners are owed
a debt of gratitude by everyone interested in semantics, regardless of the
ultimate fate of their theory.)
The benefits for cognitive science of arriving at a complete list of semantic
primes would be significant. The terms on the list would constitute a sort of
semantic analogue of Mendelejev’s (1869) Table of Elements. It would be a
universal code in terms of which all derivative concepts could be explicated.
Indeed, this explicative task is precisely the second line of research for the
NSM theorist. Just how it can work can be made vivid on the basis of a
simple example. Consider the following sentence of Polish:
to this point below.
7
(1) Janina jest szczesliwa.
Its NSM semantic analysis runs as follows: The proper name ‘Janina’ denotes
a person.5 The verb ‘jest’ serves to link the adjective ‘szczesliwa’ and the
person being designated: The person is presently in some state, the nature of
which is about to be specified. The key part of the semantic analysis concerns
the adjective ‘szczesliwa.’ In fact, this is a rather peculiar term in Polish;
one quite different from the English ‘happy’ (as Wierzbicka correctly notes).
Here is how its content can be spelled out in terms of semantic primes:
(2) Person P feels something
sometimes a person thinks something like this:
something good happened to me
I wanted this
I do not want anything more now
because of this, this person feels something good
Person P feels like this.
[Wierzbicka (1996), p. 215; also Goddard (1998)]
We arrive at an NSM analysis of the sentence (1) itself simply by substi-
tuting “Janina” for the variable P in (2). Notice that analysis nicely brings
out the difference between ‘szczesliwosc’ and apparent English translations
like ‘happiness.’ While the English term ‘happy’ is used frequently in every-
day contexts to refer to states of mild, temporary satisfaction, ‘szczesliwosc’
designates a profound sense of fulfilment (perhaps not unlike the French ‘bon-
heur’). At a minimum then, we can see that NSM analyses are extremely
useful to the comparative linguist as a tool for specifying subtle shades of
meaning.
5I am not certain whether NSM endorses a direct reference theory for proper names or
considers names covert definite descriptions. But it’s not really a crucial point for us.
8
In fact, there is more going on. Each of the words which comprise the
definition (except the variable P, obviously) is the English lexicalization of
a semantic prime. If NSM is correct, and if the analysis is truly complete,
any competent speaker whose native language is the metalanguage used here
(so English) should be able to deduce the meaning of ‘szczesliwa.’ Impor-
tantly, the explication should translate with no loss of content into any other
human language by systematically replacing the English lexicalizations of
the primes with their correspondents. Of course, that this can work is, in
part, an empirical claim. But if it were indeed right, NSM would have found
the foundations on which meaning rests. This would be a very remarkable
achievement indeed.
Discussion and Criticism
A number of criticisms to the NSM framework and its effectiveness can be
raised. One immediate worry concerns the empirical adequacy of (2). To
count as a satisfactory semantic analysis of (1), the analysis would need to
be complete. If we accept that the terms of (2) are primes and that primes
are the primitive constituents of meaning, then a reasonable test for the
completeness of (2) suggests itself. Presented with (2) a native speaker of
Polish ought to be able to unambiguously identify the word it explicates. By
the theory’s own lights, the meaning of ‘szczesliwa’ just is what is given in
(2). Yet, in point of fact, (2) fails this test. A native speaker of Polish is at
chance when trying to determine whether (2) analyzes ‘szczesliwa’ or of one
of the following synonyms (English approximations in parentheses):
• zasycona (sated)
• zadowolona6 (contented)
6This was, in fact, the favourite guess in a mini-run of the semantic reverse-engineering
experiment I conducted.
9
• zaspokojona (made to be at peace)
At best, it would appear, the analysis given in (2) is importantly incomplete;
more work is needed to identify further prime constituents of ‘szczesliwa.’
Admittedly, this need not trouble Wierzbicka overly. She has always very
reasonably maintained that NSM analyses are working empirical hypotheses.
But if the problem generalizes, it could spell trouble for NSM.
There are more fundamental worries too. The very idea of conceptual
‘primes’ is open to challenge. Assume, for the sake of argument, that some
lexical concepts truly are, in some important sense, composite.7 We were led
to posit semantic primes because, according to Arnaud, a network of con-
cepts defined exclusively in terms of other concepts yields vicious circularity.
The trouble is that Arnaud was wrong: there now exists a considerable body
of literature, inspired by de Saussure (1966), which supposes that concepts
precisely are constituted negatively. That is, they are constituted via the dif-
ference relations between themselves and all other available concepts. Worse,
there exist successful computer implementations of the Saussurian idea using
semantic networks (see discussion by Goldstone and Rogosky (2002)). Even
if such semantic networks turn out in the end not to be the correct way of
cashing out human conceptual capacities, their sheer possibility scuttles any
a priori arguments for the necessity of semantic primes.
Suppose, for the sake of argument however (and contra Saussure), that
at least some lexical concepts really can’t be fully explicated in other terms.
7Actually, even this is open to dispute. Fodor’s (1975, 1987) conceptual atomism pre-
cisely denies that any carriers of content (concepts) possess constituent structure. I omit
Fodor’s views and objections from my discussion here however because I have come to
believe that some crucial premises which the conceptual atomist relies upon are, in fact,
false. For an excellent empirical refutation of Fodor, see Goldstone and Rogosky (2002).
I hope to have a chance to address Fodor and Lepore (1992) versus Goldstone & Rogosky
(2002) in another paper.
10
It’s important to notice that even on such a scenario, NSM runs into choppy
waters. It certainly does not follow on this supposition that any concepts
are ontologically irreducible. All that follows (trivially, in fact) is that lexical
concepts are irreducible to other lexical concepts. Nothing prevents us from
supposing that lexical concepts can be explicated in some other way, perhaps
even in terms of elements which are themselves non-semantic. A plausible
candidate here might be unconscious information-bearing states of the hu-
man cognitive apparatus. In order to validly reach their desired conclusion,
Wierzbicka and Goddard must establish that the constituents of meaningful
terms need themselves to be meaningful (to us). They certainly think that
this is the case:
The NSM approach starts with the premise that semantic analysis
must be conducted in natural language, rather than in terms of
technical formalisms (abstract features, logical symbols, etc.), if
only because technical formalisms are not clear until and unless
they are explained in ordinary language.
[Goddard (1998), my emphasis]
But must we also follow them in thinking this? I don’t think we do. It
is hard to ask for a clearer avowal of what in the previous section I called
emergentism then the quote just cited. Now, emergentism with regard to
life (vitalism), consciousness (dualism), meaning or indeed any other phe-
nomenon is a respectable and substantive philosophical doctrine. But, as
such, it doesn’t come for free. Semantic emergentism must be supported by
argument or empirical evidence. Instead, Goddard takes it as a premise. By
doing so, he abandons any hope he may have had of persuading those who are
inclined to accept semantic reductionism. Nor does Goddard thereby shield
himself from criticism. By adopting a substantive philosophical thesis as an
11
unargued premise, NSM places itself in substantial jeopardy. Should seman-
tic emergentism be shown in future to be false or profoundly implausible (as
has indeed happened to vitalism and dualism), the entire enterprise collapses.
As things stand, in the absence of compelling arguments to suppose that se-
mantic properties are ontologically basic, we are not compelled to buy what
NSM is selling. In fact, we would probably be unwise to do so. (That said, I
want to reiterate that NSM semantic analysis has a useful—and perhaps in-
dispensable—role to play in cross-linguistic research. The NSM field-linguist
does valuable work by furnishing NSM-style decompositions. One day we
may need to cash out the so-called primes she employs in more ontologically
basic, non-intentional terms. But these are distinct issues.)
Formal Semantics
Characterization
The formal approach is very much the dominant, received view—at least
among philosophers and logicians—regarding how semantic theories for nat-
ural languages ought to be constructed.8 The principal ideas guiding the
formal approach can be traced via the work of Montague (1974), and David-
son (1967), through Tarski (1944) and ultimately to Frege (1953). At its core,
formal semantics develops the (intuitively very plausible) idea that to under-
stand a sentence of the vernacular is, at a minimum, to know its truth con-
ditions. If, without prior context, I inform you—a non-Polish speaker—that
it is the case that Janina jest szczesliwa, I am not being informative. But
if you manage to learn that that sentence is true if and only if Janina is
8Interestingly, there are disputes within philosophy itself—notably about the nature of
mathematics, following Benacerraf (1973)—which tacitly presuppose that by ‘semantics’
one just means formal semantics. The assumption can have implications for the sorts of
solutions one is prone to seek to the existing puzzles. These ideas are developed in my
prospectus.
12
happy, then, the argument goes, you have learned the meaning of what I
said. Formal semantics aims to work out precisely how the truth conditions
of arbitrarily complex sentences are a function of the contribution made by
their parts. A further and related goal is to trace the logical relations between
sentences. (More on that below.)
A number of the basic tools employed by formal semantics derive origi-
nally from work on the semantics of formal languages. I’d like briefly there-
fore to review how interpretations are developed for formal languages (and
why they are needed). I start with simple sentential logic and move on to
first-order predicate logic after that, before once again returning to natural
languages. The section concludes with a critical discussion of the formal
approach.
It is possible to construct a simple system of sentential logic out of the
following elements:
• Primitive Symbols. These include variables, operators and brack-
ets. The variables (A1, A2, . . . , An) range over propositions. Primitive
operators should constitute a logically complete set, as with ¬ (nega-
tion) and ∨ (disjunction). Brackets are just syntax sugar to keep things
organized.
• Rules of Formation. There are three: If A is a variable, then A is
a well-formed formula (wff). If A and B are variables, then ¬A and
A ∨B are wffs. Nothing else is well-formed.
• Definitions. There are three defined operators, in addition to the
primitive ones:
A → B =Def ¬A ∨B
A ∧B =Def ¬(¬A ∨ ¬B)
A ↔ B =Def (A → B) ∧ (B → A)
13
• Axioms. Russell and Whitehead’s logic includes the following axioms:
1. (p ∨ p) → p
2. q → (p ∨ q)
3. (p ∨ q) → (q ∨ p)
4. [p ∨ (q ∨ r)] → [q ∨ (p ∨ r)]
5. (q → r) → [(p ∨ q) → (p ∨ r)]
• Rules of Inference. There are three rules of inference:
Uniform substitution of wffs for variables.
Detachment. If A and A → B are theorems, so is B.
Replacement by definition. If A is defined as B, replace A for B or
B for A.
Each axiom is evidently a well-formed formula (wff). The rules of inference
and the definitions preseve syntactic well-formedness. Only wffs can therefore
be derived from the axioms via the rules of inference or definitions. Wffs so
derived are theorems. (This is a purely syntactic definition of theoremhood.)
A natural question to ask with regard to the set of theorems is whether they
are consistent—i.e. whether it is possible at any stage to derive φ and ¬φ
from the axioms via the rules of inference.9
Negation consistency cannot be demonstrated in syntactic terms alone.
One way to do so is to model the axiomatic system in another system, which
itself is known to be consistent (like arithmetic). Another way is to construct
a proof which shows that a contradiction cannot arise. This latter method
calls for the following additional elements:
9A good reason to try to avoid negation inconsistency is that from A ∧ ¬A and the
Duns Scotus’ rule, p → (¬p → q), one can derive any wff at all. This has the catastrophic
effect of effacing the distinction between theorems and non-theorems.
14
• Domain. This is a set consisting of (at least) two non-linguistic ele-
ments. Some possible domains include D={hat, pipe}, D={>,⊥} or
D={True, False}. Here, we can assume that the domain consists of two
elements D={1, 0}.
• Assignment Function. This function assigns one member of the
domain to each of the primitive variables in the language. If there
are n primitive variables in a given formula, there can be 2n distinct
assignment functions mapping variables to the domain.
• Model. The model is just an ordered pair consisting of the domain and
some particular assignment function: M =< D, v1 >. Its significance
will emerge in a moment.
To show the consistency of our axiomatic system, we needs to show that
nowhere in the infinite set of derivable theorems is there a formula as well as
its negation. Obviously this cannot be done by exhaustive search. Instead,
we proceed inductively. Let us call the degree of a wff the number of logical
operators it contains. Atomic formulae are of degree zero, since they do not
contain any logical operators. The negations of atomic formulae as well as
their conjunctins, disjunctions, conditionals and biconditionals are of degree
one. And so on. We can now define a valuation function which assigns each
wff, regardless of degree, to one of the elements in the domain.
• Valuation function. Given a model M, the valuation function (Vm)
extends the assignment function from wffs of degree zero to the set of all
wffs. It proceeds recursively. We assume that for degree zero wffs, the
assignment function specifies the domain assignments. It remains for
us to specify the domain assignment for wffs of the next higher degree.
In other words, we need to specify the effect of the logical operators on
domain assignment. Here’s how that’s done:
15
* Negation. Vm(¬A) = 1 ⇐⇒ Vm(A) = 0. In other words, a
degree one formula ¬A is mapped to 1 just in the case that the
degree zero formula A is mapped to 0; it is mapped to 0 otherwise.
* Conjunction. Vm(A ∧B) = 1 ⇐⇒ Vm(A) = 1 and Vm(B) = 1.
* Disjunction. Vm(A∨B) = 1 ⇐⇒ Vm(A) = 1 or Vm(B) = 1 (or
both).
* Conditional. Vm(A → B) = 1 ⇐⇒ Vm(A) = 0 or Vm(B) = 1.
* Biconditional. Vm(A ↔ B) = 1 ⇐⇒ Vm(A) = Vm(B).
Let us say that a formula A is true in a model M iff Vm(A) = 1, i.e. if
that formula maps to the domain element 1 in the model. Let us also say
that a formula is valid if it is true in all models—so regardless of how its
constituent elements are themselves mapped to members of the domain. (The
semantically valid formulae are those which are tautologous.) Using the
valuations of the logical operators just given, we can verify that each of the
axioms of our sentential logic is a tautology. Moreover, the rules of inference
preseve validity. We can be sure therefore that all theorems derived in our
system will be tautologies (i.e. will map to 1). But, in the general case, if
φ is a tautology, ¬φ is a contradiction and therefore not a theorem. We see
therefore that the initial system is indeed consistent.
Sentential logic of the kind just discussed is vastly less powerful than first-
order predicate logic (with or without identity10). To extend the approach
just outlined to the construction of interpretations for first-order predicate
logic, one must furnish an interpretation for the universal and existential
quantifiers. Here they are:
10The identity operator is needed inter alia for definite descriptions, as in “The woman
who walked on the Moon was Greek”, and for for sentences involving numerosity, such as
“There are at least seven cities in Europe” or “There are precisely three moons orbiting
Earth”.
16
* Universal Quantifier. Vm((∀x)A) = 1 ⇐⇒ Vm(A) = 1 for all
models M ′ where M ′ is defined exactly as the model M except possibly
with respect to its assignment of domain member to x.
* Existential Quantifier. Vm((∃x)A) = 1 ⇐⇒ Vm(A) = 1 for some
model M ′.
And, of course, we need also to say something about the semantics of pred-
icates. Unlike sentential logic, where variables range over full propositions,
predicate logic variables range over entities. Predicates work as functions
which take the entities as parameters and map to members of the domain
(so to truth values). Predicates can be monadic, dyadic, or n-adic, according
to their number of parameters. Here’s how that works in the abstract:
Ga is a saturated monadic function which has taken the variable a as its
parameter. Its interpretation maps it to a member of the domain: 1 if
a in fact does have the property G and 0 otherwise.
(∃x)Gx is a unsaturated monadic function. It tell us that there is there is
at least one entity in the domain which satisfies the function such that
it maps to 1. We arrive at Ga by using a as the function’s parameter.
(∀y)(∃x)Gxy is an unsaturated dyadic function. It tells us that for any
entity in the domain, there exists some entity which stands in relation
G to it.
With these resources in place, we can now halt our discussion of the semantics
of formal languages. The topic is much richer than what I have had a chance
to outline here.11 Still, I have hopefully summarized enough to explain how
the formal approach to semantics can be extended to the natural languages.
11For instance, I’ve not said anything about semantics for modal logics, an interesting
area of study. (Pre-Kripke these were dismissively known as muddle-logics as no clear and
consistent interpretation seemed possible.)
17
Much is owed in this regard to the work of Alfred Tarski. In a several
important papers published in Polish, French and German in the early 1930s,
Tarski offered a disquotational characterization of truth (for a language L).
This work helped to demystify the notion of truth and to rehabilitate it as
a philosophical topic of study.12 In brief, Tarski presupposed that we know
the meaning of our own sentences and proceeded to construct a recursive
characterization of truth on that basis. As with formal languages, one begins
by specifying, for a language L, the entities which names and singular terms
denote, as well as which entities L’s predicates are satisfied by (cf. assignment
function, above). One then constructs a list of the simple sentences of L, such
that the name of each sentence appears on the left-hand side, and its truth-
conditions on the right. For example:
(T) “Snieg jest bia ly” is true if, and only if, snow is white.
A complete such list, as well as a complete set of rules for combining sen-
tences, yields a recursive characterization of truth for L.
Donald Davidson (1967) inverses Tarski’s schema (T). Contrary to Tarski,
Davidson supposes that we understand the notion of (garden-variety, disquo-
tational) truth and constructs a recursive characterization of meaning on that
basis. In essence, a Davidsonian theory of meaning for a language L begins
by pairing all of the simple sentences of L with their truth conditions; it
builds up compound sentences recursively. In general, to know the meaning
of a sentence of a language L is to know its truth conditions.13 Of course,
12For a charitable, if critical discussion, cf. Hartry Field (1972).13I’m simplifying. In fact, Davidson is a holist about interpretation. To have a theory
of meaning for a language L is to have a theory which holistically pairs each sentence of
L with truth conditions such that words exercise a constant and systematic effect on the
meanings of sentence and such that radical disagreement betwen interpreter and interpre-
tee is minimized (the famous ‘principle of charity’). Also, one needs to fill in contextually
determined terms like ‘today’ and ‘I.’
18
we need not know whether a given sentence is true to understand it; we need
only to know what would make it true (or false).
We just saw that the semantics for predicate logic involved two sorts
of mappings: Saturated formulae (ones with no variables) are mapped to
one of the members of the nonlinguistic domain—1 or 0. Let’s call these of
semantic type < t >. Unsaturated formulae (ones which contain variables)
denote functions. That idea can be readily adapted to cope with natural
language. Noun phrases can be taken to denote individuals. We’ll say that
these are of semantic type < e >, Verb phrases, on the other hand, can be
treated like formal predicates—i.e., they can be treated as denoting n-adic
functions which, when fully satisfied, map to truth values. We’ll say that
they are of semantic type < e, t >. Let me illustrate. Take as our example
the simple sentence, “Janina czyta.” Here is how its semantic interpretation
might work:14
(3) Janina czyta.
JJaninaK = Janina (the person out in the world). Type: < e >.
JczytaK = a function mapping to 1 (in the domain) if the semantic value
of its parameter engages in the activity of reading and mapping to 0
otherwise. Hence, of type: < e, t >.
Here is the same idea expressed in a different notation:
[f: D→ {0, 1}For all x ∈ D, f(x) = 1 iff x reads
](Janina)
Things get only slightly more complicated in the case of dyadic predicates,
such as ‘loves,’ which require two parameters—one naming the lover and
14The double brackets indicate the semantic value of what they contain. This analysis,
and what follows, is based on Heim and Kratzer (1998).
19
the other naming the thing loved. Hence, noun phrases continue to denote
individuals. The dyadic predicate is, once again, a function. This time
however, it is a function which takes an individual (the loved) as its parameter
and outputs another function. That second function takes a second individual
(the lover) as a parameter and maps to a truth value. Here’s how that works
for the sentence “Janina kocha Paryz”:
(4) Janina kocha Paryz.
JJaninaK = Janina. Type: < e >.
JParyzK = Paris (the city itself). Type: < e >.
JkochaK = f: D → {g: g is a function from D to {0, 1}}For all x, y ∈ D, f(x)(y) = 1 iff y loves x. Type: < e, < e, t >>.
Alternatively, we can define a new symbol (λ) which means essentially the
same as ‘is a function.’ This lets us rewrite our two interpretations more
succinctly as follows:
JczytaK(Janina) = [λx : x ∈ D . x reads](Janina) = 1 if Janina reads
= 0 otherwise.
The function denoted by ‘czyta’ takes one parameter, the denotation of ‘Jan-
ina’, and maps to a member of the domain (1 or 0). Similarly:
JkochaK(Paryz)(Janina) = [λx ∈ D . [λy ∈ D . y loves x]](Paris)(Janina)
= 1 if Janina loves Paris
= 0 otherwise.
So much then for the bare basics of the semantics for simple declarative
natural language sentences.
20
An important difference between formal languages of the sort that
Tarski’s semantic characterization of truth applies to unproblematically and
natural languages which children acquire at their mother’s knee concerns am-
biguity. The formulae of formalized languages are precisely constructed so
as to avoid the need to map two distinct interpretations to a single syntactic
formula. This is in general, of course, not the case for sentences of natu-
ral languages where ambiguous utterances abound. Consider the following
example15:
(5) The millionaire called the governor from Texas.
The sentence can mean one of two things: Either the millionaire telephoned
a governor of some undisclosed institution while she (the millionaire) was in
the Lone-Star State. Or, the millionaire placed a call to the Governor of
the State of Texas from parts unknown. As before, we can make a little bit
of headway spelling out an interpretation of (5) by means of the following
schema:
(5T) “The millionaire called the governor from Texas.” if, and only if, the
millionaire called the governor from Texas.
The problem is that this schema alone does not help us make sense of the
ambiguity. To get a handle on that, both the left-hand side and the right-
hand side of the schema need to be made considerably more sophisticated.
What is needed on the left-hand side of the formula is an unambiguous
restatement of (5) which makes salient the relationships between the various
parts. Here, work in syntax can be very helpful. The Chomskian revolution
has made it possible to at least imagine how a complete formalization of the
syntax of natural language might be possible in the very long run. Let us
15This example is discussed, though not analyzed in detail, by Pietroski (2003b).
21
suppose for our purposes (falsely, it would appear) that X-bar rules of the sort
developed by Chomsky in his GB phase permit a correct characterization of
NL syntax. This will let us pry apart two syntacticly distinct sentence types,
a token of which one might utter by producing (5). Here is the first, involving
a Texas State Governor (or maybe a Texas-born governor) receiving a call:
(5a) IP````
NPHHH���
DP
D′
D0
The
N′
N0
millionaire
I′PPPP����
I0
+ past
VP
V′PPPP����
V0
call
NPPPPP����
DPJJ
the
N′aaaa!!!!
N0
governor
PPaaa!!!
from Texas
And here is the one which involves the millionaire telephoning from the state
of Texas to some governor or other:
(5b) IPhhhhhhh(((((((
NPHHH
���
DP
D′
D0
The
N′
N0
millionaire
I′`````
I0
+ past
VP
V′XXXXX�����
V′HHH���
V0
call
NPaaaa!!!!
the governor
PPaaa!!!
from Texas
22
Each of the two possibilities requires a distinct semantic interpretation—i.e.,
it requires that a different set of truth-conditions be inserted into the right-
hand side of schema (5T). Here, we can make use thematic relations and of
Davidson’s (and Charles Parsons’) work on the event-operator (e). (5a) can
be interpreted as:
(5a′) Past [The x: GOVERNOR(x), BENEFACTIVE(x, Texas)]
∃e(CALL(e) & AGENT(millionaire, e) & PATIENT(x, e))
Whereas (5b) would be parsed as:
(5b′) Past ∃e(CALL(e) & AGENT(millionaire, e) & PATIENT(governor, e)
& SOURCE(e, Texas))
By inserting a Chomskian syntactic tree into the left-hand side of (5T) and
a Davidsonian event-analysis into its right-hand side, we arrive at an unam-
biguous semantic interpretation for (5). In the general case, by matching
simple sentences with their semantic interpretations, and by specifying the
rules via which both the syntactic and the semantic (left- and right-hand
side) analyses are to be extended, one arrives at a recursive characterization
of meaning for some natural language L. The devil, as ever, is in the details.
Discussion and Criticism
How plausible is it that FS is capable (in its current form) of providing us
with a materially adequate explanation of the nature of meaning? There have
already been hints of trouble. We saw NSM offer an explanation (albeit a
flawed one) of the difference between the Polish adjective ‘szczesliwa’ and the
rough English translation ‘happy.’ Our FS analysis glossed over the point.16
16Unlike NSM, FS would spell out the difference between the two notions extensionally:
the set of entities which satisfies the one predicate is distinct from the set of entities which
satisfies the other. They are therefore semantically distinct. Of course, one can sensibly
wonder whether this is an explanation or just a redescription of the facts in set theory.
23
Moreover, NSM was able to recognize that while we certainly do use words to
make truth-evaluable claims, we also cajole, request, apologize, give orders,
and, in general, put language to a variety of other uses (cf. Wittgenstein
(1953)). FS does not seem readily equipped to accommodate this.17 Having
noted these apparent shortcomings, I will leave them aside here. Instead,
I want to focus on one recent criticism of FS due to Prof. Paul Pietroski
(2003a,b).
Natural languages are, fairly obviously, human capacities. A theory of
meaning for a natural language pairs up linguistic signals with interpretations
(whatever those may turn out to be in the end). Competent speakers of
natural languages are able to effect such a pairing relatively effortlessly in real
time, even if neither they nor we know quite how the underlying mechanisms
work. It seems reasonable to think that a theory of meaning for a natural
language ought therefore to be a theory of understanding—i.e., a theory
which explains how it is that people come to pair signals with interpretations.
Pietroski (2003b) argues that, contrary to appearances, FS fares poorly as a
theory of understanding, in this sense. Consider the following example:
(6) France is a hexagonal republic.
In the right context, this sentence can express a true idea: Suppose, for in-
stance, that someone jokingly suggests that countries’ political constitutions
can be read off their geographic shapes. Suppose also that, on their story,
hexagons are monarchies. One can imagine tokening (6) so as offer a counter-
example to the theory. Now, notice that on a Davidson-inspired (so also FS)
17One could perhaps replace truth-conditions in the meaning schemas with felicity-
conditions in the style of Austin (1962). Thus, in the case of declarative sentences, the
felicity-conditions would just reduce to truth-conditions, whereas in the case of speech-acts
with a different illocutionary force, we might need a significantly complex analysis than
that offered by the Davidsonian event-analysis alone.
24
account, ignoring the complexities of the left-hand side, the meaning schema
for (6) is essentially this:
(6T) “France is a hexagonal republic” if and only if
(∃x)[(Fx ∧ (∀y)(Fy → y = x)) ∧Hx ∧Rx]
It would seem then that the theory requires that there be a real, robust
entity in the world which simultaneously satisfies two predicates—viz. being a
republic, as well as being hexagonal. Now, undoubtedly, France is a republic.
Likewise, there is a sense in which France is hexagonal, much as Italy is boot-
shaped. But it is doubtful whether there exists a single entity in the world
with both of these properties. Rather, it seems more plausible that we are
able simultaneously to think of France under two descriptions. Consider what
happens when we add even more complexity into the mix:
(7) France scored two goals in the quarter-final.
To suggest that there exists an entity which is hexagonal, a republic and
scored two goals is to adopt desperate measures. What sort of entity might
that be? Geographical? Socio-economic? Socio-geographico-sportive? Piet-
roski asks whether perhaps it would not be more sensible to question whether
theories of meaning (which explain how we come to understand sentences
like (6) and (7)) ought really be theories of truth. Sure, we can express true
thoughts using our words. But, in general, the truth of our utterances is
a massive interaction effect involving our intentions, contexts, listeners and
(importantly) the world. To ask a theory of understanding for human lan-
guages to be a theory of truth is to ask it to be a theory of everything (a
point Chomsky (2000) also makes). Instead, Pietroski argues, we can hang
on to the idea that theories of meaning should have the general form of a
Tarski-style recursive definition without requiring that the right-hand side
of the schema specify anything as full-blooded as a set of truth conditions.
25
Instead, one might think of lexical items, such as ‘France’ as certain kinds of
‘recipes’ for constructing representations.18. “France is hexagonal.” leads to
the construction of one kind of mental representation. “France is a repub-
lic” leads to another. The notion [France] constructed in each instance is
different. But the two are such as to be able to be combined in (6), even if
“France is hexagonal, scored two goals, and a republic” begins to stretch our
mental resources.
Let me close with yet another problem for FS. If one rejects Pietroski’s
criticisms and hangs on to the idea that a theory of understanding (not only
has the formal character of, but) really is a theory of truth, one still needs to
explain the fundamental posits FS averts to. We saw above how meanings can
be characterized in terms of sets (or, equivalently of functions and satisfiers).
However, the ontological status of sets, and of abstract entities generally, was
left undetermined. We know that they are abstract entities. But just what
those might be, whether they truly exist, and whether they can be reduced
to better-understood natural entities is unclear. This is not the place to get
into the philosophy of mathematics. But the worries are worth keeping in
mind, if only to remind us that the explenanda in terms of which FS explains
meaning are not themselves well understood. The approach is, in this sense
at least, tacitly committed to a sort of emergentism regarding meaning.
Conceptualist Semantics
Characterization
Jackendoff’s (1983, 1987, 2002) conceptualist semantics (CS) differs consid-
erably from both of the approaches considered so far. Unlike NSM, CS
acknowledges the fundamental importance of Chomskian work on the syn-
18This idea is not unlike certain proposals alive in current cognitive psychology; cp.
Barsalou (1999)
26
tax of natural language. In this respect at least, CS is not unlike some of
the sophisticated work being done in formal semantics (Larson and Segal
(1995)). CS differs from FS however in not attempting to hook linguistic
items up with referents ‘out in the world’—at least not in any straightfor-
ward way (see below). CS is at once a radical extension of the Chomskian
programme and an attempt to integrate work in linguistics with the rest of
cognitive science. In this section I outline how this is supposed to work.
In the next, I present some objections which—by Jackendoff’s (2003) own
admission—remain a source of worry for CS.
CS begins by taking on board the key insights of the Chomskian pro-
gramme. Let me take quick moment to lay those out: The normal human
child is born with a ‘language acquisition device’, or a dedicated cognitive
apparatus for acquiring the language to which the child is exposed. Acquir-
ing a language is not a matter of hypothesis formation and disconfirmation,
as empiricist theories suppose. An unbounded set of possible grammars is
compatible with the finite data available to the learner; empiricist theories
have a hard time explaining why most candidate theories are never so much
as entertained. Repeated studies have shown that, in fact, children are highly
resistant to negative evidence, including explicit correction. Hypothesis dis-
confirmation does not seem to play a role. As a matter of fact however,
children’s syntactic performance does come into line with that of others in
their linguistic community. And, of course, children are able to generalize:
they are able to construct an unbounded number of novel, grammatical utter-
ances. The sole plausible explanation is that language constitutes an innate
cognitive mechanism, and that the child’s apparent ‘learning’ of her native
tongue is really the setting a finite number of innately-specified parameters.
(For a more careful exposition of this rather familiar story along with its
analogues in semantics, see Pinker (1989).)
The proper object of study for linguistics (or, at least so far as syntax is
27
concerned) is the linguistic competence of an idealized mature speaker. In
essence, one asks: what must the algorithm which the mind/brain’s dedicated
syntactic-module implements be like for the performance of the competent
speaker to be what it is (abstracting from inessential, individual peculiari-
ties, memory constraints and the like). Language understood as this sort of
dedicated, mental algorithm is sometimes called i-language: it is individual,
internal and intensional. Admittedly, this picture contrasts rather sharply
with the perhaps more commonsenseical notion of language as a publically
observable set of marks. Rather, what we typically think of as ‘languages’
(so Hungarian, Tagalog, Putonghua) are rough-and-ready abstractions. In
fact, there exist a number of more-or-less similar, mutually intelligible, com-
munity practices. These linguistic community practices arise as the effects
of the interaction between speakers’ i-languages. (Chomsky (2000).)
Chomskian mentalism has been a very rich source of insight into the syn-
tax of natural languages. Yet Chomsky himself has remained sceptical about
the possibility of developing a semantic theory (hence the mutual hostility
between NSM and Chomskian linguistics). The worry, it seems, has been
that it is next to impossible to draw a principled line between our under-
standing of language and our understanding of the world.19 And if this sort
of distinction cannot reliably be drawn then semantics threatens to become a
sort of grab-bag ‘theory of everything.’ One can see why a hard-nosed syntac-
tician might view that outcome as ultimately threatening the entire linguistic
enterprise. (This, I take it, was the motivation for Chomsky’s attacks on the
‘generative semantics’ work in the 1970s.)
Jackendoff maintains, contra Chomsky, that a bounded, rigorous and non-
trivial semantics is possible after all. Suppose, he argues, that we accept the
important contribution to the understanding of human languages made my
19Davidson too argued against there being a distinction between one’s conceptual scheme
and one’s knowledge of language.
28
Chomskian syntax (whether GB or Minimalist). Suppose we accept mental-
ism and nativism, as well as the focus on i-language. One of the key founding
assumptions of the Chomskian program has been the combinatorial structure
of syntax: complex syntactic structures are built up recursively out of sim-
pler constituents.20 An unargued assumption at the core of the program has
been that syntax, and syntax alone, displays this sort of structure. But, as
things stand, that assumption flies in the face of current linguistic practise.
As a matter of fact, phonologists have pursued generative theories of their
own, quite independently of syntax, for several decades. The phonological
structure of language is understood today to obey proprietary combinatorial
principles, ones quite independent of the compositional principles at work in
syntax. Jackendoff (2002, 2003) urges that linguistic theory should catch up
to linguistic practise. We should explicitly recognize that syntax, phonology
and semantics each constitute distinct combinatorial modules of the language
faculty, and that each executes its own computations over proprietary (un-
conscious) ‘representations’. On such a view, the study of semantics (much
like the study of syntax) involves theorizing about the sorts of computational
data-structures and processes which take place in the relevant portion of the
mind/brain.
Figure 1 (below) illustrates the overall architecture of the semantic fac-
ulty, as currently envisaged by CS. It also shows how semantics is thought
to link to related mental faculties—perception and motor faculties in partic-
ular—as well as to the world at large. Before discussing the inner workings
of the semantic faculty as pictured there, let me ward off a possible misun-
derstanding. The term ‘representation’ is often used to convey that some
entity stands in for some other entity for some purpose in so far as an ob-
server is concerned. Jackendoff carefully avoids appealing to this notion;
the data-structures which inhabit the cognitive system, as he envisages it,
20We saw an emphasis on combinatorial structure in our discussion of FS, above.
29
typically do not stand in any interesting one-to-one relation to the external
world. Nor is there a little homunculus in the brain to whom things are rep-
resented (Dennett (1991)). Rejecting representations may seem like a radical
move. Philosophers have taken CS to task in part for being solipsistic for
this reason. In fact, there is a well-established precedent in the cognitive lit-
erature for dispensing with inner representations. More importantly, there is
an existence proof: Rodney Brooks’ (1991) robots are capable of complex in-
teractions with their environment (including collecting pop-cans left around
the MIT lab) without the benefit of inner representations. The activation
vectors at the higher levels of the robots’ hierarchical subsumption architec-
ture do not correspond to any features in the world at large. They do not
represent anything out there at all. I read Jackendoff to be suggesting that
the computational states within the higher perceptual and semantic mod-
ules are very much in the same boat. They are fine-grained enough to allow
complex responses to emerging situations, and that’s what really matters.21
Caveats in place, let us return to Figure 1. CS semantic representations
are data structures in the unconscious mind/brain. They are formed accord-
ing to their own rules of formation, constituting a largely autonomous level
of processing. This level acts as an interface between syntactic structures,
on the one hand, and the perceptual and motor modules on the other. The
syntactic structures for concrete entities display perceptual features which en-
code salient observable aspects. In addition, they display inferential features
which tie them to other syntactic structures. (Syntactic structures repre-
21While on this topic, let me mention also that Goddard (1998) suggests that Jack-
endoff’s semantic elements are ‘abstract’ and therefore not to be identified with ordinary
word meanings. Goddard is right to say that the data structures manipulated by the CS
semantic faculty are not ordinary word meanings if the latter are accessible to conscious-
ness (as they often are). He is wrong however to suggest that they are abstract entities,
if by that one understands what philosophers typically do: acausal, atemporal, platonic
objects (cf. Burgess and Rosen (1997)).
30
senting abstract entities obviously lack the perceptual features, but possess
inferential features nonetheless.) The semantic module acts as a sort of in-
terpretive interface between syntax and the outputs of the perceptual and
motor faculties. Jackendoff argues that our experienced reality (though not,
of course, the world as such!) results from the output of the brain’s per-
ceptual modalities. In some sense, CS pushes the world of experience “into
the brain” (Jackendoff 2002; p.303). The job of semantics, as CS sees it, is
to connect the world as experienced by us with the structures which inhabit
the language faculty. One way to understand that is to view the semantic
faculty as mediating between the sorts of data-structures furnished by high-
level vision (see Biederman’s 1995 geon theory for instance) and the rest of
the language faculty. Semantic structures thus do not represent elements of
the external situation in any direct way. The structures which inhabit the
semantic module do not have a meaning; rather, they do all of the things that
meanings are supposed to do. They are meanings (Jackendoff 2002; p.306).22
One may wonder what happens to the lexicon on the picture just pro-
posed. Rather than positing a dedicated lexicon in which individual lexical
items are stored as tough in a box, CS posits only interfaces between each
of the three generative modules. In some sense, the lexicon just consists in
these interfaces. The smooth functioning of the overall system gives rise to
the appearance that there exist coherent data-units (words) with a phonol-
ogy, syntactic role and a semantics. Under the hood however, there are only
regularities in the way that the three modules come to interact. This permits
Jackendoff to explain situations where the coordination comes apart slightly
or fully. (On this account, ‘pushing up daisies’ and other idioms have a
22Cognitive semantics has sometimes been accused of vacuously translating from one
language (say, English) into another language, ‘semantic-markerese’ (cf. Lewis (1972), for
example). By tying the semantic structures to perception and to action, as he has done,
Jackendoff (2003) blunts the force of this criticism. Ironically, FS has no parallel way out.
32
normal syntax and phonology but interface as a unit with a single semantic
item: [DIE]. Likewise, nonsense phrases like “twas brillig and the slithy toves
did gyre and gimble in the wabe” have a phonology and perhaps even some
syntax but fail to interface with any semantic structures.)
It might be helpful to close by contrasting a CS analysis of a natural
language sentence with those given in previous sections. Recall example (5):
(5) The millionaire called the governor from Texas.
Analyzing a sentence of the vernacular is a matter of lining up its phono-
logical, syntactic and semantic representations. We discussed the syntactic
representations corresponding to this sentence (above).23 Giving the two cor-
responding semantic representations is a matter of decomposing the sentence
into its semantic constituents. According to CS, the human cognitive sys-
tem—both within the semantic module, as well as outside—has available to
it a small repertoire of conceptual types. We employ these to make sense of
our physical surroundings, plan, imagine, and so on. The conceptual types
also play an important role in constructing semantic interpretations on the
basis of information available from the syntax module. Among the primitive
types used by CS are at least the following: [thing] [place] [direction]
[action] [event] [manner] [amount]. A fist step toward analyzing (5) is
furnishing it with a gloss in terms of these semantic types.
Recall that (5) is ambiguous between two possible readings; we will need
two distinct semantic analyses. Let us begin with the CS gloss on the situa-
tion involving a Texan public official being contacted by a well-to-do caller:
(5α′)
23My handle on phonology is too tenuous to attempt an analysis. Asking someone else
to do it for me would be cheating. Sorry.
33
CALL(
[MILLIONAIRE
Thing
]),(
SENATOR(
ORIGIN([
TEXAS
Place
])
State
)Thing
)Event
The sentence is, as one might expect, fundamentally about an event, as in-
dicated by the category label in the main, outer square bracket. The event
type, the call, takes two parameters: the caller and the callee. Both indi-
viduals are represented in the CS analysis. How the mind/brain manages to
‘refer’ to these two worldly entities—in other words how it manages to pick
out and track those entities in the world—is left, appropriately enough, to
the cognitive psychology of perception and of attention. One further fact
matters for semantics: In this case, the callee is further modified by having
their place of origin specified.
Here is the contrasting case, where the calling is being done from Texas:
(5β′)
CALL
FROM([
TEXAS
Place
])
Path
([
MILLIONAIRE
Thing
]),(
[SENATOR
Thing
])
Event
Again, the overall content concerns an event. In this case however, the call
is coming from Texas, as indicated by the [path] information. One might
envisage filling in the path details along the lines suggested by the work of
Talmy (2001) or Regier (1996). In other words, what is being labelled [path]
here, might stand in for some particular set of activations of a structured
connectionist network, trained to be sensitive to location. (The details do
34
not matter much for us. It is important though that there are ways of cashing
the labels out in computational terms.)
Note that sentence (5) can’t mean that the millionaire is from Texas
because the semantic construct which expresses this possibility cannot be
built up on the basis of the two possible syntactic analyses (5a′ and 5b′). That
is, the mind/brain of the hearer cannot construct this (otherwise perfectly
sensible) idea from the resources provided.
Discussion and Criticism
CS has its share of critics. Some dismiss it as hopelessly vague. And they
have a point. Lined up next to the rigorous-looking formalizations which FS
theorists produce, current versions of CS look rather flimsy by comparison.
Proponents of NSM have been critical of Jackendoff too. Wierzbicka (2003,
conference presentation) repeatedly emphasized that CS has not succeeded in
producing a single full semantic description of any natural language sentence
or sentence fragment. From this, she concluded that her own project was the
best—indeed, the only—well-worked out alternative to formalist approaches.
In my view, criticisms from both sources are somewhat disingenuous. I have
already argued that until formal semantics can constitute a theory of human
understanding, it cannot count as an explanation of meaning. For its part,
NSM fails to explain semantic notions in non-semantic terms and therefore
it too can be accused of not yet having provided a single full explanation
of a semantic phenomenon. Wierzbicka may be right that CS has not pro-
duced a single semantic decomposition which takes us right from the level of
the syntax-semantics interface, via semantic structures, through molecular
semantic primitives, down to the cognitive psychology of perception. But
isn’t that a bit much to hope for so early in the game?
Instead of focusing on criticisms which FS and NSM level at conceptualist
semantics, let me here reiterate those raised by Jerzykiewicz and Scott (2003).
35
Essentially, they amount to the charge that, as it stands, the theory of ref-
erence on which CS relies entails the fallacy of psychologism. Psychologism,
recall, is the attempt to account for the necessary truths of mathematics in
terms of contingent psychological facts. Frege (1953) raised seminal objec-
tions to that sort of project.24 Most philosophers since then have regarded
psychologistic theories as patent non-starters.25 This does not necessarily
mean that Jackendoff is wrong (in fact, I don’t think that he is). It does
mean though that for CS to be fully defensible, it must provide an explicit
discussion and defence of psychologism, showing either that the doctrine is
not a fallacy or that the charge does not apply.
What then is the problem? In essence, Jackendoff’s (2002) account of
abstract objects (Section 10.9.3) looks like it’s on shaky ground. As we just
saw, on the CS account, conceptual structures within the generative semantic
module are not themselves interpreted—they do not have a semantics. They
just are the semantics of natural language. The fine-grained data-structures
that inhabit the semanticmodule interface richly with perceptual modali-
ties and with motor outputs, while individually not necessarily representing
anything in the world as such. The familiar appearance that words refer
to entities and events can be explained—for concrete referents, at least—in
terms of the relationship between semantic constructs and the outputs of
perceptual faculties. It is these outputs that we consciously experience as
our ‘world’. In the case of abstract objects (like beliefs, mortgages, obli-
gations, and numbers) which manifestly lack perceptual features, the the-
ory makes only slightly different provisions: the data-structures that encode
them possess inferential rather than perceptual features. Interfaces to syntax
24For a very good historical and sociological account of the psychologism controversy,
yet one sensitive to philosophical detail, see Kusch (1995). For a Frege-inspired attack on
Chomsky, see Dartnall (2000).25For a recent example of this, see Burgess and Rosen (1997)
36
and phonology treat all conceptual structures similarly, regardless of whether
their constitutive features are exclusively inferential or, inpart, perceptual.
So, in effect, Jackendoff’s reductive, naturalistic theory of concepts rejects
platonism and identifies abstract objects with the cognitive structures that
express them.
The paradigm cases of abstract objects are mathematical and logical en-
tities. It is odd therefore that Jackendoff does not discuss such entities ex-
plicitly. If the CS account of abstract objects is to work at all, it must
work for them. The trouble is that CS entails psychologism, the view that
the necessary truths of mathematics and logic are to be accounted for in
terms of contingent facts about human cognition. According to psycholo-
gism, 2 + 2 = 4 is a fact of human psychology, not a fact that is independent
of human beings. Frege (1953) raised seminal objections to this doctrine and
today psychologism is typically viewed as a patent fallacy. There have been
several notable attempts to defend psychology-inspired theories of the nature
of mathematical objects (Kitcher (1983), Maddy (1990) among them). But
these have not, it seems, met with much success. The good news is that there
is room for discussion. Haack (1978) points out that it is far from obvious
whether Frege’s objections continue to apply to modern, cognitive theories.
Frege’s target was the introspectionist psychology of the day, and (Jackendoff
1987, 2002) carefully avoids this approach. It may, therefore, be possible to
articulate a theory of abstract objects consonant with CS, yet responsible to
the philosophical literature.
To get off the ground, a CS-inspired account of abstract entities must
cope with a number of challenges. Mathematics is an odd domain and math-
ematical judgements are unique in a number of respects. A good theory has
to explain at least the following three features of mathematical judgements:
• Universality. Some norms derive their authority from community stan-
dards. Those norms are no less real for their conventional nature (traf-
37
fic rules come to mind), but they are only true by agreement. By
way of contrast, norms governing the behavior of abstract logical and
mathematical entities are universal (a point stressed by Nagel (1997)).
Community standards derive their authority from the norms, and not
vice-versa. Even people with untutored intuitions can come to rec-
ognize the truth of a law of logic or mathematics, though they may
require quite a bit of reflection to do so. CS needs an explanation of
how some abstract objects (which are supposed to be mental entities)
come to possess these inferential features. Are they innate? If so, Jack-
endoff’s appears to be committed to a version of Fodor’s language of
thought hypothesis, in spite of having explicitly rejected Fodor. Are
they learned? If so, the poverty of stimulus problem rears its ugly head.
• Objectivity. Logic, geometry and mathematics are not uninterpreted
formal systemsthat people happen to universally assent to regardless
of which community they inhabit. Formal interpretations of physical
phenomena permit predictions concerning the behaviour of objective
reality even in contexts vastly beyond the scope of actual (or pos-
sible) human experience. Many researchers have commented on the
‘unreasonable’ effectiveness of applied mathematics, even in contexts
where the original mathematical tools were developed for purely formal
reasons. How does mathematical reasoning manage to preserve truth
about distant contexts if mathematical objects are merely psychological
data structures with local inferential features? In other words, quite
apart from its universality, how, on the psychologistic account, does
mathematics come by its objectivity?
• Error. It is tempting to account for the validity of logical inference
in terms of the way that (normal, healthy) cognitive systems actually
reason. But we can make mistakes regarding the properties of abstract
38
objects. Even professional mathematicians occasionally draw false in-
ferences about mathematical objects. And a real feeling of surprise
and discovery can accompany mathematical innovation—that moment
when humanity discovers that we have all been conceiving of some
mathematical construct incorrectly all along. The intuition that math-
ematical objects can have properties quite different from those imputed
to them, even by professionals, fuels platonist intuitions (Godel 1947).
Validity cannot merely consist in a conformity with the way people ac-
tually reason—it is a property of arguments that conform to the way
we ought to reason.
How psychologism can account for this remains uncertain. Jackendoff (pp.
330-332) suggests several mechanisms of social “tuning” that can serve to
establish (universal) norms within a community—norms against which error
may be judged and the appearance of objectivity can arise. So when Joe
mistakes a platypus fora duck (p. 329), his error is relative to the impres-
sions of the rest of his community. “Objective” fact and the appearance of
universality is established by communityconsensus. Unfortunately, this ac-
count does quite poorly with logic and mathematics. A mathematical or
logical discovery happens when one member of the community realizes that
something is wrong with the way the community conceptualizes some aspect
of the field, and demonstrates that error to the other members of the com-
munity. The issue here is how a whole community can be shown to be in
error when the objective reality against which the error is judged is mere
community consensus. Platonism has an obvious solution to this issue, one
involving acausal abstract objects, but since CS does not countenance those,
it will have to work hard for one.
None of this demonstrates that CS is incoherent. Rather, it is intended as
a signpost toward research which could usefully be done to bolster CS, as well
as to undermine some of what is currently done in several key branches of
39
anti-naturalistic philosophy, including much of the philosophy of mathemat-
ics. But, in any case, until CS offers a compelling solution to the problems
just identified, it cannot be accepted as an adequate account of the nature of
semantics. CS, much like the other two approaches discussed above, is not
(in its current form) an acceptable theory of meaning for cognitive science.26
Conclusions
Each of the three projects outlined here is compatible with naturalism (as
defined at the outset), provided we are willing to make some important sac-
rifices. In the case of NSM, we need to acknowledge that semantic posits are
ontologically irreducible; they are emergent phenomena, brute facts about
the universe. Few linguists or philosophers are willing to pay this price at
the moment, though I suspect that this option would become increasingly ap-
pealing if no reductive explanation of semantics proved capable of delivering
an adequate theory. For its part, FS wheels out the mathematical machinery
of sets and functions to lay siege to meaning. The posits of mathematics are
themselves taken as ontologically basic and hence (yet again) as not standing
in need of any further explanation. Most philosophers today seem to find this
plausible (Shapiro (2004), Burgess and Rosen (1997)). A minority, whose ar-
guments I find compelling resist (Field (1989), Maddy (1990)). And, in any
case, there are other problems. (Pietroski 2003a) has argued, convincingly I
think, that a theory of understanding for natural languages must come apart
from a theory of truth. To the extent that FS supposes that a theory of
understanding just is a theory of truth, it runs into trouble. Finally, as we
saw, Jackendoff (2002) offers a seemingly plausible theory of understanding
for natural languages. CS is a reductive account that makes a serious effort
26Let me reiterate that the ideas expressed in the above section were developed jointly
with Sam Scott.
40
to integrate insights from phonology, syntax, semantics, psycholinguistics,
neuroscience as well as philosophy. To buy into CS however, we seem to
need to give up our best account of the nature of logic and mathematics.
Most philosophers would find this price prohibitive. It seems then that there
are good reasons for rejecting all three of the views discussed here.
Let me close by placing my own chips on the board. If I were a betting
man, I’d lay odds on CS. I think that CS will, in the long run, subsume
NSM by showing how what is best in that approach—the semantic analy-
ses—are to be cashed out in generative terms. I suspect moreover that CS
will eventually arrive at an adequate theory of the nature of logical and
mathematical posits, one which is broadly psychologistic yet not vulnerable
to Frege’s and Nagel’s arguments. The key there, I think, is to accept the
anti-psychologism arguments as defining criteria on any adequate theory of
logical and mathematical understanding. With enough attention to detail,
there are probably explanations to be found for how our mathematical judge-
ments achieve objectivity, universality, fallibility as well as a host of other
interesting features not discussed here. Suppose for a second that that’s pos-
sible. Prof. Pietroski has suggested that a theory of human understanding
ought to take the form of a Tarskian truth theory (even if it need not, in
fact, be a theory of truth). That proposal could, it would seem, be combined
with Jackendoff’s CS to yield a hybrid semantic theory. In other words, if a
compromise-solution to the question of the nature of mathematical objects
can be found, there need not exist any deep incompatibility between formal
semantics and conceptualist semantics. My bet is on that sort of seman-
tic theory ultimately furnishing cognitive science with a proper account of
understanding for natural languages.
41
References
Arnauld, A. and Nicole, P. (1662/1996). Logic, or, the art of thinking.
Austin, J. (1962). How to Do Things with Words. Oxford University Press.
Barsalou, L. W. (1999). Perceptual symbol systems. Behavioral and Brain
Sciences, (22):577–609.
Benacerraf, P. (1973). Mathematical truth. In Benacerraf, P. and Putnam,
H., editors, Philosophy of Mathematics: Selected Readings, pages 403–420.
Cambridge University Press.
Biederman, I. (1995). Visual object recognition. In Kosslyn, S. M. and
Osherson, D. N., editors, Visual Cognition, An Invitation to Cognitive
Science, chapter 4, pages 121–166. MIT Press, 2 edition.
Brooks, R. A. (1991). Intelligence without representation. Artificial Intelli-
gence, 47:139–159.
Burgess, J. P. and Rosen, G. (1997). A Subject with No Object: Strategies
for Nominalistic Interpretations of Mathematics. Oxford University Press.
Carnie, A. (1993). Syntax: A Generative Introduction. Blackwell Publishers.
Chalmers, D. J. (1996). The Conscious Mind: In Search of a Fundamental
Theory. Oxford University Press.
Chomsky, N. (2000). New Horizons in the Study of Language and Mind.
Cambridge University Press.
Dartnall, T. (2000). Reverse psychologism, cognition and content. Minds
and Machines, (10):31–52.
42
Davidson, D. (1967). Truth and meaning. In Martinich, A. P., editor, The
Philosophy of Language, chapter 6, pages 79–90. Oxford University Press.
de Saussure, F. (1911/1966). Course in General Linguistics. McGraw-Hill.
Dennett, D. C. (1991). Consciousness Explained. Back Bay Books.
Dennett, D. C. (1995). Darwin’s Dangerous Idea: Evolution and the Mean-
ings of Life. Simon & Schuster.
Field, H. (1972). Tarski’s theory of truth. In Lynch, M. P., editor, The
Nature of Truth, chapter 16, pages 365–396. MIT Press.
Field, H. (1989). Realism, Mathematics and Modality. Oxford University
Press.
Fodor, J. (1975). The Language of Thought. Harvard University Press.
Fodor, J. (1987). Psychosemantics: The Problem of Meaning in the Philos-
ophy of Mind. MIT Press.
Fodor, J. and Lepore, E. (1992). Holism: A Shopper’s Guide. Blackwell
Publishers.
Frege, G. (1953). The Foundations of Arithmetic, volume Original work
published 1884. Blackwell.
Goddard, C. (1998). Bad arguments against semantic primitives. Theoretical
Linguistics, 24(2-3):129–156.
Godel, K. (1947). What is cantor’s continuum problem? In Benacerraf, P.
and Putnam, H., editors, Philosophy of Mathematics: Selected Readings.
Cambridge University Press, 2nd edition.
43
Goldstone, R. L. and Rogosky, B. J. (2002). Using relations within conceptual
systems to translate across conceptual systems. Cognition, (84):295–320.
Haack, S. (1978). Philosophy of Logics. Cambridge University Press.
Hall-Partee, B. (1979). Semantics - mathematics of psychology? In Baverle,
Egli, and von Stechow, editors, Semantics from Different Points of View.
Springer.
Hegel, G. W. F. (1998/1812). Science of Logic. Humanities Press Interna-
tional.
Heim, I. and Kratzer, A. (1998). Semantics in Generative Grammar. Black-
well Publishers.
Jackendoff, R. (1983). Semantics and Cognition. MIT Press.
Jackendoff, R. (1987). Consciousness and the Computational Mind. Bradford
MIT Press.
Jackendoff, R. (2002). Foundations of Language: Brain, Meaning, Grammar,
Evolution. Oxford University Press.
Jackendoff, R. (2003). Precis of the foundations of language. Behavioral and
Brain Sciences, 26(6).
Jerzykiewicz, L. and Scott, S. (2003). Psychologism and conceptual seman-
tics: Review of Foundations of Language. Behavoral and Brain Sciences,
26(6):682–683.
Kitcher, P. (1983). The Nature of Mathematical Knowledge. Oxford Univer-
sity Press.
Kusch, M. (1995). Psychologism: a case study in the sociology of philosophical
knowledge. Routledge.
44
Larson, R. and Segal, G. (1995). Knowledge of Meaning: An Introduction to
Semantic Theory. MIT Press.
Lewis, D. (1972). General semantics. In Davidson, D. and Harman, G.,
editors, Semantics of Natural Language, pages 169–216. Reidel Publishing
Company.
Maddy, P. (1990). Realism in Mathematics. Oxford University Press.
Montague, R. (1974). Formal Philosophy. New Haven, Yale University Press.
Nagel, T. (1997). The Last Word. Oxford University Press.
Ouhalla, J. (1994). Introducing Transformational Grammar: From Principles
and Parameters to Minimalism. Arnold Press, 2nd edition.
Pietroski, P. M. (2003a). The character of natural language semantics. In
Barber, A., editor, Epistemology of Language. Oxford University Press.
Pietroski, P. M. (2003b). Meaning before truth.
Pinker, S. (1989). Learnability and Cognition: The Acquisition of Argument
Structure. MIT Press.
Regier, T. (1996). The Human Semantic Potential: Spatial Language and
Constrained Connectionism. MIT Press.
Shapiro, S. (2004). Naturalism reconsidered. In Shapiro, S., editor, Oxford
Hanbook of the Philosophy of Mathematics and Logic. Oxford University
Press.
Talmy, L. (2001). Toward a Cognitive Semantics: Volume I: Concept Struc-
turing Systems. MIT Press.
45
Tarski, A. (1944). The semantic conception of truth and the foundations
of semantics. In Martinich, A. P., editor, The Philosophy of Language,
chapter 4, pages 48–71. Oxford University Press, 2 edition.
Wierzbicka, A. (1996). Semantics, Primes and Universals. Oxford University
Press.
Wittgenstein, L. (1953). Philosophical Investigations. Blackwell Publishers.
LATEX
46