UofA
Spring 2004
Ling/Phil 596D: Topics in Linguistics
and Philosophy
Heidi Harley and Massimo
Piattelli-Palmarini
Wednesday January 28
Handout 1 (M. Piattelli-Palmarini)
Fodor and
compositionality
1 Ð From Chapter 5 of
ÒHume VariationsÓ
A ÒreductionistÓ metaphysics of
meaning is one that can provide sufficient conditions for meaning in a
vocabulary that is not itself either semantic or intentional. The EmpiricistsÕ
resemblance and association is an instance, and so are causation, information
(ˆ la Fred Dretske, and Fodor himself) and evolution (ˆ la Millikan and
Dennett). Consensus is Òmost tenuousÓ in cognitive science on which
metaphysics can really do the job of naturalizing meaning and representation.
Associationism
fails because itÕs unable Òto distinguish the intentional relations among
the contents of thoughts, from the causal relations among the thoughts
themselvesÓ (p. 115, emph. orig.) Take (as Hume does) probabilistic
reasoning. Imagine a fair die with 3 sides, two of which have a triangle on
them, and the third side has a circle. You roll it over and over. Suppose that,
as a result, certain expectations do form in the mind. Here comes the big
glitch, for the associationist: One thing is to explain why the thought Òit
will come up triangleÓ will occur twice as frequently as the thought Òit will
come up circleÓ. Quite another to explain how you can come to
have the thought Òit will come up triangle and it will come up circle in
the ratio two to oneÓ. The Òtwo to oneÓ bit has to be in the scope of
the Òthink thatÓ bit. But
associationism has no way to get this. This is a most elegant example of the
need to distinguish between causal relations among mental states from
relations (logical or whatever) among their intentional objects. Hume
tries to overcome this big glitch by means of the imagination (ÒfancyÓ).
FodorÕs Chapter 5 is devoted to explaining why this does not work. HumeÕs
imagination is a blank check, a mystery. And Fodor strikes further heavy blows
onto associationism, including (notably including) probabilistic
associationism. You need Turing, rule following, innatism, and tacit knowledge,
and syntacticity, to get out of the morass.
The
thought ÒhorseÓ followed (even invariably followed) by the thought ÒhoofÓ does
not amount to the thought Òhoofed horseÓ (horse with a hoof). (Remember
ANTI-ANTI-MISSILE from last week). Succession is not the right kind of glue to
hold complex concepts together. Contingencies in thought do not copy
contingencies in experience. ÒAssociationists have spent literally centuries in
the fruitless search for a way out of this.Ó (p. 117)
The
long and the short of it all is that Òthe least that the mind must be able to
represent is the content of its experience together with whatever higher-order
and relational properties of its experience determine the character of the
associations it forms.Ó (p.129. Emph. orig.) (For a cogent defense of this
thesis even in the domain of animal behavior and un-classical conditioning, see
the work of Charles R. Gallistel over many years)
2 Ð Chapter 6, toward
compositionality
To repeat: ÒCognitive processes
are constituted by causal interactions among mental representations, that is,
among semantically evaluable mental particulars.Ó (p. 135). ItÕs this, or we
are in total darkness. Three important, but distinct and independent,
properties of propositional attitudes (PA): systematicity, productivity,
compositionality. These are properties of the contents of the PAs.
Fodor
runs a new argument against a dispositional account of mental causation. In a
nutshell, a disposition must manifest itself here and now, in order for
it to be the explanation of a behavior here and now. An event must cause
its manifestation. So dispositions cannot be sufficient causes. ÒDispositions
manifest themselves only when something thatÕs not a disposition causes them to
do so. ItÕs not sufficient for the vase to break that itÕs fragile; something
has to happen that causes its fragility to cause it to break.Ó (p. 140.
Emph. orig.) If we want (as we do) a ÒrobustÓ mental causation, we cannot have
a dispositional account of mental states and mental processes. A mental link is
sufficient to bring about its successor in the causal chain. ÒA thought
is a bona fide mental particular and having a thought is a bona fide mental
event.Õ (ibid). Some mental events are causally sufficient for others. There is
mind/mind causation.
[A
bookmark, here, for the following weeks.
FodorÕs long fight against the primacy of dispositions presently brings
him in disagreement with Chomsky. Chomsky is adamant in insisting that I-language is an internal structure of the
speaker, not any sort of
ÒrelationÓ between the speaker and
something else external (or internal). Especially not between the speaker and
his/her own (alleged) LOT.
HeidiÕs comment: In fact, it
seems to me that the ultimate LF representation, assuming that we ever
determine what it is, just IS the LOT; distinguishing the LOT and the
I-language in this regard, I
suspect, will be impossible. (Unless LOT = (I-language +
deictic-reference-resolution system) -- but most laypeople would doubtless want
to include the deictic-reference resolution system as part of I-language, although of course Chomsky wouldnÕt.
Even ChosmkyÕs own expression
Òknowledge of languageÓ needs a caveat: It is not the case
that there is the speakerÕs mind and, distinct from it, a ÒlanguageÓ
which the speaker ÒknowsÓ. All we
have, according to Chomsky, is a complex internal (computational-derivational)
structure of the speaker. This has consequences for compositionality as
well. Chomsky finds it perplexing that Fodor may admit the compositionality of
LOT, but not of natural languages.
Though, of course, ChomskyÕs account is not at all ÒdispositionalÓ (in anything remotely
resembling Wittgenstein and Ryle, or Travis), the notion of Òinternal
structureÓ without a representationally separate LOT is different from FodorÕs unrepentant representational
theory. Fodor, however, gladly
admits that many representations are unconscious and that there is a huge
heterogeneity in the kinds of (innate) mental representations we have. Notice
too that in Minimalism, there are no representations any more. All is
derivational (we will go back to this next week, or the week after next).]
The decomposition of complex
concepts into Òsyntactic constituentsÓ is what differentiates them one
from the other (notably those that have co-extensive applications). Their
canonical decomposition is intrinsic, not contingent
HeidiÕs comment: a point rather
like what I was making on the first day: linguistic (I-language)) expressions
just are structured; the very
idea that they might not be is an incoherent one which I suspect we can blame
on the extreme incompleteness of their written representation, and to a certain
extent their spoken representation. There is no expression of I-language
without constituent structure.
Concepts are Òsemantically
evaluable, causally active, mental particularsÓ (p.144). They are Modes of
Presentation (MOPs) Òonly, psychologizedÓ. The same extension can be presented
to the mind in lots of different ways (ChomskyÕs term is Òpoints of viewÓ).
Rule-following
versus Òbeing in accord withÓ a rule. How can we trace the difference? Fodor
(rightly, in my view) dismisses awareness as a criterion (contra, for instance,
SearleÕs position, whom he does not cite). The key is causal power. MOPs
distinguish the causal power of otherwise extensionally identical concepts. How
things (or rules) are represented in the mind is what decides.
ÒHard wiredÓ rules may
be extensionally equivalent to bona fide rules, but they are
intentionally different. (Intentional, with a ÔtÕ, subsumes intensional, with
an ÔsÕ).
HeidiÕs comment: Is this
reiterating that language is an automatically triggered computational device,
not a set of rules (understood as ÔinstructionsÕ or similar) followed by a
homunculus who represents them independently?
Again, a cogent defense of atomism (see FodorÕs
previous work). Data on language acquisition make it very unlikely that the
child has to acquire PHYSICAL OBJECT or PARENT before she acquires MAMA. These
data, together with further arguments against making inferential power
constitutive of meaning (and against the holism that inevitably ensues)
strongly recommend to espouse atomism.
3 - On compositionality
proper
(The very
last 5-6 pages of the book). Systematicity and productivity are explained by
compositionality. Nay, only compositionality can explain them. Either
thought (LOT, Mentalese) or language have to be compositional. Or both.
If you can show that one of them isnÕt, then you have shown ipso facto
that the other is. Well, then, if you show that English is not
compositional, then you have shown that Mentalese is, and you have shown that
TOI (or RTM) is true.
Argument
1: When you learn English, you learn
how to translate from Mentalese to English and vice versa. You think in
Mentalese but you communicate in English. English may well borrow
systematicity, productivity and compositionality from Mentalese. Mentalese
is compositional. Is English too? This is an empirical issue (see
infra), and there are no knockdown arguments. But surely English does not Òlook
to be frightfully compositionalÓ (p.153). Everyone agrees that natural
languages are not compositional at surface form (see my first
weekÕs handout). So, really the whole burden of proof falls onto Logical Form
(LF). If you undermine the legitimacy of LF as being the semantic level,
then you undermine the compositionality of natural languages.
Argument
2: Ambiguous sentences are supposed
to map onto two or more un-ambiguous LFs. These, in turn, are supposed to map univocally onto
un-ambiguous thoughts. Do LFs represent sentences, or do they rather
represent thoughts? (The thoughts that the sentences are used to
express). Mentalese Òpoints in two different directions: on the one hand,
towards thought; on the other hand, towards languageÓ (p. 155). ThatÕs how
(sometimes at least) you manage to say what you think. What grants that
sentences are un-ambiguous at some level of representation (LF)? Ò..There is no obvious reason why
sentences should be ambiguity-free
at any level of representation; in other words, there is no obvious
reason why there shouldnÕt be really ambiguous sentences. [É] Perhaps
some sentences of L [natural languages] are ambiguous at every level of
representation that the grammar of L recognizes, or perhaps none are. If we
want to choose, we need an argumentÓ (ibid, emph orig). Not much (no really
cogent argument) justifies the claim that there is a level of representation of
sentences (LF) that is ambiguity-free.
HeidiÕs
comment: Well, this is true but in an uninteresting way, seems to me. ItÕs true
if you separate out the deixis-resolution device from L, which is probably not
too contentious. An expression like he or tomorrow will be ambiguous all the
way through to LF. What you can guarantee, however, is that once you anchor the
reference of such expressions to the context, the meaning computed by an LF
representation will be unambiguous...
Argument
3: Thoughts cannot be
ambiguous (see the case of Lincoln You can fool some of the peopleÉ). You do not open your mouth and close your eyes, and
think ambiguous thoughts. ItÕs open whether there are ambiguous sentences, qua
sentences, but not whether there are ambiguous thoughts. ÔThoughts are where
the buck stops.Ó What could disambiguate them? The impossibility
of entertaining ambiguous thoughts (for thoughts that have propositional
content) does not depend on presupposing that LF represents them.
Sentences are (sometimes at least) ambiguous. They become ambiguity-free
only under the LF hypothesis. Now, thoughts have causal roles,
but not sentences (nor propositions, being abstracta). Therefore (out of a simple Occamian
principle of minimizing the number of mysteries) letÕs conclude that LF
represents thoughts, rather than sentences. English does not look
terribly compositional because it isnÕt! (p.157) But, if it isnÕt, then
Mentalese has to be.
These arguments are presented rather tentatively. The
crux is that whatever LF represents is ipso facto compositional. Linguists, Fodor says (correctly), agree on that.
Coda.
Notes from a conversation with Jerry (November 16, 2003)
JF: Take Phonological Form (PF.) That is not, again,
at first sight (i.e. phonetically) compositional. You have all sorts of
contiguity phenomena, vowel-shifts, absorptions, rounding etc. non-articulated
phonemes etc. The canonical compositional analysis (features etc.) is
compositional, but thatÕs a theoretical
construct, a complex
algorithm, not a property of surface pronunciation. [See HalleÕs ® morphemes]. A lot of gap-filling must go on at
EVERY level. Why not at the semantic level too? Why should LF need no
gap-filling and have EVERYTHING made explicit?
Basically, the semantics of
natural languages is a chapter of a theory of communication. There is quite a
lot that we must be able to compute anyway reconstructing each otherÕs communicative intentions,
reconstructing each otherÕs THOUGHTS. We know how to do it, basically, in
virtue of the fact that we are all members of the same species. We know how to
fill the gaps (in the main, in a huge variety of situations)
Take ItÕs raining. (And all deictics in general). We obviously understand
that it means ItÕs raining here & now. Why should we suppose that the specifications here and now
are somehow tacitly present in the language, at some level of
representation (allegedly, at LF)? Why not suppose that we fill those gaps in
virtue of what we know about the situation, the speakerÕs intentions and so on,
and the standard way of expressing the thought ITÕS RAINING HERE AND NOW in
English? Something like GriceÕs conversational maxims.
MPP: In Italian, even the it is omitted. But we have every reason to believe that
itÕs there. Not in virtue of GriceÕs maxims, or any knowledge of the world, but
rather in virtue of our knowledge of language. (JF agrees). Why admit it for it, but not for here and now?
JF These are empirical questions, and I have no strong
objection, but it NEEDS ARGUMENTS AND DATA, it cannot be just taken for
granted. It cannot just be an assumption that L is compositional, and that
there is a strictly compositional level of representation (LF) of L,
such that everything is made explicit at that level. This is a very strong
hypothesis, that forces one to hypothesize hugely complicated underlying
structures, with a huge amount of silent components (functional categories of
all sorts, projections of all sorts) And he has an inherent problem with that? just
because the theory imposes that Ls are strictly compositional. A theory
that imposes so many diversified and complex posits ought to reconsider the
basic assumption that make these inevitable, i.e. the assumption of
compositionality.
HeidiÕs comment: Basically, it seems to me that the
primary reason that the generative program results in productive research is because
it has assumed compositionality for L. Linguists see an ambiguity or an
underspecification (Mary saw the man with the telescope; John wants to go), assume
that there must be something there that disambiguates or specifies it
(constituent structure in the first case, big PRO + the theory of control in
the second), and then go out and discover it, with corroborating
evidence, predictive power, and so on. Why on earth should we give up this
extremely productive line of inquiry on the off chance that L isnÕt
compositional? Jeepers! Seems soundly unscientific.
MPP Take ellipsis, something like John believes
that Mary is smart but Bill doesnÕt.
A syntactic treatment of ellipsis is surely a possibility (several interesting ideas are on the market). This gap-filling does not depend on conversational maxims, on guessing communicative intentions, or the like. ItÕs allowed (sometimes imposed) by our knowledge of language, of syntax. Punkt.
JF Yes, but these are cases of deletion under identity. ItÕs quite understandable that you may succeed in developing an algorithmic account of deletion under identity. But take all the other ubiquitous cases of ellipsis (in a more general sense), where there is no identity. They require, quite obviously, a gap-filling based on other stuff (presuppositions, understanding of communicative intentions, etc.) and these are non-algorithmic. It may well be the case that bona fide semantics, the individuation of thoughts from surface sentences of L, turns out to be algorithmic, but I doubt it. A lot more is required that is very probably non-algorithmic.
MPP In ÒHume VariationsÓ (HV), as well as in many other papers and books, you pay homage to Turing, and say that a healthy injection of Turing-computationalism, bringing to an end associationism and Wittgensteinianism (the ÒuseÓ theory of meaning), is arguably the best thing that ever happened to cognitive science. But, if so much is non-algorithmic, then the role of Turing-computationalism becomes rather minor.
JF But thought IS Turing-computable. We manage to reconstruct the underlying thoughts and then thoughts are causally related one to the other in (roughly speaking) syntax-sensitive computational ways, ˆ la Turing. The fundamental kind of causality is not associations, but Turing-computability. THAT was the crucial move.
MPP In a footnote in HV you say that the
(hypothetical) end-product of the algorithm, the entry into LOT (what really
corresponds to the thought that the sentence is uttered to convey) cannot be a
proposition, because propositions are abstract entities and, as such, cannot have any causal power. But thoughts
do have causal powers, so it cannot be propositions. Many years ago, in
your paper on propositional attitudes, you also said that if propositions are
somehow to be brought into the picture, they must be something that possesses a
syntactic structure. (JF confirms). Well, then, how can it be that the
Òend productÓ must possess some syntactic structure, and LF does possess a syntactic structure, but LF is not the ÔentryÓ into
the corresponding thought?
JF The very
existence of LF needs arguments and data. There is hardly any doubt that there
are thoughts, and that they are compositional (the systematicity, productivity
and fine-graininess of thought guarantee that). Why not have just
thoughts? Why add LF?
MPP Is this a
kind of OckamÕs razor
consideration?
JF Well,
something like that, but not just that. Why not just have an algorithm (syntax)
that is compositional and operates on surface English, computing sentence types
out of sentence tokens, and a bunch of communicative operations (Gricean
maxims, knowledge of the situation, intuitions about the speakerÕs intentions
etc.) that fill the gaps, and indeed reconstruct the thought that it is the speakerÕs intention to convey? We
really do not need anything else, unless there are convincing arguments and
empirical evidence that we must have something else (an LF level of
representation that is compositional). I (JF) see none being offered that is
convincing.
HeidiÕs comment: Does he (JF) argree that the wh-movement is interpreted via an operator-variable construction? Further, does he agree that there are wh-in-situ languages? Do such languages have a different algorithm for mapping questions to thoughts than wh-movement languages do (i.e. not an operator-variable structure?) Seems to me that without LF the learnability problem rears its ugly head again. How can a child reconstruct the algorithms that map the structures of his/er language to LOT if those algorithms are (wildly) different from language to language? If they are not (wildly) different from language to language, but rather parametrically different in constrained ways, then I submit that whatever the algorithms accomplish is equivalent to what LF is intended to (and is) accomplishing. That is, LF is a concrete proposal for how the mapping to LOT could happen. Until thereÕs a convincing alternative that does as well as LF does, and doesnÕt open the PandoraÕs box of learnability considerations, I plan to stick with LF as the only concrete proposal that comes close to doing the job (perhaps supplemented by a set of context-mapping principles that fix deictic reference).
On ambiguous sentences
MPP: LetÕs take
the famous sentence (Chomsky, then Higginbotham) I almost had my wallet
stolen. ItÕs instantly recognized as ambiguous between two meanings, but there is a third one
that emerges, slowly, under guidance or careful reflection. Now, itÕs a
property of L (of IL) that this surface form can receive
three distinct LFs, not a property of our communicational skills. It really seems that something
like LF (the computational apparatus that derives these forms) is needed, in order to access the associated
thoughts.
JF The
apparatus computes types. This token sentence can be mapped onto three
different types, this is algorithmic, but then something else is needed (some
filling of the gaps, as usual) to decide WHICH thought the speaker who uttered
that token actually intended to convey. ThatÕs where the buck stops.
HeidiÕs comment: Sure -- but that doesnÕt amount to
saying that the surface form is underdetermined. The hearer can compute 3 LFs
given the evidence of the written sentence; deciding which one is correct is
the same as the problem of deciding what with the telescope is intended to modify in She saw the man with the
telescope. Surely, surely that
problem doesnÕt call into question the compositionality of the 3 LFs that could
match the surface string. The string canÕt mean just anything!! And the speaker didnÕt have an ambiguity; the
speaker had one of the three LFs in mind on utterance. The problem for the
hearer is the same as deciding which lexical item to identify when the string /bQNk/ is uttered. Does the
existence of that problem
mean that language isnÕt compositional? What about the homophony of kills (3sgV), kills (plN), killÕs (genitiveN)? Are we not
justified in positing homophony for -s here, or does it all come down to
Gricean maxims, hearer-driven heuristics that the speaker knows the hearer will
use? This all seems very odd to me...
JF Think about the acquisition of the lexicon. It requires a theory of mind, understanding other peopleÕs intentions, understanding the relevant aspects of the situation etc. Quite a lot of gap-filling. It is causally dependent on what I have called MOPS (Òmodes of presentationÓ). Why should sentences be any different? There are sentential MOPS too. Ways of representing situations and communicative intentions and so forth. Maybe what we really need are MOPS, rather than LFs.
MPP Well, the classic experiments by Lila Gleitman show that syntactic frames are also essential for acquiring lexical meanings (X is gorping Y, versus X and Y are gorping).
JF The data are not so clear as they appeared initially. Lila is putting strong brakes on those conclusions, these days. She is seriously reconsidering that whole story. But, look, it may well be. I am not excluding that it may be the case, though, it seems to me, better arguments and better data are needed to affirm that syntactic frames are so frequently so essential for lexical acquisition.
Further considerations (not discussed with JF yet)
ÒCompositionality
is, par excellence, a property of representationsÓ (HV, p. 157, emphasis in the original). Sure, but
itÕs also a property of derivations
(i.e. of operators), and in minimalism, these days, there are no representations
(arguably), only derivations. And these are compositional all the way, up to and including LF, but not just at
LF. Merge is, by its nature, compositional. And so are ÒphasesÓ (modulo a
better specification of what they are), and so are the derivations by phases.
It really sounds a bit extravagant (in need of arguments and evidence, exactly as JF says, but in the opposite
direction) to assert that:
(1)
Thoughts (sentences in
the LOT) are compositional
(2)
Syntax is compositional
Syntactic derivations are (to put it mildly) relevant to the individuation of
the corresponding thought
BUT
(3)
The interface between NS
and LOT (something like LF) is not itself compositional.
This amounts,
in my (MPP) opinion, to the insertion of a huge puzzle where none needs to be.
Heidi: Yes --
or regressing to ÔI donÕt like the LF solution to this problem, so letÕs throw
it out and wait for something better to come alongÕ....
In HV, as we
saw last week. Jerry says that composite concepts are not ÒjustÓ decomposable
into basic components, but decomposable into canonical components. The
composite concept A BOY AND A GIRL has A BOY and has A GIRL as canonical
components, but not ÒAND AÓ as a component (surely not as a canonical
component). This is quite Ok, but it sounds to me as syntactically
based. A BOY is a DP, and so is A GIRL, but ÒAND AÓ is not a syntactic
constituent. There seems to be a
clear correspondence between syntactic constituents and conceptual (i.e.
semantic) constituents. If L is not compositional, how can that be? This is, I
(MPP) think, an old puzzle in FodorÕs theory. In his book on concepts, Jerry
says that BROWN COW is a composite concept (unlike BROWN and COW) because its
is composed of two mono-morphemic units, i.e. words (sic). How any such
consideration (with which I fully agree) can be offered while
maintaining that natural languages are not compositional, I fail to see.
To be
continued.