UofA

Spring 2004

Ling/Phil 596D: Topics in Linguistics and Philosophy

Heidi Harley and Massimo Piattelli-Palmarini

Compositionality

 

Wednesday January 28

Handout 1 (M. Piattelli-Palmarini)

 

          Fodor and compositionality

 

1 Ð From Chapter 5 of ÒHume VariationsÓ

A ÒreductionistÓ metaphysics of meaning is one that can provide sufficient conditions for meaning in a vocabulary that is not itself either semantic or intentional. The EmpiricistsÕ resemblance and association is an instance, and so are causation, information (ˆ la Fred Dretske, and Fodor himself) and evolution (ˆ la Millikan and Dennett). Consensus is Òmost tenuousÓ in cognitive science on which metaphysics can really do the job of naturalizing meaning and representation.

            Associationism fails because itÕs unable Òto distinguish the intentional relations among the contents of thoughts, from the causal relations among the thoughts themselvesÓ (p. 115, emph. orig.) Take (as Hume does) probabilistic reasoning. Imagine a fair die with 3 sides, two of which have a triangle on them, and the third side has a circle. You roll it over and over. Suppose that, as a result, certain expectations do form in the mind. Here comes the big glitch, for the associationist: One thing is to explain why the thought Òit will come up triangleÓ will occur twice as frequently as the thought Òit will come up circleÓ. Quite another to explain how you can come to have the thought Òit will come up triangle and it will come up circle in the ratio two to oneÓ. The Òtwo to oneÓ bit has to be in the scope of the Òthink thatÓ bit. But associationism has no way to get this. This is a most elegant example of the need to distinguish between causal relations among mental states from relations (logical or whatever) among their intentional objects. Hume tries to overcome this big glitch by means of the imagination (ÒfancyÓ). FodorÕs Chapter 5 is devoted to explaining why this does not work. HumeÕs imagination is a blank check, a mystery. And Fodor strikes further heavy blows onto associationism, including (notably including) probabilistic associationism. You need Turing, rule following, innatism, and tacit knowledge, and syntacticity, to get out of the morass.

 

            The thought ÒhorseÓ followed (even invariably followed) by the thought ÒhoofÓ does not amount to the thought Òhoofed horseÓ (horse with a hoof). (Remember ANTI-ANTI-MISSILE from last week). Succession is not the right kind of glue to hold complex concepts together. Contingencies in thought do not copy contingencies in experience. ÒAssociationists have spent literally centuries in the fruitless search for a way out of this.Ó (p. 117)

            The long and the short of it all is that Òthe least that the mind must be able to represent is the content of its experience together with whatever higher-order and relational properties of its experience determine the character of the associations it forms.Ó (p.129. Emph. orig.) (For a cogent defense of this thesis even in the domain of animal behavior and un-classical conditioning, see the work of Charles R. Gallistel over many years)

 

2 Ð Chapter 6, toward compositionality

To repeat: ÒCognitive processes are constituted by causal interactions among mental representations, that is, among semantically evaluable mental particulars.Ó (p. 135). ItÕs this, or we are in total darkness. Three important, but distinct and independent, properties of propositional attitudes (PA): systematicity, productivity, compositionality. These are properties of the contents of the PAs.

            Fodor runs a new argument against a dispositional account of mental causation. In a nutshell, a disposition must manifest itself here and now, in order for it to be the explanation of a behavior here and now. An event must cause its manifestation. So dispositions cannot be sufficient causes. ÒDispositions manifest themselves only when something thatÕs not a disposition causes them to do so. ItÕs not sufficient for the vase to break that itÕs fragile; something has to happen that causes its fragility to cause it to break.Ó (p. 140. Emph. orig.) If we want (as we do) a ÒrobustÓ mental causation, we cannot have a dispositional account of mental states and mental processes. A mental link is sufficient to bring about its successor in the causal chain. ÒA thought is a bona fide mental particular and having a thought is a bona fide mental event.Õ (ibid). Some mental events are causally sufficient for others. There is mind/mind causation.

 

[A bookmark, here, for the following weeks.  FodorÕs long fight against the primacy of dispositions presently brings him in disagreement with Chomsky. Chomsky is adamant  in insisting that I-language is an internal structure of the speaker,  not any sort of ÒrelationÓ between the speaker  and something else external (or internal). Especially not between the speaker and his/her own (alleged) LOT.

 

HeidiÕs comment: In fact, it seems to me that the ultimate LF representation, assuming that we ever determine what it is, just IS the LOT; distinguishing the LOT and the I-language in this regard,  I suspect, will be impossible. (Unless LOT = (I-language + deictic-reference-resolution system) -- but most laypeople would doubtless want to include the deictic-reference resolution system as part of I-language,  although of course Chomsky wouldnÕt.

 

Even ChosmkyÕs own expression Òknowledge  of  languageÓ needs  a caveat: It is not the case that there is the speakerÕs mind and, distinct from it, a ÒlanguageÓ which the speaker ÒknowsÓ.  All we have, according to Chomsky, is a complex internal (computational-derivational) structure of the speaker. This has consequences for compositionality as well. Chomsky finds it perplexing that Fodor may admit the compositionality of LOT, but not of natural languages.  Though, of course, ChomskyÕs account  is not at all ÒdispositionalÓ (in anything remotely resembling Wittgenstein and Ryle, or Travis), the notion of Òinternal structureÓ without a representationally separate LOT  is different from FodorÕs unrepentant representational theory.  Fodor, however, gladly admits that many representations are unconscious and that there is a huge heterogeneity in the kinds of (innate) mental representations we have. Notice too that in Minimalism, there are no representations any more. All is derivational (we will go back to this next week, or the week after next).]

 

The decomposition of complex concepts into Òsyntactic constituentsÓ is what differentiates them one from the other (notably those that have co-extensive applications). Their canonical decomposition is intrinsic, not contingent

 

HeidiÕs comment: a point rather like what I was making on the first day: linguistic (I-language)) expressions just are structured; the very idea that they might not be is an incoherent one which I suspect we can blame on the extreme incompleteness of their written representation, and to a certain extent their spoken representation. There is no expression of I-language without constituent structure.

 

Concepts are Òsemantically evaluable, causally active, mental particularsÓ (p.144). They are Modes of Presentation (MOPs) Òonly, psychologizedÓ. The same extension can be presented to the mind in lots of different ways (ChomskyÕs term is Òpoints of viewÓ).

            Rule-following versus Òbeing in accord withÓ a rule. How can we trace the difference? Fodor (rightly, in my view) dismisses awareness as a criterion (contra, for instance, SearleÕs position, whom he does not cite). The key is causal power. MOPs distinguish the causal power of otherwise extensionally identical concepts. How things (or rules) are represented in the mind is what decides.

 ÒHard wiredÓ rules may be extensionally equivalent to bona fide rules, but they are intentionally different. (Intentional, with a ÔtÕ, subsumes intensional, with an ÔsÕ).

 

HeidiÕs comment: Is this reiterating that language is an automatically triggered computational device, not a set of rules (understood as ÔinstructionsÕ or similar) followed by a homunculus who represents them independently?

 

            Again, a cogent defense of atomism (see FodorÕs previous work). Data on language acquisition make it very unlikely that the child has to acquire PHYSICAL OBJECT or PARENT before she acquires MAMA. These data, together with further arguments against making inferential power constitutive of meaning (and against the holism that inevitably ensues) strongly recommend to espouse atomism.

 

3 - On compositionality proper

(The very last 5-6 pages of the book). Systematicity and productivity are explained by compositionality. Nay, only compositionality can explain them. Either thought (LOT, Mentalese) or language have to be compositional. Or both. If you can show that one of them isnÕt, then you have shown ipso facto that the other is. Well, then, if you show that English is not compositional, then you have shown that Mentalese is, and you have shown that TOI (or RTM) is true.

            Argument 1: When you learn English, you learn how to translate from Mentalese to English and vice versa. You think in Mentalese but you communicate in English. English may well borrow systematicity, productivity and compositionality from Mentalese. Mentalese is compositional. Is English too? This is an empirical issue (see infra), and there are no knockdown arguments. But surely English does not Òlook to be frightfully compositionalÓ (p.153). Everyone agrees that natural languages are not compositional at surface form (see my first weekÕs handout). So, really the whole burden of proof falls onto Logical Form (LF). If you undermine the legitimacy of LF as being the semantic level, then you undermine the compositionality of natural languages.

            Argument 2: Ambiguous sentences are supposed to map onto two or more un-ambiguous  LFs. These, in turn, are supposed to map univocally onto un-ambiguous thoughts. Do LFs represent sentences, or do they rather represent thoughts? (The thoughts that the sentences are used to express). Mentalese Òpoints in two different directions: on the one hand, towards thought; on the other hand, towards languageÓ (p. 155). ThatÕs how (sometimes at least) you manage to say what you think. What grants that sentences are un-ambiguous at some level of representation (LF)?  Ò..There is no obvious reason why sentences  should be ambiguity-free at any level of representation; in other words, there is no obvious reason why there shouldnÕt be really ambiguous sentences. [É] Perhaps some sentences of L [natural languages] are ambiguous at every level of representation that the grammar of L recognizes, or perhaps none are. If we want to choose, we need an argumentÓ (ibid, emph orig). Not much (no really cogent argument) justifies the claim that there is a level of representation of sentences (LF) that is ambiguity-free.

 

HeidiÕs comment: Well, this is true but in an uninteresting way, seems to me. ItÕs true if you separate out the deixis-resolution device from L, which is probably not too contentious. An expression like he or tomorrow will be ambiguous all the way through to LF. What you can guarantee, however, is that once you anchor the reference of such expressions to the context, the meaning computed by an LF representation will be unambiguous...

 

            Argument 3: Thoughts cannot be ambiguous (see the case of Lincoln You can fool some of the peopleÉ). You do not open your mouth and close your eyes, and think ambiguous thoughts. ItÕs open whether there are ambiguous sentences, qua sentences, but not whether there are ambiguous thoughts. ÔThoughts are where the buck stops.Ó What could disambiguate them? The impossibility of entertaining ambiguous thoughts (for thoughts that have propositional content) does not depend on presupposing that LF represents them. Sentences are (sometimes at least) ambiguous. They become ambiguity-free only under the LF hypothesis. Now, thoughts have causal roles, but not sentences (nor propositions, being abstracta).  Therefore (out of a simple Occamian principle of minimizing the number of mysteries) letÕs conclude that LF represents thoughts, rather than sentences. English does not look terribly compositional because it isnÕt! (p.157) But, if it isnÕt, then Mentalese has to be.

            These arguments are presented rather tentatively. The crux is that whatever LF represents is ipso facto compositional.  Linguists, Fodor says (correctly), agree on that.

 

Coda. Notes from a conversation with Jerry (November 16, 2003)

JF: Take Phonological Form (PF.) That is not, again, at first sight (i.e. phonetically) compositional. You have all sorts of contiguity phenomena, vowel-shifts, absorptions, rounding etc. non-articulated phonemes etc. The canonical compositional analysis (features etc.) is compositional, but thatÕs a theoretical  construct,  a complex algorithm, not a property of surface pronunciation. [See HalleÕs ® morphemes]. A lot of gap-filling must go on at EVERY level. Why not at the semantic level too? Why should LF need no gap-filling and have EVERYTHING made explicit?

Basically, the semantics of natural languages is a chapter of a theory of communication. There is quite a lot that we must be able to compute anyway reconstructing each otherÕs communicative intentions, reconstructing each otherÕs THOUGHTS. We know how to do it, basically, in virtue of the fact that we are all members of the same species. We know how to fill the gaps (in the main, in a huge variety of situations)

Take ItÕs raining. (And all deictics in general). We obviously understand that it means ItÕs raining here & now. Why should we suppose that the specifications here and now are somehow tacitly present in the language, at some level of representation (allegedly, at LF)? Why not suppose that we fill those gaps in virtue of what we know about the situation, the speakerÕs intentions and so on, and the standard way of expressing the thought ITÕS RAINING HERE AND NOW in English? Something like GriceÕs conversational maxims.

 

MPP: In Italian, even the it is omitted. But we have every reason to believe that itÕs there. Not in virtue of GriceÕs maxims, or any knowledge of the world, but rather in virtue of our knowledge of language. (JF agrees). Why admit it for it, but not for here and now?

 

JF These are empirical questions, and I have no strong objection, but it NEEDS ARGUMENTS AND DATA, it cannot be just taken for granted. It cannot just be an assumption that L is compositional, and that there is a strictly compositional level of representation (LF) of L, such that everything is made explicit at that level. This is a very strong hypothesis, that forces one to hypothesize hugely complicated underlying structures, with a huge amount of silent components (functional categories of all sorts, projections of all sorts) And he has an inherent problem with that? just because the theory imposes that Ls are strictly compositional. A theory that imposes so many diversified and complex posits ought to reconsider the basic assumption that make these inevitable, i.e. the assumption of compositionality. 

 

HeidiÕs comment: Basically, it seems to me that the primary reason that the generative program results in productive research is because it has assumed compositionality for L. Linguists see an ambiguity or an underspecification (Mary saw the man with the telescope; John wants to go), assume that there must be something there that disambiguates or specifies it (constituent structure in the first case, big PRO + the theory of control in the second), and then go out and discover it, with corroborating evidence, predictive power, and so on. Why on earth should we give up this extremely productive line of inquiry on the off chance that L isnÕt compositional? Jeepers! Seems soundly unscientific.

 

MPP Take ellipsis, something like John believes that Mary is smart but Bill doesnÕt.

A syntactic treatment of ellipsis is surely a possibility (several interesting ideas are on the market). This gap-filling does not depend on conversational maxims, on guessing communicative intentions, or the like. ItÕs allowed (sometimes imposed) by our knowledge of language, of syntax. Punkt.

 

JF Yes, but these are cases of deletion under identity. ItÕs quite understandable that you may succeed in developing an algorithmic account of deletion under identity. But take all the other ubiquitous cases of ellipsis (in a more general sense), where there is no identity. They require,  quite obviously, a gap-filling based on other stuff (presuppositions, understanding of communicative intentions, etc.) and these are non-algorithmic. It may well be the case that bona fide semantics, the individuation of thoughts from surface sentences of L, turns out to be algorithmic, but I doubt it. A lot more is required that is very probably non-algorithmic.

 

MPP In ÒHume VariationsÓ (HV), as well as in many other papers and books, you pay homage to Turing, and say that a healthy injection of Turing-computationalism, bringing to an end associationism and Wittgensteinianism (the ÒuseÓ theory of meaning), is arguably the best thing that ever happened to cognitive science.  But, if so much is non-algorithmic, then the role of Turing-computationalism becomes rather minor.

 

JF But thought IS Turing-computable. We manage to reconstruct the underlying thoughts and then thoughts are causally related one to the other in (roughly speaking) syntax-sensitive computational ways, ˆ la Turing. The fundamental kind of causality is not associations, but Turing-computability. THAT was the crucial move.

 

On propositions

 

MPP In a footnote in HV you say that the (hypothetical) end-product of the algorithm, the entry into LOT (what really corresponds to the thought that the sentence is uttered to convey) cannot be a proposition, because propositions are abstract  entities and, as such, cannot have any causal power. But thoughts do have causal powers, so it cannot be propositions. Many years ago, in your paper on propositional attitudes, you also said that if propositions are somehow to be brought into the picture, they must be something that possesses a syntactic structure. (JF confirms). Well, then, how can it be that the Òend productÓ must possess some syntactic structure,  and LF does possess a syntactic structure,  but LF is not the ÔentryÓ into the corresponding thought?

 

JF The very existence of LF needs arguments and data. There is hardly any doubt that there are thoughts, and that they are compositional (the systematicity, productivity and fine-graininess of thought guarantee that). Why not have just thoughts? Why add LF?

 

MPP Is this a kind of OckamÕs razor  consideration?

 

JF Well, something like that, but not just that. Why not just have an algorithm (syntax) that is compositional and operates on surface English, computing sentence types out of sentence tokens, and a bunch of communicative operations (Gricean maxims, knowledge of the situation, intuitions about the speakerÕs intentions etc.) that fill the gaps, and indeed reconstruct  the thought that it is the speakerÕs intention to convey? We really do not need anything else, unless there are convincing arguments and empirical evidence that we must have something else (an LF level of representation that is compositional). I (JF) see none being offered that is convincing.

 

HeidiÕs comment: Does he (JF) argree that the wh-movement is interpreted via an operator-variable construction? Further, does he agree that there are wh-in-situ languages? Do such languages have a different algorithm for mapping questions to thoughts than wh-movement languages do (i.e. not an operator-variable structure?) Seems to me that without LF the learnability problem rears its ugly head again. How can a child reconstruct the algorithms that map the structures of his/er language to LOT if those algorithms are (wildly) different from language to language? If they are not (wildly) different from language to language, but rather parametrically different in constrained ways, then I submit that whatever the algorithms accomplish is equivalent to what LF is intended to (and is) accomplishing. That is, LF is a concrete proposal for how the mapping to LOT could happen. Until  thereÕs a convincing alternative that does as well as LF does, and doesnÕt open the PandoraÕs box of learnability considerations, I plan to stick with LF as the only concrete proposal that comes close to doing the job (perhaps supplemented by a set of context-mapping principles that fix deictic reference).

 

On ambiguous sentences

 

MPP: LetÕs take the famous sentence (Chomsky, then Higginbotham) I almost had my wallet stolen.  ItÕs instantly recognized as ambiguous between  two meanings, but there is a third one that emerges, slowly, under guidance or careful reflection. Now, itÕs a property of L (of IL) that this surface form can receive three distinct LFs, not a property of our communicational  skills. It really seems that something like LF (the computational apparatus that derives these forms) is needed,  in order to access the associated thoughts.

 

JF The apparatus computes types. This token sentence can be mapped onto three different types, this is algorithmic, but then something else is needed (some filling of the gaps, as usual) to decide WHICH thought the speaker who uttered that token actually intended to convey. ThatÕs where the buck stops.

 

HeidiÕs comment: Sure -- but that doesnÕt amount to saying that the surface form is underdetermined. The hearer can compute 3 LFs given the evidence of the written sentence; deciding which one is correct is the same as the problem of deciding what with the telescope is intended to modify in She saw the man with the telescope. Surely, surely that problem doesnÕt call into question the compositionality of the 3 LFs that could match the surface string. The string canÕt mean just anything!! And the speaker didnÕt have an ambiguity; the speaker had one of the three LFs in mind on utterance. The problem for the hearer is the same as deciding which lexical item to identify when the string /bQNk/ is uttered. Does the existence of that problem mean that language isnÕt compositional? What about the homophony of kills (3sgV), kills (plN), killÕs (genitiveN)? Are we not justified in positing homophony for -s here, or does it all come down to Gricean maxims, hearer-driven heuristics that the speaker knows the hearer will use? This all seems very odd to me...

 

Modes of presentation (MOPS)

 

JF Think about the acquisition of the lexicon. It requires a theory of mind, understanding other peopleÕs intentions, understanding the relevant aspects of the situation etc. Quite a lot of gap-filling. It is causally dependent on what I have called MOPS (Òmodes of presentationÓ). Why should sentences be any different? There are sentential MOPS too. Ways of representing situations and communicative intentions and so forth. Maybe what we really need are MOPS, rather than LFs.

 

MPP Well, the classic experiments by Lila Gleitman show that syntactic frames are also essential for acquiring lexical meanings (X is gorping Y, versus X and Y are gorping).

 

JF The data are not so clear as they appeared initially. Lila is putting strong brakes on those conclusions, these days. She is seriously reconsidering that whole story. But, look, it may well be. I am not excluding that it may be the case,  though, it seems to me, better arguments and better data are needed to affirm that syntactic frames are so frequently so essential for lexical acquisition.

 

Further considerations (not discussed with JF yet)

 

ÒCompositionality is, par excellence,  a property of representationsÓ (HV, p. 157, emphasis in the original). Sure, but itÕs also a property of derivations (i.e. of operators), and in minimalism, these days, there are no representations (arguably), only derivations. And these are compositional all the way,  up to and including LF, but not just at LF. Merge is, by its nature, compositional. And so are ÒphasesÓ (modulo a better specification of what they are), and so are the derivations by phases. It really sounds a bit extravagant (in need of arguments and evidence,  exactly as JF says, but in the opposite direction) to assert that:

(1)                                                                                                                                                           Thoughts (sentences in the LOT) are compositional

(2)                                                                                                                                                           Syntax is compositional Syntactic derivations are (to put it mildly) relevant to the individuation of the corresponding thought

BUT

(3)                                                                                                                                                           The interface between NS and LOT (something like LF) is not itself compositional.

This amounts, in my (MPP) opinion, to the insertion of a huge puzzle where none needs to be.

 

Heidi: Yes -- or regressing to ÔI donÕt like the LF solution to this problem, so letÕs throw it out and wait for something better to come alongÕ....

 

In HV, as we saw last week. Jerry says that composite concepts are not ÒjustÓ decomposable into basic components, but decomposable into canonical components. The composite concept A BOY AND A GIRL has A BOY and has A GIRL as canonical components, but not ÒAND AÓ as a component (surely not as a canonical component). This is quite Ok, but it sounds to me as syntactically based. A BOY is a DP, and so is A GIRL, but ÒAND AÓ is not a syntactic constituent.  There seems to be a clear correspondence between syntactic constituents and conceptual (i.e. semantic) constituents. If L is not compositional, how can that be? This is, I (MPP) think, an old puzzle in FodorÕs theory. In his book on concepts, Jerry says that BROWN COW is a composite concept (unlike BROWN and COW) because its is composed of two mono-morphemic units, i.e. words (sic). How any such consideration (with which I fully agree) can be offered while maintaining that natural languages are not compositional, I fail to see.

To be continued.