Semantic Properties Of The Mental Representations example essay topic
It can be characterized as the conjunction of the following three theses (A), (B) and (C): (A) Representational Theory of Mind (RTM): (cf. Field 1978: 37, Fodor 1987: 17) (1) Representational Theory of Thought: For each propositional attitude A, there is a unique and distinct (i.e. dedicated) [1] psychological relation R, and for all propositions P and subjects S, S As that P if and only if there is a mental representation #P# such that (a) S bears R to #P#, and (b) #P# means that P. (2) Representational Theory of Thinking: Mental processes, thinking in particular, consists of causal sequences of of mental representations. (B) Mental representations, which, as per (A 1), constitute the direct 'objects' of propositional attitudes, belong to a representational or symbolic system which is such that (cf. Fodor and Pylyshyn 1988: 12-3) (1) representations of the system have a combinatorial syntax and semantics: structurally complex (molecular) representations are systematically built up out of structurally simple (atomic) constituents, and the semantic content of a molecular representation is a function of the semantic content of its atomic constituents together with its syntactic / formal structure, and (2) the operations on representations (constituting, as per (A 2), the domain of mental processes, thinking) are causally sensitive to the syntactic / formal structure of representations defined by this combinatorial syntax.
(C) Functionalist Materialism. Mental representations so characterized are, at some suitable level, functionally entities that are realized by the physical properties of the subject having propositional attitudes (if the subject is an organism, then the realizing properties are presumably the neuro physiological properties in the brain or the central nervous system of the organism). The relation R in (A 1), when RTM is combined with (B), is meant to be understood as a computational / functional relation. The idea is that each attitude is identified with a characteristic computational / functional role played by the mental sentence that is the direct 'object' of that kind of attitude. (Scare quotes are necessary because it is more appropriate to reserve 'object' for a proposition as we have done above, but as long as we keep this in mind, it is harmless to use it in this way for LOT sentences.) For instance, what makes a certain mental sentence an (occur rent) belief might be that it is characteristically the output of perceptual output systems and input to an inferential system that interacts decision-theoretically with desires to produce further sentences or actions.
Or equivalently, we may think of belief sentences as those that are accessible only to certain sorts of computational operations appropriate for beliefs, but not to others. Similarly, desire-sentences (and sentences for other attitudes) may be characterized by a different set of operations that define a characteristic computational role for them. In the literature it is customary to use the metaphor of a 'belief-box' (cf. Schiffer 1981) as a blanket term to cover whatever specific computational role belief sentences turn out to have in the mental economy of their possessors. (Similarly for 'desire-box', etc.) The Language of Thought Hypothesis is so-called because of (B): token mental representations are like sentences in a language in that they have a syntactically and semantically regimented constituent structure. Put differently, mental representations that are the direct 'objects' of attitudes are structurally complex symbols whose complexity lends itself to a syntactic and semantic analysis.
This is also why the LOT is sometimes called Mentalese. It is (B 2) that makes LOTH a species of the so-called Computational Theory of Mind (CTM). This is why LOTH is sometimes called the Computational / Representational Theory of Mind or Thought (RTM / C RTT) (cf. Rey 1991, 1997). Indeed, LOTH seems to be the most natural product when RTM is combined with a view that would treat mental processes or thinking as computational when computation is understood traditionally or classically (this is a recent term emphasizing the contrast with connection ist processing, which we will discuss later).
When someone believes that P, there is a trivial sense in which the immediate 'object' of her belief, what she believes, can be said to be a complex symbol, according to LOTH, a sentence in her LOT physically realized in the neurophysiology of her brain, that has both syntactic structure and a semantic content, namely the proposition that P. So, contrary to the orthodox view that takes the belief relation as a dyadic relation between an agent and a proposition, LOTH takes it to be a triadic relation among an agent, a Mentalese sentence, and a proposition. The Mentalese sentence can then be said to have the proposition as its semantic / intentional content. It is only in this (perhaps indirect) sense can it be said that what is believed is a proposition, and thus the object of the attitude in more common philosophical parlance. This triadic view seems to have an advantage over the orthodox dyadic view in that it is a puzzle in the dyadic view how what are thought to be purely physical organisms can stand in direct relation to abstract objects like propositions in such a way as to influence their causal powers.
According to folk psychology, it is because those states have the propositional content they do that they have the causal powers they do. LOTH makes this relatively non-mysterious by introducing a physical intermediary that is capable of having the relevant causal powers in virtue of its syntactic structure that encodes its semantic content. Another advantage of this is that the thought processes can be causally guided by the syntactic forms of the sentences in a way that respect their semantic contents. This is the virtue of (B) to which we " ll come back below. Mainly because of these features, LOTH is said to scientifically vindicate folk psychology if it turns out to be true.
Status of LOTH LOTH has primarily been advanced as an empirical thesis (although some have argued for the truth of LOTH on a priori or conceptual grounds, given the natural conceptual contours of folk psychology -- see Davies 1989, 1991; Lycan 1993; Rey 1995). It is not meant to be taken as an analysis of what the folk mean (or, for that matter, what the scientists ought to mean) when they talk about various propositional attitudes and their role in thinking. In this regard, LOT theorists typically view themselves as engaged in some sort of a proto-science, or at least in some empirical research program continuous with scientific psychology or more generally with empirical inquiry. Indeed, as we will see in more detail below, when Jerry Fodor first explicitly articulated and elaborated LOTH in some considerable detail in his (1975), he basically defended it on the ground that it was assumed by our best scientific theories or models in cognitive psychology and psycholinguistics. This empirical status accorded to LOTH should be kept firmly in mind when assessing its plausibility and especially its prospects in the light of new evidence and developments in scientific psychology. When viewed this way, LOTH is not, strictly speaking, committed to preserving the folk taxonomy of the mental states in any very exact way.
Notions like belief, desire, hope, fear, etc. are folk notions and, as such, it may not be utterly plausible to expect (arguments aside) that a scientific psychology will preserve the exact contours of these concepts. On the contrary, there is every reason to believe that scientific counterparts of these notions will carve the mental space somewhat differently. For instance, it has been noted that the folk notion of belief harbors many distinctions. It is noted for example that it has both a disposition al and an occur rent sense. In the occur rent sense, it seems to mean something like consciously entertaining and accepting a thought (proposition) as true.
There is quite a bit of literature and controversy on the disposition al sense. [2] Beliefs are also capable of being explicitly stored in long term memory as opposed to being merely disposition al or tacit. Compare, for instance: I believe that there was a big surprise party for my 24th birthday vs. I have always believed that lions don't eat their food with forks and knives, or that 13652/4 = 3413, even though until now these latter two thoughts had never occurred to me. There is furthermore the issue of degree of belief: while I may believe that George will come to dinner with his new girlfriend even though I wouldn't bet on it, you, thinking that you know him better than I do, may nevertheless go to the wall for it.
It is unlikely that there will be one single construct of scientific psychology that will exactly correspond to the folk notion of belief in all these ways. For LOTH to vindicate folk psychology it is sufficient that a scientific psychology with a LOT architecture come up with scientifically grounded psychological states that are recognizably like the propositional attitudes of folk psychology, and that play more or less similar roles in psychological explanations. [3] Scope of LOTH LOTH is an hypothesis about the nature of thought and thinking with propositional content. As such, it may or may not be applicable to other aspects of mental life. Officially, it is silent about the nature of some mental phenomena such as experience, qualia, [4] sensory processes, mental images, visual and auditory imagination, sensory memory, perceptual pattern-recognition capacities, dreaming, hallucinating, etc. To be sure, many LOT theorists hold views about these aspects of mental life that make it seem that they are also to be explained by something similar to LOTH.
[5] For instance, Fodor seems to think that many modular input systems (Fodor 1983) have their own LOT to the extent to which they can be explained in representational and computational terms. Indeed, many contemporary psychological models treat perceptual input systems in just these terms. [6] There is indeed some evidence that this kind of treatment is appropriate for many perceptual processes. But it is to be kept in mind that a system may employ representations and be computational without necessarily satisfying any or both of the clauses in (B) above in any full-fledged way. Just think of finite automat a theory where there are plenty of examples of a computational process defined over states or symbols which lack full-blown syntactic and / or semantic structural complexity. Whether sensory or perceptual processes are to be treated within the framework of full-blown LOTH is again an open empirical question.
It may well be that the answer to this question is affirmative. If so, there may be more than one LOT realized in different subsystems or mechanisms in the mind / brain. So LOTH is not committed to there being a single representational system realized in the brain, nor is it committed to the claim that all mental representations are complex or language-like, nor would it be falsified if it turns out that most aspects of mental life other than the ones involving propositional attitudes don't require a LOT. Similarly, there is strong evidence that the mind also exploits an image-like representational medium for certain kinds of mental tasks. [7] LOTH is non-committal about the existence of an image-like representational system for many mental tasks other than the ones involving propositional attitudes. But it is committed to the claim that propositional thought and thinking cannot be successfully accounted for in its entirety in purely imagistic terms.
It claims that a combinatorial sentential syntax is necessary for propositional attitudes and a purely imagistic medium is not an adequate medium to capture that. [8] There are in fact some interesting and difficult issues surrounding these claims. The adequacy of an imagistic system seems to turn on the nature of syntax at the sentential level. For instance, Fodor, in Chapter 4 of his (1975) book, allows that many lexical items in one's LOT may be image-like; he introduces the notion of a mental image / picture under description to avoid some obvious inadequacies of pictures (e. g., what makes a picture a picture of a fat woman rather than a pregnant one, or vice versa, etc. ). This is an attempt to combine discursive and imagistic representational elements at the lexical level.
There may even be a well defined sense in which pictures can be combined to produce structurally complex pictures (as in British Empiricism: image-like simple ideas are combined to produce complex ideas, e. g., the idea of a unicorn) But what is absolutely essential for LOTH, and what Fodor insists on, is the claim that there is no adequate way in which a purely image-like system can capture what is involved in making judgments, i. e., in judging propositions to be true. This seems to require a discursive syntactic approach at the sentential level. The general problem here is the inadequacy of pictures or image-like representations to express propositions. I can judge that the blue box is on top of the red one without judging that the red box is under the blue one. I can judge that Mary kisses John without judging that John kisses Mary, and so on for indefinitely many such cases, concrete as well as abstract.
It is hard to see how images or pictures can do that without using any syntactic structure or discursive elements, to say nothing of judging, e. g., conditionals, disjunctive or negative propositions, quantification's, negative existential's, etc. [9] Moreover, there are difficulties with imagistic representations arising from demands on processing representations. As we will see below, (B 2) turns out to provide the foundations for one of the most important arguments for LOTH: it makes it possible to mechanize thinking understood as a semantically coherent thought process, which, as per (A 2), consists of a causal sequence of of mental representations. It is not clear, however, how an equivalent of (B 2) could be provided for images or pictures in order to accommodate operations defined over them, even if something like an equivalent of (B 1) could be given. On the other hand, there are truly promising attempts to integrate discursive symbolic theorem-proving with reasoning with image-like symbols. They achieve impressive efficiency in theorem-proving or in any deductive process defined over the expressions of such an integrated system.
Such attempts, if they prove to be generalizable to psychological theorizing, are by no means threats to LOTH; on the contrary, such systems have every features to make them a species of a LOT system: they satisfy (B). [10] Nativism and LOTH In the book (1975) in which Fodor introduced the LOTH, he also argued that all concepts are innate. As a result, the connection between LOTH and an implausibly strong version of conceptual nativism looked very much internal. This historical coincidence has led some people to think that LOTH is essentially committed to a very strong version of nativism, so strong in fact that it seems to make a reductio of itself (see, for instance, P.S. Church land 1986, Putnam 1988, Clark 1994). The gist of his argument was that since learning concepts is a form of hypothesis formation and confirmation, it requires a system of mental representations in which formation and confirmation of hypotheses are to be carried out, but then there is a non-trivial sense in which one already has (albeit potentially) the resources to express the extension of the concepts to be learned. However, it should be emphasized that LOTH is not committed to such a strong version of nativism, especially about concepts.
It is certainly plausible to assume that LOTH will turn out to have some empirically (as well as theoretically / a priori) motivated nativist commitments about the structural organization and dynamic management of the entire representational system. But this much is to be expected especially in the light of recent empirical findings and trends. This, however, does not constitutes a reductio. It is an open empirical question how much nativism is true about concepts, and LOTH should be so taken as to be capable of accommodating whatever turns out to be true in this matter. LOTH, therefore, when properly conceived, is independent of any specific proposal about conceptual nativism. [11] Naturalism and LOTH One of the most attractive features of LOTH is that it is a central component of an ongoing research program in philosophy of psychology to naturalize the mind, to give a theoretical framework in which the mind could naturally be seen as part of the physical world without postulating irreducibly psychic entities, events, processes or properties.
Fodor, the most ardent defender of LOTH, once identified the major mysteries in philosophy of mind thus: How could anything material have conscious states? How could anything material have semantical properties? How could anything material be rational? (where this means something like: how could the state transitions of a physical system preserve semantical properties? ).
(1991: 285, Reply to Devitt) LOTH is a full-blown attempt to give a naturalist answer to the third question, an attempt to solve at least part of the problem underlying the second one, and is almost completely silent about the first. [12] According to RTM, propositional attitudes are relations to meaningful mental representations whose constitute the domain of thinking. This much can, in principle, be granted by an intentional realist who would nevertheless reject LOTH. Indeed, there are plenty of theorists who accept RTM in some suitable form (and also happily accept (C) in many cases) but reject LOTH either by explicitly rejecting (B) or simply by remaining neutral about it. Among some of the prominent people who choose the former option are Searle (1984, 1990, 1992), Stal naker (1984), Lewis (1972), Bar wise and Perry (1983). [13] Some who opt for the latter include Loan (1982 a, 1982 b), Dretske (1981); Armstrong (1980), and many contemporary functionalists.
[14] But RTM per se doesn't so much propose a naturalistic solution to intentionality and mechanization of thinking as simply assert a framework to emphasize intentional realism and, perhaps, with (C), a declaration of a commitment to naturalism or physicalism at best. How, then, is the addition of (B) supposed to help? Let us first try to see in a bit more detail what the problem is supposed to be in the first place to which (B) is proposed as a solution. So let us start by reflecting on thinking and see what it is about thinking that makes it a mystery in Fodor's list. This will give rise to one of the most powerful (albeit still non demonstrative) arguments for LOTH. The Problem of Thinking RTM's second clause (A 2), in effect, says that thinking is at least the of states that are (a) intentional (i.e. have representational / propositional content) and (b) causally connected.
But, surely, thinking is more. There could be a causally connected series of intentional states that makes no sense at all. Thinking, therefore, is causally proceeding from states to states that would make semantic sense: the transitions among states must preserve some of their semantic properties to count as thinking. In the ideal case, this property would be the truth value of the states.
But in most cases, any interesting intentional property like warranted ness, degree of confirmation, semantic coherence given a certain practical context like satisfaction of goals in a specific context, etc. would do. In general, it is hard to spell out what this requirement of 'making sense' comes to. The intuitive idea, however, should be clear. Thinking is not proceeding from thoughts to thoughts in arbitrary fashion: thoughts that are causally connected are in some fashion semantically connected too. If this were not so, there would be little point and gain in thinking. Thinking couldn't serve any useful purpose.
Call this general phenomenon, then, the semantic coherence of causally connected thought processes. LOTH is offered as a solution to this puzzle: how is thinking, conceived this way, physically possible? This is the problem of thinking, thus the problem of mechanization of rationality in Fodor's version. How does LOTH propose to solve this problem and bring us one big step closer to the naturalization of the mind? Syntactic Engine Driving a Semantic Engine: Computation The two most important achievements of 20th century that are at the foundations of LOTH as well as most of modern Artificial Intelligence (AI) research and the so-called information processing approaches to cognition (practically almost all of contemporary cognitive psychology) are (i) the developments in modern symbolic (formal) logic, and (ii) Alan Turing's idea of a Turing Machine and Turing computability. It is putting these two ideas together that gives LOTH its enormous explanatory power within a naturalistic framework.
Modern logic showed that most of deductive reasoning can be formalized, i.e. most semantic relations among symbols can be entirely captured by the symbols' formal / syntactic properties and the relations among them. And Turing showed, roughly, that if a process has a formally specifiable character then it can be mechanized. So we can appreciate the implications of (i) and (ii) for the philosophy of psychology in this way: if thinking consists in processing representations physically realized in the brain (in the way the internal data structures are realized in a computer) and these representations form a formal system, i.e. a language with its proper combinatorial syntax (and semantics) and a set of derivations rules formally defined over the syntactic features of those representations (allowing for specific but extremely powerful programs to be written in terms of them), then the problem of thinking, as I described it above, can in principle be solved in completely naturalistic terms, thus the mystery surrounding how a physical device can ever have semantically coherent state transitions (processes) can be removed. Thus, given the commitment to naturalism, the hypothesis that the brain is a kind of computer trafficking in representations in virtue of their syntactic properties is the basic idea of LOTH (and the AI vision of cognition). Computers are environments in which symbols are manipulated in virtue of their formal features, but what is thus preserved are their semantic properties, hence the semantic coherence of symbolic processes.
Slightly paraphrasing Haugeland (cf. 1985: 106), who puts the same point nicely in the form of a motto: THE FORMALIST MOTTO: If you take care of the syntax of a representational system, its semantics will take care of itself. This is in virtue of the mimicry or mirroring relation between the semantic and formal properties of symbols. As Dennett once put it in describing LOTH, we can view the thinking brain as a syntactically driven engine preserving semantic properties of its processes, i.e. driving a semantic engine. What is so nice about this picture is that if LOTH is true we have a naturalistically adequate causal treatment of thinking that respect the semantic properties of the thoughts involved: it is in virtue of the physically coded syntactic / formal features that thoughts cause each other while the coherence of their semantic properties is preserved precisely in virtue of this. Whether or not LOTH actually turns out to be empirically true in the details or in its entire vision of rational thinking, this picture of a syntactic engine driving a semantic one can at least be taken to be an important philosophical demonstration of how Descartes' challenge can be met (cf.
Rey 1997: chp. 8). Descartes claimed that rationality in the sense of having the power 'to act in all the contingencies of life in the way in which our reason makes us act' cannot possibly be possessed by a purely physical device: 'The rational soul... could not be in any way extracted from the power of matter... but must... be expressly created' (1637/1970: 117-18). Descartes was completely puzzled by just this rational character and semantic coherence of thought processes so much so that he failed to even imagine a possible mechanistic explication of it. He thus was forced to appeal to Divine creation. But we can now see / imagine at least a possible mechanistic / naturalistic scenario.
[15] Intentionality and LOTH But where do the semantic properties of the mental representations come from in the first place? How can they mean anything? This is Brentano's challenge to a naturalist. Brentano's bafflement was with the intentionality of the human mind, its apparently mysterious power to represent things, events, properties in the world.
He thought that nothing physical can have this property of intentionality: 'The reference to something as an object is a distinguishing characteristic of all mental phenomena. No physical phenomenon exhibits anything similar' (Brentano 1874/1973: 97). This problem of intentionality is the second problem or mystery in Fodor's list that I quoted above. I said that LOTH officially offers only a partial solution to it and perhaps proposes a framework within which the remainder of the solution can be couched and elaborated in a naturalistically acceptable way.
As characterized at the beginning, RTM contains a clause (A 1 b) that says that the immediate object of a propositional attitude that P is a mental representation #P# that means that P. Again, (B 1) attributes a compositional semantics to the syntactically complex symbols belonging to one's LOT that are, as per (C), physically realized in the brain of a thinking organism. According to LOTH, the semantic content of propositional attitudes is inherited from the semantic content of the mental symbols. So Brentano's questions for a LOT theorist becomes: how do the symbols in one's LOT get their meanings in the first place? There are two levels or stages at which this question can be raised and answered: (1) At the level of atomic (simple) symbols: how do the atomic symbols represent what they do? (2) At the level of molecular (phrasal complexes or sentences) symbols: how do molecular symbols represent what they do? There have been at least two major lines LOT theorists have taken regarding these questions.
The one that is least committal might perhaps be usefully described as the official position regarding LOTH's treatment of intentionality. Most LOT theorists seem to have taken this line. The official line doesn't propose any theory about the first stage, but simply assumes that the first question can be answered in a naturalistically acceptable way. In other words, officially LOTH simply assumes that the atomic symbols / expressions in one's LOT have whatever meanings they have. [16] But, the official line continues, LOTH has a lot to say about the second stage, the stage where the semantic contents are computed or assigned to complex (molecular) symbols on the basis of their combinatorial syntax or grammar together with whatever meanings atomic symbols are assumed to have in the first stage. This procedure is familiar from a Tarski-style [17] definition of truth conditions of sentences.
The truth-value of complex sentences in propositional logic are completely determined by the truth-values of the atomic sentences they contain together with the rules fixed by the truth-tables of the connectives occurring in the complex sentences. Example: 'P and Q' is true just in case both 'P' and 'Q' are true, but false otherwise. This process is similar but more complex in first-order languages, and even more so for natural languages -- in fact, we don't have a completely working compositional semantics for the latter at the moment. So, if we have a semantic interpretation of atomic symbols (if we have symbols whose reference and extension are fixed at the first stage by whatever naturalistic mechanism turns out to govern it), then the combinatorial syntax will take over and effectively determine the semantic interpretation (truth-conditions) of the complex sentences they are constituents of.
So officially LOTH would only contribute to a complete naturalization project if there is a naturalistic story at the atomic level. Early Fodor (1975, 1978, 1978 a, 1980), for instance, envisaged a science of psychology which, among other things, would reasonably set for itself the goal of discovering the combinatorial syntactic principles of LOT and the computational rules governing its operations, without worrying much about semantic matters, especially about how to fix the semantics of atomic symbols (he probably thought that this was not a job for LOTH). Similarly, Field (1978) is very explicit about the combinatorial rules for assigning truth-conditions to the sentences of the internal code. In fact, Field's major argument for LOTH is that, given a naturalistic causal theory of reference for atomic symbols, about which he is optimistic (Field 1972), it is the only naturalistic theory that has a chance of solving Brentano's puzzle.
For the moment, this is not much more than a hope, but, according to the LOT theorist, it is a well-founded hope based on a number of theoretical and empirical assumptions and data. Furthermore, it is a framework defining a naturalistic research program in which there have been promising successes. [18] As I said, this official and, in a way, least committal line has been overall the more standard way of conceiving LOTH's role in the project of naturalizing intentionality. But some have gone beyond it and explored the ways in which the resources of LOTH can be exploited even in answering the first question (1) about the semantics of atomic symbols.
Now, there is a weak version of an answer to (1) on the part of LOTH and a strong version. On the weak version, LOTH may be un tendentiously viewed as inevitably providing some of the resources in giving the ultimate naturalistic theory in naturalizing the meaning of atomic symbols. The basic idea is that whatever the ultimate naturalistic theory turns out to be true about atomic expressions, computation as conceived by LOTH will be part of it. For instance, it may be that, as with no mic co variation theories of meaning (Fodor 1987, 1990 a; Dretske 1981), the meaning of an atomic predicate may consist in its potential to get token ed in the presence of (or, in causal response to) something that instantiates the property the predicate is said to express. A natural way of explicating this potential may partly but ultimately rely on certain computational principles the symbol may be subjected to within a LOT framework, or principles that in some sense govern the 'behavior' of the symbol.
Insofar as computation is naturalistically understood in the way LOTH proposes, a complete answer to the first question about the semantics of atomic symbols may plausibly involve an explicatory appeal to computation within a system of symbols. This is the weak version because it doesn't see LOTH as proposing a complete solution to the first question (1) above, but only helping it. A strong version would have it that LOTH provides a complete naturalistic solution to both questions: given the resources of LOTH we don't need to look any further to meet Brentano's challenge. The basic idea lies in so-called functional or conceptual role semantics, according to which a concept is the concept it is precisely in virtue of the particular causal / functional potential it has in interacting with other concepts. Each concept may be thought of as having a certain distinctive set of epistemic / semantic relations or liaisons to other concepts. We can conceive of this set as determining a certain 'conceptual role' for each concept.
We can then take these roles to determine the semantic identity of concepts: concepts are the concepts they are because they have the conceptual roles they have; that is to say, among other things, concepts represent whatever they do precisely in virtue of these roles. The idea then is to reduce each conceptual role to causal / functional role of atomic symbols (now conceived as primitive terms in LOTH), and then use the resources of LOTH to reduce it in turn to computational role. Since computation is naturalistically well-defined, the argument goes, and since causal interactions between thoughts and concepts can be understood completely in terms of computation, we can completely naturalize intentionality if we can successfully treat meanings as arising out of thoughts / concepts ' internal interactions with each other. In other words, the strong version of LOTH would claim that atomic symbols in LOT have the content they do in virtue of their potential for causal interactions with other tokens, and cashing out this potential in mechanical / naturalistic terms is what, among other things, LOTH is for.
LOTH then comes as a naturalistic rescuer for conceptual role semantics. It is not clear whether any one holds this strong version of LOTH in this rather naive form. But certainly some people have elaborated the basic idea in quite subtle ways, for which Cummins (1989: chp. 8) is perhaps the best example. (But also see Block 1986 and Field 1978.) But even in the best hands, the proposal turns out to be very problematic and full of difficulties nobody seems to know how to straighten out. In fact, some of the most ardent critics of taking LOTH as incorporating a functional role semantics turn out to be some of the most ardent defenders of LOTH understood in a weak, non-committal sense we have explored above -- see Fodor (1987: chp.
3), Fodor and Lepore (1991), Fodor's attack (1978 b) on AI's way of doing procedural semantics is also relevant here. Haugeland (1981), Searle (1980, 1984), and Putnam (1988) quite explicitly take LOTH to involve a program for providing a complete semantic account of mental symbols, which they then attack accordingly. [19] As indicated previously, LOTH is almost completely silent about consciousness and the problem of qualia, the third mystery in Fodor's list in the quote above. But the naturalist's hope is that this problem too will be solved, if not by LOTH, then by something else.
On the other hand, it is important to emphasize that LOTH is neutral about the of consciousness / qualia. If it turns out that qualia cannot be naturalized, this would by no means show that LOTH is false or defective in some way. In fact, there are people who seem to think that LOTH may well turn out to be true even though qualia can perhaps not be naturalized (e. g., Block 1980, Chalmers 1996, McGinn 1991). Finally, it should be emphasized that LOTH has no particular commitment to every symbolic activity's being conscious. Conscious thoughts and thinking may be the tip of a computational iceberg. Nevertheless, there are ways in which LOTH can be helpful for an account of state consciousness that seeks to explain a thought's being conscious in terms of a higher order thought which is about the first order thought.
So, to the extent to which thought and thinking are conscious, to that extent LOTH can perhaps be viewed as providing some of the necessary resources for a naturalistic account of state consciousness -- for elaboration see Rosenthal (1997) and Lycan (1997). Arguments for LOTH We have already seen two major arguments, perhaps the historically most important ones, for LOTH: First, we have noted that if LOTH is true then all the essential features of the common sense conception of propositional attitudes will be explicated in a naturalistic framework which is likely to be co-opted by scientific cognitive psychology, thus vindicating folk psychology. Second, we discussed that, if true, LOTH would solve one of the mysteries about thinking minds: how is thinking (as characterized above) possible? How is rationality mechanically possible? Then we have also seen a third argument that LOTH would partially contribute to the project of naturalizing intentionality by offering an account of how the semantic properties of whole attitudes are fixed on the basis of their atomic constituents. But there have been many other arguments for LOTH.
In this section, I will try to describe only those arguments that have been historically more influential and controversial. Argument from Contemporary Cognitive Psychology When Fodor first formulated LOTH with significant elaboration in his (1975), he introduced his major argument for it along with its initial formulation in the first chapter. It was basically this: our best scientific theories and models of different aspects of higher cognition assume a framework that requires a computational / representational medium for them to be true. More specifically, he analyzes the basic form of the information processing models developed to account for three types of cognitive phenomena: perception as the fixation of perceptual beliefs, concept learning as hypothesis formation and confirmation, and decision making as a form of representing and evaluating the consequences of possible actions carried out by the agent in a situation with a p reordered set of preferences. He rightly points out that all these models treat mental processes as computational processes defined over representations. Then he draws what seems to be the obvious conclusion: if these models are right in at least treating mental processes as computational, even if not in detail, then there must be a LOT over which they are defined, hence LOTH.
In Fodor's (1975), the arguments for different aspects of LOTH are diffused and the emphasis, with the book's slogan 'no computation without representation', is put on the RTM rather than on (B) or (C). But all the elements are surely there. Argument from the Productivity of Thought People seem to be capable of entertaining an infinite number of thoughts, at least in principle, although they in fact entertain only a finite number of them. Indeed adults who speak a natural language are capable of understanding sentences they have never heard uttered before. Here is one: there is a big lake of melted gold on the dark side of the moon. I bet that you never heard this sentence before, and yet, you have no difficulty in understanding it: it is one you " re in fact likely to believe false.
But this sentence was arbitrary, there are infinitely many such sentences I can in principle utter and you can in principle understand. But understanding a sentence is to entertain the thought / proposition it expresses. So there are in principle infinitely many thoughts you are capable of entertaining. This is sometimes expressed by saying that we have an unbounded competence in entertaining different thoughts, even though we have a bounded performance. But this unbounded capacity is to be achieved by finite means. For instance, storing an infinite number of representations in our heads is out of the question: we are finite beings.
If human cognitive capacities (capacities to entertain an unbounded number of thoughts, or to have attitudes towards an unbounded number of propositions) are productive in this sense, how is this to be explained on the basis of resources? The explanation LOTH offers is straightforward: postulate a representational system that satisfies at least (B 1). Indeed, recursion is the only known way to produce an infinite number of symbols from a finite base. In fact, given LOTH, productivity of thought as a competence mechanism seems to be guaranteed. [20] Argument from the Systematicity and Compositional ity of Thought Systematicity of thought consists in the empirical fact that the ability to entertain certain thoughts is intrinsically connected to the ability to entertain certain others.
Which ones? Thoughts that are related in a certain way. In what way? There is a certain initial difficulty in answering such questions. I think, partly because of this, Fodor (1987) and Fodor and Pylyshyn (1988), who are the original defenders of this kind of argument, first argue for the of language production and understanding: the ability to produce / understand certain sentences is intrinsically connected to the ability to produce / understand certain others. Given that a mature speaker is able to produce / understand a certain sentence in her native language, by psychological law, there always appear to be a cluster of other sentences that she is able to produce / understand.
For instance, you don't seem to find speakers who know how to express in their native language the fact that John loves the girl but not the fact that the girl loves John. This is apparently so, moreover, for expressions of any n-place relation. Fodor and Pylyshyn bring out the force of this psychological fact by comparing learning languages the way we actually do with learning a language by memorizing a huge phrase book. In the phrase book model, there is nothing to prevent someone learning how to say 'John loves the girl' without learning how to say 'the girl loves John. ' In fact, that is exactly the way some information booklets prepared for tourists help them to cope with their new social environment. You might, for example, learn from a phrase book how to say 'I'd like to have a cup of coffee with sugar and milk' in Turkish without knowing how to say / understand absolutely anything else in Turkish.
In other words, the phrase book model of learning a language allows arbitrarily punctate linguistic capabilities. In contrast, a speaker's knowledge of her native language is not punctate, it is systematic. Accordingly, you do not find, by nosological necessity, native speakers whose linguistic capacities are punctate. Now, how is this empirical truth (in fact, a law-like generalization) to be explained?
Obviously if this is a general nosological fact, then learning one's native language cannot be modeled on the phrase book model. What is the alternative? The alternative is well known. Native speakers master the grammar and vocabulary of their language.
But this is just to say that sentences are not atomic, but have syntactic constituent structure. If you have a vocabulary, the grammar tells you how to combine systematically the words into sentences. Hence, in this way, if you know how to construct a particular sentence out of certain words, you automatically know how to construct many others. If you view all sentences as atomic, then, as Fodor and Pylyshyn say, the of language production / understanding is a mystery, but if you acknowledge that sentences have syntactic constituent structure, of linguistic capacities is what you automatically get; it is guaranteed. This is the orthodox explanation of linguistic. From here, according to Fodor and Pylyshyn, establishing the of thought as a nosological fact is one step away.
If it is a law that the ability to understand a sentence is systematically connected to the ability to understand many others, then it is similarly a law that the ability to think a thought is systematically connected to the ability to think many others. For to understand a sentence is just to think the thought / proposition it expresses. Since, according to RTM, to think a certain thought is just to token a representation in the head that expresses the relevant proposition, the ability to token certain representations is systematically connected to the ability to token certain others. But then, this fact needs an adequate explanation too. The classical explanation LOTH offers is to postulate a system of representations with combinatorial syntax exactly as in the case of the explanation of the linguistic. This is what (B 1) offers.
[21] This seems to be the only explanation that does not make the of thought a miracle, and thus argues for the LOT hypothesis. However, thought is not only systematic but also compositional: systematically connected thoughts are also always semantically related in such a way that the thoughts so related seem to be composed out of the same semantic elements. For instance, the ability to think 'John loves the girl' is connected to the ability to think 'the girl loves John' but not to, say, 'protons are made up of quarks' or to '2+2 = 4. ' Why is this so? The answer LOTH gives is to postulate a combinatorial semantics in addition to a combinatorial syntax, where an atomic constituent of a mental sentence makes (approximately) the same semantic contribution to any complex mental expression in which it occurs. This is what Fodor and Pylyshyn call 'the principle of '.
[22] In brief, it is an argument for LOTH that it offers a cogent and principled solution to the and of cognitive capacities by postulating a system of representations that has a combinatorial syntax and semantics, i. e., a system of representations that satisfies at least (B 1). Argument from the Systematicity of Thinking (Inferential Coherence) Systematicity of thought does not seem to be restricted solely to the systematic ability to entertain certain thoughts. If the system of mental representations does have a combinatorial syntax, then there is a set of rules, syntactic formation rules, so to speak, that govern the construction of well-formed expressions in the system. It is this fact, (B 1), that guarantees that if you can form a mental sentence on the basis of certain rules, then you can also form many others on the basis of the same rules.
The rules of combinatorial syntax determine the syntactic or formal structure of complex mental representations. This is the formative (or, formational) aspect of. But inferential thought processes seem to be systematic too: the ability to make certain inferences is intrinsically connected to the ability to make certain many others. For instance, you do not find minds that can infer 'A' from 'A&B' but cannot infer 'C' from 'A&B&C. ' It seems to be a psychological fact that inferential capacities come in clusters that are homogeneous in certain aspects. How is this fact (i. e., the inferential or transformational) to be explained? As we have seen, the explanation LOTH offers depends on the exploitation of the notion of logical form or syntactic structure determined by the combinatorial syntax postulated for the representational system.
The combinatorial syntax not only gives us a criterion of well-formed ness for mental expressions, but it also defines the logical form or syntactic structure for each well-formed expression. The classical solution to inferential is to make the mental operations on representations sensitive to their form or structure, i.e. to insist on (B 2). Since, from a syntactic view point, similarly formed expressions will have similar forms, it is possible to define a single operation which will apply to only certain expressions that have a certain form, say, only to conjunctions, or disjunctions. This allows the LOT theorist to give homogeneous explanations of what appear to be homogeneous classes of inferential capacities. This is one of the greatest virtues of LOTH, hence provides an argument for it.
The solution LOTH offers for what I called the problem of thinking, above, is connected to the argument here because the two phenomena are connected in a deep way. Thinking requires that the logic o-semantic properties of a particular thought process (say, inferring that John is happy from knowing that if John is at the beach then John is happy and coming to realize that John is indeed at the beach) be somehow causally implicated in the process. The of inferential thought processes then is based on the observation that if the agent is capable of making that particular inference, then she is capable of making many other somehow similarly organized inferences. But the idea of similar organization in this context seems to demand some sort of classification of thoughts independently of their particular content. But what can the basis of such a classification be? The only basis seems to be the logic o-syntactic properties of thoughts, their form.
Although it feels a little uneasy to talk about syntactic properties of thoughts common- understood, it seems that they are forced upon us by the very attempt to understand their semantic properties: how, for instance, could we explain the semantic content of the thought that if John is at the beach then he is happy without somehow appealing to its being a conditional? This is the point of contact between the two phenomena. Especially when the demands of naturalism are added to this picture, inferring a LOT ( = a representational system satisfying B) realized in the brain becomes almost irresistible. Indeed Rey (1995) doesn't resist and claims that, given the above observations, LOTH can be established on the basis of arguments that are not 'merely empirical'. I leave it to the reader to evaluate whether mere critical reflection on our concepts of thought and thinking could, all by itself, establish LOTH. [23] Objections to LOTH There have been numerous arguments against LOTH.
Some of them are directed more specifically against the Representational Theory of Mind (A), some against functionalist materialism (C). Here I will concentrate only on those arguments specifically targeting (B) -- the most controversial component of LOTH. Regress Arguments against LOTH These arguments rely on the explanations offered by LOTH defenders for certain aspects of natural languages. In particular, many LOT theorists advert to LOTH to explain (1) how natural languages are learned, (2) how natural languages are understood, or (3) how the utterances in such languages can be meaningful.
For instance, according to Fodor (1975), natural languages are learned by forming and confirming hypotheses about the translation of natural language sentences into Mentalese such as: 'Snow is white' is true in English if and only if P, where 'P' is a sentence in one's LOT. But to be able to do that, one needs a representational medium in which to form and confirm hypotheses. The LOT is such a medium. Again, natural languages are understood because, roughly, such an understanding consists in translating their sentences into one's Mentalese. Similarly, natural language utterances are meaningful in virtue of the meanings of corresponding Mentalese sentences. The basic complaint is that in each of these cases, either the explanations generate a regress because the same sort of explanations ought to be given for how the LOT is learned, understood or can be meaningful, or else they are gratuitous because if a successful explanation can be given for LOT that does not generate a regress then it could and ought to be given for the natural language phenomena without introducing a LOT (see, e.g. Blackburn 1984).
Fodor's response in (1975) is (1) that LOT is not learned, it's innate, (2) that it's understood in a different sense than the sense involved in natural language comprehension, and (3) that LOT sentences acquire their meanings not in virtue of another meaningful language but in a completely different way, perhaps by standing in some sort of causal relation to what they represent (see above) or by having certain computational profiles. For many who have a Wittgenstein ian bent, these replies are not likely to be very convincing. But here the issues tend to concern RTM rather than (B). Laurence and Margolis (1997) point out that the regress arguments depend on the assumption that LOTH is introduced only to explain (1) - (3). If it can be shown that there are lots of other empirical phenomena for which the LOTH provides good explanations, then the regress arguments fail because LOTH then would not be gratuitous. In fact, as we have seen above, there are plenty of such phenomena.
But still it is important to realize that the sort of explanations proposed for the understanding of one's LOT (computational use / activity of LOT sentences with certain meanings) and how LOT sentences can be meaningful (computational roles and / or no mic relations with the world) cannot be given for (1) - (3): it's unclear, for example, what it would be like to give a computational role and / or no mic relation account for the meanings of natural language utterances. Propositional Attitudes without Explicit Representations Dennett in his review of Fodor's (1975) has raised the following objection (cf. Fodor 1987: 21-3 for a similar discussion): In a recent conversation with the designer of a chess-playing program I heard the following criticism of a rival program: 'it thinks it should get its queen out early. ' This ascribes a propositional attitude to the program in a very useful and predictive way, for as the designer went on to say, one can usefully count on chasing that queen around the board. But for all the many levels of explicit representation to be found in that program, nowhere is anything roughly synonymous with 'I should get my queen out early' explicitly token ed. The level of analysis to which the designer's remark belongs describes features of the program that are, in an entirely innocent way, emergent properties of the computational processes that have 'engineering reality.
' I see no reason to believe that the relation between belief-talk and psychological talk will be any more direct. (Dennett 1981: 107) The objection, as Fodor (1987: 22) points out, isn't that the program has a disposition al, or potential, belief that it will get its queen out early. Rather, the program actually operates on this belief. There appear to be lots of other examples: e.g. in reasoning we pretty often follow certain inference rules like modus pone ns, disjunctive syllogism, etc. without necessarily explicitly representing them..