January 21, 2010

-page 75-

American psychologist B. F. Skinner became famous for his pioneering research on learning and behavior. During his 60-year career, Skinner discovered important principles of operant conditioning, a type of learning that involves reinforcement and punishment. A strict behaviorist, Skinner believed that operant conditioning could explain even the most complex of human behaviors.


Many American psychologists were quick to adopt behaviorism, and animal laboratories were set up all over the country. Aiming to predict and control behavior, the behaviorists strategy was to vary a stimulus in the environment and observe an organisms response. They saw no need to speculate about mental processes inside the head. For example, Watson argued that thinking was simply talking within one’ self silently. He believed that thinking could be studied by recording the movement of certain muscles in the throat.

American psychologist B. F. Skinner designed an apparatus, now called a Skinner box, that allowed him to formulate important principles of animal learning. An animal placed inside the box is rewarded with a small bit of food each time it makes the desired response, such as pressing a lever or pecking a key. A device outside the box records the animals responses.

The most forceful leader of behaviorism was B. F. Skinner, an American psychologist who began studying animal learning in the 1930s. Skinner coined the term reinforcement and invented a new research apparatus called the Skinner box for use in testing animals. Based on his experiments with rats and pigeons, Skinner identified a number of basic principles of learning. He claimed that these principles explained not only the behavior of laboratory animals, but also accounted for how human beings learn new behaviors or change existing behaviors. He concluded that nearly all behavior is shaped by complex patterns of reinforcement in a persons environment, a process that he called operant conditioning (also referred to as instrumental conditioning). Skinners views on the causes of human behavior made him one of the most famous and controversial psychologists of the 20th century.

Operant conditioning, pioneered by American psychologist B. F. Skinner, is the process of shaping behavior by means of reinforcement and punishment. This illustration shows how a mouse can learn to maneuver through a maze. The mouse is rewarded with food when it reaches the first turn in the maze (A). Once the first behavior becomes ingrained, the mouse is not rewarded until it makes the second turn (B). After many times through the maze, the mouse must reach the end of the maze to receive its reward ©.

Skinner and others applied his findings to modify behavior in the workplace, the classroom, the clinic, and other settings. In World War II (1939-1945), for example, he worked for the U.S. government on a top-secret project in which he trained pigeons to guide an armed glider plane toward enemy ships. He also invented the first teaching machine, which allowed students to learn at their own pace by solving a series of problems and receiving immediate feedback. In his popular book Walden Two (1948), Skinner presented his vision of a behaviorist utopia, in which socially adaptive behaviors are maintained by rewards, or positive reinforcements. Throughout his career, Skinner held firm to his belief that psychologists should focus on the prediction and control of behavior.

Faced with a choice between psychoanalysis and behaviorism, many psychologists in the 1950s and 1960s sensed a void in psychologies conception of human nature. Freud had drawn attention to the darker forces of the unconscious, and Skinner was interested only in the effects of reinforcement on observable behavior. Humanistic psychology was born out of a desire to understand the conscious mind, free will, human dignity, and the capacity for self-reflection and growth. An alternative to psychoanalysis and behaviorism, humanistic psychology became known as the third force.

The humanistic movement was led by American psychologists Carl Rogers and Abraham Maslow. According to Rogers, all humans are born with a drive to achieve their full capacity and to behave in ways that are consistent with their true selves. Rogers, a psychotherapist, developed person-centered therapy, a nonjudgmental, nondirective approach that helped clients clarify their sense of whom they are in an effort to facilitate their own healing process. At about the same time, Maslow theorized that all people are motivated to fulfill a hierarchy of needs. At the bottom of the hierarchy are basic physiological needs, such as hunger, thirst, and sleep. Further up the hierarchy are needs for safety and security, needs for belonging and love, and esteem-related needs for status and achievement. Once these needs are met, Maslow believed, people strive for self-actualization, the ultimate state of personal fulfillment. Though, Maslow legitimately gave to a position: The musician must orchestrate within the harmonies of a symphony, an artist must paint, justly as a poet must poetize, if he is finally to be at peace with himself. What a man can be, he must be.

Swiss psychologist Jean Piaget based his early theories of intellectual development on his questioning and observation of his own children. From these and later studies, Piaget concluded that all children pass through a predictable series of cognitive stages.

From the 1920s through the 1960s, behaviorism dominated psychology in the United States. Eventually, however, psychologists began to move away from strict behaviorism. Many became increasingly interested in cognition, a term used to describe all the mental processes involved in acquiring, storing, and using knowledge. Such processes include perception, memory, thinking, problem solving, imagining, and language. This shift in emphasis toward cognition had such a profound influence on psychology that it has often been called the cognitive revolution. The psychological study of cognition became known as cognitive psychology.

One reason for psychologists renewed interest in mental processes was the invention of the computer, which provided an intriguing metaphor for the human mind. The hardware of the computer was likened to the brain, and computer programs provided a step-by-step model of how information from the environment is input, stored, and retrieved to produce a response. Based on the computer metaphor, psychologists began to formulate information-processing models of human thought and behavior.

In the 1950s American linguist Noam Chomsky proposed that the human brain is especially constructed to detect and reproduce language and that the ability to form and understand language is innate to all human beings. According to Chomsky, young children learn and apply grammatical rules and vocabulary as they are exposed to them and do not require initial formal teaching.

The pioneering work of Swiss psychologist Jean Piaget also inspired psychologists to study cognition. During the 1920s, while administering intelligence tests in schools, Piaget became interested in how children think. He designed various tasks and interview questions to reveal how children of different ages reason about time, nature, numbers, causality, morality, and other concepts. Based on his many studies, Piaget theorized that from infancy to adolescence, children advance through a predictable series of cognitive stages.

The cognitive revolution also gained momentum from developments in the study of language. Behaviorist B.F. Skinner had claimed that language is acquired according to the laws of operant conditioning, in much the same way that rats learn to press a bar for food pellets. In 1959, however, American linguist Noam Chomsky charged that Skinners account of language development was wrong. Chomsky noted that children all over the world start to speak at roughly the same age and proceed through roughly the same stages without being explicitly taught or rewarded for the effort. According to Chomsky, the human capacity for learning language is innate. He theorized that the human brain is hardwired for language as a product of evolution. By pointing to the primary importance of biological dispositions in the development of language, Chomskys theory dealt a serious blow to the behaviorist assumption that all human behaviors are formed and maintained by reinforcement.

Before psychology became established in science, it was popularly associated with extrasensory perception (ESP) and other paranormal phenomena (phenomena beyond the laws of science). Today, these topics lie outside the traditional scope of scientific psychology and fall within the domain of parapsychology. Psychologists note that thousands of studies have failed to demonstrate the existence of paranormal phenomena.

Grounded in the conviction that mind and behavior must be studied using statistical and scientific methods, but psychology has become a highly respected and socially useful discipline. Psychologists now study importantly and sensitive topics such as the similarities and differences between men and women, racial and ethnic diversity, sexual orientation, marriage and divorce, abortion, adoption, intelligence testing, sleep and sleep disorders, obesity and dieting, and the effects of psychoactive drugs such as methylphenidate (Ritalin) and fluoxetine (Prozac).

In the last few decades, researchers have made significant breakthroughs in understanding the brain, mental processes, and behavior. This section of the article provides examples of contemporary research in psychology: the plasticity of the brain and nervous system, the nature of consciousness, memory distortions, competence and rationality, genetic influences on behavior, infancy, the nature of intelligence, human motivation, prejudice and discrimination, the benefits of psychotherapy, and the psychological influences on the immune system.

Psychologists once believed that the neural circuits of the adult brain and nervous system were fully developed and no longer subject to change. Then in the 1980s and 1990s a series of provocative experiments showed that the adult brain has flexibility, or plasticity - a capacity to change as a result of usage and experience.

These experiments showed that adult rats flooded with visual stimulation formed new neural connections in the brains visual cortex, where visual signals are interpreted. Likewise, those trained to run an obstacle course formed new connections in the cerebellum, where balance and motor skills are coordinated. Similar results with birds, mice, and monkeys have confirmed the point: Experience can stimulate the growth of new connections and mold the brains neural architecture.

Once the brain reaches maturity, the number of neurons does not increase, and any neurons that are damaged are permanently disabled. But the plasticity of the brain can greatly benefit people with damage to the brain and nervous system. Organisms can compensate for loss by strengthening old neural connections and sprouting new ones. That is why people who suffer strokes are often able to recover their lost speech and motor abilities.

In 1860 German physicist Gustav Fechner theorized that if the human brain were divided into right and left halves, each side would have its own stream of consciousness. Modern medicine has actually allowed scientists to investigate this hypothesis. People who suffer from life-threatening epileptic seizures sometimes undergo a radical surgery that severs the corpus callosum, a bridge of nerve tissue that connects the right and left hemispheres of the brain. After the surgery, the two hemispheres can no longer communicate with each other.

Scientists have long considered the nature of consciousness without producing a fully satisfactory definition. In the early 20th century American philosopher and psychologist William James suggested that consciousness is a mental process involving both attention to external stimuli and short-term memory. Later scientific explorations of consciousness mostly expanded upon James work. In this article from a 1997 special issue of Scientific American, Nobel laureate Francis Crick, who helped determine the structure of a DNA, and fellow biophysicist Christof Koch explain how experiments on vision might deepen our understanding of consciousness.

Beginning in the 1960s American neurologist Roger Sperry and others tested such split-brain patients in carefully designed experiments. The researchers found that the hemispheres of these patients seemed to function independently, almost as if the subjects had two brains. In addition, they discovered that the left hemisphere was capable of speech and language, but not the right hemisphere. For example, when split-brain patients saw the image of an object flashed in their left visual field (thus sending the visual information to the right hemisphere), they were incapable of naming or describing the object. Yet they could easily point to the correct object with their left hand (which is controlled by the right hemisphere). As Sperrys colleague Michael Gazzaniga stated, Each half brain seemed to work and function outside of the conscious realm of the other.

Other psychologists interested in consciousness have examined how people are influenced without their awareness. For example, research has demonstrated that under certain conditions in the laboratory, people can be fleetingly affected by subliminal stimuli, sensory information presented so rapidly or faintly that it falls below the threshold of awareness. (Note, however, that scientists have discredited claims that people can be importantly influenced by subliminal messages in advertising, rock music, or other media.) Other evidence for influence without awareness comes from studies of people with a type of amnesia that prevents them from forming new memories. In experiments, these subjects are unable to recognize words they previously viewed in a list, but they are more likely to use those words later in an unrelated task. In fact, memory without awareness is normal, as when people come up with an idea they think is original, only later to realize that they had inadvertently borrowed it from another source.

Cognitive psychologists have often likened human memory to a computer that encodes, stores, and retrieves information. It is now clear, however, that remembering is an active process and that people construct and alters memories according to their beliefs, wishes, needs, and information received from outside sources.

Without realizing it, people sometimes create memories that are false. In one study, for example, subjects watched a slide show depicting a car accident. They saw either a STOP sign or a YIELD sign in the slides, but afterward they were asked a question about the accident that implied the presence of the other sign. Influenced by this suggestion, many subjects recalled the wrong traffic sign. In another study, people who heard a list of sleep-related words (bed, yawn) or music-related words (jazz, instruments) were often convinced moments later that they had also heard the words sleep or music - words that fit the category but were not on the list. In a third study, researchers asked college students to recall their high-school grades. Then the researchers checked those memories against the students actual transcripts. The students recalled most grades correctly, but most of the errors inflated their grades, particularly when the actual grades were low. When scientists distinguish between human beings and other animals, they point to our larger cerebral cortex (the outer part of the brain) and to our superior intellect - as seen in the abilities to acquire and store large amounts of information, solve problems, and communicate through the use of language.

In recent years, however, those studying human cognition has found that people are often less than rational and accurate in their performance. Some researchers have found that people are prone to forgetting, and worse, that memories of past events are often highly distorted. Others have observed that people often violate the rules of logic and probability when reasoning about real events, as when gamblers overestimate the odds of winning in games of chance. One reason for these mistakes is that we commonly rely on cognitive heuristics, mental shortcuts that allow us to make judgments that are quick but often in error. To understand how heuristics can lead to mistaken assumptions, imagine offering people a lottery ticket containing six numbers out of a pool of the numbers 1 through 40. If given a choice between the tickets 6-39-2-10-24-30 and 1-2-3-4-5-6, most people select the first ticket, because it has the appearance of randomness. Yet out of the 3,838,380 possible winning combinations, both sequences are equally likely.

One of the oldest debates in psychology, and in philosophy, concerns whether individual human traits and abilities are predetermined from birth or due to ones upbringing and experiences. This debate is often termed the nature-nurture debate. A strict genetic (nature) position states that people are predisposed to become sociable, smart, cheerful, or depressed according to their genetic blueprint. In contrast, a strict environmental (nurture) position says that people are shaped by parents, peers, cultural institutions, and life experiences.

Research shows that the more genetically related a person is to someone with schizophrenia, the greater the risk that person has of developing the illness. For example, children of one parent with schizophrenia have a 13 percent chance of developing the illness, whereas children of two parents with schizophrenia have a 46 percent chance of developing the disorder.

Researchers can estimate the role of genetic factors in two ways: (1) twin studies and (2) adoption studies. Twin studies compare identical twins with fraternal twins of the same sex. If identical twins (who share all the same genes) are more similar to each other on a given trait than are same-sex fraternal twins (who share only about half of the same genes), then genetic factors are assumed to influence the trait. Other studies compare identical twins who are raised together with identical twins who are separated at birth and raised in different families. If the twins raised together are more similar to each other than the twins raised apart, childhood experiences are presumed to influence the trait. Sometimes researchers conduct adoption studies, in which they compare adopted children to their biological and adoptive parents. If these children display traits that resemble those of their biological relatives more than their adoptive relatives, genetic factors are assumed to play a role in the trait.

In recent years, several twin and adoption studies have shown that genetic factors play a role in the development of intellectual abilities, temperament and personality, vocational interests, and various psychological disorders. Interestingly, however, this same research indicates that at least 50 percent of the variation in these characteristics within the population is attributable to factors in the environment. Today, most researchers agree that psychological characteristics spring from a combination of the forces of nature and nurture.

Helpless to survive on their own, newborn babies nevertheless possess a remarkable range of skills that aid in their survival. Newborns can see, hear, taste, smell, and feel pain; vision is the least developed sense at birth but improves rapidly in the first months. Crying communicates their need for food, comfort, or stimulation. Newborns also have reflexes for sucking, swallowing, grasping, and turning their head in search of their mothers nipple.

In 1890 William James described the newborns experience as one great blooming, buzzing confusion. However, with the aid of sophisticated research methods, psychologists have discovered that infants are smarter than was previously known.

A period of dramatic growth, infancy lasts from birth to around 18 months of age. Researchers have found that infants are born with certain abilities designed to aid their survival. For example, newborns show a distinct preference for human faces over other visual stimuli.

To learn about the perceptual world of infants, researchers measure infants head movements, eye movements, facial expressions, brain waves, heart rate, and respiration. Using these indicators, psychologists have found that shortly after birth. Infants show a distinct preference for the human face over other visual stimuli. Also suggesting that newborns are tuned into the face as a social object is the fact that within 72 hours of birth, they can mimic adults who purse the lips or stick out the tongue - a rudimentary form of imitation. Newborns can distinguish between their mothers voice and that of another woman. And at two weeks old, nursing infants are more attracted to the body odor of their mother and other breast-feeding females than to that of other women. Taken together, these findings show that infants are equipped at birth with certain senses and reflexes designed to aid their survival.

In 1905 French psychologist Alfred Binet and colleague Théodore Simon devised one of the first tests of general intelligence. The test sought to identify French children likely to have difficulty in school so that they could receive special education. An American version of Binets tests, the Stanford-Binet Intelligence Scale, is still used today.

In 1905 French psychologist Alfred Binet devised the first major intelligence test for the purpose of identifying slow learners in school. In doing so, Binet assumed that intelligence could be measured as a general intellectual capacity and summarized in a numerical score, or intelligence quotient (IQ). Consistently, testing has revealed that although each of us is more skilled in some areas than in others, a general intelligence underlies our more specific abilities.

Intelligence tests often play a decisive role in determining whether a person is admitted to college, graduate school, or professional school. Thousands of people take intelligence tests every year, but many psychologists and education experts question whether these tests are an accurate way of measuring who will succeed or fail in school and later in life. In this 1998 Scientific American article, psychology and education professor Robert J. Sternberg of Yale University in New Haven, Connecticut, presents evidence against conventional intelligence tests and proposes several ways to improve testing.

Today, many psychologists believe that there is more than one type of intelligence. American psychologist Howard Gardner proposed the existence of multiple intelligence. Each linked to a separate system within the brain. He theorized that there are seven types of intelligence: linguistic, logical-mathematical, spatial, musical, bodily-kinesthetic, interpersonal, and intrapersonal. American psychologist Robert Sternberg suggested a different model of intelligence, consisting of three components: analytic (school smarts, as measured in academic tests), creative (a capacity for insight), and practical (street smarts, or the ability to size up and adapt to situations). Psychologists from all branches of the discipline study the topic of motivation, an inner state that moves an organism toward the fulfillment of some goal. Over the years, different theories of motivation have been proposed. Some theories state that people are motivated by the need to satisfy physiological needs, whereas others state that people seek to maintain an optimum level of bodily arousal (not too little and not too much). Still other theories focus on the ways in which people respond to external incentives such as money, grades in school, and recognition. Motivation researchers study a wide range of topics, including hunger and obesity, sexual desire, the effects of reward and punishment, and the needs for power, achievement, social acceptance, love, and self-esteem.

In 1954 American psychologist Abraham Maslow proposed that all people are motivated to fulfill a hierarchical pyramid of needs. At the bottom of Maslows pyramids are needs essential to survival, such as the needs for food, water, and sleep. The need for safety follows these physiological needs. According to Maslow, higher-level needs become important to us only after our more basic needs are satisfied. These higher needs include the need for love and belongingness, the need for esteem, and the need for self-actualization (in Maslows theory, a state in which people realize their greatest potential).

The view that the role of sentences in inference gives a more important key to their meaning than their external relations to things in the world. The meaning of a sentence becomes its place in a network of inferences that it legitimates. Also, known as its functional role semantics, procedural semantic or conceptual role semantics. As these view bear some relation to the coherence theory of truth, and suffers from the same suspicion that divorces meaning from any clear association with things in the world.

Paradoxes rest upon the assumption that analysis is a relation with concept, then are involving entities of other sorts, such as linguistic expressions, and that in true analysis, analysand and analysandum are one and the same concept. However, these assumptions are explicit in the British philosopher George Edward Moore, but some of Moores remarks hint at a solution that a statement of an analysis is a statement partially taken about the concept involved and partly about the verbal expression used to express it. Moore is to suggest that he thinks of a solution of this sort is bound to be right, however, facts to suggest one because he cannot reveal of any way in which the analysis can be as part of the expression.

Elsewhere, the possibility clearly does set of apparent incontrovertible premises giving unacceptable or contradictory conclusions. To solve a paradox will involve showing that either these hidden flaws in the premises, or what the reasoning is erroneous, or that the apparently unacceptable conclusion can, in fact, be tolerable. Paradoxes are therefore important in philosophy, for until one is solved it shows that there is something about our reasoning and our concepts that we do not understand. Famous families of paradoxes include the semantic paradoxes and Zenos paradoxes. At the beginning of the 20th century, Russells paradox and other set-theoretic paradoxes of set theory, while the Sorites paradox has led to the investigation of the semantics of vagueness, and fuzzy logic. Paradoxes are under their other titles. Much as there is as much as a puzzle arising when someone says p but I do not believe that p. What is said is not contradictory, since (for many instances of p) both parts of it could br. true. But the person nevertheless violates one presupposition of normal practice, namely that you assert something only if you believe it: By adding that you do not believe what you just said you undo the natural significance of the original act saying it.

Furthermore, the moral philosopher and epistemologist Bernard Bolzano (1781-1848), whose logical work was based on a strong sense of there being an ontological underpinning of science and epistemology, lying in a theory of the objective entailments masking up the structure of scientific theories. His ability to challenge wisdom and come up with startling new ideas, as a Christian philosopher whether than from any position of mathematical authority, that for considerations of infinity, Bolzanos significant work was Paradoxin des Unenndlichen, written in retirement and translated into the English as Paradoxes of the Infinite. Here, Bolzano considered directly the points that had concerned Galileo - the conflicting result that seem to emerge when infinity is studied. Certainly most of the paradoxical statements encountered in the mathematical domain . . . are propositions which either immediately contain the idea of the infinite, or at least in some way or other depends upon that idea for their attempted proof.

Continuing, Bolzano looks at two possible approaches to infinity. One is simply the case of setting up a sequence of numbers, such as the whole numbers, and saying that as it cannot conceivably be said to have a last term, it is inherently infinite - not finite. It is easy enough to show that the whole numbers do not have a point at which they stop, giving a name to that last number whatever it might have been to call it ultimate. Then what’s wrong with ultimate + 1? Why is that not a whole number?

The second approach to infinity, which Bolzano ascribes in Paradoses of the Infinite to some philosophers . . . Taking this approach describe his first conception of infinity as the bad infinity. Although the German philosopher Friedrich George Hegal (1770-1831) applies the conceptual form of infinity and points that it is, rather, the basis for a substandard infinity that merely reaches toward the absolute, but never reaches it. In Paradoses of the Infinite, he calls this form of potential infinity as a variable quantity knowing no limit to its growth (a definition adopted, even by many mathematicians) . . . always growing int the infinite and never reaching it. As far as Hegel and his colleagues were concerned, using this uprush, there was no need for a real infinity beyond some unreachable absolute. Instead we deal with a variable quality that is as big as we need it to be, or often in calculus as small as we need it to be, without ever reaching the absolute, ultimate, truly infinite.

Bolzano argues, though, that there is something else, an infinity that does not have this whatever you need it to be elasticity. In fact a truly infinite quantity (for example, the length of a straight line unbounded in either direction, meaning: The magnitude of the spatial entity containing all the points determined solely by their abstractly conceivable relation to two fixed points) does not by any means need to be variable, and in an adduced example it is in fact not variable. Conversely, it is quite possible for a quantity merely capable of being taken greater than we have already taken it, and of becoming larger than any pre-assigned (finite) quantity, nevertheless to mean at all times merely finitely, which holds in particular of every numerical quantity 1, 2, 3, 4, 5.

In other words, for Bolzano there could be a true infinity that was not a variable something that was only bigger than anything you might specify. Such a true infinity was the result of joining two pints together and extending that line in both directions without stopping. And what is more, he could separate off the demands of calculus, using a finite quality without ever bothering with the slippery potential infinity. Here was both a deeper understanding of the nature of infinity and the basis on which are built in his safe infinity free calculus.

This use of the inexhaustible follows on directly from most Bolzanos criticism of the way that ∞we used as a variable something that would be bigger than anything you could specify, but never quite reached the true, absolute infinity. In Paradoxes of the Infinity Bolzano points out that is possible for a quantity merely capable of becoming larger than any other one pre-assigned (finite) quantity, nevertheless to remain at all times merely finite.

Bolzano intended tis as a criticism of the way infinity was treated, but Professor Jacquette sees it instead of a way of masking use of practical applications like calculus without the need for weasel words about infinity.

By replacing ∞with ¤ we do away with one of the most common requirements for infinity, but is there anything left that map out to the real world? Can we confine infinity to that pure mathematical other world, where anything, however unreal, can be constructed, and forget about it elsewhere? Surprisingly, this seems to have been the view, at least at one point in time, even of the German mathematician and founder of set-theory Georg Cantor (1845-1918), himself, whose comments in 1883, that only the finite numbers are real.

Keeping within the lines of reason, both the Cambridge mathematician and philosopher Frank Plumpton Ramsey (1903-30) and the Italian mathematician G. Peano (1858-1932) have been to distinguish logical paradoxes and that depend upon the notion of reference or truth (semantic notions), such are the postulates justifying mathematical induction. It ensures that a numerical series is closed, in the sense that nothing but zero and its successors can be numbers. In that any series satisfying a set of axioms can be conceived as the sequence of natural numbers. Candidates from set theory include the Zermelo numbers, where the empty set is zero, and the successor of each number is its unit set, and the von Neuman numbers, where each number is the set of all smaller numbers. A similar and equally fundamental complementarity exists in the relation between zero and infinity. Although the fullness of infinity is logically antithetical to the emptiness of zero, infinity can be obtained from zero with a simple mathematical operation. The division of many numbers by zero is infinity, while the multiplication of any number by zero is zero.

With the set theory developed by the German mathematician and logician Georg Cantor. From 1878 to 1807, Cantor created a theory of abstract sets of entities that eventually became a mathematical discipline. A set, as he defined it, is a collection of definite and distinguished objects in thought or perception conceived as a whole.

Cantor attempted to prove that the process of counting and the definition of integers could be placed on a solid mathematical foundation. His method was to repeatedly place the elements in one set into one-to-one correspondence with those in another. In the case of integers, Cantor showed that each integer (1, 2, 3, . . . n) could be paired with an even integer (2, 4, 6, . . . n), and, therefore, that the set of all integers was equal to the set of all even numbers.

Amazingly, Cantor discovered that some infinite sets were large than others and that infinite sets formed a hierarchy of greater infinities. After this failed attempt to save the classical view of logical foundations and internal consistency of mathematical systems, it soon became obvious that a major crack had appeared in the seemingly sold foundations of number and mathematics. Meanwhile, an impressive number of mathematicians began to see that everything from functional analysis to the theory of real numbers depended on the problematic character of number itself.

While, in the theory of probability Ramsey was the first to show how a personalised theory could be developed, based on precise behavioural notions of preference and expectation. In the philosophy of language, Ramsey was one of the first thinkers to accept a redundancy theory of truth, which hr combined with radical views of the function of many kinds of propositions. Neither generalizations nor causal propositions, nor those treating probability or ethics, describe facts, but each has a different specific function in our intellectual economy.

Ramsey advocates that of a sentence generated by taking all the sentences affirmed in a scientific theory that use some term, e.g., quark. Replacing the term by a variable, and existentially quantifying into the result. Instead of saying quarks have such-and-such properties, Ramsey postdated that the sentence as saying that there is something that has those properties. If the process is repeated, the sentence gives the topic-neutral structure of the theory, but removes any implications that we know what the term so treated denote. It leaves open the possibility of identifying the theoretical item with whatever it is that best fits the description provided. Nonetheless, it was pointed out by the Cambridge mathematician Newman that if the process is carried out for all except the logical bones of the theory, then by the Löwenheim-Skolem theorem, the result will be interpretable in any domain of sufficient cardinality, and the content of the theory may reasonably be felt to have been lost.

It seems, that the most taken of paradoxes in the foundations of set theory as discovered by Russell in 1901. Some classes have themselves as members: The class of all abstract objects, for example, is an abstract object, whereby, others do not: The class of donkeys is not itself a donkey. Now consider the class of all classes that are not members of themselves, is this class a member of itself, that, if it is, then it is not, and if it is not, then it is.

The paradox is structurally similar to easier examples, such as the paradox of the barber. Such one like a village having a barber in it, who shaves all and only the people who do not have in themselves. Who shaves the barber? If he shaves himself, then he does not, but if he does not shave himself, then he does not. The paradox is actually just a proof that there is no such barber or in other words, that the condition is inconsistent. All the same, it is no too easy to say why there is no such class as the one Russell defines. It seems that there must be some restriction on the kind of definition that are allowed to define classes and the difficulty that of finding a well-motivated principle behind any such restriction.

The French mathematician and philosopher Henri Jules Poincaré (1854-1912) believed those paradoses like those of Russell nd the barber were due to such as the predicative definitions, and therefore proposed banning them. But, it tuns out that classical mathematics required such definitions at too many points for the ban to be easily absolved. Having, in turn, as forwarded by Poincaré and Russell, was that in order to solve the logical and semantic paradoxes it would have to ban any collection (set) containing members that can only be defined by means of the collection taken as a whole. It is, effectively by all occurring principles into which have an adopting vicious regress, as to mark the definition for which involves no such failure. There is frequently room for dispute about whether regresses are benign or vicious, since the issue will hinge on whether it is necessary to reapply the procedure. The cosmological argument is an attempt to find a stopping point for what is otherwise seen as an infinite regress, and, to ban of the predicative definitions.

The investigation of questions that arise from reflection upon sciences and scientific inquiry, are such as called of a philosophy of science. Such questions include, what distinctions in the methods of science? Is there a clear demarcation between scenes and other disciplines, and how do we place such enquires as history, economics or sociology? And scientific theories probable or more in the nature of provisional conjecture? Can the be verified or falsified? What distinguished good from bad explanations? Might there be one unified since, embracing all the special science? For much of the 20th century their questions were pursued in a high abstract and logical framework it being supposed that as general logic of scientific discovery that a general logic of scientific discovery a justification might be found. However, many now take interests in a more historical, contextual and sometimes sociological approach, in which the methods and successes of a science at a particular time are regarded less in terms of universal logical principles and procedure, and more in terms of their availability to methods and paradigms as well as the social context.

In addition, to general questions of methodology, there are specific problems within particular sciences, giving subjects as biology, mathematics and physics.

The intuitive certainty that sparks aflame the dialectic awarenesses for its immediate concerns are either of the truths or by some other in an object of apprehensions, such as a concept. Awareness as such, has to its amounting quality value the place where philosophically understanding of the source of our knowledge are, however, in covering the sensible apprehension of things and pure intuition it is that which stricture sensation into the experience of things accent of its direction that orchestrates the celestial overture into measures in space and time.

The notion that determines how something is seen or evaluated of the status of law and morality especially associated with St Thomas Aquinas and the subsequent scholastic tradition. More widely, any attempt to cement the moral and legal order together with the nature of the cosmos or how the nature of human beings, for which sense it is also found in some Protestant writers, and arguably derivative from a Platonic view of ethics, and is implicit in ancient Stoicism. Law stands above and apart from the activities of human lawmaker, it constitutes an objective set of principles that can be seen true by natural light or reason, and (in religion versions of the theory) that express Gods will for creation. Non-religious versions of the theory substitute objective conditions for human flourishing as the source of constraints upon permissible actions and social arrangements. Within the natural law tradition, different views have been held about the relationship between the rule of law about God s will, for instance the Dutch philosopher Hugo Grothius (1583-1645), similarly takes upon the view that the content of natural law is independent of any will, including that of God, while the German theorist and historian Samuel von Pufendorf (1632-94) takes the opposite view, thereby facing the problem of one horn of the Euthyphro dilemma, that simply states, that its dilemma arises from whatever the source of authority is supposed to be, for in which do we care about the general good because it is good, or do we just call good things that we care about. Wherefore, by facing the problem that may be to assume of a strong form, in which it is claimed that various facts entail values, or a weaker form, from which it confines itself to holding that reason by itself is capable of discerning moral requirements that are supped of binding to all human bings regardless of their desires

Although the morality of people send the ethical amount, from which the same thing, is that there is a usage that restricts morality to systems such as that of the German philosopher and founder of ethical philosophy Immanuel Kant (1724-1804), based on notions such as duty, obligation, and principles of conduct, reserving ethics for more than the Aristotelian approach to practical reasoning based on the notion of a virtue, and generally avoiding the separation of moral considerations from other practical considerations. The scholarly issues are complex, with some writers seeing Kant as more Aristotelian and Aristotle as, ore involved in a separate sphere of responsibility and duty, than the simple contrast suggests. Some theorists see the subject in terms of a number of laws (as in the Ten Commandments). The status of these laws may be test they are the edicts of a divine lawmaker, or that they are truths of reason, knowable deductively. Other approaches to ethics (e.g., eudaimonism, situation ethics, virtue ethics) eschew general principles as much as possible, frequently disguising the great complexity of practical reasoning. For Kantian notion of the moral law is a binding requirement of the categorical imperative, and to understand whether they are equivalent at some deep level. Kants own applications of the notion are not always convincing, as for one cause of confusion in relating Kants ethics to theories such additional expressivism is that it is easy, but mistaken, to suppose that the categorical nature of the imperative means that it cannot be the expression of sentiment, but must derive from something unconditional or necessary such as the voice of reason.

For whichever reason, the mortal being makes of its presence to the future of weighing of that which one must do, or that which can be required of one. The term carries implications of that which is owed (due) to other people, or perhaps in onself. Universal duties would be owed to persons (or sentient beings) as such, whereas special duty in virtue of specific relations, such for being the child of someone, or having made someone a promise. Duty or obligation is the primary concept of deontological approaches to ethics, but is constructed in other systems out of other notions. In the system of Kant, a perfect duty is one that must be performed whatever the circumstances: Imperfect duties may have to give way to the more stringent ones. In another way, perfect duties are those that are correlative with the right to others, imperfect duties are not. Problems with the concept include the ways in which due needs to be specified (a frequent criticism of Kant is that his notion of duty is too abstract). The concept may also suggest of a regimented view of ethical life in which we are all forced conscripts in a kind of moral army, and may encourage an individualistic and antagonistic view of social relations.

The most generally accepted account of externalism and/or internalism, that this distinction is that a theory of justification is internalist if only if it requiem that all of the factors needed for a belief to be epistemologically justified for a given person be cognitively accessible to that person, internal to his cognitive perceptive, and externalist, if it allows that at least some of the justifying factors need not be thus accessible, so that thy can be external to the believers cognitively perceptive, beyond any such given relations. However, epistemologists often use the distinction between internalist and externalist theories of epistemic justification without offering any very explicit explication.

The externalist/internalist distinction has been mainly applied to theories of epistemic justification: It has also been applied in a closely related way to accounts of knowledge and in a rather different way to accounts of belief and thought contents.

The internalist requirement of cognitive accessibility can be interpreted in at least two ways: A strong version of internalism would require that the believer actually be aware of the justifying factor in order to be justified: While a weaker version would require only that he be capable of becoming aware of them by focussing his attentions appropriately, but without the need for any change of position, new information, etc. Though the phrase cognitively accessible suggests the weak interpretation, the main intuitive motivation for internalism, viz. the idea that epistemic justification requires that the believer actually have in his cognitive possession a reason for thinking that the belief is true, and would require the strong interpretation.

Perhaps, the clearest example of an internalist position would be a Foundationalist view according to which foundational beliefs pertain to immediately experienced states of mind and other beliefs are justified by standing in cognitively accessible logical or inferential relations to such foundational beliefs. Such a view could count as either a strong or a weak version of internalism, depending on whether actual awareness of the justifying elements or only the capacity to become aware of them is required. Similarly, a coherent view could also be internalist, if both the beliefs or other states with which a justification belief is required to cohere and the coherence relations themselves are reflectively accessible.

It should be carefully noticed that when internalism is construed in this way, it is neither necessary nor sufficient by itself for internalism that the justifying factors literally be internal mental states of the person in question. Not necessary, necessary, because on at least some views, e.g., a direct realist view of perception, something other than a mental state of the believer can be cognitively accessible: Not sufficient, because there are views according to which at least some mental states need not be actual (strong version) or even possible (weak version) objects of cognitive awareness. Also, on this way of drawing the distinction, a hybrid view, according to which some of the factors required for justification must be cognitively accessible while others need not and in general will not be, would count as an externalist view. Obviously too, a view that was externalist in relation to a strong version of internalism (by not requiring that the believer actually be aware of all justifying factors) could still be internalist in relation to a weak version (by requiring that he at least be capable of becoming aware of them).

The most prominent recent externalist views have been versions of reliabilism, whose requirements for justification is roughly that the belief be produced in a way or via a process that makes of objectively likely that the belief is true. What makes such a view externalist is the absence of any requirement that the person for whom the belief is justified have any sort of cognitive access to the relations of reliability in question. Lacking such access, such a person will in general have no reason for thinking that the belief is true or likely to be true, but will, on such an account, nonetheless be epistemically justified in according it. Thus such a view arguably marks a major break from the modern epistemological tradition, stemming from Descartes, which identifies epistemic justification with having a reason, perhaps even a conclusive reason for thinking that the belief is true. An epistemologist working within this tradition is likely to feel that the externalist, than offering a competing account of the same concept of epistemic justification with which the traditional epistemologist is concerned, has simply changed the subject.

The main objection to externalism rests on the intuitive certainty that the basic requirement for epistemic justification is that the acceptance of the belief in question be rational or responsible in relation to the cognitive goal of truth, which seems to require in turn that the believer actually be dialectally aware of a reason for thinking that the belief is true (or, at the very least, that such a reason be available to him). Since the satisfaction of an externalist condition is neither necessary nor sufficient for the existence of such a cognitively accessible reason, it is argued, externalism is mistaken as an account of epistemic justification. This general point has been elaborated by appeal to two sorts of putative intuitive counter-examples to externalism. The first of these challenges the necessity of belief which seem intuitively to be justified, but for which the externalist conditions are not satisfied. The standard examples in this sort are cases where beliefs are produced in some very nonstandard way, e.g., by a Cartesian demon, but nonetheless, in such a way that the subjective experience of the believer is indistinguishable from that of someone whose beliefs are produced more normally. The intuitive claim is that the believer in such a case is nonetheless epistemically justified, as much so as one whose belief is produced in a more normal way, and hence that externalist account of justification must be mistaken.

Perhaps the most striking reply to this sort of counter-example, on behalf of a cognitive process is to be assessed in normal possible worlds, i.e., in possible worlds that are actually the way our world is common-seismically believed to be, than in the world which contains the belief being judged. Since the cognitive processes employed in the Cartesian demon cases are, for which we may assume, reliable when assessed in this way, the reliabilist can agree that such beliefs are justified. The obvious, to a considerable degree of bringing out the issue of whether it is or not an adequate rationale for this construal of reliabilism, so that the reply is not merely a notional presupposition guised as having representation.

The correlative way of elaborating on the general objection to justificatory externalism challenges the sufficiency of the various externalist conditions by citing cases where those conditions are satisfied, but where the believers in question seem intuitively not to be justified. In this context, the most widely discussed examples have to do with possible occult cognitive capacities, like clairvoyance. Considering the point in application once, again, to reliabilism, the claim is that to think that he has such a cognitive power, and, perhaps, even good reasons to the contrary, is not rational or responsible and therefore not epistemically justified in accepting the belief that result from his clairvoyance, despite the fact that the Reliabilist condition is satisfied.

One sort of response to this latter sorts of objection is to bite the bullet and insist that such believers are in fact justified, dismissing the seeming intuitions to the contrary as latent internalist prejudice. A more widely adopted response attempts to impose additional conditions, usually of a roughly internalist sort, which will rule out the offending example, while stopping far of a full internalism. But, while there is little doubt that such modified versions of externalism can handle particular cases, as well enough to avoid clear intuitive implausibility, the usually problematic cases that they cannot handle, and also whether there is and clear motivation for the additional requirements other than the general internalist view of justification that externalist are committed to reject.

A view in this same general vein, one that might be described as a hybrid of internalism and externalism holds that epistemic justification requires that there is a justificatory factor that is cognitively accessible to the believer in question (though it need not be actually grasped), thus ruling out, e.g., a pure reliabilism. At the same time, however, though it must be objectively true that beliefs for which such a factor is available are likely to be true, in addition, the fact need not be in any way grasped or cognitively accessible to the believer. In effect, of the premises needed to argue that a particular belief is likely to be true, one must be accessible in a way that would satisfy at least weak internalism, the internalist will respond that this hybrid view is of no help at all in meeting the objection and has no belief nor is it held in the rational, responsible way that justification intuitively seems to require, for the believer in question, lacking one crucial premise, still has no reason at all for thinking that his belief is likely to be true.

An alternative to giving an externalist account of epistemic justification, one which may be more defensible while still accommodating many of the same motivating concerns, is to give an externalist account of knowledge directly, without relying on an intermediate account of justification. Such a view will obviously have to reject the justified true belief account of knowledge, holding instead that knowledge is true belief which satisfies the chosen externalist condition, e.g., a result of a reliable process (and perhaps, further conditions as well). This makes it possible for such a view to retain internalist account of epistemic justification, though the centrality of that concept to epistemology would obviously be seriously diminished.

Such an externalist account of knowledge can accommodate the commonsense conviction that animals, young children, and unsophisticated adults posses knowledge, though not the weaker conviction (if such a conviction does exist) that such individuals are epistemically justified in their beliefs. It is also at least less vulnerable to internalist counter-examples of the sort discussed, since the intuitions involved there pertain more clearly to justification than to knowledge. What is uncertain is what ultimate philosophical significance the resulting conception of knowledge is supposed to have. In particular, does it have any serious bearing on traditional epistemological problems and on the deepest and most troubling versions of scepticism, which seems in fact to be primarily concerned with justification, the an knowledge?`

A rather different use of the terms internalism and externalism has to do with the issue of how the content of beliefs and thoughts is determined: According to an internalist view of content, the content of such intention states depends only on the non-relational, internal properties of the individuals mind or grain, and not at all on his physical and social environment: While according to an externalist view, content is significantly affected by such external factors and suggests a view that appears of both internal and external elements is standardly classified as an external view.

As with justification and knowledge, the traditional view of content has been strongly internalist in character. The main argument for externalism derives from the philosophy of language, more specifically from the various phenomena pertaining to natural kind terms, indexicals, etc. that motivate the views that have come to be known as direct reference theories. Such phenomena seem at least to show that the belief or thought content that can be properly attributed to a person is dependant on facts about his environment -, e.g., whether he is on Earth or Twin Earth, what is fact pointing at, the classificatory criteria employed by expects in his social group, etc. - not just on what is going on internally in his mind or brain.

An objection to externalist account of content is that they seem unable to do justice to our ability to know the content of our beliefs or thought from the inside, simply by reflection. If content is depending on external factors pertaining to the environment, then knowledge of content should depend on knowledge of these factors - which will not in general be available to the person whose belief or thought is in question.

The adoption of an externalist account of mental content would seem to support an externalist account of justification, by way that if part or all of the content of a belief inaccessible to the believer, then both the justifying status of other beliefs in relation to that content and the status of that content ss justifying further beliefs will be similarly inaccessible, thus contravening the internalist requirement for justification. An internalist must insist that there are no justification relations of these sorts, that our internally associable content can either be justified or justly anything else: But such a response appears lame unless it is coupled with an attempt to show that the externalist account of content is mistaken.

For the best of all possibilities that explains inference had first been formulated by Princeton philosopher Gilber Harnam. The idea in that when we have an easy to excavate, the idea is that when we have a best explanation of some phenomenon, we are entitled to repose confidence in it simply on that account. Sometimes thought to br. the lynchpin of scientific method, the principle is not easy to formulate and has come under attack, notably since our best current explanation of something may only be the best of a bad lot. There exist cases in which the best explanation is still not all that convincing, so other considerations than pure explanation success seem to play a role.

The philosopher Bas van Fraassen (The Scientific Image, 1980) explicated further in that constructive empiricism divides science into observation statement and theory statement. It holds that the latter are capable of strict truth and falsity, but maintains that the appropriate attitude is not to believe them, but only to accept them ss best as empirically adequate. It is often regarded as a variety of pragmatism or instrumentalism, although more orthodox varieties of those position deny that theoretic statements have truth-value. A related vie is held by the German philosopher Hans Valhinger (1854-1933), who, however, thinks that we ca be sure that theoretical statements are actually false. In other words, theories are useful because they enable us to cope with what would otherwise be the unmanageable complexity of things. The doctrine bears some affinity to pragmatism, but differs in that Vaihlinger thinks that our useful theories are nevertheless really false.

Lectures published as Pragmatism, became the new name for old ways of thinking (1907) summed up Jamess original contributions to the theory called pragmatism, a term first used by the American logician C. S. Peirce. James generalized the pragmatic method, developing it from a critique of the logical basis of the sciences into a basis for the evaluation of all experience. He maintained that the meaning of ideas is found only in terms of their possible consequences. If consequences are lacking, ideas are meaningless. James contended that this is the method used by scientists to define their terms and to test their hypotheses, which, if meaningful, entail predictions. The hypotheses can be considered true if the predicted events take place. On the other hand, most metaphysical theories are meaningless, because they entail no testable predictions. Meaningful theories, James argued, are instruments for dealing with problems that arise in experience.

According to James pragmatism, then, truth is that which works. One determines what works by testing propositions in experience. In so doing, one finds that certain propositions become true. As James put it, Truth is something that happens to an idea in the process of its verification; it is not a static property. This does not mean, however, that anything can be true. The true is only the expedient in the way of our thinking, just as the right is only the expedient in the way of our behaving, James maintained. One cannot believe whatever one want to believe, because such self-centered beliefs would not work out.

James was opposed to absolute metaphysical systems and argued against doctrines that describe reality as a unified, monolithic whole. In Essays in Radical Empiricism (1912), he argued for a pluralistic universe, denying that the world can be explained in terms of an absolute force or scheme that determines the interrelations of things and events. He held that the interrelations, whether they serve to hold things together or apart, are just as real as the things themselves.

By the end of his life, James had become world-famous as a philosopher and psychologist. In both fields, he functioned more as an originator of new thought than as a founder of dogmatic schools. His pragmatic philosophy was further developed by American philosopher John Dewey and others; later studies in physics by Albert Einstein made the theories of interrelations advanced by James appear prophetic.

Nevertheless, in science a way of looking at a field, this, however, as accorded through the lines of force whose evident physical reality o of the intervening medium whereby, the task of the philosopher of science has often been posed in terms of demarcating good or scientific theories from bad, unscientific ones wherefore he property of a statement or theory that it is capable of being refuted by experience. It a proper annunciation is accredited within the philosophy falsifiability, wherein scientific theory, as opposed to unfalsifaiability and noticeably psychoanalysis and historical materialism. The philosopher of science Raimund Karl Popper (1902-1994) whose idea was that it could be a positive virtue is a scientific theory that it is bold, conjectural, and goes beyond the evidence, but that it has to be capable of facing possible refutation. If every way that things turn out is compatible with it, then it is no longer a scientific theory, but for instance, an ideology or article of faith. Popper argued that the central virtue of science, as opposed to pseudo-science, is not that it puts forward the hypotheses that are confirmed by evidence. That is, they genuinely face the possibility of test and rejection through not confirming to evidences so gathered, that is to give no accent of te extent in which it is rational to rely upon scientific theory, however, the actual picture of acceptance and rejection of scientific hypotheses is more complex than Popper suggests. As perhaps, the view that everyday attributions of intention beliefs, and, meanings to other persons proceed via tacit use of a theory that enables one to construct of these interpretations as an explanation of their doing.

The view is commonly held along with functionalism according to which psychological states are theoretical entities, identified by te network of their causes ad effects. The theory-theory has different implications, depending upon which feature of theories is being stressed. Theories may be thought of as capable of as yielding predication and explanations, as achieved by a process of theorizing as answering to empirical evidence, that its principle describable without them, as liable to be overturned by newer and better theorise, and so on. The main problem with is with our understanding of others as the outcome of a piece of theorising is the non-existence of medium, in which his theory can be couched as the chid learns simultaneously with the minds language.

Our understanding of others is not gained by the tacit use of a theory, enabling us to infer what thoughts or intentions explain their action, but by reliving the situation in their moccasins or from their point of view, not thereby understanding what they experienced and thought, and therefore expressed. Understanding others is achieved when we can ourselves deliberate as they did, and hear their words if theory are our own. The suggestion is a modern development of the verstehe, the tradition associated with the German philosopher, literary critic and historian Wilhelm Dilthey (1833-1911).

In addition, the hypnotise especially associated with the American philosopher of mind J.A. Fodor, who affirmed strongly that mental processing occurs in a language different from ones ordinary native language, but underlying and explaining our competence with it. The idea is a development of the Chomskyan notion if an innate universal grammar, it is a way of drawing three analogies between the workings of the brain or mind and those of a standard computer, since computer programs are linguistically complex sets of instruction whose execution explains the surface behaviour of the computer. As an explanation of ordinary language learning and competence the hypothecs of thought have not found universal favour. It apparently only explains ordinary representational powers by invoking innate things of the same sort, and it invites the image of the learning infant translating the language surrounding and in back into an innate language whose own powers are a mysterious biological given.

In philosophy, the doctrine that all existence is resolvable into matter or into an attribute or effect of matter. According to this doctrine, matter is the ultimate reality, and the phenomenon of consciousness is explained by physicochemical changes in the nervous system. Materialism is thus the antithesis of idealism, in which the supremacy of mind is affirmed and matter is characterized as an aspect or objectification of mind. Extreme or absolute materialism is known as materialistic monism. According to the mind-stuff theory of monism, as expounded by the British metaphysician W. K. Clifford, in his Elements of Dynamic (1879-87), matters and minds are consubstantial, each being merely an aspect of the other. Philosophical materialism is ancient and has had numerous formulations. The early Greek philosophers subscribed to a variant of materialism known as hylozoism, according to which matter and life are identical. Related to hylozoism is the doctrine of hylotheism, in which matter is held to be divine, or the existence of God is disavowed apart from matter. Cosmological materialism is a term used to characterize a materialistic interpretation of the universe.

Antireligious materialism is motivated by a spirit of hostility toward the theological dogmas of organized religion, particularly those of Christianity. Notable among the exponents of antireligious materialism was the 18th-century French philosophers Denis Diderot, Paul Henri de’Holbach, and Julien Offroy de La Mettrie. According to historical materialism, as set forth in the writings of Karl Marx, Friedrich Engels, and Vladimir Ilich Lenin, in every historical epoch the prevailing economic system by which the necessities of life are produced determines the form of societal organization and the political, religious, ethical, intellectual, and artistic history of the epoch.

In modern times philosophical materialism has been largely influenced by the doctrine of evolution and may indeed be said to have been assimilated in the wider theory of evolution. Supporters of the theory of evolution go beyond the mere antithesis or atheism of materialism and seek positively to show how the diversities and differences in creation are the result of natural as opposed to supernatural processes.

The Philosophy of Mind is the branch of philosophy that considers mental phenomena such as sensation, perception, thought, belief, desire, intention, memory, emotion, imagination, and purposeful action. These phenomena, which can be broadly grouped as thoughts and experiences, are features of human beings; many of them are also found in other animals. Philosophers are interested in the nature of each of these phenomena as well as their relationships to one another and to physical phenomena, such as motion.

Philosophers use the characteristics of inward accessibility, subjectivity, intentionality, goal-directedness, creativity and freedom, and consciousness to distinguish mental phenomena from physical phenomena.

Perhaps the most important characteristic of mental phenomena is that they are inwardly accessible, or available to us through introspection. We each know our own minds - our sensations, thoughts, memories, desires, and fantasies - in a direct sense, by internal reflection. We also know our mental states and mental events in a way that no one else can. In other words, we have privileged access to our own mental states.

Certain mental phenomena, those we generally call experiences, have a subjective nature - that is, they have certain characteristics we become aware of when we reflect. For instance, there is something it is like to feel pain, or have an itch, or see something red. These characteristics are subjective in that they are accessible to the subject of the experience, the person who has the experience, but not to others.

Other mental phenomena, which we broadly refer to as thoughts, have a characteristic philosophers call intentionality. Intentional thoughts are about other thoughts or objects, which are represented as having certain properties or for being related to one another in a certain way. The belief that California is west of Nevada, for example, is about California and Nevada and represents the former being west of the latter. Although we have privileged access to our intentional states, many of them do not seem to have a subjective nature, at least not in the way that experiences do.

A number of mental phenomena appear to be connected to one another as elements in an intelligent, goal-directed system. The system works as follows: First, our sense organs are stimulated by events in our environment; next, by virtue of these stimulations, we perceive things about the external world; finally, we use this information, as well as information we have remembered or inferred, to guide our actions in ways that further our goals. Goal-directedness seems to accompany only mental phenomena.

Another important characteristic of mind, especially of human minds, is the capacity for choice and imagination. Rather than automatically converting past influences into future actions, individual minds are capable of exhibiting creativity and freedom. For instance, we can imagine things we have not experienced and can act in ways that no one expects or could predict.

Mental phenomena are conscious, and consciousness may be the closest term we have for describing what is special about mental phenomena. Minds are sometimes referred to as consciousness, yet it is difficult to describe exactly what consciousness is. Although consciousness is closely related to inward accessibility and subjectivity, these very characteristics seem to hinder us in reaching an objective scientific understanding of it.

Although philosophers have written about mental phenomena since ancient times, the philosophy of mind did not garner much attention until the work of French philosopher René Descartes in the 17th century. Descartes work represented a turning point in thinking about mind by making a strong distinction between bodies and minds, or the physical and the mental. This duality between mind and body, known as Cartesian dualism, has posed significant problems for philosophy ever since.

Descartes believed there are two basic kinds of things in the world, a belief known as substance dualism. For Descartes, the principles of existence for these two groups of things - bodies and minds - are completely different from one another: Bodies exist by being extended in space, while minds exist by being conscious. According to Descartes, nothing can be done to give a body thought and consciousness. No matter how we shape a body or combine it with other bodies, we cannot turn the body into a mind, a thing that is conscious, because being conscious is not a way of being extended.

For Descartes, a person consists of a human body and a human mind causally interacting with one another. For example, the intentions of a human being may have caused that persons limbs to move. In this way, the mind can affect the body. In addition, the sense organs of a human being maybe affected by light, pressure, or sound, external sources, which in turn affect the brain, affecting mental states. Thus, the body may affect the mind. Exactly how mind can affect body, and vice versa, is a central issue in the philosophy of mind, and is known as the mind-body problem. According to Descartes, this interaction of mind and body is peculiarly intimate. Unlike the interaction between a pilot and his ship, the connection between mind and body more closely resembles two substances that have been thoroughly mixed together.

In response to the mind-body problem arising from Descartes theory of substance dualism, a number of philosophers have advocated various forms of substance monism, the doctrine that there is ultimately just one kind of thing in reality. In the 18th century, Irish philosopher George Berkeley claimed there were no material objects in the world, only minds and their ideas. Berkeley thought that talk about physical objects was simply a way of organizing the flow of experience. Near the turn of the 20th century, American psychologist and philosopher William James proposed another form of substance monism. James claimed that experience is the basic stuff from which both bodies and minds are constructed.

Most philosophers of mind today are substance monists of a third type: They are materialists who believe that everything in the world is basically material, or a physical object. Among materialists, there is still considerable disagreement about the status of mental properties, which are conceived as properties of bodies or brains. Materialists who are property dualists believe that mental properties are an additional kind of property or attribute, not reducible to physical properties. Property dualists have the problem of explaining how such properties can fit into the world envisaged by modern physical science, according to which there are physical explanations for all things.

Materialists who are property monists believe that there is ultimately only one type of property, although they disagree on whether or not mental properties exist in material form. Some property monists, known as reductive materialists, hold that mental properties exist simply as a subset of relatively complex and nonbasic physical properties of the brain. Reductive materialists have the problem of explaining how the physical states of the brain can be inwardly accessible and have a subjective character, as mental states do. Other property monists, known as Eliminative materialists, consider the whole category of mental properties to be a mistake. According to them, mental properties should be treated as discredited postulates of an outmoded theory. Eliminative materialism is difficult for most people to accept, since we seem to have direct knowledge of our own philosophy of mind concerns itself with a number of specialized problems. For which the mind-body problem, becomes an important note of issues, that includes, of those impounded by personal identity, immortality, and artificial intelligence, least of mention, the capacity to deal flexibly and effectively with theoretical problems, it must be doubted to whether one thing can think or enabling if one can be of thought, as equally manifested in whatever logical, theoretical, mathematical, etc., successes we achieve. Nor is there much confidence left that intelligence tests measure any such subject’s capacity to take intelligence tests, often of very specific and culturally peculiar kinds. For the question whether humans alone posses intelligence.

During much of Western history, the mind has been identified with the soul as presented in Christian Theology. According to Christianity, the soul is the source of a persons identity and is usually regarded as immaterial; thus, it is capable of enduring after the death of the body. Descartes conception of the mind as a separate, nonmaterial substance fits well with this understanding of the soul. In Descartes view, we are aware of our bodies only as the cause of sensations and other mental phenomena. Consequently our personal essence is composed more fundamentally of mind and the preservation of the mind after death would constitute our continued existence.

The mind conceived by materialist forms of substance monism does not fit as neatly with this traditional concept of the soul. With materialism, once a physical body is destroyed, nothing enduring remains. Some philosophers think that a concept of personal identity can be constructed that permits the possibility of life after death without appealing to separate immaterial substances. Following in the tradition of 17th-century British philosopher John Locke, these philosophers propose that a person consists of a stream of mental events linked by memory. These link of memory, rather than a single underlying substance, provide the unity of a single consciousness through time. Immortality is conceivable if we think of these memory links as connecting a later consciousness in heaven with an earlier one on earth.

The field of artificial intelligence also raises interesting questions for the philosophy of mind. People have designed machines that mimic or model many aspects of human intelligence, and there are robots currently in use whose behavior is described in terms of goals, beliefs, and perceptions. Such machines are capable of behavior that, were it exhibited by a human being, would surely be taken to be free and creative. As an example, in 1996 an IBM computer named Deep Blue won a chess game against Russian world champion Garry Kasparov under international match regulations. Moreover, it is possible to design robots that have some sort of privileged access to their internal states. Philosophers disagree over whether such robots truly think or simply appear to think and whether such robots should be considered to be conscious.

Philosophy, a speculative world-view which asserts that basic reality is constantly in a process of flux and change. Indeed, reality is identified with pure process. Concepts such as creativity, freedom, novelty, emergence, and growth are fundamental explanatory categories for process philosophy. This metaphysical perspective is to be contrasted with a philosophy of substance, the view that a fixed and permanent reality underlies the changing or fluctuating world of ordinary experience. Whereas substance philosophy emphasizes static being, process philosophy emphasizes dynamically becoming.

Although process philosophy is as old as the 6th-century Bc Greek philosopher, Heraclitus, renewed interest in it was stimulated in the 19th century by the theory of evolution. Key figures in the development of modern process philosophy were the British philosophers’ Herbert Spencer, Samuel Alexander, and Alfred North Whitehead, the American philosophers Charles S. Peirce and William James, and the French philosophers Henri Bergson and Pierre Teilhard de Chardin. Whiteheads Process and Reality: An Essay in Cosmology (1929) is generally considered the most important systematic expression of process philosophy.

A contemporary theology has been strongly influenced by process philosophy. The American theologian Charles Hartshorne, for instance, rather than interpreting God as an unchanging absolute, emphasizes Gods sensitive and caring relationship with the world. A personal God enters into relationships in such a way that he is affected by the relationships, and to be affected by relationships is to change. So God too is in the process of growth and development.

Neurophysiology, is the study of how nerve cells, or neurons, receive and transmit information. Two types of phenomena are involved in processing nerve signals: electrical and chemical. Electrical events propagate a signal within a neuron, and chemical processes transmit the signal from one neuron to another neuron or to a muscle cell.

The signals conveying everything that human beings sense and think, and every motion they make, follows nerve pathways in the human body as waves of ions (atoms or groups of atoms that carries electric charges). Australian physiologist Sir John Eccles discovered many of the intricacies of this electrochemical signaling process, particularly the pivotal step in which a signal is conveyed from one nerve cell to another. He shared the 1963 Nobel Prize in physiology or medicine for this work, which he described in a 1965 Scientific American article.

A neuron is a long cell that has a thick central area containing the nucleus; it also has one long process called an axon and one or more short, bushy processes called dendrites. Dendrites receive impulses from other neurons. (The exceptions are sensory neurons, such as those that transmit information about temperature or touch, in which the signal is generated by specialized receptors in the skin.) These impulses are propagated electrically along the cell membrane to the end of the axon. At the tip of the axon the signal is chemically transmitted to an adjacent neuron or muscle cell.

Like all other cells, neurons contain charged ions: potassium and sodium (positively charged) and chlorine (negatively charged). Neurons differ from other cells in that they are able to produce a nerve impulse. A neuron is polarized - that is, it has an overall negative charge inside the cell membrane because of the high concentration of chlorine ions and low concentration of potassium and sodium ions. The concentration of these same ions is exactly reversed outside the cell. This charge differential represents stored electrical energy, sometimes referred to as membrane potential or resting potential. The negative charge inside the cell is maintained by two features. The first is the selective permeability of the cell membrane, which is more permeable to potassium than sodium. The second feature is sodium pumps within the cell membrane that actively pump sodium out of the cell. When depolarization occurs, this charge differential across the membrane is reversed, and a nerve impulse is produced.

Depolarization is a rapid change in the permeability of the cell membrane. When sensory input or any other kind of stimulating current is received by the neuron, the membrane permeability is changed, allowing a sudden influx of sodium ions into the cell. The high concentration of sodium, or action potential, changes the overalls charge within the cell from negative too positively. The locals change in ion concentration triggers similar reactions along the membrane, propagating the nerve impulse. After a brief period called the refractory period, during which the ionic concentration returned to resting potential, the neuron can repeat this process.

Nerve impulses travel at different speeds, depending on the cellular composition of a neuron. Where speed of impulse is important, as in the nervous system, axons are insulated with a membranous substance called myelin. The insulation provided by myelin maintains the ionic charge over long distances. Nerve impulses are propagated at specific points along the myelin sheath; these points are called the nodes of Ranvier. Examples of myelinated axons are those in sensory nerve fibers and nerves connected to skeletal muscles. In non-militated cells, the nerve impulse is propagated more diffusely.

When the electrical signal reaches the tip of an axon, it stimulates small presynaptic vesicles in the cell. These vesicles contain chemicals called neurotransmitters, which are released into the microscopic space between neurons (the synaptic cleft). The neurotransmitters attach to specialized receptors on the surface of the adjacent neuron. This stimulus causes the adjacent cell to depolarize and propagate an action potential of its own. The duration of a stimulus from a neurotransmitter is limited by the breakdown of the chemicals in the synaptic cleft and the reuptake by the neuron that produced them. Formerly, each neuron was thought to make only one transmitter, but recent studies have shown that some cells make two or more.

Philosophy of Mind considers mental phenomena such as sensation, perception, thought, belief, desire, intention, memory, emotion, imagination, and purposeful action. These phenomena, which can be broadly grouped as thoughts and experiences, are features of human beings; many of them are also found in other animals. Philosophers are interested in the nature of each of these phenomena as well as their relationships to one another and to physical phenomena, such as motion.

In the 17th century, French philosopher René Descartes proposed that only two substances ultimately exist; mind and body. Yet, if the two are entirely distinct, as Descartes believed, how can one substance interact with the other? How, for example, is the intention of a human mind able to cause movement in the persons limbs? The issue of the interaction between mind and body is known in philosophy as the mind-body problem.

Many fields other than a philosophy share an interest in the nature of mind. In religion, the nature of mind is connected with various conceptions of the soul and the possibility of life after death. In many abstract theories of mind there is considerable overlap between philosophy and the science of psychology. Once part of philosophy, psychology split off and formed a separate branch of knowledge in the 19th century. While psychology used scientific experiments to study mental states and events, philosophy uses reasoned arguments and thought experiments in seeking to understand the concepts that underlie mental phenomena. Also influenced by philosophy of mind is the field of artificial intelligence (AI), which endeavors to develop computers that can mimic what the human mind can do. Cognitive science attempts to integrate the understanding of mind provided by philosophy, psychology, AI, and other disciplines. Finally, all of these fields benefit from the detailed understanding of the brain that has emerged through neuroscience in the late 20th century.

Philosophers use the characteristics of inward accessibility, subjectivity, intentionality, goal-directedness, creativity and freedom, and consciousness to distinguish mental phenomena from physical phenomena.

Perhaps the most important characteristic of mental phenomena is that they are inwardly accessible, or available to us through introspection. We each know our own minds - our sensations, thoughts, memories, desires, and fantasies - in a direct sense, by internal reflection. We also know our mental states and mental events in a way that no one else can. In other words, we have privileged access to our own mental states.

Certain mental phenomena, those we generally call experiences, have a subjective nature - that is, they have certain characteristics we become aware of when we reflect. For instance, there is something it is like to feel pain, or have an itch, or see something red. These characteristics are subjective in that they are accessible to the subject of the experience, the person who has the experience, but not to others.

Other mental phenomena, which we broadly refer to as thoughts, have a characteristic philosophers call intentionality. Intentional thoughts are about other thoughts or objects, which are represented as having certain properties or for being related to one another in a certain way. The belief that California is west of Nevada, for example, is about California and Nevada and represents the former for being west of the latter. Although we have privileged access to our intentional states, many of them do not seem to have a subjective nature, at least not in the way that experiences do.

A number of mental phenomena appear to be connected to one another as elements in an intelligent, goal-directed system. The system works as follows: First, our sense organs are stimulated by events in our environment; next, by virtue of these stimulations, we perceive things about the external world; finally, we use this information, as well as information we have remembered or inferred, to guide our actions in ways that further our goals. Goal-directedness seems to accompany only mental phenomena.

Another important characteristic of mind, especially of human minds, is the capacity for choice and imagination. Rather than automatically converting past influences into future actions, individual minds are capable of exhibiting creativity and freedom. For instance, we can imagine things we have not experienced and can act in ways that no one expects or could predict.

Scientists have long considered the nature of consciousness without producing a fully satisfactory definition. In the early 20th century American philosopher and psychologist William James suggested that consciousness is a mental process involving both attention to external stimuli and short-term memory. Later scientific explorations of consciousness mostly expanded upon James work. In this article from a 1997 special issue of Scientific American, Nobel laureate Francis Crick, who helped determine the structure of DNA, and fellow biophysicists Christof Koch explains how experiments on vision might deepen our understanding of consciousness.

Mental phenomena are conscious, and consciousness may be the closest term we have for describing what is special about mental phenomena. Minds are sometimes referred to as consciousness, yet it is difficult to describe exactly what consciousness is. Although consciousness is closely related to inward accessibility and subjectivity, these very characteristics seem to hinder us in reaching an objective scientific understanding of it.

Although philosophers have written about mental phenomena since ancient times, the philosophy of mind did not garner much attention until the work of French philosopher René Descartes in the 17th century. Descartes work represented a turning point in thinking about mind by making a strong distinction between bodies and minds, or the physical and the mental. This duality between mind and body, known as Cartesian dualism, has posed significant problems for philosophy ever since.

In response to the mind-body problem arising from Descartes theory of substance dualism, a number of philosophers have advocated various forms of substance monism, the doctrine that there is ultimately just one kind of thing in reality. In the 18th century, Irish philosopher George Berkeley claimed there were no material objects in the world, only minds and their ideas. Berkeley thought that talk about physical objects was simply a way of organizing the flow of experience. Near the turn of the 20th century, American psychologist and philosopher William James proposed another form of substance monism. James claimed that experience is the basic stuff from which both bodies and minds are constructed.

Most philosophers of mind today are substance monists of a third type: They are materialists who believe that everything in the world is basically material, or a physical object. Among materialists, there is still considerable disagreement about the status of mental properties, which are conceived as properties of bodies or brains. Materialists who are property dualists believe that mental properties are an additional kind of property or attribute, not reducible to physical properties. Property dualists have the problem of explaining how such properties can fit into the world envisaged by modern physical science, according to which there are physical explanations for all things.

Materialists who are property monists believe that there is ultimately only one type of property, although they disagree on whether or not mental properties exist in material form. Some property monists, known as reductive materialists, hold that mental properties exist simply as a subset of relatively complex and nonbasic physical properties of the brain. Reductive materialists have the problem of explaining how the physical states of the brain can be inwardly accessible and have a subjective character, as mental states do. Other property monists, known as Eliminative materialists, consider the whole category of mental properties to be a mistake. According to them, mental properties should be treated as discredited postulates of an outmoded theory. Eliminative materialism is difficult for most people to accept, since we seem to have direct knowledge of our own mental phenomena by introspection and because we use the general principles we understand about mental phenomena to predict and explain the behavior of others.

Philosophy of mind seeks to answer such questions as, is mind distinct from matter? Can we define what it is to be conscious, and can we give principled reasons for deciding whether other creatures are conscious, or whether machines might be made so that they are conscious? What are thinking, feeling, experience remembering? Is it useful to divide the function of the mind up, separating memory from intelligence, or rationality from sentiment, or do mental functions form an integrated whole? The dominant philosophies of mind in the current western tradition include varieties of physicalism and functionalism. Ours’ concerns itself with a number of specialized problems. In addition to the mind-body problem, important issues include those of personal identity, as well as immortality, and artificial intelligence.

During much of Western history, the mind has been identified with the soul as presented in Christian Theology. According to Christianity, the soul is the source of a persons identity and is usually regarded as immaterial; thus, it is capable of enduring after the death of the body. Descartes conception of the mind as a separate, nonmaterial substance fits well with this understanding of the soul. In Descartes view, we are aware of our bodies only as the cause of sensations and other mental phenomena. Consequently our personal essence is composed more fundamentally of mind and the preservation of the mind after death would constitute our continued existence.

The mind conceived by materialist forms of substance monism does not fit as neatly with this traditional concept of the soul. With materialism, once a physical body is destroyed, nothing enduring remains. Some philosophers think that a concept of personal identity can be constructed that permits the possibility of life after death without appealing to separate immaterial substances. Following in the tradition of 17th-century British philosopher John Locke, these philosophers propose that a person consists of a stream of mental events linked by memory. These link of memory, rather than a single underlying substance, provide the unity of a single consciousness through time. Immortality is conceivable if we think of these memory links as connecting a later consciousness in heaven with an earlier one on earth.

No simple, agreed-upon definition of consciousness exists. Attempted definitions tend to be tautological (for example, consciousness defined as awareness) or merely descriptive (for example, consciousness described as sensations, thoughts, or feelings). Despite this problem of definition, the subject of consciousness has had a remarkable history. At one time the primary subject matter of psychology, consciousness as an area of study suffered an almost total demise, later reemerging to become a topic of current interest.

Most of the philosophical discussions of consciousness arose from the mind-body issues posed by the French philosopher and mathematician René Descartes in the 17th century. Descartes asked: Is the mind, or consciousness, independent of matter? Is consciousness extended (physical) or unextended (nonphysical)? Is consciousness determinative, or is it determined? English philosophers such as John Locke equated consciousness with physical sensations and the information they provide, whereas European philosophers such as Gottfried Wilhelm Leibniz and Immanuel Kant gave a more central and active role to consciousness.

The philosopher who most directly influenced subsequent exploration of the subject of consciousness was the 19th-century German educator Johann Friedrich Herbart, who wrote that ideas had quality and intensity and that they may inhibit or facilitate one another. Thus, ideas may pass from states of reality (consciousness) to states of tendencies (unconsciousness), with the dividing line between the two states being described as the threshold of consciousness. This formulation of Herbart clearly presages the development, by the German psychologist and physiologist Gustav Theodor Fechner, of the psychophysical measurement of sensation thresholds, and the later development by Sigmund Freud of the concept of the unconscious.

The experimental analysis of consciousness dates from 1879, when the German psychologist Wilhelm Max Wundt started his research laboratory. For Wundt, the task of psychology was the study of the structure of consciousness, which extended well beyond sensations and included feelings, images, memory, attention, duration, and movement. Because early interest focused on the content and dynamics of consciousness, it is not surprising that the central methodology of such studies was introspection; that is, subjects reported on the mental contents of their own consciousness. This introspective approach was developed most fully by the American psychologist Edward Bradford Titchener at Cornell University. Setting his task as that of describing the structure of the mind, Titchener attempted to detail, from introspective self-reports, the dimensions of the elements of consciousness. For example, taste was dimensionalized into four basic categories: sweet, sour, salt, and bitter. This approach was known as structuralism.

By the 1920s, however, a remarkable revolution had occurred in psychology that was to essentially remove considerations of consciousness from psychological research for some 50 years: Behaviorism captured the field of psychology. The main initiator of this movement was the American psychologist John Broadus Watson. In a 1913 article, Watson stated, I believe that we can write a psychology and never use the terms consciousness, mental states, mind . . . imagery and the like. Psychologists then turned almost exclusively to behavior, as described in terms of stimulus and response, and consciousness was totally bypassed as a subject. A survey of eight leading introductory psychology texts published between 1930 and the 1950s found no mention of the topic of consciousness in five texts, and in two it was treated as a historical curiosity.

Both Analytic and Linguistic philosophy, are 20th-century philosophical movements, and dominate the bettermost of Britain and the United States since World War II, that aims to clarify language and analyze the concepts expressed in it. The movement has been given a variety of designations, including linguistic analysis, logical empiricism, logical positivism, Cambridge analysis, and Oxford philosophy. The last two labels are derived from the universities in England where this philosophical method has been particularly influential. Although no specific doctrines or tenets are accepted by the movement as a whole, analytic and linguistic philosophers agree that the proper activity of philosophy is clarifying language, or, as some prefer, clarifying concepts. The aim of this activity is to settle philosophical disputes and resolve philosophical problems, which, it is argued, originates in linguistic confusion.

A considerable diversity of views exists among analytic and linguistic philosophers regarding the nature of conceptual or linguistic analysis. Some have been primarily concerned with clarifying the meaning of specific words or phrases as an essential step in making philosophical assertions clear and unambiguous. Others have been more concerned with determining the general conditions that must be met for any linguistic utterance to be meaningful; their intent is to establish a criterion that will distinguish between meaningful and nonsensical sentences. Still other analysts have been interested in creating formal, symbolic languages that are mathematical in nature. Their claim is that philosophical problems can be more effectively dealt with once they are formulated in a rigorous logical language.

By contrast, many philosophers associated with the movement have focused on the analysis of ordinary, or natural, language. Difficulties arise when concepts such as time and freedom, for example, are considered apart from the linguistic context in which they normally appear. Attention to language as it is ordinarily used is the key, it is argued, to resolving many philosophical puzzles.

Many experts believe that philosophy as an intellectual discipline originated with the works of Plato, one of the most celebrated philosophers in history. The Greek thinker had an immeasurable influence on Western thought. However, Platos expression of ideas in the form of dialogues - the dialectical method, used most famously by his teacher Socrates - has led to difficulties in interpreting some of the finer points of his thoughts. The issue of what exactly Plato meant to say is addressed in the following excerpt by author R.M. Hare.

Linguistic analysis as a method of philosophy is as old as the Greeks. Several of the dialogues of Plato, for example, are specifically concerned with clarifying terms and concepts. Nevertheless, this style of philosophizing has received dramatically renewed emphasis in the 20th century. Influenced by the earlier British empirical tradition of John Locke, George Berkeley, David Hume, and John Stuart Mill and by the writings of the German mathematician and philosopher Gottlob Frége, the 20th-century English philosophers G. E. Moore and Bertrand Russell became the founders of this contemporary analytic and linguistic trend. As students together at the University of Cambridge, Moore and Russell rejected Hegelian idealism, particularly as it was reflected in the work of the English metaphysician F. H. Bradley, who held that nothing is completely real except the Absolute. In their opposition to idealism and in their commitment to the view that careful attention to language is crucial in philosophical inquiry. They set the mood and style of philosophizing for much of the 20th century English-speaking world.

For Moore, philosophy was first and foremost analysis. The philosophical task involves clarifying puzzling propositions or concepts by indicating less puzzling propositions or concepts to which the originals are held to be logically equivalent. Once this task has been completed, the truth or falsity of problematic philosophical assertions can be determined more adequately. Moore was noted for his careful analyses of such puzzling philosophical claims as time is unreal, analyses that then aided in determining the truth of such assertions.

Russell, strongly influenced by the precision of mathematics, was concerned with developing an ideal logical language that would accurately reflect the nature of the world. Complex propositions, Russell maintained, can be resolved into their simplest components, which he called atomic propositions. These propositions refer to atomic facts, the ultimate constituents of the universe. The metaphysical views based on this logical analysis of language and the insistence that meaningful propositions must correspond to facts constitute what Russell called logical atomism. His interest in the structure of language also led him to distinguish between the grammatical form of a proposition and its logical form. The statements John is good and John is tall have the same grammatical form but different logical forms. Failure to recognize this would lead one to treat the property goodness as if it were a characteristic of John in the same way that the property tallness is a characteristic of John. Such failure results in philosophical confusion.

Austrian-born philosopher Ludwig Wittgenstein was one of the most influential thinkers of the 20th century. With his fundamental work, Tractatus Logico-philosophicus, published in 1921, he became a central figure in the movement known as analytic and linguistic philosophy.

Russells work in mathematics attracted at Cambridge, the Austrian philosopher Ludwig Wittgenstein, who became a central figure in the analytic and linguistic movement. In his first major work, Tractatus Logico-philosophicus (1921, translated, 1922), in which he first presented his theory of language, Wittgenstein argued that all philosophy is a critique of language and that philosophy aims at the logical clarification of thoughts. The results of Wittgensteins analysis resembled Russells logical atomism. The world, he argued, is ultimately composed of simple facts, which it is the purpose of language to picture. To be meaningful, statements about the world must be reducible to linguistic utterances that have a structure similar to the simple facts pictured. In this early Wittgensteinian analysis, only propositions that picture facts - the propositions of science - are considered factually meaningful. Metaphysical, theological, and ethical sentences were judged to be factually meaningless.

Influenced by Russell, Wittgenstein, Ernst Mach, and others, a group of philosophers and mathematicians in Vienna in the 1920s initiated the movement known as logical positivism: Led by Moritz Schlick and Rudolf Carnap, the Vienna Circle initiated one of the most important chapters in the history of analytic and linguistic philosophy. According to the positivist, the task of philosophy is the clarification of meaning, not the discovery of new facts (the job of the scientists) or the construction of comprehensive accounts of reality (the misguided pursuit of traditional metaphysics).

The positivist divided all meaningful assertions into two classes: analytic propositions and empirically verifiable ones. Analytic propositions, which include the propositions of logic and mathematics, are statements the truth or falsity of which depended altogether on the meanings of the terms constituting the statement. An example would be the proposition two plus two equals four. The second class of meaningful propositions includes all statements about the world that can be verified, at least in principle, by sense experience. Indeed, the meaning of such propositions is identified with the empirical method of their verification. This verifiability theory of meaning, the positivist concluded, would demonstrate that scientific statements are legitimate factual claims and that metaphysical, religious, and ethical sentences are factually empties. The ideas of logical positivism were made popular in England by the publication of A.J. Ayers Language, Truth and Logic in 1936.

The positivist verifiability theory of meaning came under intense criticism by philosophers such as the Austrian-born British philosopher Karl Popper. Eventually this narrow theory of meaning yielded to a broader understanding of the nature of language. Again, an influential figure was Wittgenstein. Repudiating many of his earlier conclusions in the Tractatus, he initiated a new line of thought culminating in his posthumously published Philosophical Investigations (1953; translated, 1953). In this work, Wittgenstein argued that once attention is directed to the way language is actually used in ordinary discourse, the variety and flexibility of language become clear. Propositions do much more than simply picture facts.

This recognition led to Wittgensteins influential concept of language games. The scientist, the poet, and the theologian, for example, are involved in different language games. Moreover, the meaning of a proposition must be understood in its context, that is, in terms of the rules of the language game of which that proposition is a part. Philosophy, concluded Wittgenstein, is an attempt to resolve problems that arise as the result of linguistic confusion, and the key to the resolution of such problems is ordinary language analysis and the proper use of language.

Additional contributions within the analytic and linguistic movement include the work of the British philosophers Gilbert Ryle, John Austin, and P. F. Strawson and the American philosopher W. V. Quine. According to Ryle, the task of philosophy is to restate systematically misleading expressions in forms that are logically more accurate. He was particularly concerned with statements the grammatical form of which suggests the existence of nonexistent objects. For example, Ryle is best known for his analysis that has of a mental capacity of language, language that misleadingly suggests that the mind is an entity in the same way as the body.

Austin maintained that one of the most fruitful starting points for philosophical inquiry is attention to the extremely fine distinctions drawn in ordinary language. His analysis of language eventually led to a general theory of speech acts, that is, to a description of the variety of activities that an individual may be performing when something is uttered.

Strawson is known for his analysis of the relationship between formal logic and ordinary language. The complexity of the latter, he argued, is inadequately represented by formal logic. A variety of analytic tools, therefore, are needed in addition to logic in analyzing ordinary language.

Quine discussed the relationship between language and ontology. He argued that language systems tend to commit their users to the existence of certain things. For Quine, the justification for speaking one way rather than another is a thoroughly pragmatic one.

The commitment to language analysis as a way of pursuing philosophy has continued as a significant contemporary dimension in philosophy. A division also continues to exist between those who prefer to work with the precision and rigor of symbolic logical systems and those who prefer to analyze ordinary language. Although few contemporary philosophers maintain that all philosophical problems are linguistic, the view continues to be widely held that attention to the logical structure of language and to how language is used in everyday discourse can often aid in resolving philosophical problems.

A loose title for various philosophies that emphasize certain common themes, the individual, the experience of choice, and the absence of rational understandings of the universe, with a consequent dread or sense of absurdity i9n human life however, Existentialism is a philosophical movement or tendency, emphasizing individual existence, freedom, and choice, that influenced many diverse writers in the 19th and 20th centuries.

Because of the diversity of positions associated with existentialism, the term is impossible to define precisely. Certain themes common to virtually all existentialist writers can, however, be identified. The term itself suggests one major theme: the stress on concrete individual existence and, consequently, on subjectivity, individual freedom, and choice.

Most philosophers since Plato have held that the highest ethical good are the same for everyone; insofar as one approaches moral perfection, one resembles other morally perfect individuals. The 19th-century Danish philosopher Søren Kierkegaard, who was the first writer to call himself existential, reacted against this tradition by insisting that the highest good for the individual are to find his or her own unique vocation. As he wrote in his journal, I must find a truth that is true for me . . . the idea for which I can live or die. Other existentialist writers have echoed Kierkegaards belief that one must choose ones own way without the aid of universal, objective standards. Against the traditional view that moral choice involves an objective judgment of right and wrong, Existentialists have argued that no objective, rational basis can be found for moral decisions. The 19th-century German philosopher Friedrich Nietzsche further contended that the individual must decide which situations are to count as moral situations.

All Existentialists have followed Kierkegaard in stressing the importance of passionate individual action in deciding questions of both morality and truth. They have insisted, accordingly, that personal experience and acting on ones own convictions are essential in arriving at the truth. Thus, the understanding of a situation by someone involved in that situation is superior to that of a detached, objective observer. This emphasis on the perspective of the individual agent has also made Existentialists suspicious of systematic reasoning. Kierkegaard, Nietzsche, and other existentialist writers have been deliberately unsystematic in the exposition of their philosophies, preferring to express themselves in aphorisms, dialogues, parables, and other literary forms. Despite their antirationalist position, however, most Existentialists cannot be said to be irrationalists in the sense of denying all validity to rational thought. They have held that rational clarity is desirable wherever possible, but that the most important questions in life are not accessible to reason or science. Furthermore, they have argued that even science is not as rational as is commonly supposed. Nietzsche, for instance, asserted that the scientific assumption of an orderly universe is for the most part a useful fiction.

Perhaps the most prominent theme in existentialist writing is that of choice. Humanities primary distinction, in the view of most Existentialists, is the freedom to choose. Existentialists have held that human beings do not have a fixed nature, or essence, as other animals and plants do; each human being makes choices that create his or her own nature. In the formulation of the 20th-century French philosopher Jean-Paul Sartre, existence precedes essence. Choice is therefore central to human existence, and it is inescapable; even the refusal to choose is a choice. Freedom of choice entails commitment and responsibility. Because individuals are free to choose their own path, Existentialists have argued, they must accept the risk and responsibility of following their commitment wherever it leads.

Kierkegaard held that it is spiritually crucial to recognize that one experiences not only a fear of specific objects but also a feeling of general apprehension, which he called dread. He interpreted it as Gods way of calling each individual to make a commitment to a personally valid way of life. The word anxiety (German Angst) has a similarly crucial role in the work of the 20th-century German philosopher Martin Heidegger; anxiety leads to the individuals confrontation with nothingness and with the impossibility of finding ultimate justification for the choices he or she must make. In the philosophy of Sartre, the word nausea is used for the individuals recognition of the pure contingency of the universe, and the word anguish is used for the recognition of the total freedom of choice that confronts the individual at every moment.

Existentialism as a distinct philosophical and literary movement belongs to the 19th and 20th centuries, but elements of existentialism can be found in the thought (and life) of Socrates, in the Bible, and in the work of many premodern philosophers and writers.

The first to anticipate the major concerns of modern existentialism was the 17th-century French philosopher Blaise Pascal. Pascal rejected the rigorous rationalism of his contemporary René Descartes, asserting, in his Pensées (1670), that a systematic philosophy that presumes to explain God and humanity is a form of pride. Like later existentialist writers, he saw human life in terms of paradoxes: The human self, which combines mind and body, is itself a paradox and contradiction.

Kierkegaard, generally regarded as the founder of modern existentialism, reacted against the systematic absolute idealism of the 19th-century German philosopher Georg Wilhelm Friedrich Hegel, who claimed to have worked out a total rational understanding of humanity and history. Kierkegaard, on the contrary, stressed the ambiguity and absurdity of the human situation. The individuals response to this situation must be to live a totally committed life, and this commitment can only be understood by the individual who has made it. The individual therefore must always be prepared to defy the norms of society for the sake of the higher authority of a personally valid way of life. Kierkegaard ultimately advocated a leap of faith into a Christian way of life, which, although incomprehensible and full of risk, was the only commitment he believed could save the individual from despair.

Danish religious philosopher Søren Kierkegaard rejected the all-encompassing, analytical philosophical systems of such 19th-century thinkers as German philosopher G.W.F. Hegel. Instead, Kierkegaard focused on the choices the individual must make in all aspects of his or her life, especially the choice to maintain religious faith. In Fear and Trembling (1846; translated 1941), Kierkegaard explored the concept of faith through an examination of the biblical story of Abraham and Isaac, in which God demanded that Abraham demonstrate his faith by sacrificing his son.

One of the most controversial works of 19th-century philosophy, Thus Spake Zarathustra (1883-1885) articulated German philosopher Friedrich Nietzsches theory of the Übermensch, a term translated as Superman or Overman. The Superman was an individual who overcame what Nietzsche termed the slave morality of traditional values, and lived according to his own morality. Nietzsche also advanced his idea that God is dead, or that traditional morality was no longer relevant in peoples lives. In this passage, the sage Zarathustra came down from the mountain where he had spent the last ten years alone to preach to the people.

Nietzsche, who was not acquainted with the work of Kierkegaard, influenced subsequent existentialist thought through his criticism of traditional metaphysical and moral assumptions and through his espousal of tragic pessimism and the life-affirming individual will that opposes itself to the moral conformity of the majority. In contrast to Kierkegaard, whose attack on conventional morality led him to advocate a radically individualistic Christianity, Nietzsche proclaimed the death of God and went on to reject the entire Judeo-Christian moral tradition in favor of a heroic pagan ideal.

The modern philosophy movements of phenomenology and existentialism have been greatly influenced by the thought of German philosopher Martin Heidegger. According to Heidegger, humankind has fallen into a crisis by taking a narrow, technological approach to the world and by ignoring the larger question of existence. People, if they wish to live authentically, must broaden their perspectives. Instead of taking their existence for granted, people should view themselves as part of Being (Heideggers terms for that which underlies all existence).

Heidegger, like Pascal and Kierkegaard, reacted against an attempt to put philosophy on a conclusive rationalistic basis - in this case the phenomenology of the 20th-century German philosopher Edmund Husserl. Heidegger argued that humanity finds itself in an incomprehensible, indifferent world. Human beings can never hope to understand why they are here; instead, each individual must choose a goal and follow it with passionate conviction, aware of the certainty of death and the ultimate meaninglessness of ones life. Heidegger contributed to existentialist thought an original emphasis on being and ontology (see Metaphysics) as well as on language.

Twentieth-century French intellectual Jean-Paul Sartre helped to develop existential philosophy through his writings, novels, and plays. Much as much of Sartres’ activities that affords the use of more words than needed to express an idea it finds in his work as it is focused on the predicament bewildered upon such plexuities that the dilemma of choice is faced by free individuals and on the particular and, yet, peculiar challenge of creating meaning by acting responsibly in an indifferent world. In stating that man is condemned to be free, Sartre reminds us of the responsibility that accompanies human decisions.

Sartre first gave the term existentialism general currency by using it for his own philosophy and by becoming the leading figure of a distinct movement in France that became internationally influential after World War II. Sartres philosophy is explicitly atheistic and pessimistic; he declared that human beings require a rational basis for their lives but are unable to achieve one, and thus human life is a futile passion. Sartre nevertheless insisted that his existentialism is a form of humanism, and he strongly emphasized human freedom, choice, and responsibility. He eventually tried to reconcile these existentialist concepts with a Marxist analysis of society and history.

Although existentialist thought encompasses the uncompromising atheism of Nietzsche and Sartre and the agnosticism of Heidegger, its origin in the intensely religious philosophies of Pascal and Kierkegaard foreshadowed its profound influence on 20th-century theologies. The 20th-century German philosopher Karl Jaspers, although he rejected explicit religious doctrines, influenced a contemporary theology through his preoccupation with transcendence and the limits of human experience. The German Protestant theologians Paul Tillich and Rudolf Bultmann, the French Roman Catholic theologian Gabriel Marcel, the Russian Orthodox philosopher Nikolay Berdyayev, and the German Jewish philosopher Martin Buber inherited many of Kierkegaards concerns, especially that a personal sense of authenticity and commitment is essential to religious faith.

Renowned as one of the most important writers in world history, 19th-century Russian author Fyodor Dostoyevsky wrote psychologically intense novels which probed the motivations and moral justifications for his characters actions. Dostoyevsky commonly addressed themes such as the struggle between good and evil within the human soul and the idea of salvation through suffering. The Brothers Karamazov (1879-1880), generally considered Dostoyevskys best work, interlaces religious exploration with the story of a families violent quarrels over a woman and a disputed inheritance.

Also, Maurice Merleau-Ponty (1908-1961), the French existentialist philosopher, whose phenomenological studies of the role of the body in perception and society opened a new field of philosophical investigation. He taught at the University of Lyon, at the Sorbonne, and, after 1952, at the Collège de France. His first important work was The Structure of Comportment (1942; translated 1963), a critique of behaviorism. His major work, Phenomenology of Perception (1945; translated 1962), is a detailed study of perception, influenced by the German philosopher Edmund Husserls phenomenology and by Gestalt psychology. In it, he argues that science presupposes an original and unique perceptual relation to the world that cannot be explained or even described in scientific terms. This book can be viewed as a critique of cognitivism - the view that the working of the human mind can be understood in terms of rules or programs. It is also a telling critique of the existentialism of his contemporary, Jean-Paul Sartre, showing how human freedom is never total, as Sartre claimed, but is limited by our embodiment.

With Sartre and Simone de Beauvoir, Merleau-Ponty founded an influential postwar French journal, Les Temps Modernes. His brilliant and timely essays on art, film, politics, psychology, and religion, first published in this journal, were later collected in Sense and Nonsense (1948; translated 1964). At the time of his death, he was working on a book, The Visible and the Invisible (1964; translated 1968), arguing that the whole perceptual world has the sort of organic unity he had earlier.

A number of existentialist philosophers used literary forms to convey their thought, and existentialism has been as vital and as extensive a movement in literature as in philosophy. The 19th-century Russian novelist Fyodor Dostoyevsky is probably the greatest existentialist literary figure. In Notes from the Underground (1864), the alienated antihero rages against the optimistic assumptions of rationalist humanism. The view of human nature that emerges in this and other novels of Dostoyevsky is that it is unpredictable and perversely self-destructive; only Christian love can save humanity from itself, but such love cannot be understood philosophically. As the character Alyosha says in The Brothers Karamazov (1879-80), We must love life more than the meaning of it.

The opening lines of Russian novelist Fyodor Dostoyevskys Notes from Underground (1864) - I am a sick man . . . I am a spiteful man - are among the most famous in 19th-century literature. Published five years after his release from prison and involuntary, military service in Siberia, Notes from Underground is a sign of Dostoyevskys rejection of the radical social thinking he had embraced in his youth. The unnamed narrator is antagonistic in tone, questioning the readers sense of morality as well as the foundations of rational thinking. In this excerpt from the beginning of the novel, the narrator describes himself, derisively referring to himself as an overly conscious intellectual.

In the 20th century, the novels of the Austrian Jewish writer Franz Kafka, such as The Trial (1925; translated 1937) and The Castle (1926; translated 1930), present isolated men confronting vast, elusive, menacing bureaucracies; Kafkas themes of anxiety, guilt, and solitude reflect the influence of Kierkegaard, Dostoyevsky, and Nietzsche. The influence of Nietzsche is also discernible in the novels of the French writers André Malraux and in the plays of Sartre. The work of the French writer Albert Camus is usually associated with existentialism because of the prominence in it of such themes as the apparent absurdity and futility of life, the indifference of the universe, and the necessity of engagement in a just cause. Existentialist themes are also reflected in the theater of the absurd, notably in the plays of Samuel Beckett and Eugene Unesco. In the United States, the influence of existentialism on literature has been more indirect and diffuse, but traces of Kierkegaards thought can be found in the novels of Walker Percy and John Updike, and various existentialist themes are apparent in the work of such diverse writers as Norman Mailer, John Barth, and Arthur

The problem of defining knowledge in terms of true belief plus some favoured relation between the believer and the facts began with Platos view in the Theaetetus, that knowledge is true belief plus some logos, an epistemology is to begin of holding the Foundations of knowledge, a special branches of philosophy that addresses the philosophical problems surrounding the theory of knowledge. Epistemology is concerned with the definition of knowledge and related concepts, the sources and criteria of knowledge, the kinds of knowledge possible and the degree to which each is certain, and the exact relation among the one who knows and the object known.

Thirteenth-century Italian philosopher and theologian Saint Thomas Aquinas attempted to synthesize Christian belief with a broad range of human knowledge, embracing diverse sources such as Greek philosopher Aristotle and Islamic and Jewish scholars. His thought exerted lasting influence on the development of Christian theology and Western philosophy. Author Anthony Kenny examines the complexities of Aquinas’ concepts of substance and accident.

In the 5th century Bc, the Greek Sophists questioned the possibility of reliable and objective knowledge. Thus, a leading Sophist, Georges, argued that nothing really exists, that if anything did exist it could not be known, and that if knowledge were possible, it could not be communicated. Another prominent Sophist, Protagoras, maintained that no persons opinions can be said to be more correct than anothers, because each is the sole judge of his or her own experience. Plato, following his illustrious teacher Socrates, tried to answer the Sophists by postulating the existence of a world of unchanging and invisible forms, or ideas, about which it is possible to have exact and certain knowledge. The things one sees and touches, they maintained, are imperfect copies of the pure forms studied in mathematics and philosophy. Accordingly, only the abstract reasoning of these disciplines yields genuine knowledge, whereas reliance on sense perception produces vague and inconsistent opinions. They concluded that philosophical contemplation of the unseen world of forms is the highest goal of human life.

Aristotle followed Plato in regarding abstract knowledge as superior to any other, but disagreed with him as to the proper method of achieving it. Aristotle maintained that almost all knowledge is derived from experience. Knowledge is gained either directly, by abstracting the defining traits of a species, or indirectly, by deducing new facts from those already known, in accordance with the rules of logic. Careful observation and strict adherence to the rules of logic, which were first set down in systematic form by Aristotle, would help guard against the pitfalls the Sophists had exposed. The Stoic and Epicurean schools agreed with Aristotle that knowledge originates in sense perception, but against both Aristotle and Plato they maintained that philosophy is to be valued as a practical guide to life, rather than as an end in itself.

After many centuries of declining interest in rational and scientific knowledge, the Scholastic philosopher Saint Thomas Aquinas and other philosophers of the Middle Ages helped to restore confidence in reason and experience, blending rational methods with faith into a unified system of beliefs. Aquinas followed Aristotle in regarding perception as the starting point and logic as the intellectual procedure for arriving at reliable knowledge of nature, but he considered faith in scriptural authority as the main source of religious belief.

From the 17th to the late 19th century, the main issue in epistemology was reasoning versus sense perception in acquiring knowledge. For the rationalists, of whom the French philosopher René Descartes, the Dutch philosopher Baruch Spinoza, and the German philosopher Gottfried Wilhelm Leibniz were the leaders, the main source and final test of knowledge was deductive reasoning based on self-evident principles, or axioms. For the empiricist, beginning with the English philosophers Francis Bacon and John Locke, the main source and final test of knowledge was sense perception.

Bacon inaugurated the new era of modern science by criticizing the medieval reliance on tradition and authority and also by setting down new rules of scientific method, including the first set of rules of inductive logic ever formulated. Locke attacked the rationalist belief that the principles of knowledge are intuitively self-evident, arguing that all knowledge is derived from experience, either from experience of the external world, which stamps sensations on the mind, or from internal experience, in which the mind reflects on its own activities. Human knowledge of external physical objects, he claimed, is always subject to the errors of the senses, and he concluded that one cannot have absolutely certain knowledge of the physical world.

Irish-born philosopher and clergyman George Berkeley (1685-1753) argued that everything that human being conceived of exists as an idea in a mind, a philosophical focus which is known as idealism. Berkeley reasoned that because one cannot control ones thoughts, they must come directly from a larger mind: that of God. In this excerpt from his Treatise Concerning the Principles of Human Knowledge, written in 1710, Berkeley explained why he believed that it is impossible . . . that there should be any such thing as an outward object.

The Irish philosopher George Berkeley agreed with Locke, that knowledge comes through ideas, but he denied Lockes belief that a distinction can be made between ideas and objects. The British philosopher David Hume continued the empiricist tradition, but he did not accept Berkeleys conclusion that knowledge was of ideas only. He divided all knowledge into two kinds: knowledge of relations of ideas - that is, the knowledge found in mathematics and logic, which is exact and certain but provide no information about the world; and knowledge of matters of fact - that is, the knowledge derived from sense perception. Hume argued that most knowledge of matters of fact depends upon cause and effect, and since no logical connection exists between any given cause and its effect, one cannot hope to know any future matter of fact with certainty. Thus, the most reliable laws of science might not remain true - a conclusion that had a revolutionary impact on philosophy.

The German philosopher Immanuel Kant tried to solve the crisis precipitated by Locke and brought to a climax by Hume; his proposed solution combined elements of rationalism with elements of empiricism. He agreed with the rationalists that one can have exact and certain knowledge, but he followed the empiricist in holding that such knowledge is more informative about the structure of thought than about the world outside of thought. He distinguished three kinds of knowledge: analytical a priori, which is exact and certain but uninformative, because it makes clear only what is contained in definitions; synthetic a posteriori, which conveys information about the world learned from experience, but is subject to the errors of the senses; and synthetic a priori, which is discovered by pure intuition and is both exact and certain, for it expresses the necessary conditions that the mind imposes on all objects of experience. Mathematics and philosophy, according to Kant, provide this last. Since the time of Kant, one of the most frequently argued questions in philosophy has been whether or not such a thing as synthetic a priori knowledge really exists.

During the 19th century, the German philosopher Georg Wilhelm Friedrich Hegel revived the rationalist claim that absolutely certain knowledge of reality can be obtained by equating the processes of thought, of nature, and of history. Hegel inspired an interest in history and a historical approach to knowledge that was further emphasized by Herbert Spencer in Britain and by the German school of historicism. Spencer and the French philosopher Auguste Comte brought attention to the importance of sociology as a branch of knowledge, and both extended the principles of empiricism to the study of society.

The American school of pragmatism, founded by the philosophers Charles Sanders Peirce, William James, and John Dewey at the turn of this century, carried empiricism further by maintaining that knowledge is an instrument of action and that all beliefs should be judged by their usefulness as rules for predicting experiences.

In the early 20th century, epistemological problems were discussed thoroughly, and subtle shades of difference grew into rival schools of thought. Special attention was given to the relation between the act of perceiving something, the object directly perceived, and the thing that can be said to be known as a result of the perception. The phenomenalists contended that the objects of knowledge are the same as the objects perceived. The neorealists argued that one has direct perceptions of physical objects or parts of physical objects, rather than of ones own mental states. The critical realists took a middle position, holding that although one perceives only sensory data such as colors and sounds, these stand for physical objects and provide knowledge thereof.

A method for dealing with the problem of clarifying the relation between the act of knowing and the object known was developed by the German philosopher Edmund Husserl. He outlined an elaborate procedure that he called phenomenology, by which one is said to be able to distinguish the way things appear to be from the way one thinks they really are, thus gaining a more precise understanding of the conceptual foundations of knowledge.

Scientific knowledge reveals that scientific knowledge and method did not spring full-blown from the minds of the ancient Greeks any more than language and culture emerged fully in the minds of Homo sapiens sapient. Scientific knowledge is an extension of ordinary language into greater levels of abstraction and precision through reliance upon geometric and numerical relationships. We speculate that the seeds of the scientific imagination were planted in ancient Greece, as opposed to Chinese or Babylonian culture, partly because the social, political and an economic climate in Greece was more open because the social, political, and economic climates in Greece was more open to the pursuit of knowledge with marginal cultural utility. Another important factor was that the special character of Homeric religion allowed the Greeks to invent a conceptual framework that would prove useful in future scientific investigations, but it was only after this inheritance from Greek philosophy was wed to some essential features of Judeo-Christian beliefs about the origin of the cosmos that the paradigm for classical physics emerged.

During the second quarter of the 20th century, two schools of thought emerged, each indebted to the Austrian philosopher Ludwig Wittgenstein. The first of these schools, logical empiricism, or logical positivism, had its origins in Vienna, Austria, but it soon spread to England and the United States. The logical empiricist insisted that there is only one kind of knowledge: scientific knowledge; that any valid knowledge claim must be verifiable in experience; and hence that much that had passed for philosophy was neither true nor false but literally meaningless. Finally, following Hume and Kant, a clear distinction must be maintained between analytic and synthetic statements. The so-called verifiability criterion of meaning has undergone changes as a result of discussions among the logical empiricists themselves, as well as their critics, but has not been discarded. More recently, the sharp distinction between the Analytic and the synthetic has been attacked by a number of philosophers, chiefly by American philosopher W.V.O. Quine, whose overall approach is in the pragmatic tradition.

The latter of these recent schools of thought, generally referred to as linguistic analysis, or ordinary language philosophy, seem to break with traditional epistemology. The linguistic analysts undertake to examine the actual way key epistemological terms are used - terms such as knowledge, perception, and probability - and to formulate definitive rules for their use in order to avoid verbal confusion. British philosopher John Langshaw Austin argued, for example, that to say a statement was truly added but nothing to the statement except a promise by the speaker or writer. Austin does not consider truth a quality or property attaching to statements or utterances. However, the ruling thought is that it is only through a correct appreciation of the role and point of this language that we can come to a better conception of what the language is about, and avoid the oversimplifications and distortion we are apt to bring to its subject matter.

Linguistics is the scientific study of language. It encompasses the description of languages, the study of their origin, and the analysis of how children acquire language and how people learn languages other than their own. Linguistics is also concerned with relationships between languages and with the ways languages change over time. Linguists may study language as a thought process and seek a theory that accounts for the universal human capacity to produce and understand language. Some linguists examine language within a cultural context. By observing talk, they try to determine what a person needs to know in order to speak appropriately in different settings, such as the workplace, among friends, or among family. Other linguists focus on what happens when speakers from different language and cultural backgrounds interact. Linguists may also concentrate on how to help people learn another language, using what they know about the learners first language and about the language being acquired.

Although there are many ways of studying language, most approaches belong to one of the two main branches of linguistics: descriptive linguistics and comparative linguistics.

Descriptive linguistics is the study and analysis of spoken language. The techniques of descriptive linguistics were devised by German American anthropologist Franz Boas and American linguist and anthropologist Edward Sapir in the early 1900s to record and analyze Native American languages. Descriptive linguistics begins with what a linguist hears native speakers say. By listening to native speakers, the linguist gathered a body of data and analyzed it in order to identify distinctive sounds, called phonemes. Individual phonemes, such as /p/ and /b/, are established on the grounds that substitution of one for the other changes the meaning of a word. After identifying the entire inventory of sounds in a language, the linguist looks at how these sounds combine to create morphemes, or units of sound that carry meaning, such as the words push and bush. Morphemes may be individual words such as push; root words, such as a berry in blueberry; or prefixes (pre-in previews) and suffixes (-nesses in openness).

The linguists next step is to see how morphemes combine into sentences, obeying both the dictionary meaning of the morpheme and the grammatical rules of the sentence. In the sentence She pushed the bush, the morpheme she, a pronoun, is the subject; push, a transitive verb, is the verb the, a definite article, is the determiner; and bush, a noun, is the object. Knowing the function of the morphemes in the sentence enables the linguist to describe the grammar of the language. The scientific procedures of phonemics (finding phonemes), morphology (discovering morphemes), and syntaxes (describing the order of morphemes and their function) provided descriptive linguists with a way to write down grammars of languages never before written down or analyzed. In this way they can begin to study and understand these languages.

Comparative linguistics is the study and analysis, by means of written records, of the origins and relatedness of different languages. In 1786 Sir William Jones, a British scholar, asserted that Sanskrit, Greek, and Latin was related to one another and had descended from a common source. He based this assertion on observations of similarities in sounds and meanings among the three languages. For example, the Sanskrit word bhratar for brother resembles the Latin word frater, the Greek word phrater, (and the English word brother).

Other scholars went on to compare Icelandic with Scandinavian languages, and Germanic languages with Sanskrit, Greek, and Latin. The correspondences among languages, known as genetic relationships, came to be represented on what comparative linguists refer to as family trees. Family trees established by comparative linguists include the Indo-European, relating Sanskrit, Greek, Latin, German, English, and other Asian and European languages; the Algonquian, relating Fox, Cree, Menomini, Ojibwa, and other Native North American languages; and the Bantu, relating Swahili, Xhosa, Zulu, Kikuyu, and other African languages.

Comparative linguists also look for similarities in the way words are formed in different languages. Latin and English, for example, change the form of a word to express different meanings, as when the English verb goes interchangeably to go and gone, only to express a past action. Chinese, on the other hand, has no such inflected forms; the verb remains the same while other words indicate the time (as, in go store tomorrow). In Swahili, prefixes, suffixes, and infixes (additions in the body of the word) combine with a root word to change its meaning. For example, a single word might expressibly go when something was done, by whom, to whom, and in what manner.

Some comparative linguists reconstruct hypothetical ancestral languages known as proto-languages, which they use to demonstrate relatedness among contemporary languages. A proto-language is not intended to depict a real language, however, and does not represent the speech of ancestors of people speaking modern languages. Unfortunately, some groups have mistakenly used such reconstructions in efforts to demonstrate the ancestral homeland of those people.

Comparative linguists have suggested that certain basic words in a language do not change over time, because people are reluctant to introduce new words for such constants as arm, eye, or mother. These words are termed culture free. By comparing lists of culture-free words in languages within a family, linguists can derive the percentage of related words and use a formula to figure out when the languages separated from one another.

By the 1960s comparativists were no longer satisfied with focusing on origins, migrations, and the family tree method. They challenged as unrealistic the notion that an earlier language could remain sufficiently isolated for other languages to be derived exclusively from it over a period of time. Today comparativists seek to understand the more complicated reality of language history, taking language contact into account. They are concerned with universal characteristics of language and with comparisons of grammars and structures.

The field of linguistics both borrows from and lends its own theories and methods to other disciplines. The many subfields of linguistics have expanded our understanding of languages. Linguistic theories and methods are also used in other fields of study. These overlapping interests have led to the creation of several cross-disciplinary fields.

Socio-linguistics is the study of patterns and variations in language within a society or community. It focuses on the way people use language to express social class, group status, gender, or ethnicity, and it looks at how they make choices about the form of language they use. It also examines the way people use language to negotiate their role in society and to achieve positions of power. For example, sociolinguistic studies have found that the way a New Yorker pronounces the phoneme /r/ in an expression such as fourth floor can indicate the persons social class. According to one study, people aspiring to move from the lower middle classes to the upper middle class attach prestige to pronouncing the /r/. Sometimes they even overcorrect their speech, pronouncing a /r/ where those whom they wish to copy may not.

Some sociolinguists believe that analyzing such variables as the use of a particular phoneme can predict the direction of language change. Change, they say, moves toward the variable associated with power, prestige, or other quality having high social value. Other sociolinguists focus on what happens when speakers of different languages interact. This approach to language change emphasizes the way languages mix rather than the direction of change within a community. The goal of Sociolinguistics is to understand communicative competence - what people need to know to use the appropriate language for a given social setting.

Psycholinguistics merge the fields of psychology and linguistics to study how people process language and how language use is related to underlying mental processes. Studies of childrens language acquisition and of second-language acquisition are psycholinguistic in nature. Psycholinguists work to develop models for how language is processed and understood, using evidence from studies of what happens when these processes go awry. They also study language disorders such as aphasia (impairment of the ability to use or comprehend words) and dyslexia (impairment of the ability to make out written language).

Computational linguistics involves the use of computers to compile linguistic data, analyze languages, translate from one language to another, and develop and test models of language processing. Linguists use computers and large samples of actual language to analyze the relatedness and the structure of languages and to look for patterns and similarities. Computers also aid in stylistic studies, information retrieval, various forms of textual analysis, and the construction of dictionaries and concordances. Applying computers to language studies has resulted in machine translation systems and machines that recognize and produce speech and text. Such machines facilitate communication with humans, including those who are perceptually or linguistically impaired.

Applied linguistics employs linguistic theory and methods in teaching and in research on learning a second language. Linguists look at the errors people make as they learn another language and at their strategies for communicating in the new language at different degrees of competence. In seeking to understand what happens in the mind of the learner, applied linguists recognize that motivation, attitude, learning style, and personality affect how well a person learns another language.

Anthropological linguistics, also known as linguistic anthropology, uses linguistic approaches to analyze culture. Anthropological linguists examine the relationship between a culture and its language. The way cultures and languages have changed over time, and how different cultures and languages are related to one another. For example, the present English use of family and given names arose in the late 13th and early 14th centuries when the laws concerning registration, tenure, and inheritance of property were changed.

Philosophical linguistics examines the philosophy of language. Philosophers of language search for the grammatical principles and tendencies that all human languages share. Among the concerns of linguistic philosophers is the range of possible word order combinations throughout the world. One finding is that 95 percent of the worlds languages use a subject-verb-object (SVO) order as English does (She pushed the bush.). Only 5 percent use a subject-object-verb (SOV) order or verb-subject-object (VSO) order.

Neurolinguistics are the study of how language is processed and represented in the brain. Neurolinguists seek to identify the parts of the brain involved with the production and understanding of language and to determine where the components of language (phonemes, morphemes, and structure or syntax) are stored. In doing so, they make use of techniques for analyzing the structure of the brain and the effects of brain damage on language.

Speculation about language goes back thousands of years. Ancient Greek philosophers speculated on the origins of language and the relationship between objects and their names. They also discussed the rules that govern language, or grammar, and by the 3rd century Bc they had begun grouping words into parts of speech and devising names for different forms of verbs and nouns.

In India religion provided the motivation for the study of language nearly 2500 years ago. Hindu priests noted that the language they spoke had changed since the compilation of their ancient sacred texts, the Vedas, starting about 1000 Bc. They believed that for certain religious ceremonies based upon the Vedas to succeed, they needed to reproduce the language of the Vedas precisely. Panini, an Indian grammarian who lived about 400 Bc, produced the earliest work describing the rules of Sanskrit, the ancient language of India.

The Romans used Greek grammars as models for their own, adding commentary on Latin style and usage. Statesman and orator Marcus Tullius Cicero wrote on rhetoric and style in the 1st century Bc. Later grammarians Aelius Donatus (4th century AD) and Priscian (6th century AD) produced detailed Latin grammars. Roman works served as textbooks and standards for the study of language for more than 1000 years.

It was not until the end of the 18th century that language was researched and studied in a scientific way. During the 17th and 18th centuries, modern languages, such as French and English, replaced Latin as the means of universal communication in the West. This occurrence, along with developments in printing, meant that many more texts became available. At about this time, the study of phonetics, or the sounds of a language, began. Such investigations led to comparisons of sounds in different languages; in the late 18th century the observation of correspondences among Sanskrit, Latin, and Greek gave birth to the field of Indo-European linguistics.

During the 19th century, European linguists focused on philology, or the historical analysis and comparison of languages. They studied written texts and looked for changes over time or for relationships between one language and another.

American linguist, writer, teacher, and political activist Noam Chomsky is considered the founder of transformational-generative linguistic analysis, which revolutionized the field of linguistics. This system of linguistics treats grammar as a theory of language - that is, Chomsky believes that in addition to the rules of grammar specific to individual languages, there are universal rules common to all languages that indicate that the ability to form and understand language is innate to all human beings. Chomsky also is well known for his political activism - he opposed United States involvement in Vietnam in the 1960s and 1970s and has written various books and articles and delivered many lectures in an attempt to educate and empower people on various political and social issues.

In the early 20th century, linguistics expanded to include the study of unwritten languages. In the United States linguists and anthropologists began to study the rapidly disappearing spoken languages of Native North Americans. Because many of these languages were unwritten, researchers could not use historical analysis in their studies. In their pioneering research on these languages, anthropologists Franz Boas and Edward Sapir developed the techniques of descriptive linguistics and theorized on the ways in which language shapes our perceptions of the world.

An important outgrowth of descriptive linguistics is a theory known as structuralism, which assumes that language is a system with a highly organized structure. Structuralism began with publication of the work of Swiss linguist Ferdinand de Saussure in Cours de linguistique générale (1916; Course in General Linguistics, 1959). This work, compiled by Saussures students after his death, is considered the foundation of the modern field of linguistics. Saussure made a distinction between actual speech, or spoken language, and the knowledge underlying speech that speakers share about what is grammatical. Speech, he said, represents instances of grammar, and the linguist’s task is to find the underlying rules of a particular language from examples found in speech. To the Structuralists, grammar is a set of relationships that account for speech, rather than a set of instances of speech, as it is to the descriptivist.

Once linguists began to study language as a set of abstract rules that somehow account for speech, other scholars began to take an interest in the field. They drew analogies between language and other forms of human behavior, based on the belief that a shared structure underlies many aspects of a culture. Anthropologists, for example, became interested in a Structuralists approach to the interpretation of kinship systems and analysis of myth and religion. American linguist Leonard Bloomfield promoted structuralism in the United States.

Saussures ideas also influenced European linguistics, most notably in France and Czechoslovakia (now the Czech Republic). In 1926 Czech linguist Vilem Mathesius founded the Linguistic Circle of Prague, a group that expanded the focus of the field to include the context of language use. The Prague circle developed the field of phonology, or the study of sounds, and demonstrated that universal features of sounds in the languages of the world interrelate in a systematic way. Linguistic analysis, they said, should focus on the distinctiveness of sounds rather than on the ways they combine. Where descriptivists tried to locate and describe individual phonemes, such as /b/ and /p/, the Prague linguists stressed the features of these phonemes and their interrelationships in different languages. In English, for example, the voice distinguishes between the similar sounds of /b/ and /p/, but these are not distinct phonemes in a number of other languages. An Arabic speaker might pronounce the cities Pompei and Bombay the same way.

As linguistics developed in the 20th century, the notion became prevalent that language is more than speech - specifically, that it is an abstract system of interrelationships shared by members of a speech community. Structural linguistics led linguists to look at the rules and the patterns of behavior shared by such communities. Whereas structural linguists saw the basis of language in the social structure, other linguists looked at language as a mental process.

The 1957 publication of Syntactic Structures by American linguist Noam Chomsky initiated what many view as a scientific revolution in linguistics. Chomsky sought a theory that would account for both linguistic structure and the creativity of language - the fact that we can create entirely original sentences and understand sentences never before uttered. He proposed that all people have an innate ability to acquire language. The task of the linguist, he claimed, is to describe this universal human ability, known as language competence, with a grammar from which the grammars of all languages could be derived. The linguist would develop this grammar by looking at the rules children use in hearing and speaking their first language. He termed the resulting model, or grammar, a transformational-generative grammar, referring to the transformations (or rules) that generate (or account for) language. Certain rules, Chomsky asserted, are shared by all languages and form part of a universal grammar, while others are language specific and associated with particular speech communities. Since the 1960s much of the development in the field of linguistics has been a reaction to or against Chomskys theories.

At the end of the 20th century, linguists used the term grammar primarily to refer to a subconscious linguistic system that enables people to produce and comprehend an unlimited number of utterances. Grammar thus accounts for our linguistic competence. Observations about the actual language we use, or language performance, are used to theorize about this invisible mechanism known as grammar.

The orientation toward the scientific study of language led by Chomsky has had an impact on nongenerative linguists as well. Comparative and historically oriented linguists are looking for the various ways linguistic universals show up in individual languages. Psycholinguists, interested in language acquisition, are investigating the notion that an ideal speaker-hearer is the origin of the acquisition process. Sociolinguists are examining the rules that underlie the choice of language variants, or codes, and allow for switching from one code to another. Some linguists are studying language performance - the way people use language - see how it reveals a cognitive ability shared by all human beings. Others seek to understand animal communication within such a framework. What mental processes enable chimpanzees to make signs and communicate with one another and how do these processes differ from those of humans?

A written bibliographic note in gratification To Ludwig Wittgenstein (1889-1951), an Austrian-British philosopher, who was one of the most influential thinkers of the 20th century, particularly noted for his contribution to the movement known as analytic and linguistic philosophy.

Born in Vienna on April 26, 1889, Wittgenstein was raised in a wealthy and cultured family. After attending schools in Linz and Berlin, he went to England to study engineering at the University of Manchester. His interest in pure mathematics led him to Trinity College, University of Cambridge, to study with Bertrand Russell. There he turned his attention to philosophy. By 1918 Wittgenstein had completed his Tractatus Logico-philosophicus (1921; translated 1922), a work he then believed provided the final solution to philosophical problems. Subsequently, he turned from philosophy and for several years taught elementary school in an Austrian village. In 1929 he returned to Cambridge to resume his work in philosophy and was appointed to the faculty of Trinity College. Soon he began to reject certain conclusions of the Tractatus and to develop the position reflected in his Philosophical Investigations (pub. posthumously 1953; translated 1953). Wittgenstein retired in 1947; he died in Cambridge on April 29, 1951. A sensitive, intense man who often sought solitude and was frequently depressed, Wittgenstein abhorred pretense and was noted for his simple style of life and dress. The philosopher was forceful and confident in personality, however, and he exerted considerable influence on those with whom he came in contact.

Wittgensteins’ philosophical life may be divided into two distinct phases: an early period, represented by the Tractatus, and a later period, represented by the Philosophical Investigations. Throughout most of his life, however, Wittgenstein consistently viewed philosophy as linguistic or conceptual analysis. In the Tractatus he argued that philosophy aims at the logical clarification of thoughts. In the Philosophical Investigations, however, he maintained that philosophy is a battle against the bewitchment of our intelligence by means of language.

Language, Wittgenstein argued in the Tractatus, is composed of complex propositions that can be analyzed into fewer complex propositions until one arrives at simple or elementary propositions. Correspondingly, the world is composed of complex facts that can be analyzed into fewer complex facts until one arrives at simple, or atomic, facts. The world is the totality of these facts. According to Wittgensteins picture theory of meaning, it is the nature of elementary propositions logically to picture atomic facts, or states of affairs. He claimed that the nature of language required elementary propositions, and his theory of meaning required that there be atomic facts pictured by the elementary propositions. On this analysis, only propositions that picture facts - the propositions of science - are considered cognitively meaningfully. Metaphysical and ethical statements are not meaningful assertions. The logical positivist associated with the Vienna Circle were greatly influenced by this conclusion.

Wittgenstein came to believe, however, that the narrow view of language reflected in the Tractatus was mistaken. In the Philosophical Investigations he argued that if one actually looks to see how language is used, the variety of linguistic usage becomes clear. Words are like tools, and just as tools serve different functions, so linguistic expressions serve many functions. Although some propositions are used to picture facts, others are used to command, question, pray, thank, curse, and so on. This recognition of linguistic flexibility and variety led to Wittgensteins concept of a language game and to the conclusion that people play different language games. The scientist, for example, is involved in a different language game than the theologian. Moreover, the meaning of a proposition must be understood in terms of its context, that is, in terms of the rules of the game of which that proposition is a part. The key to the resolution of philosophical puzzles is the therapeutic process of examining and describing language in use.

Analytic and Linguistic Philosophy, is a product out of the 20th-century philosophical movement, and dominant in Britain and the United States since World War II, that aims to clarify language and analyze the concepts expressed in it. The movement has been given a variety of designations, including linguistic analysis, logical empiricism, logical positivism, Cambridge analysis, and Oxford philosophy. The last two labels are derived from the universities in England where this philosophical method has been particularly influential. Although no specific doctrines or tenets are accepted by the movement as a whole, analytic and linguistic philosophers agree that the proper activity of philosophy is clarifying language, or, as some prefer, clarifying concepts. The aim of this activity is to settle philosophical disputes and resolve philosophical problems, which, it is argued, originates in linguistic confusion.

A considerable diversity of views exists among analytic and linguistic philosophers regarding the nature of conceptual or linguistic analysis. Some have been primarily concerned with clarifying the meaning of specific words or phrases as an essential step in making philosophical assertions clear and unambiguous. Others have been more concerned with determining the general conditions that must be met for any linguistic utterance to be meaningful; their intent is to establish a criterion that will distinguish between meaningful and nonsensical sentences. Still other analysts have been interested in creating formal, symbolic languages that are mathematical in nature. Their claim is that philosophical problems can be more effectively dealt with once they are formulated in a rigorous logical language.

By contrast, many philosophers associated with the movement have focused on the analysis of ordinary, or natural, language. Difficulties arise when concepts such as time and freedom, for example, are considered apart from the linguistic context in which they normally appear. Attention to language as it is ordinarily used is the key, it is argued, to resolving many philosophical puzzles.

Linguistic analysis as a method of philosophy is as old as the Greeks. Several of the dialogues of Plato, for example, are specifically concerned with clarifying terms and concepts. Nevertheless, this style of philosophizing has received dramatically renewed emphasis in the 20th century. Influenced by the earlier British empirical tradition of John Locke, George Berkeley, David Hume, and John Stuart Mill and by the writings of the German mathematician and philosopher Gottlob Frége, the 20th-century English philosophers G. E. Moore and Bertrand Russell became the founders of this contemporary analytic and linguistic trend. As students together at the University of Cambridge, Moore and Russell rejected Hegelian idealism, particularly as it was reflected in the work of the English metaphysician F. H. Bradley, who held that nothing is completely real except the Absolute. In their opposition to idealism and in their commitment to the view that careful attention to language is crucial in philosophical inquiry. They set the mood and style of philosophizing for much of the 20th century English-speaking world.

For Moore, philosophy was first and foremost analysis. The philosophical task involves clarifying puzzling propositions or concepts by indicating fewer puzzling propositions or concepts to which the originals are held to be logically equivalent. Once this task has been completed, the truth or falsity of problematic philosophical assertions can be determined more adequately. Moore was noted for his careful analyses of such puzzling philosophical claims as time is unreal, analyses that then aided in determining the truth of such assertions.

Russell, strongly influenced by the precision of mathematics, was concerned with developing an ideal logical language that would accurately reflect the nature of the world. Complex propositions, Russell maintained, can be resolved into their simplest components, which he called atomic propositions. These propositions refer to atomic facts, the ultimate constituents of the universe. The metaphysical views based on this logical analysis of language and the insistence that meaningful propositions must correspond to facts constitute what Russell called logical atomism. His interest in the structure of language also led him to distinguish between the grammatical form of a proposition and its logical form. The statements John is good and John is tall have the same grammatical form but different logical forms. Failure to recognize this would lead one to treat the property goodness as if it were a characteristic of John in the same way that the property tallness is a characteristic of John. Such failure results in philosophical confusion.

Russells work in mathematics attracted to Cambridge. The Austrian philosopher Ludwig Wittgenstein, who became a central figure in the analytic and linguistic movement. In his first major work, Tractatus Logico-philosophicus (1921; translated 1922), in which he first presented his theory of language, Wittgenstein argued that all philosophy is a critique of language and that philosophy aims at the logical clarification of thoughts. The results of Wittgensteins analysis resembled Russells logical atomism. The world, he argued, is ultimately composed of simple facts, which it is the purpose of language to picture. To be meaningful, statements about the world must be reducible to linguistic utterances that have a structure similar to the simple facts pictured. In this early Wittgensteinian analysis, only propositions that picture facts - the propositions of science - are considered factually meaningful. Metaphysical, theological, and ethical sentences were judged to be factually meaningless continues to exist between.

Influenced by Russell, Wittgenstein, Ernst Mach, and others, a group of philosophers and mathematicians in Vienna in the 1920s initiated the movement known as logical positivism. Led by Moritz Schlick and Rudolf Carnap, the Vienna Circle initiated one of the most important chapters in the history of analytic and linguistic philosophy. According to the positivist, the task of philosophy is the clarification of meaning, not the discovery of new facts (the job of the scientists) or the construction of comprehensive accounts of reality (the misguided pursuit of traditional metaphysics).

The positivist divided all meaningful assertions into two classes: analytic propositions and empirically verifiable ones. Analytic propositions, which include the propositions of logic and mathematics, are statements the truth or falsity of which depend together on the meanings of the terms constituting the statement. An example would be the proposition two plus two equals four. The second class of meaningful propositions includes all statements about the world that can be verified, at least in principle, by sense experience. Indeed, the meaning of such propositions is identified with the empirical method of their verification. This verifiability theory of meaning, the positivist concluded, would demonstrate that scientific statements are legitimate factual claims and that metaphysical, religious, and ethical sentences are factually empties. The ideas of logical positivism were made popular in England by the publication of A.J. Ayers Language, Truth and Logic in 1936.

The positivist verifiability theory of meaning came under intense criticism by philosophers such as the Austrian-born British philosopher Karl Popper. Eventually this narrow theory of meaning yielded to a broader understanding of the nature of language. Again, an influential figure was Wittgenstein. Repudiating many of his earlier conclusions in the Tractatus, he initiated a new line of thought culminating in his posthumously published Philosophical Investigations (1953; translated 1953). In this work, Wittgenstein argued that once attention is directed to the way language is actually used in ordinary discourse, the variety and flexibility of language become clear. Propositions do much more than simply picture facts.

This recognition led to Wittgensteins influential concept of language games. The scientist, the poet, and the theologian, for example, are involved in different language games. Moreover, the meaning of a proposition must be understood in its context, that is, in terms of the rules of the language game of which that proposition is a part. Philosophy, concluded Wittgenstein, is an attempt to resolve problems that arise as the result of linguistic confusion, and the key to the resolution of such problems is ordinary language analysis and the proper use of language.

Additional contributions within the analytic and linguistic movement include the work of the British philosophers Gilbert Ryle, John Austin, and P. F. Strawson and the American philosopher W. V. Quine. According to Ryle, the task of philosophy is to restate systematically misleading expressions in forms that are logically more accurate. He was particularly concerned with statements the grammatical form of which suggests the existence of nonexistent objects. For example, Ryle is best known for his analysis of mentalistic language, language that misleadingly suggests that the mind is an entity in the same way as the body.

Austin maintained that one of the most fruitful starting points for philosophical inquiry is attention to the extremely fine distinctions drawn in ordinary language. His analysis of language eventually led to a general theory of speech acts, that is, to a description of the variety of activities that an individual may be performing when something is uttered.

Strawson is known for his analysis of the relationship between formal logic and ordinary language. The complexity of the latter, he argued, is inadequately represented by formal logic. A variety of analytic tools, therefore, are needed in addition to logic in analyzing ordinary language.

Quine discussed the relationship between language and ontology. He argued that language systems tend to commit their users to the existence of certain things. For Quine, the justification for speaking one way rather than another is a thoroughly pragmatic one.

The commitment to language analysis as a way of pursuing philosophy has continued as a significant contemporary dimension in philosophy. A division also those who prefer to work with the precision and rigor of symbolic logical systems and those who prefer to analyze ordinary language. Although few contemporary philosophers maintain that all philosophical problems are linguistic, the view continues to be widely held that attention to the logical structure of language and to how language is used in everyday discourse can often aid in resolving philosophical problems

Are terms of logical calculus are also called a formal language, and a logical system? A system in which explicit rules are provided to determining (1) which are the expressions of the system (2) which sequence of expressions count as well formed (well-forced formulae) (3) which sequence would count ss proofs. A system may include axioms for which leaves terminate a proof, however, it shows of the prepositional calculus and the predicated calculus.

Its most immediate of issues surrounding certainty are especially connected with those concerning scepticism. Although Greek scepticism entered on the value of enquiry and questioning, scepticism is now the denial that knowledge or even rational belief is possible, either about some specific subject-matter, e.g., ethics, or in any area whatsoever. Classical scepticism, springs from the observation that the best method in some area seems to fall short of giving us contact with the truth, e.g., there is a gulf between appearances and reality, it frequently cites the conflicting judgements that our methods deliver, with the result that questions of truth become undefinable. In classic thought the various examples of this conflict were systemized in the tropes of Aenesidemus. So that, the scepticism of Pyrrho and the new Academy was a system of argument and inasmuch as opposing dogmatism, and, particularly the philosophical system building of the Stoics.

As it has come down to us, particularly in the writings of Sextus Empiricus, its method was typically to cite reasons for finding our issue undesirable (sceptics devoted particular energy to undermining the Stoics conception of some truths as delivered by direct apprehension or some katalepsis). As a result the sceptic concludes eposhé, or the suspension of belief, and then go on to celebrate a way of life whose object was ataraxia, or the tranquillity resulting from suspension of belief.

Fixed by its will for and of itself, the mere mitigated scepticism which accepts every day or commonsense belief, is that, not the delivery of reason, but as due more to custom and habit. Nonetheless, it is self-satisfied at the proper time, however, the power of reason to give us much more. Mitigated scepticism is thus closer to the attitude fostered by the accentuations from Pyrrho through to Sextus Expiricus. Despite the fact that the phrase Cartesian scepticism is sometimes used. Descartes himself was not a sceptic, however, in the method of doubt uses a sceptical scenario in order to begin the process of finding a general distinction to mark its point of knowledge. Descartes trusts in categories of clear and distinct ideas, not far removed from the phantasiá kataleptikê of the Stoics.

For many sceptics had traditionally held that knowledge requires certainty, artistry. And, of course, they claim that certain knowledge is not possible. In part, nonetheless, of the principle that every effect its a consequence of an antecedent cause or causes. For causality to be true it is not necessary for an effect to be predictable as the antecedent causes may be numerous, too complicated, or too interrelated for analysis. Nevertheless, in order to avoid scepticism, this participating sceptic has generally held that knowledge does not require certainty. Except for alleged cases of things that are evident for one just by being true. It has often been thought, that any thing known must satisfy certain criteria as well for being true. It is often taught that anything is known must satisfy certain standards. In so saying, that by deduction or induction, there will be criteria specifying when it is. As these alleged cases of self-evident truths, the general principle specifying the sort of consideration that will make such standards in the apparent or justly conclude in accepting it warranted to some degree.

Besides, there is another view - the absolute globular view that we do not have any knowledge whatsoever. In whatever manner, it is doubtful that any philosopher seriously entertains of an absolute scepticism. Even the Pyrrhonist sceptics, who held that we should refrain from accenting to any non-evident standards that no such hesitancy about asserting to the evident, the non-evident is any belief that requires evidences because it is warranted.

René Descartes (1596-1650), in his sceptical guise, never doubted the content of his own ideas. Its challenging logic, inasmuch as of whether they corresponded to anything beyond ideas.

All the same, Pyrrhonism and Cartesian form of a virtual globular scepticism, in having been held and defended, that of assuming that knowledge is some form of true, sufficiently warranted belief, it is the warranted condition that provides the truth or belief conditions, in that of providing the grist for the sceptics mill about. The Pyrrhonist will suggest that no non-evident, empirically deferring the sufficiency of giving in but warranted. Whereas, a Cartesian sceptic will agree that no empirical standard about anything other than ones own mind and its contents is sufficiently warranted, because there are always legitimate grounds for doubting it. Whereby, the essential difference between the two views concerns the stringency of the requirements for a belief being sufficiently warranted to take account of as knowledge.

A Cartesian requires certainty. A Pyrrhonist merely requires that the standards in case are more warranted then its negation.

Cartesian scepticism, occuruing by an unduly influence with which Descartes agues for scepticism, nnd his reply holds, in that we do not have any knowledge of any empirical standards, in that of anything beyond the contents of our own minds. The reason is roughly in the position that there is a legitimate doubt about all such standards, only because there is no way to justifiably deny that our senses are being stimulated by some sense, for which it is radically different from the objects which we normally think, in whatever manner they affect our senses. Therefrom, if the Pyrrhonist are the agnostic, the Cartesian sceptic is the atheist.

Because the Pyrrhonist require much less of a belief in order for it to be confirmed as knowledge than do the Cartesian, the argument for Pyrrhonism are much more difficult to construct. A Pyrrhonist must show that there is no better set of reasons for believing to any standards, of which are in case that any knowledge learnt of the mind is understood by some of its forms, that has to require certainty.

The underlying latencies that are given among the many derivative contributions as awaiting their presence to the future that of specifying to the theory of knowledge, is, but, nonetheless, the possibility to identify a set of shared doctrines, however, identity to discern two broad styles of instances to discern, in like the manner, these two styles of pragmatism, clarify the innovation that a Cartesian approval is fundamentally flawed, nonetheless, of responding very differently but not fordone.

Repudiating the requirements of absolute certainty or knowledge, insisting on the connection of knowledge with activity, as, too, of pragmatism of a reformist distributing knowledge upon the legitimacy of traditional questions about the truth-unconditionals of our cognitive practices, and sustain a conception of truth objectives, enough to give those questions that undergo of a gathering in their own purposive latencies, yet we are given to the spoken word for which a dialectic awareness sparks the flame from the ambers of fire.

Pragmatism of a determinant revolution, by contrast, relinquishing the objectivity of youth, acknowledges no legitimate epistemological questions over and above those that are naturally kindred of our current cognitive conviction.

It seems clear that certainty is a property that can be assembled to either a person or a belief. We can say that a person, S is certain, or we can say that its descendable alinement are aligned as of p, is certain. The two uses can be connected by saying that S has the right to be certain just in case the value of p is sufficiently verified.

In defining certainty, it is crucial to note that the term has both an absolute and relative sense. More or less, we take a proposition to be certain when we have no doubt about its truth. We may do this in error or unreasonably, but objectively a proposition is certain when such absence of doubt is justifiable. The sceptical tradition in philosophy denies that objective certainty is often possible, or ever possible, either for any proposition at all, or for any proposition from some suspect family (ethics, theory, memory, empirical judgement etc.) a major sceptical weapon is the possibility of upsetting events that Can cast doubt back onto what were hitherto taken to be certainties. Others include reminders of the divergence of human opinion, and the fallible source of our confidence. Fundamentalist approaches to knowledge look for a basis of certainty, upon which the structure of our system is built. Others reject the metaphor, looking for mutual support and coherence, without foundation.

However, in moral theory, the view that there are inviolable moral standards or absolute variable human desires or policies or prescriptions.

In spite of the notorious difficulty of reading Kantian ethics, a hypothetical imperative embeds a command which is in place only given some antecedent desire or project: If you want to look wise, stay quiet. The injunction to stay quiet only applies to those with the antecedent desire or inclination. If one has no desire to look wise the injunction cannot be so avoided: It is a requirement that binds anybody, regardless of their inclination. It could be represented as, for example, tell the truth (regardless of whether you want to or not). The distinction is not always signalled by presence or absence of the conditional or hypothetical form: If you crave drink, don’t become a bartender may be regarded as an absolute injunction applying to anyone, although only activated in case of those with the stated desire.

In Grundlegung zur Metaphsik der Sitten (1785), Kant discussed five forms of the categorical imperative: (1) the formula of universal law: act only on that maxim through which you can at the same times will that it should become universal law: (2) the formula of the law of nature: act as if the maxim of your action were to become through your will a universal law of nature: (3) the formula of the end-in-itself: act in such a way that you always treat humanity, whether in your own person or in the person of any other, never simply as a means, but always at the same time as an end: (4) the formula of autonomy, or considering the will of every rational being as a will which makes universal law: (5) the formula of the Kingdom of Ends, which provides a model for the systematic union of different rational beings under common laws.

Even so, a proposition that is not a conditional p. Moreover, the affirmative and negative, modern opinion is wary of this distinction, since what appears categorical may vary notation. Apparently, categorical propositions may also turn out to be disguised conditionals: X is intelligent (categorical?) = if X is given a range of tasks she performs them better than many people (conditional?) The problem. Nonetheless, is not merely one of classification, since deep metaphysical questions arise when facts that seem to be categorical and therefore solid, come to seem by contrast conditional, or purely hypothetical or potential.

A limited area of knowledge or endeavour to which pursuits, activities and interests are a central representation held to a concept of physical theory. In this way, a field is defined by the distribution of a physical quantity, such as temperature, mass density, or potential energy y, at different points in space. In the particularly important example of force fields, such as gravitational, electrical, and magnetic fields, the field value at a point is the force which a test particle would experience if it were located at that point. The philosophical problem is whether a force field is to be thought of as purely potential, so the presence of a field merely describes the propensity of masses to move relative to each other, or whether it should be thought of in terms of the physically real modifications of a medium, whose properties result in such powers that is, are force fields purely potential, fully characterized by dispositional statements or conditionals, or are they categorical or actual? The former option seems to require within ungrounded dispositions, or regions of space that differs only in what happens if an object is placed there. The law-like shape of these dispositions, apparent for example in the curved lines of force of the magnetic field, may then seem quite inexplicable. To atomists, such as Newton it would represent a return to Aristotelian entelechies, or quasi-psychological affinities between things, which are responsible for their motions. The latter option requires understanding of how forces of attraction and repulsion can be grounded in the properties of the medium.

The basic idea of a field is arguably present in Leibniz, who was certainly hostile to Newtonian atomism. Although his equal hostility to action at a distance muddies the water. It is usually credited to the Jesuit mathematician and scientist Joseph Boscovich (1711-87) and Immanuel Kant (1724-1804), both of whom influenced the scientist Faraday, with whose work the physical notion became established. In his paper On the Physical Character of the Lines of Magnetic Force (1852). Faraday was to suggest several criteria for assessing the physical reality of lines of force, such as whether they are affected by an intervening material medium, whether the motion depends on the nature of what is placed at the receiving end. As far as electromagnetic fields go, Faraday himself inclined to the view that the mathematical similarity between heat flow, currents, and electro-magnetic lines of force was evidence for the physical reality of the intervening medium.

Once, again, our mentioning recognition for which its case value, whereby its view is especially associated the American psychologist and philosopher William James (1842-1910), that the truth of a statement can be defined in terms of a utility of accepting it. Communicated, so much as a dispiriting position for which its place of valuation may be viewed as an objection. Since there are things that are false, as it may be useful to accept, and conversely there are things that are true and that it may be damaging to accept. Nevertheless, there are deep connections between the idea that a representation system is accorded, and the likely success of the projects in progressive formality, by its possession. The evolution of a system of representation either perceptual or linguistic, seems bounded to connect successes with everything adapting or with utility in the modest sense. The Wittgenstein doctrine stipulates the meaning of use that upon the nature of belief and its relations with human attitude, emotion and the idea that belief in the truth on one hand, the action of the other. One way of binding with cement, wherefore the connection is found in the idea that natural selection becomes much as much in adapting us to the cognitive creatures, because beliefs have effects, they work. Pragmatism can be found in Kants doctrine, and continued to play an influencing role in the theory of meaning and truth.

James, (1842-1910), although with characteristic generosity exaggerated in his debt to Charles S. Peirce (1839-1914), he charted that the method of doubt encouraged people to pretend to doubt what they did not doubt in their hearts, and criticize its individualists insistence, that the ultimate test of certainty is to be found in the individuals personalized consciousness.

From his earliest writings, James understood cognitive processes in teleological terms. Though, he held us in the satisfactory interests. His will to Believe doctrine, the view that we are sometimes justified in believing beyond the evidential relics upon the notion that a beliefs benefits are relevant to its justification. His pragmatic method of analysing philosophical problems, for which requires that we find the meaning of terms by examining their application to objects in experimental situations, similarly reflects the teleological approach in its attention to consequences.

Such an approach, however, sets James theory of meaning apart from verification, dismissive of metaphysics. Unlike the verification alist, who takes cognitive meaning to be a matter only of consequences in sensory experience. James took pragmatic meaning to include emotional and matter responses. Moreover, his ,metaphysical standard of value, not a way of dismissing them as meaningless. It should also be noted that in a greater extent, circumspective moments James did not hold that even his broad set of consequences were exhaustive of a terms meaning. Theism, for example, he took to have antecedently, definitional meaning, in addition to its varying degree of importance and chance upon an important pragmatic meaning.

James theory of truth reflects upon his teleological conception of cognition, by considering a true belief to be one which is compatible with our existing system of beliefs, and leads us to satisfactory interaction with the world.

However, Peirces famous pragmatist principle is a rule of logic employed in clarifying our concepts and ideas. Consider the claim the liquid in a flask is an acid, if, we believe this, we except that it would turn red: We accept an action of ours to have certain experimental results. The pragmatic principle holds that listing the conditional expectations of this kind, in that we associate such immediacy with applications of a conceptual representation that provides a complete and orderly sets clarification of the concept. This is relevant ti the logic of abduction: Clarificationists using the pragmatic principle provides all the information about the content of a hypothesis that is relevantly to decide whether it is worth testing.

To a greater extent, and what is most important, is the famed apprehension of the pragmatic principle, in so that, Pierces’ account of reality: When we take something to be rea that by this single case, we think it is fated to be agreed upon by all who investigate the matter to which it stand, in other words, if I believe that it is really the case that P, then I except that if anyone were to inquire depthfully into the finding its measure into whether p, they would arrive at the belief that p. It is not part of the theory that the experimental consequences of our actions should be specified by a warranted empiricist vocabulary - Peirce insisted that perceptual theories are abounding in latency. Even so, nor is it his view that the collected conditionals do or not clarify a concept as all analytic. In addition, in later writings, he argues that the pragmatic principle could only be made plausible to someone who accepted its metaphysical realism: It requires that would-bees are objective and, of course, real.

If realism itself can be given a fairly quick clarification, it is more difficult to chart the various forms of supposition, for they seem legendary. Other opponents deny that the entitles posited by the relevant discourse that exist or at least exists: The standard example is idealism, which reality is somehow mind-curative or mind-co-ordinated - that real object comprising the external world are not independently of eloping minds, but only exist as in some way correlative to the mental operations. The doctrine assembled of idealism enters on the conceptual note that reality as we understand this as meaningful and reflects the working of mindful purposes. And it construes this as meaning that the inquiring mind itself makes of a formative constellations and not of any mere understanding of the nature of the real bit even the resulting charger we attribute to it.

Wherefore, the term its most straightforwardly used when qualifying another linguistic form of Grammatik: a real x may be contrasted with a fake, a failed x, a near x, and so on. To trat something as real, without qualification, is to suppose it to be part of the actualized world. To reify something is to suppose that we have committed by some indoctrinated treatise, as that of a theory. The central error in thinking of reality and the totality of existence is to think of the unreal as a separate domain of things, perhaps, unfairly to that of the benefits of existence.

Such that non-existence of all things, as the product of logical confusion of treating the term nothing as itself a referring expression instead of a quantifier. (Stating informally as a quantifier is an expression that reports of a quantity of times that a predicate is satisfied in some class of things, i.e., in a domain.) This confusion leads the unsuspecting to think that a sentence such as Nothing is all around us talks of a special kind of thing that is all around us, when in fact it merely denies that the predicate is all around us has appreciations. The feelings that lad some philosophers and theologians, notably Heidegger, to talk of the experience of. Nothing, is not properly the experience of anything, but rather the failure of a hope or expectations that there would be something of some kind at some point. This may arise in quite everyday cases, as when one finds that the article of functions one expected to see as usual, in the corner has disappeared. The difference between existentialist and analytic philosophy, on the point of what, whereas the former is afraid of nothing, and the latter thinks that there is nothing to be afraid of.

A rather different set of concerns arise when actions are specified in terms of doing nothing, saying nothing may be an admission of guilt, and doing nothing in some circumstances may be tantamount to murder. Still, other substitutional problems arise over conceptualizing empty space and time.

Whereas, the standard opposition between those who affirm and those who deny, the real existence of some kind of thing or some kind of fact or state of affairs. Almost any area of discourse may be the focus of this dispute: The external world, the past and future, other minds, mathematical objects, possibilities, universals, moral or aesthetic properties are examples. There be to one influential suggestion, as associated with the British philosopher of logic and language, and the most determinative of philosophers centred round Anthony Dummett (1925), to which is borrowed from the intuitivistic critique of classical mathematics, and suggested that the unrestricted use of the principle of bivalence is the trademark of realism. However, this ha to overcome counter-examples both ways: Although Aquinas wads a moral realist, he held that moral really was not sufficiently structured to make true or false every moral claim. Unlike Kant who believed that he could use the law of bivalence happily in mathematics, precisely because it wad only our own construction. Realism can itself be subdivided: Kant, for example, combines empirical realism (within the phenomenal world the realist says the right things - surrounding objects really exist and independent of us and our mental stares) with transcendental idealism (the phenomenal world as a whole reflects the structures imposed on it by the activity of our minds as they render it intelligible to us). In modern philosophy the orthodox oppositions to realism has been from philosopher such as Goodman, who, impressed by the extent to which we perceive the world through conceptual and linguistic lenses of our own making.

Assigned to the modern treatment of existence in the theory of quantification is sometimes put by saying that existence is not a predicate. The idea is that the existential quantifies themselves as an operator on a predicate, indicating that the property it expresses has instances. Existence is therefore treated as a second-order property, or a property of properties. It is fitting to say, that in this it is like number, for when we say that these things of a kind, we do not describe the thing (ad we would if we said there are red things of the kind), but instead attribute a property to the kind itself. The parallelled numbers are exploited by the German mathematician and philosopher of mathematics Gottlob Frége in the dictum that affirmation of existence is merely denied of the number nought. A problem, nevertheless, proves accountable for its crated by sentences like This exists, where some particular thing is undirected, such that a sentence seems to express a contingent truth (for this insight has not existed), yet no other predicate is involved. This exists is, and unlike Tamed tigers exist, where a property is said to have an instance, for the word this and does not locate a property, but only an individual.

Possible worlds seem able to differ from each other purely in the presence or absence of individuals, and not merely in the distribution of exemplification of properties.

The philosophical ponderance over which to set upon the unreal, as belonging to the domain of Being. Nonetheless, there is little for us that can be said with the philosophers study. So it is not apparent that there can be such a subject for being by itself. Nevertheless, the concept had a central place in philosophy from Parmenides to Heidegger. The essential question of why is there something and not of nothing? Prompting over logical reflection on what it is for a universal to have an instance, nd as long history of attempts to explain contingent existence, by which id to reference and a necessary ground.

In the transition, ever since Plato, this ground becomes a self-sufficient, perfect, unchanging, and external something, identified with the Good or God, but whose relation with the everyday world remains obscure. The celebrated argument for the existence of God first propounded by Anselm in his Proslogin. The argument by defining God as something than which nothing greater can be conceived. God then exists in the understanding since we understand this concept. However, if, He only existed in the understanding something greater could be conceived, for a being that exists in reality is greater than one that exists in the understanding. Bu then, we can conceive of something greater than that than which nothing greater can be conceived, which is contradictory. Therefore, God cannot exist on the understanding, but exists in reality.

An influential argument (or family of arguments) for the existence of God, finding its premisses are that all natural things are dependent for their existence on something else. The totality of dependent brings must then itself depend upon a non-dependent, or necessarily existent bring of which is God. Like the argument to design, the cosmological argument was attacked by the Scottish philosopher and historian David Hume (1711-76) and Immanuel Kant.

Its main problem, nonetheless, is that it requires us to make sense of the notion of necessary existence. For if the answer to the question of why anything exists is that some other tings of a similar kind exists, the question merely arises gain. So, the God that ends the question must exist necessarily: It must not be an entity of which the same kinds of questions can be raised. The other problem with the argument is attributing concern and care to the deity, not for connecting the necessarily existent being it derives with human values and aspirations.

The ontological argument has been treated by modern theologians such as Barth, following Hegel, not so much as a proof with which to confront the unconverted, but as an explanation of the deep meaning of religious belief. Collingwood, regards the argument s proving not that because our idea of God is that of id quo maius cogitare viequit, therefore God exists, but proving that because this is our idea of God, we stand committed to belief in its existence. Its existence is a metaphysical point or absolute presupposition of certain forms of thought.

In the 20th century, modal versions of the ontological argument have been propounded by the American philosophers Charles Hertshorne, Norman Malcolm, and Alvin Plantinge. One version is to define something as unsurpassably great, if it exists and is perfect in every possible world. Then, to allow that it is at least possible that an unsurpassable great being existed. This means that there is a possible world in which such a being exists. However, if it exists in one world, it exists in all (for the fact that such a being exists in a world that entails, in at least, it exists and is perfect in every world), so, it exists necessarily. The correct response to this argument is to disallow the apparently reasonable concession that it is possible that such a being exists. This concession is much more dangerous than it looks, since in the modal logic, involved from possibly necessarily p, we can device necessarily p. A symmetrical proof starting from the assumption that it is possibly that such a being not exist would derive that it is impossible that it exists.

The doctrine that it makes an ethical difference of whether an agent actively intervenes to bring about a result, or omits to act in circumstances in which it is foreseen, that as a result of the omission the same result occurs. Thus, suppose that I wish you dead. If I act to bring about your death, I am a murderer, however, if I happily discover you in danger of death, and fail to act to save you, I am not acting, and therefore, according to the doctrine of acts and omissions not a murderer. Critics implore that omissions can be as deliberate and immoral as I am responsible for your food and fact to feed you. Only omission is surely a killing, Doing nothing can be a way of doing something, or in other worlds, absence of bodily movement can also constitute acting negligently, or deliberately, and defending on the context ,may be a way of deceiving, betraying, or killing. Nonetheless, criminal law offers to find its conveniences, from which to distinguish discontinuous intervention, for which is permissible, from bringing about result, which may not be, if, for instance, the result is death of a patient. The question is whether the difference, if there is one, is, between acting and omitting to act be discernibly or defined in a way that bars a general moral might.

The double effect of a principle attempting to define when an action that had both good and bad results is morally permissible. I one formation such an action is permissible if (1) The action is not wrong in itself, (2) the bad consequences is not that which is intended (3) the good is not itself a result of the bad consequences, and (4) the two consequential effects are commensurate. Thus, for instance, I might justifiably bomb an enemy factory, foreseeing but intending that the death of nearby civilians, whereas bombing the death of nearby civilians intentionally would be disallowed. The principle has its roots in Thomist moral philosophy, accordingly. St. Thomas Aquinas (1225-74), held that it is meaningless to ask whether a human being is two tings (soul and body) or, only just as it is meaningless to ask whether the wax and the shape given to it by the stamp are one: On this analogy the sound is ye form of the body. Life after death is possible only because a form itself doe not perish (pricking is a loss of form).

And is, therefore, in some sense available to reactivate a new body. Therefore, not who survives body death, but it may be resurrected in the same personalized body that becomes reanimated by the same form, that which Aquinass account, as a person has no privileged self-understanding, we understand ourselves as we do everything else, by way of sense experience and abstraction, and knowing the principle of our own lives is an achievement, not as a given. Difficultly at this point led the logical positivist to abandon the notion of an epistemological foundation together, and to flirt with the coherence theory of truth, it is widely accepted that trying to make the connection between thought and experience through basic sentence s depends on an untenable myth of the given

The special way that we each have of knowing our own thoughts, intentions, and sensationalist have brought in the many philosophical behaviorist and functionalist tendencies, that have found it important to deny that there is such a special way , arguing the way that I know of my own mind inasmuch as the way that I know of yours, e.g., by seeing what I say when asked. Others, however, point out that the behaviour of reporting the result of introspection in a particular and legitimate kind of behavioural access that deserves notice in any account of historically human psychology. The historical philosophy of reflection upon the astute of history, or of historical, thinking, finds the term was used in the 18th century, e.g., by Volante was to mean critical historical thinking as opposed to the mere collection and repetition of stories about the past. In Hegelian, particularly by conflicting elements within his own system, however, it came to man universal or world history. The Enlightenment confidence was being replaced by science, reason, and understanding that gave history a progressive moral thread, and under the influence of the German philosopher, whom is in spreading Romanticism, came Gottfried Herder (1744-1803),and, Immanuel Kant, this idea took it further to hold, so that philosophy of history cannot be the detecting of a grand system, the unfolding of the evolution of human nature as witnessed in successive stages (the progress of rationality or of Spirit). This essential speculative philosophy of history is given a extra Kantian twist in the German idealist Johann Fichte, in whom the extra association of temporal succession with logical implication introduces the idea that concepts themselves are the dynamic engine of historical change. The idea is readily intelligible in that their world of nature and of thought become identified. The work of Herder, Kant, Flichte and Schelling is synthesized by Hegel: History has a plot, as too, this to the moral development of man, equates with freedom within the state, this in turn is the development of thought, or a logical development in which various necessary moment in the life of the concept are successively achieved and improved upon. Hegels method is at its most successful, when the object is the history of ideas, and the evolution of thinking may march in steps with logical oppositions and their resolution encounters red by various systems of thought.

Within the revolutionary communism, Karl Marx (1818-83) and the German social philosopher Friedrich Engels (1820-95), there emerges a rather different kind of story, based upon Hefls progressive structure not laying the achievement of the goal of history to a future in which the political condition for freedom comes to exist, so that economic and political fears than reason is in the engine room. Although, it is such that speculations upon the history may that it be continued to be written, notably: late examples, by the late 19th century large-scale speculation of tis kind with the nature of historical understanding, and in particular with a comparison between the ,methos of natural science and with the historians. For writers such as the German neo-Kantian Wilhelm Windelband and the German philosopher and literary critic and historian Wilhelm Dilthey, it is important to show that the human sciences such as history is objective and legitimate, nonetheless they are in some way deferent from the enquiry of the scientist. Since the subjective-matter is the past thought and actions of human brings, what is needed and actions of human beings, past thought and actions of human beings, what is needed is an ability to re-live that past thought, knowing the deliberations of past agents, as if they were the historians own. The most influential British writer on this theme was the philosopher and historian George Collingwood (1889-1943) whose, The Idea of History (1946), contains an extensive defence of the verstehe approach, but it is nonetheless, the explanation from their actions, however, by re-living the situation as our understanding that understanding others is not gained by the tactic use of a theory, enabling us to infer what thoughts or intentionality experienced, again, the matter to which the subjective-matters of past thoughts and actions , as I have a human ability of knowing the deliberations of past agents as if they were the historians own. The immediate question of the form of historical explanation, and the fact that general laws have other than no place or any apprentices in the order of a minor place in the human sciences, it is also prominent in thoughts about distinctiveness as to regain their actions, but by re-living the situation in or thereby understanding of what they experience and thought.

By comparison, the moral philosopher and epistemologist Bernard Bolzano (1781-1848) argues, though, that there is something else, an infinity that doe not have this whatever you need it to be elasticity. In fact a truly infinite quantity (for example, the length of a straight line unbounded in either direction, meaning: The magnitude of the spatial entity containing all the points determined solely by their abstractly conceivable relation to two fixed points) does not by any means need to be variable, and in adduced example it is in fact not variable. Conversely, it is quite possible for a quantity merely capable of being taken greater than we have already taken it, and of becoming larger than any pre-assigned (finite) quantity, nevertheless it is to mean, in that of all times is merely finite, which holds in particular of every numerical quantity 1, 2, 3, 4, 5.

In other words, for Bolzano there could be a true infinity that was not a variable something that was only bigger than anything you might specify. Such a true infinity was the result of joining two pints together and extending that line in both directions without stopping. And what is more, he could separate off the demands of calculus, using a finite quality without ever bothering with the slippery potential infinity. Here was both a deeper understanding of the nature of infinity and the basis on which are built in his safe infinity free calculus.

This use of the inexhaustible follows on directly from most Bolzanos criticism of the way that ∞ we used as a variable something that would be bigger than anything you could specify, but never quite reached the true, absolute infinity. In Paradoxes of the Infinity Bolzano points out that is possible for a quantity merely capable of becoming larger than any one pre-assigned (finite) quantity, nevertheless to remain at all times merely finite.

Bolzano intended tis as a criticism of the way infinity was treated, but Professor Jacquette sees it instead of a way of masking use of practical applications like calculus without the need for weasel words about infinity.

By replacing ∞ with ¤ we do away with one of the most common requirements for infinity, but is there anything left that map out to the real world? Can we confine infinity to that pure mathematical other world, where anything, however unreal, can be constructed, and forget about it elsewhere? Surprisingly, this seems to have been the view, at least at one point in time, even of the German mathematician and founder of set-theory Georg Cantor (1845-1918), himself, whose comments in 1883, that only the finite numbers are real.

Keeping within the lines of reason, both the Cambridge mathematician and philosopher Frank Plumpton Ramsey (1903-30) and the Italian mathematician G. Peano (1858-1932) have been to distinguish logical paradoxes and that depend upon the notion of reference or truth (semantic notions), such are the postulates justifying mathematical induction. It ensures that a numerical series is closed, in the sense that nothing but zero and its successors can be numbers. In that any series satisfying a set of axioms can be conceived as the sequence of natural numbers. Candidates from set theory include the Zermelo numbers, where the empty set is zero, and the successor of each number is its unit set, and the von Neuman numbers, where each number is the set of all smaller numbers. A similar and equally fundamental complementarity exists in the relation between zero and infinity. Although the fullness of infinity is logically antithetical to the emptiness of zero, infinity can be obtained from zero with a simple mathematical operation. The division of many numbers by zero is infinity, while the multiplication of any number by zero is zero.

With the set theory developed by the German mathematician and logician Georg Cantor. From 1878 to 1807, Cantor created a theory of abstract sets of entities that eventually became a mathematical discipline. A set, as he defined it, is a collection of definite and distinguished objects in thought or perception conceived as a whole.

Cantor attempted to prove that the process of counting and the definition of integers could be placed on a solid mathematical foundation. His method was to repeatedly place the elements in one set into one-to-one correspondence with those in another. In the case of integers, Cantor showed that each integer (1, 2, 3, . . . n) could be paired with an even integers (2, 4, 6, . . . n), and, therefore, that the set of all integers was equal to the set of all even numbers.

Amazingly, Cantor discovered that some infinite sets were large than others and that infinite sets formed a hierarchy of greater infinities. After this failed attempt to save the classical view of logical foundations and internal consistency of mathematical systems, it soon became obvious that a major crack had appeared in the seemingly sold foundations of number and mathematics. Meanwhile, an impressive number of mathematicians began to see that everything from functional analysis to the theory of real numbers depended on the problematic character of number itself.

While, in the theory of probability Ramsey was the first to show how a personalised theory could be developed, based on precise behavioural notions of preference and expectation. In the philosophy of language, Ramsey was one of the first thinkers to accept a redundancy theory of truth, which hr combined with radical views of the function of man y kinds of propositions. Neither generalizations nor causal propositions, nor those treating probability or ethics, describe facts, but each has a different specific function in our intellectual economy.

Ramsey advocates that of a sentence generated by taking all the sentence affirmed in a scientific theory that use some term, e.g., quark. Replacing the term by a variable, and existentially quantifying into the result. Instead of saying quarks have such-and-such properties, Ramsey postdated that the sentence as saying that there is something that has those properties. If the process is repeated, the sentence gives the topic-neutral structure of the theory, but removes any implications that we know what the term so treated denote. It leaves open the possibility of identifying the theoretical item with whatever it is that best fits the description provided. Nonetheless, it was pointed out by the Cambridge mathematician Newman that if the process is carried out for all except the logical bones of the theory, then by the Löwenheim-Skolem theorem, the result will be interpretable in any domain of sufficient cardinality, and the content of the theory may reasonably be felt to have been lost.

It seems, that the most taken of paradoxes in the foundations of set theory as discovered by Russell in 1901. Some classes have themselves as members: The class of all abstract objects, for example, is an abstract object, whereby, others do not: The class of donkeys is not itself a donkey. Now consider the class of all classes that are not members of themselves, is this class a member of itself, that, if it is, then it is not, and if it is not, then it is.

The paradox is structurally similar to easier examples, such as the paradox of the barber. Such one like a village having a barber in it, who shaves all and only the people who do not have in themselves. Who shaves the barber? If he shaves himself, then he does not, but if he does not shave himself, then he does not. The paradox is actually just a proof that there is no such barber or in other words, that the condition is inconsistent. All the same, it is no to easy to say why there is no such class as the one Russell defines. It seems that there must be some restriction on the kind of definition that are allowed to define classes and the difficulty that of finding a well-motivated principle behind any such restriction.

The French mathematician and philosopher Henri Jules Poincaré (1854-1912) believed that paradoses like those of Russell nd the barber were due to such as the impredicative definitions, and therefore proposed banning them. But, it tuns out that classical mathematics required such definitions at too many points for the ban to be easily absolved. Having, in turn, as forwarded by Poincaré and Russell, was that in order to solve the logical and semantic paradoxes it would have to ban any collection (set) containing members that can only be defined by means of the collection taken as a whole. It is, effectively by all occurring principles into which have an adopting vicious regress, as to mark the definition for which involves no such failure. There is frequently room for dispute about whether regresses are benign or vicious, since the issue will hinge on whether it is necessary to reapply the procedure. The cosmological argument is an attempt to find a stopping point for what is otherwise seen as being an infinite regress, and, to ban of the predicative definitions.

The investigation of questions that arise from reflection upon sciences and scientific inquiry, are such as called of a philosophy of science. Such questions include, what distinctions in the methods of science? s there a clear demarcation between scenes and other disciplines, and how do we place such enquires as history, economics or sociology? And scientific theories probable or more in the nature of provisional conjecture? Can the be verified or falsified? What distinguished good from bad explanations? Might there be one unified since, embracing all the special science? For much of the 20th century there questions were pursued in a highly abstract and logical framework it being supposed that as general logic of scientific discovery that a general logic of scientific discovery a justification might be found. However, many now take interests in a more historical, contextual and sometimes sociological approach, in which the methods and successes of a science at a particular time are regarded less in terms of universal logical principles and procedure, and more in terms of their availability to methods and paradigms as well as the social context.

In addition, to general questions of methodology, there are specific problems within particular sciences, giving subjects as biology, mathematics and physics.

The intuitive certainty that sparks aflame the dialectic awarenesses for its immediate concerns are either of the truth or by some other in an object of apprehensions, such as a concept. Awareness as such, has to its amounting quality value the place where philosophical understanding of the source of our knowledge are, however, in covering the sensible apprehension of things and pure intuition it is that which stricture sensation into the experience of things accent of its direction that orchestrates the celestial overture into measures in space and time.

The notion that determines how something is seen or evaluated of the status of law and morality especially associated with St Thomas Aquinas and the subsequent scholastic tradition. More widely, any attempt to cement the moral and legal order together with the nature of the cosmos or how the nature of human beings, for which sense it is also found in some Protestant writers, and arguably derivative from a Platonic view of ethics, and is implicit in ancient Stoicism. Law stands above and apart from the activities of human lawmaker, it constitutes an objective set of principles that can be seen true by natural light or reason, and (in religion versions of the theory) that express Gods will for creation. Non-religious versions of the theory substitute objective conditions for human flourishing as the source of constraints upon permissible actions and social arrangements. Within the natural law tradition, different views have been held about the relationship between the rule of law about God s will, for instance the Dutch philosopher Hugo Grothius (1583-1645), similarly takes upon the view that the content of natural law is independent of any will, including that of God, while the German theorist and historian Samuel von Pufendorf (1632-94) takes the opposite view, thereby facing the problem of one horn of the Euthyphro dilemma, that simply states, that its dilemma arises from whatever the source of authority is supposed to be, for in which do we care about the general good because it is good, or do we just call good things that we care about. Wherefore, by facing the problem that may be to assume of a strong form, in which it is claimed that various facts entail values, or a weaker form, from which it confines itself to holding that reason by itself is capable of discerning moral requirements that are supped of binding to all human bings regardless of their desires

Although the morality of people send the ethical amount from which the same thing, is that there is a usage that restricts morality to systems such as that of the German philosopher and founder of ethical philosophy Immanuel Kant (1724-1804), based on notions such as duty, obligation, and principles of conduct, reserving ethics for more than the Aristotelian approach to practical reasoning based on the notion of a virtue, and generally avoiding the separation of moral considerations from other practical considerations. The scholarly issues are complex, with some writers seeing Kant as more Aristotelian and Aristotle as, ore involved in a separate sphere of responsibility and duty, than the simple contrast suggests. Some theorists see the subject in terms of a number of laws (as in the Ten Commandments). The status of these laws may be test they are the edicts of a divine lawmaker, or that they are truths of reason, knowable deductively. Other approaches to ethics (e.g., eudaimonism, situation ethics, virtue ethics) eschew general principles as much as possible, frequently disguising the great complexity of practical reasoning. For Kantian notion of the moral law is a binding requirement of the categorical imperative, and to understand whether they are equivalent at some deep level. Kants own applications of the notion are not always convincing, as for one cause of confusion in relating Kants ethics to theories such additional expressivism is that it is easy, but mistaken, to suppose that the categorical nature of the imperative means that it cannot be the expression of sentiment, but must derive from something unconditional or necessary such as the voice of reason.

For which ever reason, the mortal being makes of its presence to the future of weighing of that which one must do, or that which can be required of one. The term carries implications of that which is owed (due) to other people, or perhaps in onself. Universal duties would be owed to persons (or sentient beings) as such, whereas special duty in virtue of specific relations, such as being the child of someone, or having made someone a promise. Duty or obligation is the primary concept of deontological approaches to ethics, but is constructed in other systems out of other notions. In the system of Kant, a perfect duty is one that must be performed whatever the circumstances: Imperfect duties may have to give way to the more stringent ones. In another way, perfect duties are those that are correlative with the right to others, imperfect duties are not. Problems with the concept include the ways in which due needs to be specified (a frequent criticism of Kant is that his notion of duty is too abstract). The concept may also suggest of a regimented view of ethical life in which we are all forced conscripts in a kind of moral army, and may encourage an individualistic and antagonistic view of social relations.

The most generally accepted account of externalism and/or internalism, that this distinction is that a theory of justification is internalist if only if it requiem that all of the factors needed for a belief to be epistemologically justified for a given person be cognitively accessible to that person, internal to cognitive perceptivity, and externalist, if it allows that at least some of the justifying factors need not be thus accessible, so that thy can be external to the believers cognitive perceptive, beyond any such given relations. However, epistemologists often use the distinction between internalist and externalist theories of epistemic justification without offering any very explicit explication.

The externalist/internalist distinction has been mainly applied to theories of epistemic justification: It has also been applied in a closely related way to accounts of knowledge and in a rather different way to accounts of belief and thought contents.

The internalist requirement of cognitive accessibility can be interpreted in at least two ways: A strong version of internalism would require that the believer actually be aware of the justifying factor in order to be justified: While a weaker version would require only that he be capable of becoming aware of them by focussing his attentions appropriately, but without the need for any change of position, new information, etc. Though the phrase cognitively accessible suggests the weak interpretation, the main intuitive motivation for internalism, viz the idea that epistemic justification requires that the believer actually have in his cognitive possession a reason for thinking that the belief is true, and would require the strong interpretation.

Perhaps, the clearest example of an internalist position would be a Foundationalist view according to which foundational beliefs pertain to immediately experienced states of mind and other beliefs are justified by standing in cognitively accessible logical or inferential relations to such foundational beliefs. Such a view could count as either a strong or a weak version of internalism, depending on whether actual awareness of the justifying elements or only the capacity to become aware of them is required. Similarly, a coherent view could also be internalist, if both the beliefs or other states with which a justification belief is required to cohere and the coherence relations themselves are reflectively accessible.

It should be carefully noticed that when internalism is construed in this way, it is neither necessary nor sufficient by itself for internalism that the justifying factors literally be internal mental states of the person in question. Not necessary, necessary, because on at least some views, e.g., a direct realist view of perception, something other than a mental state of the believer can be cognitively accessible: Not sufficient, because there are views according to which at least some mental states need not be actual (strong version) or even possible (weak version) objects of cognitive awareness. Also, on this way of drawing the distinction, a hybrid view, according to which some of the factors required for justification must be cognitively accessible while others need not and in general will not be, would count as an externalist view. Obviously too, a view that was externalist in relation to a strong version of internalism (by not requiring that the believer actually be aware of all justifiable factors) could still be internalist in relation to a weak version (by requiring that he at least be capable of becoming aware of them).

The most prominent recent externalist views have been versions of reliabilism, whose requirements for justification is roughly that the belief be produced in a way or via a process that makes of objectively likely that the belief is true. What makes such a view externalist is the absence of any requirement that the person for whom the belief is justified have any sort of cognitive access to the relations of reliability in question. Lacking such access, such a person will in general have no reason for thinking that the belief is true or likely to be true , but will, on such an account, nonetheless be epistemically justified in according it. Thus such a view arguably marks a major break from the modern epistemological tradition, stemming from Descartes, which identifies epistemic justification with having a reason, perhaps even a conclusive reason for thinking that the belief is true. An epistemologist working within this tradition is likely to feel that the externalist, than offering a competing account of the same concept of epistemic justification with which the traditional epistemologist is concerned, has simply changed the subject.

The main objection to externalism rests on the intuitive certainty that the basic requirement for epistemic justification is that the acceptance of the belief in question be rational or responsible in relation to the cognitive goal of truth, which seems to require in turn that the believer actually be dialectally aware of a reason for thinking that the belief is true (or, at the very least, that such a reason be available to him). Since the satisfaction of an externalist condition is neither necessary nor sufficient for the existence of such a cognitively accessible reason, it is argued, externalism is mistaken as an account of epistemic justification. This general point has been elaborated by appeal to two sorts of putative intuitive counter-examples to externalism. The first of these challenges the necessity of belief which seem intuitively to be justified, but for which the externalist conditions are not satisfied. The standard examples in this sort are cases where beliefs are produced in some very nonstandard way, e.g., by a Cartesian demon, but nonetheless, in such a way that the subjective experience of the believer is indistinguishable from that of someone whose beliefs are produced more normally. The intuitive claim is that the believer in such a case is nonetheless epistemically justified, as much so as one whose belief is produced in a more normal way, and hence that externalist account of justification must be mistaken.

Perhaps the most striking reply to this sort of counter-example, on behalf of a cognitive process is to be assessed in normal possible worlds, i.e., in possible worlds that are actually the way our world is common-seismically believed to be, than in the world which contains the belief being judged. Since the cognitive processes employed in the Cartesian demon cases are, for which we may assume, reliable when assessed in this way, the reliability can agree that such beliefs are justified. The obvious, to a considerable degree of bringing out the issue of whether it is or not an adequate rationale for this construal of Reliabilism, so that the reply is not merely a notional presupposition guised as having representation.

The correlative way of elaborating on the general objection to justificatory externalism challenges the sufficiency of the various externalist conditions by citing cases where those conditions are satisfied, but where the believers in question seem intuitively not to be justified. In this context, the most widely discussed examples have to do with possible occult cognitive capacities, like clairvoyance. Considering the point in application once, again, to Reliabilism, the claim is that to think that he has such a cognitive power, and, perhaps, even good reasons to the contrary, is not rational or responsible and therefore not epistemically justified in accepting the belief that result from his clairvoyance, despite the fact that the Reliabilist condition is satisfied.

One sort of response to this latter sorts of objection is to bite the bullet and insist that such believers are in fact justified, dismissing the seeming intuitions to the contrary as latent internalist prejudice. A more widely adopted response attempts to impose additional conditions, usually of a roughly internalized sort, which will rule out the offending example, while stopping far of a full internalism. But, while there is little doubt that such modified versions of externalism can handle particular cases, as well enough to avoid clear intuitive implausibility, the usually problematic cases that they cannot handle, and also whether there is and clear motivation for the additional requirements other than the general internalist view of justification that externalist are committed to reject.

A view in this same general vein, one that might be described as a hybrid of internalism and externalism holds that epistemic justification requires that there is a justificatory factor that is cognitively accessible to the believer in question (though it need not be actually grasped), thus ruling out, e.g., a pure Reliabilism. At the same time, however, though it must be objectively true that beliefs for which such a factor is available are likely to be true, in addition, the fact need not be in any way grasped or cognitively accessible to the believer. In effect, of the premises needed to argue that a particular belief is likely to be true, one must be accessible in a way that would satisfy at least weakly internalised. The internalist will respond that this hybrid view is of no help at all in meeting the objection and has no belief nor is it held in the rational, responsible way that justification intuitively seems to require, for the believer in question, lacking one crucial premise, still has no reason at all for thinking that his belief is likely to be true.

An alternative to giving an externalist account of epistemic justification, one which may be more defensible while still accommodating many of the same motivating concerns, is to give an externalist account of knowledge directly, without relying on an intermediate account of justification. Such a view will obviously have to reject the justified true belief account of knowledge, holding instead that knowledge is true belief which satisfies the chosen externalist condition, e.g., a result of a reliable process (and perhaps, further conditions as well). This makes it possible for such a view to retain internalist account of epistemic justification, though the centrality of that concept to epistemology would obviously be seriously diminished.

Such an externalist account of knowledge can accommodate the commonsense conviction that animals, young children, and unsophisticated adults posses knowledge, though not the weaker conviction (if such a conviction does exists) that such individuals are epistemically justified in their beliefs. It is also at least less vulnerable to internalist counter-examples of the sort discussed, since the intuitions involved there pertain more clearly to justification than to knowledge. What is uncertain is what ultimate philosophical significance the resulting conception of knowledge is supposed to have. In particular, does it have any serious bearing on traditional epistemological problems and on the deepest and most troubling versions of scepticism, which seems in fact to be primarily concerned with justification, the an knowledge?`

A rather different use of the terms internalism and externalism has to do with the issue of how the content of beliefs and thoughts is determined: According to an internalist view of content, the content of such intention states depends only on the non-relational, internal properties of the individuals mind or grain, and not at all on his physical and social environment: While according to an externalist view, content is significantly affected by such external factors and suggests a view that appears of both internal and external elements are standardly classified as an external view.

As with justification and knowledge, the traditional view of content has been strongly internalized in character. The main argument for externalism derives from the philosophy of language, more specifically from the various phenomena pertaining to natural kind terms, indexicals, etc. that motivate the views that have come to be known as direct reference theories. Such phenomena seem at least to show that the belief or thought content that can be properly attributed to a person is dependant on facts about his environment, e.g., whether he is on Earth or Twin Earth, what is fact pointing at, the classificatory criteria employed by expects in his social group, etc. - not just on what is going on internally in his mind or brain.

An objection to externalist account of content is that they seem unable to do justice to our ability to know the content of our beliefs or thought from the inside, simply by reflection. If content is depend on external factors pertaining to the environment, then knowledge of content should depend on knowledge of these factors - which will not in general be available to the person whose belief or thought is in question.

The adoption of an externalized account of mental content would seem to support an externalist account of justification, by way that if part or all of the content of a belief inaccessible to the believer, then both the justifying status of other beliefs in relation to that content and the status of that content justifying the beliefs will be similarly inaccessible, thus contravening the internalist requirement for justification. An internalist must insist that there are no justifiable relations of these sorts, that our internally associable content can either be justified or justly anything else: But such a response appears lame unless it is coupled with an attempt to show that the externalist account of content is mistaken.

A great deal of philosophical effort has been lavished on the attempt to naturalize content, i.e. to explain in non-semantic,. Non-intentional terms what it is for something to be represental (have content) and what it is for something to have some particular content rather than some other. There appear to be only four types of theory that have been proposed: Theories that ground representation in (1) similarity, (2) converance, (3) functional role, (4) teleology.

Similarly, theories hold that ‘r’ represents ‘x’ in virtue of being similar to ‘x’. This has seemed hopeless to most as a theory of mental representation because it appears to require that things in the brain must share properties with the things they represent: To represent a cat as furry appears to require something furry in the brain. Perhaps, a notion of similarity that is naturalistic and does not involve property sharing can be worked out, but it is not obvious how.

Covariance theories hold that ‘r’s’ represent ‘x’ is grounded in the fact that ‘r’s’ occurrence canaries with that of ‘x’. This is most compelling he n one thinks about detection systems, the firing a neural structures in the visual system is said to represent vertical orientations, if its firing varies with the occurrence of vertical lines in the visual field of perceptivity..

Functional role theories hold that ‘r’s’ represent ‘x’ is grounded in the functional role ‘r’ has in the representing system, i.e., on the relations imposed by specific cognitive processes imposed by specific cognitive processes between ‘r’ and other representations in the system’s repertoire. Functional role theories take their cue from such common-sense ideas as that people cannot believe that cats are furry if they did not know that cats are animals or that fur is like hair.

Teleological theories hold that ‘r’ represent ‘x’ if it is ‘r’s’ function to indicate, i.e., covary with ‘x’. Teleological theories differ depending on the theory of functions they import. Perhaps the most important distinction is that between historical theories of functions. Historical theories individate functional states (hence contents) in a way that is sensitive to the historical development of the state, i.e., to factors such as the way the state was ‘learned’, or the way it evolved. An historical theory might hold that the function of ‘r’ is to indicate ‘x’ only if the capacity to token ‘r’ was developed (selected, learned) because it indicates ‘x’. Thus, a state physically indistinguishable from ‘r’s’ historical origins would not represent ‘x’ according to historical theories.

Theories of representational content may be classified according to whether they are atomistic or holistic and according to whether they are externalistic or internalistic, whereby, emphasizing the priority of a whole over its parts. Furthermore, in the philosophy of language, this becomes the claim that the meaning of an individual word or sentence can only be understood in terms of its relation to an indefinitely larger body of language, such as a whole theory, or even a whole language or form of life. In the philosophy of mind a mental state similarly may be identified only in terms of its relations with others. Moderate holism may allow the other things besides these relationships also count; extreme holism would hold that a network of relationships is all that we have. A holistic view of science holds that experience only confirms or disconfirms large bodies of doctrine, impinging at the edges, and leaving some leeway over the adjustment that it requires.

Once, again, in the philosophy of mind and language, the view that what is thought, or said, or experienced, is essentially dependent on aspects of the world external to the mind of the subject. The view goes beyond holding that such mental states are typically caused by external factors, to insist that they could not have existed as they now do without the subject being embedded in an external world of a certain kind. It is these external relations that make up the essence or identify of the mental state. Externalism is thus opposed to the Cartesian separation of the mental from the physical, since that holds that the mental could in principle exist as it does even if there were no external world at all. Various external factors have been advanced as ones on which mental content depends, including the usage of experts, the linguistic, norms of the community. And the general causal relationships of the subject. In the theory of knowledge, externalism is the view that a person might know something by being suitably situated with respect to it, without that relationship being in any sense within his purview. The person might, for example, be very reliable in some respect without believing that he is. The view allows that you can know without being justified in believing that you know.

However, atomistic theories take a representation’s content to be something that can be specified independent entity of that representation’ s relations to other representations. What the American philosopher of mind, Jerry Alan Fodor (1935-) calls the crude causal theory, for example, takes a representation to be a
cow
- a menial representation with the same content as the word ‘cow’ - if its tokens are caused by instantiations of the property of being-a-cow, and this is a condition that places no explicit constraints on how
cow
’s must or might relate to other representations. Holistic theories contrasted with atomistic theories in taking the relations a representation bears to others to be essential to its content. According to functional role theories, a representation is a
cow
if it behaves like a
cow
should behave in inference.

Internalist theories take the content of a representation to be a matter determined by factors internal to the system that uses it. Thus, what Block (1986) calls ‘short-armed’ functional role theories are internalist. Externalist theories take the content of a representation to be determined, in part at least, by factors external to the system that uses it. Covariance theories, as well as telelogical theories that invoke an historical theory of functions, take content to be determined by ‘external’ factors. Crossing the atomist-holistic distinction with the internalist-externalist distinction.

Externalist theories (sometimes called non-individualistic theories) have the consequence that molecule for molecule identical cognitive systems might yet harbour representations with different contents. This has given rise to a controversy concerning ‘narrow’ content. If we assume some form of externalist theory is correct, then content is, in the first instance ‘wide’ content, i.e., determined in part by factors external to the representing system. On the other hand, it seems clear that, on plausible assumptions about how to individuate psychological capacities, internally equivalent systems must have the same psychological capacities. Hence, it would appear that wide content cannot be relevant to characterizing psychological equivalence. Since cognitive science generally assumes that content is relevant to characterizing psychological equivalence, philosophers attracted to externalist theories of content have sometimes attempted to introduce ‘narrow’ content, i.e., an aspect or kind of content that is equivalent internally equivalent systems. The simplest such theory is Fodor’s idea (1987) that narrow content is a function from contents (i.e., from whatever the external factors are) to wide contents.

All the same, what a person expresses by a sentence is often a function of the environment in which he or she is placed. For example, the disease I refer to by the term like ‘arthritis’, or the kind of tree I refer to as a ‘Maple’ will be defined by criteria of which I know next to nothing. This raises the possibility of imagining two persons in rather different environments, but in which everything appears the same to each of them. The wide content of their thoughts and sayings will be different if the situation surrounding them is appropriately different: ‘situation’ may include the actual objects they perceive or the chemical or physical kinds of object in the world they inhabit, or the history of their words, or the decisions of authorities on what counts as an example, of one of the terms they use. The narrow content is that part of their thought which remains identical, through their identity of the way things appear, regardless of these differences of surroundings. Partisans of wide content may doubt whether any content in this sense narrow, partisans of narrow content believe that it is the fundamental notion, with wide content being explicable in terms of narrow content plus context.

Even so, the distinction between facts and values has outgrown its name: it applies not only to matters of fact vs, matters of value, but also to statements that something is, vs. statements that something ought to be. Roughly, factual statements - ‘is statements’ in the relevant sense - represent some state of affairs as obtaining, whereas normative statements - evaluative, and deontic ones - attribute goodness to something, or ascribe, to an agent, an obligation to act. Neither distinction is merely linguistic. Specifying a book’s monetary value is making a factual statement, though it attributes a kind of value. ‘That is a good book’ expresses a value judgement though the term ‘value’ is absent (nor would ‘valuable’ be synonymous with ‘good’). Similarly, ‘we are morally obligated to fight’ superficially expresses an is-statement, and ‘By all indications it ough to rain’ makes a kind of ought-claim; but the former is an ought-statement, the latter an (epistemic) is-statement.

Theoretical difficulties also beset the distinction. Some have absorbed values into facts holding that all value is instrumental, roughly, to have value is to contribute - in a factual analysable way - to something further which is (say) deemed desirable. Others have suffused facts with values, arguing that facts (and observations) are ‘theory-impregnated’ and contending that values are inescapable to theoretical choice. But while some philosophers doubt that fact/value distinctions can be sustained, there persists a sense of a deep difference between evaluating, or attributing an obligation and, on the other hand, saying how the world is.

Fact/value distinctions, may be defended by appeal to the notion of intrinsic value, value a thing has in itself and thus independently of its consequences. Roughly, a value statement (proper) is an ascription of intrinsic value, one to the effect that a thing is to some degree good in itself. This leaves open whether ought-statements are implicitly value statements, but even if they imply that something has intrinsic value - e.g., moral value - they can be independently characterized, say by appeal to rules that provide (justifying) reasons for action. One might also ground the fact value distinction in the attridinal (or even motivational) component apparently implied by the making of valuational or deontic judgements: Thus, ‘it is a good book, but that is no reason for a positive attribute towards it’ and ‘you ought to do it, but there is no reason to’ seem inadmissible, whereas, substituting, ‘an expensive book’ and ‘you will do it’ yields permissible judgements. One might also argue that factual judgements are the kind which are in principle appraisable scientifically, and thereby anchor the distinction on the factual side. This line is plausible, but there is controversy over whether scientific procedures are ‘value-free’ in the required way.

Philosophers differ regarding the sense, if any, in which epistemology is normative (roughly, valuational). But what precisely is at stake in this controversy is no clearly than the problematic fact/value distinction itself. Must epistemologists as such make judgements of value or epistemic responsibility? If epistemology is naturalizable, then even epistemic principles simply articulate under what conditions - say, appropriate perceptual stimulations - a belief is justified, or constitutes knowledge. Its standards of justification, then would be like standards of, e.g., resilience for bridges. It is not obvious, however, that there appropriate standards can be established without independent judgements that, say, a certain kind of evidence is good enough for justified belief (or knowledge). The most plausible view may be that justification is like intrinsic goodness, though it supervenes on natural properties, it cannot be analysed wholly in factual statements.

Thus far, belief has been depicted as being all-or-nothing, however, as a resulting causality for which we have grounds for thinking it true, and, all the same, its acceptance is governed by epistemic norms, and, least of mention, it is partially subject to voluntary control and has functional affinities to belief. Still, the notion of acceptance, like that of degrees of belief, merely extends the standard picture, and does not replace it.

Traditionally, belief has been of epistemological interest in its propositional guise: ‘S’ believes that ‘p’, where ‘p’ is a reposition towards which an agent, ‘S’ exhibits an attitude of acceptance. Not all belief is of this sort. If I trust you to say, I believe you. And someone may believe in Mr. Radek, or in a free-market economy, or in God. It is sometimes supposed that all belief is ‘reducible’ to propositional belief, belief-that. Thus, my believing you might be thought a matter of my believing, is, perhaps, that what you say is true, and your belief in free markets or God, is a matter of your believing that free-market economies are desirable or that God exists.

Several philosophers have followed St, Thomas Aquinas (1225-74), in supposing that to believe in God is simply to believe that certain truths hold while others argue that belief-in is a distinctive attitude, on that includes essentially an element of trust. More commonly, belief-in has been taken to involve a combination of propositional belief together with some further attitude.

The moral philosopher Richard Price (1723-91) defends the claim that there are different sorts of belief-in, some, but not all reducible to beliefs-that. If you believe in God, you believe that God exists, that God is good, you believe that God is good, etc. But according to Price, your belief involves, in addition, a certain complex pro-attitude toward its object. Even so, belief-in outruns the evidence for the corresponding belief-that. Does this diminish its rationality? If belief-in presupposes belief-that, it might be thought that the evidential standards for the former must be, at least, as high as standards for the latter. And any additional pro-attitude might be thought to require a further layer of justification not required for cases of belief-that.

Belief-in may be, in general, less susceptible to alternations in the face of unfavourable evidence than belief-that. A believer who encounters evidence against God’s existence may remain unshaken in his belief, in part because the evidence does not bear on his pro-attitude. So long as this ids united with his belief that God exists, the reasonably so - in a way that an ordinary propositional belief that would not.

Knowledge and belief, according to most epistemologists, knowledge entails belief, so that I cannot know that such and such is the case unless I believe that such and such is the case. Others think this entailment thesis can be rendered more accurately if we substitute for belief some closely related attitude. For instance, several philosophers would prefer to say that knowledge entail psychological certainties (Prichard, 1950 and Ayer, 1956) or conviction (Lehrer, 1974) or acceptance (Lehrer, 1989). Nonetheless, there are arguments against all versions of the thesis that knowledge requires having a belief-like attitude toward the known. These arguments are given by philosophers who think that knowledge and belief (or a facsimile) are mutually incompatible (the incomparability thesis), or by ones who say that knowledge does not entail belief, or vice versa, so that each may exist without the other, but the two may also coexist (the separability thesis).

The incompatibility thesis is sometimes traced to Plato (429-347 Bc) in view of his claim that knowledge is infallible while belief or opinion is fallible ('Republic' 476-9). But this claim would not support the thesis. Belief might be a component of an infallible form of knowledge in spite of the fallibility of belief. Perhaps, knowledge involves some factor that compensates for the fallibility of belief.

A. Duncan-Jones (1939: Also Vendler, 1978) cite linguistic evidence to back up the incompatibility thesis. He notes that people often say 'I do not believe she is guilty. I know she is' and the like, which suggest that belief rule out knowledge. However, as Lehrer (1974) indicates, the above exclamation is only a more emphatic way of saying 'I do not just believe she is guilty, I know she is' where 'just' makes it especially clear that the speaker is signalling that she has something more salient than mere belief, not that she has something inconsistent with belief, namely knowledge. Compare: 'You do not hurt him, you killed him.'

H.A. Prichard (1966) offers a defence of the incompatibility thesis that hinges on the equation of knowledge with certainty (both infallibility and psychological certitude) and the assumption that when we believe in the truth of a claim we are not certain about its truth. Given that belief always involves uncertainty while knowledge never dies, believing something rules out the possibility of knowing it. Unfortunately, however, Prichard gives 'us' no goods reason to grant that states of belief are never ones involving confidence. Conscious beliefs clearly involve some level of confidence, to suggest that we cease to believe things about which we are completely confident is bizarre.

A.D. Woozley (1953) defends a version of the separability thesis. Woozley's version, which deals with psychological certainty rather than belief per se, is that knowledge can exist in the absence of confidence about the item known, although might also be accompanied by confidence as well. Woozley remarks that the test of whether I know something is 'what I can do, where what I can do may include answering questions.' On the basis of this remark he suggests that even when people are unsure of the truth of a claim, they might know that the claim is true. We unhesitatingly attribute knowledge to people who give correct responses on examinations even if those people show no confidence in their answers. Woozley acknowledges, however, that it would be odd for those who lack confidence to claim knowledge. It would be peculiar to say, I am unsure that for whatever reason my answer is true: Still, I know it is correct But this tension Woozley explains using a distinction between conditions under which we are justified in making a claim (such as a claim to know something), and conditions under which the claim we make is true. While 'I know such and such' might be true even if I am unsure whether such and such holds, nonetheless it would be inappropriate for me to claim that I know that such and such unless I were sure of the truth of my claim.

Colin Radford (1966) extends Woozley's defence of the separability thesis. In Radford's view, not only is knowledge compatible with the lack of certainty, it is also compatible with a complete lack of belief. He argues by example. In one example, Jean has forgotten that he learned some English history year's priori and yet he is able to give several correct responses to questions such as 'When did the Battle of Hastings occur?' Since he forgot that he took history, he considers the correct response to be no more than guesses. Thus, when he says that the Battle of Hastings took place in 1066 he would deny having the belief that the Battle of Hastings took place in 1066. A disposition he would deny being responsible (or having the right to be convincing) that 1066 was the correct date. Radford would none the less insist that Jean know when the Battle occurred, since clearly be remembering the correct date. Radford admits that it would be inappropriate for Jean to say that he knew when the Battle of Hastings occurred, but, like Woozley he attributes the impropriety to a fact about when it is and is not appropriate to claim knowledge. When we claim knowledge, we ought, at least to believe that we have the knowledge we claim, or else our behaviour is 'intentionally misleading'.

Those that agree with Radford's defence of the separability thesis will probably think of belief as an inner state that can be detected through introspection. That Jean lack's beliefs about English history is plausible on this Cartesian picture since Jean does not find himself with any beliefs about English history when ne seek them out. One might criticize Radford, however, by rejecting that Cartesian view of belief. One could argue that some beliefs are thoroughly unconscious, for example. Or one could adopt a behaviourist conception of belief, such as Alexander Bain's (1859), according to which having beliefs is a matter of the way people are disposed to behave (and has not Radford already adopted a behaviourist conception of knowledge?) Since Jean gives the correct response when queried, a form of verbal behaviour, a behaviourist would be tempted to credit him with the belief that the Battle of Hastings occurred in 1066.

D.M. Armstrong (1873) takes a different tack against Radford. Jean does know that the Battle of Hastings took place in 1066. Armstrong will grant Radfod that point, in fact, Armstrong suggests that Jean believe that 1066 is not the date the Battle of Hastings occurred, for Armstrong equates the belief that such and such is just possible but no more than just possible with the belief that such and such is not the case. However, Armstrong insists, Jean also believes that the Battle did occur in 1066. After all, had Jean been mistaught that the Battle occurred in 1066, and subsequently 'guessed' that it took place in 1066, we would surely describe the situation as one in which Jean's false belief about the Battle became unconscious over time but persisted of a memory trace that was causally responsible for his guess. Out of consistency, we must describe Radford's original case as one that Jean's true belief became unconscious but persisted long enough to cause his guess. Thus, while Jean consciously believes that the Battle did not occur in 1066, unconsciously he does believe it occurred in 1066. So after all, Radford does not have a counterexample to the claim that knowledge entails belief.

Armstrong's response to Radford was to reject Radford's claim that the examinee lacked the relevant belief about English history. Another response is to argue that the examinee lacks the knowledge Radford attributes to him (cf. Sorenson, 1982). If Armstrong is correct in suggesting that Jean believes both that 1066 is and that it is not the date of the Battle of Hastings, one might deny Jean knowledge on the grounds that people who believe the denial of what they believe cannot be said t know the truth of their belief. Another strategy might be to compare the examine case with examples of ignorance given in recent attacks on externalist accounts of knowledge (needless to say. Externalists themselves would tend not to favour this strategy). Consider the following case developed by BonJour (1985): For no apparent reason, Samantha believes that she is clairvoyant. Again, for no apparent reason, she one day comes to believe that the President is in New York City, even though she has every reason to believe that the President is in Washington, D.C. In fact, Samantha is a completely reliable clairvoyant, and she has arrived at her belief about the whereabouts of the President thorough the power of her clairvoyance. Yet surely Samanthas belief is completely irrational. She is not justified in thinking what she does. If so, then she does not know where the President is. But Radford's examinee is unconventional. Even if Jean lacks the belief that Radford denies him, Radford does not have an example of knowledge that is unattended with belief. Suppose that Jean's memory had been sufficiently powerful to produce the relevant belief. As Radford says, in having every reason to suppose that his response is mere guesswork, and he has every reason to consider his belief false. His belief would be an irrational one, and hence one about whose truth Jean would be ignorant.

Least has been of mention to an approaching view from which 'perception' basis upon itself as a fundamental philosophical topic both for its central place in ant theory of knowledge, and its central place un any theory of consciousness. Philosophy in this area is constrained by a number of properties that we believe to hold of perception, (1) It gives 'us' knowledge of the world around 'us,' (2) We are conscious of that world by being aware of 'sensible qualities': Colour, sounds, tastes, smells, felt warmth, and the shapes and positions of objects in the environment. (3) Such consciousness is effected through highly complex information channels, such as the output of the three different types of colour-sensitive cells in the eye, or the channels in the ear for interpreting pulses of air pressure as frequencies of sound. (4) There ensues even more complex neurophysiological coding of that information, and eventually higher-order brain functions bring it about that we interpreted the information so received. (Much of this complexity has been revealed by the difficulties of writing programs enabling computers to recognize quite simple aspects of the visual scene.) The problem is to avoid thinking of here being a central, ghostly, conscious self, fed information in the same way that a screen if fed information by a remote television camera. Once such a model is in place, experience will seem like a veil getting between 'us' and the world, and the direct objects of perception will seem to be private items in an inner theatre or sensorium. The difficulty of avoiding this model is epically cute when we considered the secondary qualities of colour, sound, tactile feelings and taste, which can easily seem to have a purely private existence inside the perceiver, like sensation of pain. Calling such supposed items names like 'sense-data' or 'percepts' exacerbates the tendency, but once the model is in place, the first property, that perception gives 'us' knowledge of the world and its surrounding surfaces, is quickly threatened, for there will now seem little connection between these items in immediate experience and any independent reality. Reactions to this problem include 'scepticism' and 'idealism.'

A more hopeful approach is to claim that the complexities of (3) and (4) explain how we can have direct acquaintance of the world, than suggesting that the acquaintance we do have been at best indirect. It is pointed out that perceptions are not like sensation, precisely because they have a content, or outer-directed nature. To have a perception is to be aware of the world for being such-and-such a way, than to enjoy a mere modification of sensation. But such direct realism has to be sustained in the face of the evident personal (neurophysiological and other) factors determining how we perceive. One approach is to ask why it is useful to be conscious of what we perceive, when other aspects of our functioning work with information determining responses without any conscious awareness or intervention. A solution to this problem would offer the hope of making consciousness part of the natural world, than a strange optional extra.

Furthering, perceptual knowledge is knowledge acquired by or through the senses and includes most of what we know. We cross intersections when we see the light turn green, head for the kitchen when we smell the roast burning, squeeze the fruit to determine its ripeness, and climb out of bed when we hear the alarm ring. In each case we come to know something-that the light has turned green, that the roast is burning, that the melon is overripe, and that it is time to get up-by some sensory means. Seeing that the light has turned green is learning something-that, the light has turned green-by use of the eyes. Feeling that the melon is overripe is coming to know a fact-that the melon is overripe-by one's sense to touch. In each case the resulting knowledge is somehow based on, derived from or grounded in the sort of experience that characterizes the sense modality in question.

Much as much of our perceptual knowledge is indirect, dependent or derived. By this I mean that the facts we describe ourselves as learning, as coming to know, by perceptual means are pieces of knowledge that depend on our coming to know something else, some other fact, in a more direct way. We see, by the gauge, that we need gas, see, by the newspapers, that our team has lost again, see, by her expression, that she is nervous. This derived or dependent sort of knowledge is particularly prevalent in the cases of vision, but it occurs, to a lesser degree, in every sense modality. We install bells and other noise-makers so that we calm for example, hear (by the bell) that someone is at the door and (by the alarm) that its time to get up. When we obtain knowledge in this way, it is clear that unless one sees-hence, comes to know something about the gauge (that it says) and (hence, know) that one is described as coming to know by perceptual means. If one cannot hear that the bell is ringing, one cannot-in at least in this way-hear that one's visitors have arrived. In such cases one sees (hears, smells, etc.) that 'a' is 'F', coming to know thereby that 'a' is 'F', by seeing (hearing, etc.) that some other condition, 'b's' being 'G', obtains when this occurs, the knowledge (that 'a' is 'F') is derived from, or dependent on, the more basic perceptual knowledge that 'b' is 'G'.

And finally, the representational Theory of mind, (which goes back at least to Aristotle) takes as its starting point commonsense mental states, such as thoughts, beliefs, desires, perceptions and images. Such states are said to have 'intentionality' - they are about or refer to things, and may be evaluated with respect to properties like consistency, truth, appropriateness and accuracy. (For example, the thought that cousins are not related is inconsistent, the belief that Elvis is dead is true, the desire to eat the moon is inappropriate, a visual experience of a ripe strawberry as red is accurate, an image of George W. Bush with dreadlocks is inaccurate.)

The Representational Theory of Mind, defines such intentional mental states as relations to mental representations, and explains the intentionality of the former in terms of the semantic properties of the latter. For example, to believe that Elvis is dead is to be appropriately related to a mental representation whose propositional content is that Elvis is dead. (The desire that Elvis be dead, the fear that he is dead, the regret that he is dead, etc., involve different relations to the same mental representation.) To perceive a strawberry is to have a sensory experience of some kind which is appropriately related to (e.g., caused by) the strawberry Representational theory of mind also understands mental processes such as thinking, reasoning and imagining as sequences of intentional mental states. For example, to imagine the moon rising over a mountain is to entertain a series of mental images of the moon (and a mountain). To infer a proposition q from the proposition’s p and if 'p' then 'q' is (among other things) to have a sequence of thoughts of the form 'p', 'if p' then 'q', 'q'.

Contemporary philosophers of mind have typically supposed (or at least hoped) that the mind can be naturalized -, i.e., that all mental facts have explanations in the terms of natural science. This assumption is shared within cognitive science, which attempts to provide accounts of mental states and processes in terms (ultimately) of features of the brain and central nervous system. In the course of doing so, the various sub-disciplines of cognitive science (including cognitive and computational psychology and cognitive and computational neuroscience) postulate a number of different kinds of structures and processes, many of which are not directly implicated by mental states and processes as commonsensical conceived. There remains, however, a shared commitment to the idea that mental states and processes are to be explained in terms of mental representations.

In philosophy, recent debates about mental representation have centred around the existence of propositional attitudes (beliefs, desires, etc.) and the determination of their contents (how they come to be about what they are about), and the existence of phenomenal properties and their relation to the content of thought and perceptual experience. Within cognitive science itself, the philosophically relevant debates have been focussed on the computational architecture of the brain and central nervous system, and the compatibility of scientific and commonsense accounts of mentality.

Intentional Realists such as Dretske (e.g., 1988) and Fodor (e.g., 1987) note that the generalizations we apply in everyday life in predicting and explaining each other's behaviour (often collectively referred to as 'folk psychology') are both remarkably successful and indispensable. What a person believes, doubts, desires, fears, etc. is a highly reliable indicator of what that person will do. We have no other way of making sense of each other's behaviour than by ascribing such states and applying the relevant generalizations. We are thus committed to the basic truth of commonsense psychology and, hence, to the existence of the states its generalizations refer to. (Some realists, such as Fodor, also hold that commonsense psychology will be vindicated by cognitive science, given that propositional attitudes can be construed as computational relations to mental representations.)

Intentional Eliminativists, such as Churchland, (perhaps) Dennett and (at one time) Stich argue that no such things as propositional attitudes (and their constituent representational states) are implicated by the successful explanation and prediction of our mental lives and behaviour. Churchland denies that the generalizations of commonsense propositional-attitude psychology are true. He (1981) argues that folk psychology is a theory of the mind with a long history of failure and decline, and that it resists incorporation into the framework of modern scientific theories (including cognitive psychology). As such, it is comparable to alchemy and phlogiston theory, and ought to suffer a comparable fate. Commonsense psychology is false, and the states (and representations) it postulates simply don't exist. (It should be noted that Churchland is not an eliminativist about mental representation tout court.

Dennett (1987) grants that the generalizations of commonsense psychology are true and indispensable, but denies that this is sufficient reason to believe in the entities they appear to refer to. He argues that to give an intentional explanation of a system's behaviour is merely to adopt the 'intentional stance' toward it. If the strategy of assigning contentful states to a system and predicting and explaining its behaviour (on the assumption that it is rational -, i.e., that it behaves as it should, given the propositional attitudes it should have in its environment) is successful, then the system is intentional, and the propositional-attitude generalizations we apply to it are true. But there is nothing more to having a propositional attitude than this.

Though he has been taken to be thus claiming that intentional explanations should be construed instrumentally, Dennett (1991) insists that he is a 'moderate' realist about propositional attitudes, since he believes that the patterns in the behaviour and behavioural dispositions of a system on the basis of which we (truly) attribute intentional states to it are objectively real. In the event that there are two or more explanatorily adequate but substantially different systems of intentional ascriptions to an individual, however, Dennett claims there is no fact of the matter about what the system believes (1987, 1991). This does suggest an irrealism at least with respect to the sorts of things Fodor and Dretske take beliefs to be; though it is not the view that there is simply nothing in the world that makes intentional explanations true.

(Davidson 1973, 1974 and Lewis 1974 also defend the view that what it is to have a propositional attitude is just to be interpretable in a particular way. It is, however, not entirely clear whether they intend their views to imply irrealism about propositional attitudes.). Stich (1983) argues that cognitive psychology does not (or, in any case, should not) taxonomize mental states by their semantic properties at all, since attribution of psychological states by content is sensitive to factors that render it problematic in the context of a scientific psychology. Cognitive psychology seeks causal explanations of behaviour and cognition, and the causal powers of a mental state are determined by its intrinsic 'structural' or 'syntactic' properties. The semantic properties of a mental state, however, are determined by its extrinsic properties -, e.g., its history, environmental or intra-mental relations. Hence, such properties cannot figure in causal-scientific explanations of behaviour. (Fodor 1994 and Dretske 1988 are realist attempts to come to grips with some of these problems.) Stich proposes a syntactic theory of the mind, on which the semantic properties of mental states play no explanatory role.

It is a traditional assumption among realists about mental representations that representational states come in two basic varieties (Boghossian 1995). There are those, such as thoughts, which are composed of concepts and have no phenomenal ('what-it's-like') features ('qualia'), and those, such as sensory experiences, which have phenomenal features but no conceptual constituents. (Non-conceptual content is usually defined as a kind of content that states of a creature lacking concepts but, nonetheless enjoy. On this taxonomy, mental states can represent either in a way analogous to expressions of natural languages or in a way analogous to drawings, paintings, maps or photographs. (Perceptual states such as seeing that something is blue, are sometimes thought of as hybrid states, consisting of, for example, a Non-conceptual sensory experience and a thought, or some more integrated compound of sensory and conceptual components.)

Some historical discussions of the representational properties of mind (e.g., Aristotle 1984, Locke 1689/1975, Hume 1739/1978) seem to assume that Non-conceptual representations - percepts ('impressions'), images ('ideas') and the like - are the only kinds of mental representations, and that the mind represents the world in virtue of being in states that resemble things in it. On such a view, all representational states have their content in virtue of their phenomenal features. Powerful arguments, however, focussing on the lack of generality (Berkeley 1975), ambiguity (Wittgenstein 1953) and non-compositionality (Fodor 1981) of sensory and imagistic representations, as well as their unsuitability to function as logical (Frége 1918/1997, Geach 1957) or mathematical (Frége 1884/1953) concepts, and the symmetry of resemblance (Goodman 1976), convinced philosophers that no theory of mind can get by with only Non-conceptual representations construed in this way.

Contemporary disagreement over Non-conceptual representation concerns the existence and nature of phenomenal properties and the role they play in determining the content of sensory experience. Dennett (1988), for example, denies that there are such things as qualia at all; while Brandom (2002), McDowell (1994), Rey (1991) and Sellars (1956) deny that they are needed to explain the content of sensory experience. Among those who accept that experiences have phenomenal content, some (Dretske, Lycan, Tye) argue that it is reducible to a kind of intentional content, while others (Block, Loar, Peacocke) argue that it is irreducible.

The representationalist thesis is often formulated as the claim that phenomenal properties are representational or intentional. However, this formulation is ambiguous between a reductive and a non-deductive claim (though the term 'representationalism' is most often used for the reductive claim). On one hand, it could mean that the phenomenal content of an experience is a kind of intentional content (the properties it represents). On the other, it could mean that the (irreducible) phenomenal properties of an experience determine an intentional content. Representationalists such as Dretske, Lycan and Tye would assent to the former claim, whereas phenomenalists such as Block, Chalmers, Loar and Peacocke would assent to the latter. (Among phenomenalists, there is further disagreement about whether qualia are intrinsically representational (Loar) or not (Block, Peacocke).

Most (reductive) representationalists are motivated by the conviction that one or another naturalistic explanation of intentionality is, in broad outline, correct, and by the desire to complete the naturalization of the mental by applying such theories to the problem of phenomenality. (Needless to say, most phenomenalists (Chalmers is the major exception) are just as eager to naturalize the phenomenal - though not in the same way.)

The main argument for representationalism appeals to the transparency of experience. The properties that characterize what it's like to have a perceptual experience are presented in experience as properties of objects perceived: in attending to an experience, one seems to 'see through it' to the objects and properties it is experiences of. They are not presented as properties of the experience itself. If nonetheless they were properties of the experience, perception would be massively deceptive. But perception is not massively deceptive. According to the representationalist, the phenomenal character of an experience is due to its representing objective, non-experiential properties. (In veridical perception, these properties are locally instantiated; in illusion and hallucination, they are not.) On this view, introspection is indirect perception: one comes to know what phenomenal features one's experience has by coming to know what objective features it represents.

In order to account for the intuitive differences between conceptual and sensory representations, representationalists appeal to their structural or functional differences. Dretske (1995), for example, distinguishes experiences and thoughts on the basis of the origin and nature of their functions: an experience of a property 'P' is a state of a system whose evolved function is to indicate the presence of 'P' in the environment; a thought representing the property 'P', on the other hand, is a state of a system whose assigned (learned) function is to calibrate the output of the experiential system. Rey (1991) takes both thoughts and experiences to be relations to sentences in the language of thought, and distinguishes them on the basis of (the functional roles of) such sentences' constituent predicates. Lycan (1987, 1996) distinguishes them in terms of their functional-computational profiles. Tye (2000) distinguishes them in terms of their functional roles and the intrinsic structure of their vehicles: thoughts are representations in a language-like medium, whereas experiences are image-like representations consisting of 'symbol-filled arrays.' (The account of mental images in Tye 1991.)

Phenomenalists tend to make use of the same sorts of features (function, intrinsic structure) in explaining some of the intuitive differences between thoughts and experiences; but they do not suppose that such features exhaust the differences between phenomenal and non-phenomenal representations. For the phenomenalist, it is the phenomenal properties of experiences - qualia themselves - that constitute the fundamental difference between experience and thought. Peacocke (1992), for example, develops the notion of a perceptual 'scenario' (an assignment of phenomenal properties to coordinates of a three-dimensional egocentric space), whose content is 'correct' (a semantic property) if in the corresponding 'scene' (the portion of the external world represented by the scenario) properties are distributed as their phenomenal analogues are in the scenario.

Another sort of representation championed by phenomenalists (e.g., Block, Chalmers (2003) and Loar (1996)) is the 'phenomenal concept' -, a conceptual/phenomenal hybrid consisting of a phenomenological 'sample' (an image or an occurrent sensation) integrated with (or functioning as) a conceptual component. Phenomenal concepts are postulated to account for the apparent fact (among others) that, as McGinn (1991) puts it, 'you cannot form [introspective] concepts of conscious properties unless you yourself instantiate those properties.' One cannot have a phenomenal concept of a phenomenal property 'P', and, hence, phenomenal beliefs about P, without having experience of 'P', because 'P' itself is (in some way) constitutive of the concept of 'P'. (Jackson 1982, 1986 and Nagel 1974.)

Though imagery has played an important role in the history of philosophy of mind, the important contemporary literature on it is primarily psychological. In a series of psychological experiments done in the 1970s (summarized in Kosslyn 1980 and Shepard and Cooper 1982), subjects' response time in tasks involving mental manipulation and examination of presented figures was found to vary in proportion to the spatial properties (size, orientation, etc.) of the figures presented. The question of how these experimental results are to be explained has kindled a lively debate on the nature of imagery and imagination.

Kosslyn (1980) claims that the results suggest that the tasks were accomplished via the examination and manipulation of mental representations that they have spatial properties, i.e., pictorial representations, or images. Others, principally Pylyshyn (1979, 1981, 2003), argue that the empirical facts can be explained in terms exclusively of discursive, or propositional representations and cognitive processes defined over them. (Pylyshyn takes such representations to be sentences in a language of thought.)

The idea that pictorial representations are literally pictures in the head is not taken seriously by proponents of the pictorial view of imagery. The claim is, rather, that mental images represent in a way that is relevantly like the way pictures represent. (Attention has been focussed on visual imagery - hence the designation 'pictorial'; Though of course, there may imagery in other modalities - auditory, olfactory, etc. - as well.)

The distinction between pictorial and discursive representation can be characterized in terms of the distinction between analog and digital representation (Goodman 1976). This distinction has itself been variously understood (Fodor & Pylyshyn 1981, Goodman 1976, Haugeland 1981, Lewis 1971, McGinn 1989), though a widely accepted construal is that analog representation is continuous (i.e., in virtue of continuously variable properties of the representation), while digital representation is discrete (i.e., in virtue of properties a representation either has or doesn't have) (Dretske 1981). (An analog/digital distinction may also be made with respect to cognitive processes. (Block 1983.)) On this understanding of the analog/digital distinction, imagistic representations, which represent in virtue of properties that may vary continuously (such for being more or less bright, loud, vivid, etc.), would be analog, while conceptual representations, whose properties do not vary continuously (a thought cannot be more or less about Elvis: either it is or it is not) would be digital.

It might be supposed that the pictorial/discursive distinction is best made in terms of the phenomenal/nonphenomenal distinction, but it is not obvious that this is the case. For one thing, there may be nonphenomenal properties of representations that vary continuously. Moreover, there are ways of understanding pictorial representation that presuppose neither phenomenality nor analogicity. According to Kosslyn (1980, 1982, 1983), a mental representation is 'quasi-pictorial' when every part of the representation corresponds to a part of the object represented, and relative distances between parts of the object represented are preserved among the parts of the representation. But distances between parts of a representation can be defined functionally rather than spatially - for example, in terms of the number of discrete computational steps required to combine stored information about them. (Rey 1981.)

Tye (1991) proposes a view of images on which they are hybrid representations, consisting both of pictorial and discursive elements. On Tye's account, images are '(labelled) interpreted symbol-filled arrays.' The symbols represent discursively, while their arrangement in arrays has representational significance (the location of each 'cell' in the array represents a specific viewer-centred 2-D location on the surface of the imagined object)

The contents of mental representations are typically taken to be abstract objects (properties, relations, propositions, sets, etc.). A pressing question, especially for the naturalist, is how mental representations come to have their contents. Here the issue is not how to naturalize content (abstract objects can't be naturalized), but, rather, how to provide a naturalistic account of the content-determining relations between mental representations and the abstract objects they express. There are two basic types of contemporary naturalistic theories of content-determination, causal-informational and functional.

Causal-informational theories hold that the content of a mental representation is grounded in the information it carries about what does (Devitt 1996) or would (Fodor 1987, 1990) cause it to occur. There is, however, widespread agreement that causal-informational relations are not sufficient to determine the content of mental representations. Such relations are common, but representation is not. Tree trunks, smoke, thermostats and ringing telephones carry information about what they are causally related to, but they do not represent (in the relevant sense) what they carry information about. Further, a representation can be caused by something it does not represent, and can represent something that has not caused it.

The main attempts to specify what makes a causal-informational state a mental representation are Asymmetric Dependency Theories, the Asymmetric Dependency Theory distinguishes merely informational relations from representational relations on the basis of their higher-order relations to each other: informational relations depend upon representational relations, but not vice-versa. For example, if tokens of a mental state type are reliably caused by horses, cows-on-dark-nights, zebras-in-the-mist and Great Danes, then they carry information about horses, etc. If, however, such tokens are caused by cows-on-dark-nights, etc. because they were caused by horses, but not vice versa, then they represent horses.

According to Teleological Theories, representational relations are those a representation-producing mechanism has the selected (by evolution or learning) function of establishing. For example, zebra-caused horse-representations do not mean zebra, because the mechanism by which such tokens are produced has the selected function of indicating horses, not zebras. The horse-representation-producing mechanism that responds to zebras is malfunctioning.

Functional theories, hold that the content of a mental representation are well grounded in causal computational inferential relations to other mental portrayals other than mental representations. They differ on whether relata should include all other mental representations or only some of them, and on whether to include external states of affairs. The view that the content of a mental representation is determined by its inferential/computational relations with all other representations is holism; the view it is determined by relations to only some other mental states is localisms (or molecularism). (The view that the content of a mental state depends on none of its relations to other mental states is atomism.) Functional theories that recognize no content-determining external relata have been called solipsistic (Harman 1987). Some theorists posit distinct roles for internal and external connections, the former determining semantic properties analogous to sense, the latter determining semantic properties analogous to reference (McGinn 1982, Sterelny 1989)

(Reductive) representationalists (Dretske, Lycan, Tye) usually take one or another of these theories to provide an explanation of the (Non-conceptual) content of experiential states. They thus tend to be Externalists (see the next section) about phenomenological as well as conceptual content. Phenomenalists and non-deductive representationalists (Block, Chalmers, Loar, Peacocke, Siewert), on the other hand, take it that the representational content of such states is (at least in part) determined by their intrinsic phenomenal properties. Further, those who advocate a phenomenology-based approach to conceptual content (Horgan and Tiensen, Loar, Pitt, Searle, Siewert) also seem to be committed to internalist individuation of the content (if not the reference) of such states.

Generally, those who, like informational theorists, think relations to one's (natural or social) environment are (at least partially) determinative of the content of mental representations are Externalists (e.g., Burge 1979, 1986, McGinn 1977, Putnam 1975), whereas those who, like some proponents of functional theories, think representational content is determined by an individual's intrinsic properties alone, are internalists (or individualists).

This issue is widely taken to be of central importance, since psychological explanation, whether commonsense or scientific, is supposed to be both causal and content-based. (Beliefs and desires cause the behaviours they do because they have the contents they do. For example, the desire that one have a beer and the beliefs that there is beer in the refrigerator and that the refrigerator is in the kitchen may explain one's getting up and going to the kitchen.) If, however, a mental representation's having a particular content is due to factors extrinsic to it, it is unclear how its having that content could determine its causal powers, which, arguably, must be intrinsic. Some who accept the standard arguments for externalism have argued that internal factors determine a component of the content of a mental representation. They say that mental representations have both 'narrow' content (determined by intrinsic factors) and 'wide' or 'broad' content (determined by narrow content plus extrinsic factors). (This distinction may be applied to the sub-personal representations of cognitive science as well as to those of commonsense psychology.

Narrow content has been variously construed. Putnam (1975), Fodor (1982)), and Block (1986), for example, seem to understand it as something like dedictorial content (i.e., Frégean sense, or perhaps character, à la Kaplan 1989). On this construal, narrow content is context-independent and directly expressible. Fodor (1987) and Block (1986), however, have also characterized narrow content as radically inexpressible. On this construal, narrow content is a kind of proto-content, or content-determinant, and can be specified only indirectly, via specifications of context/wide-content pairings. Both, construe of as a narrow content and are characterized as functions from context to (wide) content. The narrow content of a representation is determined by properties intrinsic to it or its possessor such as its syntactic structure or its intra-mental computational or inferential role (or its phenomenology.

Burge (1986) has argued that causation-based worries about externalist individuation of psychological content, and the introduction of the narrow notion, are misguided. Fodor (1994, 1998) has more recently urged that a scientific psychology might not need narrow content in order to supply naturalistic (causal) explanations of human cognition and action, since the sorts of cases they were introduced to handle, viz., Twin-Earth cases and Frége cases, are nomologically either impossible or dismissible as exceptions to non-strict psychological laws.

The leading contemporary version of the Representational Theory of Mind, the Computational Theory of Mind, claims that the brain is a kind of computer and that mental processes are computations. According to the computational theory of mind, cognitive states are constituted by computational relations to mental representations of various kinds, and cognitive processes are sequences of such states. The computational theory of mind and the representational theory of mind, may by attempting to explain all psychological states and processes in terms of mental representation. In the course of constructing detailed empirical theories of human and animal cognition and developing models of cognitive processes’ implementable in artificial information processing systems, cognitive scientists have proposed a variety of types of mental representations. While some of these may be suited to be mental relata of commonsense psychological states, some - so-called 'subpersonal' or 'sub-doxastic' representations - are not. Though many philosophers believe that computational theory of mind can provide the best scientific explanations of cognition and behaviour, there is disagreement over whether such explanations will vindicate the commonsense psychological explanations of prescientific representational theory of mind.

According to Stich's (1983) Syntactic Theory of Mind, for example, computational theories of psychological states should concern themselves only with the formal properties of the objects those states are relations to. Commitment to the explanatory relevance of content, however, is for most cognitive scientists fundamental. That mental processes are computations, which computations are rule-governed sequences of semantically evaluable objects, and that the rules apply to the symbols in virtue of their content, are central tenets of mainstream cognitive science.

Explanations in cognitive science appeal to a many different kinds of mental representation, including, for example, the 'mental models' of Johnson-Laird 1983, the 'retinal arrays,' 'primal sketches' and '2½ -D sketches' of Marr 1982, the 'frames' of Minsky 1974, the 'sub-symbolic' structures of Smolensky 1989, the 'quasi-pictures' of Kosslyn 1980, and the 'interpreted symbol-filled arrays' of Tye 1991 - in addition to representations that may be appropriate to the explanation of commonsense

Psychological states. Computational explanations have been offered of, among other mental phenomena, belief.

The classicists hold that mental representations are symbolic structures, which typically have semantically evaluable constituents, and that mental processes are rule-governed manipulations of them that are sensitive to their constituent structure. The connectionists, hold that mental representations are realized by patterns of activation in a network of simple processors ('nodes') and that mental processes consist of the spreading activation of such patterns. The nodes themselves are, typically, not taken to be semantically evaluable; nor do the patterns have semantically evaluable constituents. (Though there are versions of Connectionism -, 'localist' versions - on which individual nodes are taken to have semantic properties (e.g., Ballard 1986, Ballard & Hayes 1984).) It is arguable, however, that localist theories are neither definitive nor representative of the Conceptionist program.

Classicists are motivated (in part) by properties thought seems to share with language. Jerry Alan Fodor's (1935-), Language of Thought Hypothesis, (Fodor 1975, 1987), according to which the system of mental symbols constituting the neural basis of thought is structured like a language, provides a well-worked-out version of the classical approach as applied to commonsense psychology. According to the language of a thought hypothesis, the potential infinity of complex representational mental states is generated from a finite stock of primitive representational states, in accordance with recursive formation rules. This combinatorial structure accounts for the properties of productivity and systematicity of the system of mental representations. As in the case of symbolic languages, including natural languages (though Fodor does not suppose either that the language of thought hypotheses explains only linguistic capacities or that only verbal creatures have this sort of cognitive architecture), these properties of thought are explained by appeal to the content of the representational units and their combinability into contentful complexes. That is, the semantics of both language and thought is compositional: the content of a complex representation is determined by the contents of its constituents and their structural configuration.

Connectionists are motivated mainly by a consideration of the architecture of the brain, which apparently consists of layered networks of interconnected neurons. They argue that this sort of architecture is unsuited to carrying out classical serial computations. For one thing, processing in the brain is typically massively parallel. In addition, the elements whose manipulation drive’s computation in Conceptionist networks (principally, the connections between nodes) are neither semantically compositional nor semantically evaluable, as they are on the classical approach. This contrast with classical computationalism is often characterized by saying that representation is, with respect to computation, distributed as opposed to local: representation is local if it is computationally basic; and distributed if it is not. (Another way of putting this is to say that for classicists mental representations are computationally atomic, whereas for connectionists they are not.)

Moreover, connectionists argue that information processing as it occurs in Conceptionist networks more closely resembles some features of actual human cognitive functioning. For example, whereas on the classical view learning involves something like hypothesis formation and testing (Fodor 1981), on the Conceptionist model it is a matter of evolving distribution of 'weight' (strength) on the connections between nodes, and typically does not involve the formulation of hypotheses regarding the identity conditions for the objects of knowledge. The Conceptionist network is 'trained up' by repeated exposure to the objects it is to learn to distinguish; and, though networks typically require many more exposures to the objects than do humans, this seems to model at least one feature of this type of human learning quite well.

Further, degradation in the performance of such networks in response to damage is gradual, not sudden as in the case of a classical information processor, and hence more accurately models the loss of human cognitive function as it typically occurs in response to brain damage. It is also sometimes claimed that Conceptionist systems show the kind of flexibility in response to novel situations typical of human cognition - situations in which classical systems are relatively 'brittle' or 'fragile.'

Some philosophers have maintained that Connectionism entails that there are no propositional attitudes. Ramsey, Stich and Garon (1990) have argued that if Conceptionist models of cognition are basically correct, then there are no discrete representational states as conceived in ordinary commonsense psychology and classical cognitive science. Others, however (e.g., Smolensky 1989), hold that certain types of higher-level patterns of activity in a neural network may be roughly identified with the representational states of commonsense psychology. Still others argue that language-of-thought style representation is both necessary in general and realizable within Conceptionist architectures, collect the central contemporary papers in the classicist/Conceptionist debate, and provides useful introductory material as well.

Whereas Stich (1983) accepts that mental processes are computational, but denies that computations are sequences of mental representations, others accept the notion of mental representation, but deny that computational theory of mind provides the correct account of mental states and processes.

Van Gelder (1995) denies that psychological processes are computational. He argues that cognitive systems are dynamic, and that cognitive states are not relations to mental symbols, but quantifiable states of a complex system consisting of (in the case of human beings) a nervous system, a body and the environment in which they are embedded. Cognitive processes are not rule-governed sequences of discrete symbolic states, but continuous, evolving total states of dynamic systems determined by continuous, simultaneous and mutually determining states of the systems components. Representation in a dynamic system is essentially information-theoretic, though the bearers of information are not symbols, but state variables or parameters.

Horst (1996), on the other hand, argues that though computational models may be useful in scientific psychology, they are of no help in achieving a philosophical understanding of the intentionality of commonsense mental states. Computational theory of mind attempts to reduce the intentionality of such states to the intentionality of the mental symbols they are relations to. But, Horst claims, the relevant notion of symbolic content is essentially bound up with the notions of convention and intention. So the computational theory of mind involves itself in a vicious circularity: the very properties that are supposed to be reduced are (tacitly) appealed to in the reduction.

To say that a mental object has semantic properties is, paradigmatically, to say that it may be about, or be true or false of, an object or objects, or that it may be true or false simpliciter. Suppose I think that you took to sniffing snuff. I am thinking about you, and if what I think of you (that they take snuff) is true of you, then my thought is true. According to representational theory of mind such states are to be explained as relations between agents and mental representations. To think that you take snuff is to token in some way a mental representation whose content is that ocelots take snuff. On this view, the semantic properties of mental states are the semantic properties of the representations they are relations to.

Linguistic acts seem to share such properties with mental states. Suppose I say that you take snuff. I am talking about you, and if what I say of you (that they take snuff) is true of them, then my utterance is true. Now, to say that you take snuff is (in part) to utter a sentence that means that you take snuff. Many philosophers have thought that the semantic properties of linguistic expressions are inherited from the intentional mental states they are conventionally used to express. On this view, the semantic properties of linguistic expressions are the semantic properties of the representations that are the mental relata of the states they are conventionally used to express.

It is also widely held that in addition to having such properties as reference, truth-conditions and truth - so-called extensional properties - expressions of natural languages also have intensional properties, in virtue of expressing properties or propositions - i.e., in virtue of having meanings or senses, where two expressions may have the same reference, truth-conditions or truth value, yet express different properties or propositions (Frége 1892/1997). If the semantic properties of natural-language expressions are inherited from the thoughts and concepts they express (or vice versa, or both), then an analogous distinction may be appropriate for mental representations.

Theories of representational content may be classified according to whether they are atomistic or holistic and according to whether they are externalistic or internalistic, whereby, emphasizing the priority of a whole over its parts. Furthermore, in the philosophy of language, this becomes the claim that the meaning of an individual word or sentence can only be understood in terms of its relation to an indefinitely larger body of language, such as à whole theory, or even a whole language or form of life. In the philosophy of mind a mental state similarly may be identified only in terms of its relations with others. Moderate holism may allow the other things besides these relationships also count ; extreme holism would hold that a network of relationships is all that we have. A holistic view of science holds that experience only confirms or disconfirms large bodies of doctrine, impinging at the edges, and leaving some leeway over the adjustment that it requires.

Once, again, in the philosophy of mind and language, the view that what is thought, or said, or experienced, is essentially dependent on aspects of the world external to the mind of the subject. The view goes beyond holding that such mental states are typically caused by external factors, to insist that they could not have existed as they now do without the subject being embedded in an external world of a certain kind. It is these external relations that make up the essence or identify of the mental state. Externalism is thus opposed to the Cartesian separation of the mental from the physical, since that holds that the mental could in principle exist as it does even if there were no external world at all. Various external factors have been advanced as ones on which mental content depends, including the usage of experts, the linguistic, norms of the community. And the general causal relationships of the subject. In the theory of knowledge, externalism is the view that a person might know something by being suitably situated with respect to it, without that relationship being in any sense within his purview. The person might, for example, be very reliable in some respect without believing that he is. The view allows that you can know without being justified in believing that you know.

However, atomistic theories take a representation’s content to be something that can be specified independent entity of that representation’ s relations to other representations. What the American philosopher of mind, Jerry Alan Fodor (1935-) calls the crude causal theory, for example, takes a representation to be a
cow
- a menial representation with the same content as the word ‘cow’ - if its tokens are caused by instantiations of the property of being-a-cow, and this is a condition that places no explicit constraints on how
cow
’s must or might relate to other representations. Holistic theories contrasted with atomistic theories in taking the relations à representation bears to others to be essential to its content. According to functional role theories, a representation is a
cow
if it behaves like a
cow
should behave in inference.

Internalist theories take the content of a representation to be a matter determined by factors internal to the system that uses it. Thus, what Block (1986) calls ‘short-armed’ functional role theories are internalist. Externalist theories take the content of a representation to be determined, in part at least, by factors external to the system that uses it. Covariance theories, as well as telelogical theories that invoke an historical theory of functions, take content to be determined by ‘external’ factors. Crossing the atomist-holistic distinction with the internalist-externalist distinction.

Externalist theories (sometimes called non-individualistic theories) have the consequence that molecule for molecule identical cognitive systems might yet harbour representations with different contents. This has given rise to a controversy concerning ‘narrow’ content. If we assume some form of externalist theory is correct, then content is, in the first instance ‘wide’ content, i.e., determined in part by factors external to the representing system. On the other hand, it seems clear that, on plausible assumptions about how to individuate psychological capacities, internally equivalent systems must have the same psychological capacities. Hence, it would appear that wide content cannot be relevant to characterizing psychological equivalence. Since cognitive science generally assumes that content is relevant to characterizing psychological equivalence, philosophers attracted to externalist theories of content have sometimes attempted to introduce ‘narrow’ content, i.e., an aspect or kind of content that is equivalent internally equivalent systems. The simplest such theory is Fodor’s idea (1987) that narrow content is a function from contents (i.e., from whatever the external factors are) to wide contents.

All the same, what a person expresses by a sentence is often a function of the environment in which he or she is placed. For example, the disease I refer to by the term like ‘arthritis’, or the kind of tree I refer to as a ‘Maple’ will be defined by criteria of which I know next to nothing. This raises the possibility of imagining two persons in rather different environments, but in which everything appears the same to each of them. The wide content of their thoughts and sayings will be different if the situation surrounding them is appropriately different: ‘situation’ may include the actual objects they perceive or the chemical or physical kinds of object in the world they inhabit, or the history of their words, or the decisions of authorities on what counts as an example, of one of the terms they use. The narrow content is that part of their thought which remains identical, through their identity of the way things appear, regardless of these differences of surroundings. Partisans of wide content may doubt whether any content in this sense narrow, partisans of narrow content believer that it is the fundamental notion, with wide content being explicable in terms of narrow content plus context.

Even so, the distinction between facts and values has outgrown its name: it applies not only to matters of fact vs, matters of value, but also to statements that something is, vs. statements that something ought to be. Roughly, factual statements - ‘is statements’ in the relevant sense - represent some state of affairs as obtaining, whereas normative statements - evaluative, and deontic ones - attribute goodness to something, or ascribe, to an agent, an obligation to act. Neither distinction is merely linguistic. Specifying a book’s monetary value is making a factual statement, though it attributes a kind of value. ‘That is a good book’ expresses a valu judgement though the term ‘value’ is absent (nor would ‘valuable’ be synonymous with ‘good’). Similarly, ‘we are morally obligated to fight’ superficially expresses a statement, and ‘By all indications it ough to rain’ makes a kind of ought-claim; but the former is an ought-statement, the latter an (epistemic) is-statement.

Theoretical difficulties also beset the distinction. Some have absorbed values into facts holding that all value is instrumental, roughly, to have value is to contribute - in a factual analysable way - to something further which is (say) deemed desirable. Others have suffused facts with values, arguing that facts (and observations) are ‘theory-impregnated’ and contending that values are inescapable to theoretical choice. But while some philosophers doubt that fact/value distinctions can be sustained, there persists a sense of a deep difference between evaluating, or attributing an obligation and, on the other hand, saying how the world is.

Fact/value distinctions, may be defended by appeal to the notion of intrinsic value, as à thing has in itself and thus independently of its consequences. Roughly, a valu statement (proper) is an ascription of intrinsic value, one to the effect that a thing is to some degree good in itself. This leaves open whether ought-statements are implicitly value statements, but even if they imply that something has intrinsic value - e.g., moral value - they can be independently characterized, say by appeal to rules that provide (justifying) reasons for action. One might also ground the fact value distinction in the attributional (or even motivational) component apparently implied by the making of valuational or deontic judgements: Thus, ‘it is a good book, but that is no reason for a positive attribute towards it’ and ‘you ought to do it, but there is no reason to’ seem inadmissible, whereas, substituting, ‘an expensive book’ and ‘you will do it’ yields permissible judgements. One might also argue that factual judgements are the kind which are in principle appraisable scientifically, and thereby anchor the distinction on the factual side. This ligne is plausible, but there is controversy over whether scientific procedures are ‘value-free’ in the required way.

Philosophers differ regarding the sense, if any, in which epistemology is normative (roughly, valuational). But what precisely is at stake in this controversy is no clearly than the problematic fact/value distinction itself. Must epistemologists as such make judgements of value or epistemic responsibility ? If epistemology is naturalizable, then even epistemic principles simply articulate under what conditions - say, appropriate perceptual stimulations - a belief is justified, or constitutes knowledge. Its standards of justification, then would be like standards of, e.g., resilience for bridges. It is not obvious, however, that there appropriate standards can be established without independent judgements that, say, a certain kind of evidence is good enough for justified belief (or knowledge). The most plausible view may be that justification is like intrinsic goodness, though it supervenes on natural properties, it cannot be analysed wholly in factuel statements.

Thus far, belief has been depicted as being all-or-nothing, however, as à resulting causality for which we have grounds for thinking it true, and, all the same, its acceptance is governed by epistemic norms, and, least of mention, it is partially subject to voluntary control and has functional affinities to belief. Still, the notion of acceptance, like that of degrees of belief, merely extends the standard picture, and does not replace it.

Traditionally, belief has been of epistemological interest in its propositional guise: ‘S’ believes that ‘p’, where ‘p’ is a reposition towards which an agent, ‘S’ exhibits an attitude of acceptance. Not all belief is of this sort. If I trust you to say, I believer you. And someone may believe in Mr. Radek, or in a free-market economy, or in God. It is sometimes supposed that all belief is ‘reducible’ to propositional belief, belief-that. Thus, my believing you might be thought a matter of my believing, is, perhaps, that what you say is true, and your belief in free markets or God, is a matter of your believing that free-market economies are desirable or that God exists.

Some philosophers have followed St. Thomas Aquinas (1225-74), in supposing that to believer in God is simply to believer that certain truths hold while others argue that belief-in is a distinctive attitude, on that includes essentially an element of trust. More commonly, belief-in has been taken to involve a combination of propositional belief together with some further attitude.

The moral philosopher Richard Price (1723-91) defends the claim that there are different sorts of belief-in, some, but not all reducible to beliefs-that. If you believer in God, you believer that God exists, that God is good, you believer that God is good, etc. But according to Price, your belief involves, in addition, a certain complex pro-attitude toward its object. Even so, belief-in outruns the evidence for the corresponding belief-that. Does this diminish its rationality ? If belief-in presupposes believes-that, it might be thought that the evidential standards for the former must be, at least, as high as standards for the latter. And any additional pro-attitude might be thought to require a further layer of justification not required for cases of belief-that.

Belief-in may be, in general, less susceptible to alternations in the face of unfavourable evidence than belief-that. A believer who encounters evidence against God’s existence may remain unshaken in his belief, in part because the evidence does not bear on his pro-attitude. So long as this ids united with his belief that God exists, and reasonably so - in a way that an ordinary propositional belief that would not.

The correlative way of elaborating on the general objection to justificatory externalism challenges the sufficiency of the various externalist conditions by citing cases where those conditions are satisfied, but where the believers in question seem intuitively not to be justified. In this context, the most widely discussed examples have to do with possible occult cognitive capacities, like clairvoyance. Considering the point in application once, again, to reliabilism, the claim is that to think that he has such a cognitive power, and, perhaps, even good reasons to the contrary, is not rational or responsible and therefore not epistemically justified in accepting the belief that result from his clairvoyance, dispite the fact that the reliablist condition is satisfied.

One sort of response to this latter sorts of an objection is to ‘bite the bullet’ and insist that such believers are in fact justified, dismissing the seeming intuitions to the contrary as latent internalist prejudice. More widely adopted response attempts to impose additional conditions, usually of a roughly internalist sort, which will rule out the offending example, while stopping far of a full internalism. But, while there is little doubt that such modified versions of externalism can handle particular cases, as well enough to avoid clear intuitive implausibility, the usually problematic cases that they cannot handle, and also whether there is and clear motivation for the additional requirements other than the general internalist view of justification that externalist is committed to reject.

A view in this same general vein, one that might be described as a hybrid of internalism and externalism holds that epistemic justification requires that there is a justificatial factor that is cognitively accessible to the believer in question (though it need not be actually grasped), thus ruling out, e.g., a pure reliabilism. At the same time, however, though it must be objectively true that beliefs for which such a factor is available are likely to be true, in addition, the fact need not be in any way grasped or cognitively accessible to the believer. In effect, of the premises needed to argue that a particular belief is likely to be true, one must be accessible in a way that would satisfy at least weak internalism, the internalist will respond that this hybrid view is of no help at all in meeting the objection and has no belief nor is it held in the rational, responsible way that justification intuitively seems to require, for the believer in question, lacking one crucial premise, still has no reason at all for thinking that his belief is likely to be true.

An alternative to giving an externalist account of epistemic justification, one which may be more defensible while still accommodating many of the same motivating concerns, is to give an externalist account of knowledge directly, without relying on an intermediate account of justification. Such a view will obviously have to reject the justified true belief account of knowledge, holding instead that knowledge is true belief which satisfies the chosen externalist condition, e.g., a result of a reliable process (and perhaps, further conditions as well). This makes it possible for such a view to retain internalist account of epistemic justification, though the centrality of that concept to epistemology would obviously be seriously diminished.

Such an externalist account of knowledge can accommodate the commonsense conviction that animals, young children, and unsophisticated adults’ posse’s knowledge, though not the weaker conviction (if such a conviction does exist) that such individuals are epistemically justified in their beliefs. It is, at least, less vulnerable to internalist counter-examples of the sort discussed, since the intuitions involved there pertain more clearly to justification than to knowledge. What is uncertain is what ultimate philosophical significance the resulting conception of knowledge, for which is accepted or advanced as true or real on the basis of less than conclusive evidence, as can only be assumed to have. In particular, does it have any serious bearing on traditional epistemological problems and on the deepest and most troubling versions of scepticism, which seems in fact to be primarily concerned with justification, and knowledge?`

A rather different use of the terms ‘internalism’ and ‘externalism’ have to do with the issue of how the content of beliefs and thoughts is determined: According to an internalist view of content, the content of such intention states depends only on the non-relational, internal properties of the individual’s mind or grain, and not at all on his physical and social environment: While according to an externalist view, content is significantly affected by such external factors and suggests a view that appears of both internal and external elements are standardly classified as an external view.

As with justification and knowledge, the traditional view of content has been strongly internalist in character. The main argument for externalism derives from the philosophy y of language, more specifically from the various phenomena pertaining to natural kind terms, indexicals, etc. that motivate the views that have come to be known as ‘direct reference’ theories. Such phenomena seem at least to show that the belief or thought content that can be properly attributed to a person is dependant on facts about his environment, e.g., whether he is on Earth or Twin Earth, what is fact pointing at, the classificatory criterion employed by expects in his social group, etc. - not just on what is going on internally in his mind or brain.

An objection to externalist account of content is that they seem unable to do justice to our ability to know the content of our beliefs or thought ‘from the inside’, simply by reflection. If content is depending on external factors pertaining to the environment, then knowledge of content should depend on knowledge of these factors - which will not in general be available to the person whose belief or thought is in question.

The adoption of an externalist account of mental content would seem to support an externalist account of justification, apart from all contentful representation is a belief inaccessible to the believer, then both the justifying statuses of other beliefs in relation to that of the same representation are the status of that content, being totally rationalized by further beliefs for which it will be similarly inaccessible. Thus, contravening the internalist requirement for justification. An internalist must insist that there are no justification relations of these sorts, that our internally associable content can also not be warranted or as stated or indicated without the deviated departure from a course or procedure or from a norm or standard in showing no deviation from traditionally held methods of justification exacting by anything else: But such a response appears lame unless it is coupled with an attempt to show that the externalist account of content is mistaken.

Philosophers and scientists alike, have often held that the simplicity or parsimony of a theory is one reason, all else being equal, to view it as true. This goes beyond the unproblematic idea that simpler theories are easier to work with and gave greater aesthetic appeal.

One theory is more parsimonious than another when it postulates fewer entities, processes, changes or explanatory principles: The simplicity of a theory depends on essentially the same consecrations, though parsimony and simplicity obviously become the same. Demanding clarification of what makes one theory simpler or more parsimonious is plausible than another before the justification of these methodological maxims can be addressed.

If we set this description problem to one side, the major normative problem is as follows: What reason is there to think that simplicity is a sign of truth? Why should we accept a simpler theory instead of its more complex rivals? Newton and Leibniz thought that the answer was to be found in a substantive fact about nature. In Principia, Newton laid down as his first Rule of Reasoning in Philosophy that nature does nothing in vain . . . for Nature is pleased with simplicity and affects not the pomp of superfluous causes. Leibniz thought that the actual world obeys simple laws because Gods' taste for simplicity influenced his decision about which world to actualize.

The tragedy of the Western mind, described by Koyré, is a direct consequence of the stark Cartesian division between mind and world. We discovered the certain principles of physical reality, said Descartes, by the prejudices of the senses, but by the light of reason, and which thus possess so great evidence that we cannot doubt of their truth. Since the real, or that which actually exists external to ourselves, was in his view only that which could be represented in the quantitative terms of mathematics, Descartes concludes that all quantitative aspects of reality could be traced to the deceitfulness of the senses.

The most fundamental aspect of the Western intellectual tradition is the assumption that there is a fundamental division between the material and the immaterial world or between the realm of matter and the realm of pure mind or spirit. The metaphysical frame-work based on this assumption is known as ontological dualism. As the word dual implies, the framework is predicated on ontology, or a conception of the nature of God or being, that assumes reality has two distinct and separable dimensions. The concept of Being as continuous, immutable and having a prior or separate existence from the world of change dates from the ancient Greek philosopher Parmenides. The same qualities were associated with the God of the Judeo-Christian tradition, and they were considerably amplified by the role played in theology by Platonic and Neo-Platonism philosophy.

Nicolas Copernicus, Galileo, Johannes Kepler, and Isaac Newton were all inheritors of a cultural tradition in which ontological dualism was a primary article of faith. Hence the idealization of the mathematical ideal as a source of communion with God, which dates from Pythagoras, provided a metaphysical foundation for the emerging natural sciences. This explains why, the creators of classical physics believed that doing physics was a form of communion with the geometrical and mathematical forms resident in the perfect mind of God. This view would survive in a modified form in what is now known as Einsteinian epistemology and accounts in no small part for the reluctance of many physicists to accept the epistemology associated with the Copenhagen Interpretation.

At the beginning of the nineteenth century, Pierre-Simon LaPlace, along with a number of other French mathematicians, advanced the view that the science of mechanics constituted a complete view of nature. Since this science, by observing its epistemology, had revealed itself to be the fundamental science, the hypothesis of God was, they concluded, entirely unnecessary.

LaPlace is recognized for eliminating not only the theological component of classical physics but the entire metaphysical component as well. The epistemology of science requires, he said, that we proceed by inductive generalizations from observed facts to hypotheses that are tested by observed conformity of the phenomena. What was unique about LaPlaces view of hypotheses was his insistence that we cannot attribute reality to them. Although concepts like force, mass, motion, cause, and laws are obviously present in classical physics, they exist in LaPlaces view only as quantities. Physics is concerned, he argued, with quantities that we associate as a matter of convenience with concepts, and the truth about nature are only the quantities.

As this view of hypotheses and the truth of nature as quantities was extended in the nineteenth century to a mathematical description of phenomena like heat, light, electricity, and magnetism. LaPlace’s assumptions about the actual character of scientific truth seemed correct. This progress suggested that if we could remove all thoughts about the nature of or the source of phenomena, the pursuit of strictly quantitative concepts would bring us to a complete description of all aspects of physical reality. Subsequently, figures like Comte, Kirchhoff, Hertz, and Poincaré developed a program for the study of nature hat was quite different from that of the original creators of classical physics.

The seventeenth-century view of physics as a philosophy of nature or as natural philosophy was displaced by the view of physics as an autonomous science that was the science of nature. This view, which was premised on the doctrine of positivism, promised to subsume all of nature with a mathematical analysis of entities in motion and claimed that the true understanding of nature was revealed only in the mathematical description. Since the doctrine of positivism assumes that the knowledge we call physics resides only in the mathematical formalism of physical theory, it disallows the prospect that the vision of physical reality revealed in physical theory can have any other meaning. In the history of science, the irony is that positivism, which was intended to banish metaphysical concerns from the domain of science, served to perpetuate a seventeenth-century metaphysical assumption about the relationship between physical reality and physical theory.

Epistemology since Hume and Kant has drawn back from this theological underpinning. Indeed, the very idea that nature is simple (or uniform) has come in for a critique. The view has taken hold that a preference for simple and parsimonious hypotheses is purely methodological: It is constitutive of the attitude we call scientific and makes no substantive assumption about the way the world is.

A variety of otherwise diverse twentieth-century philosophers of science have attempted, in different ways, to flesh out this position. Two examples must suffice here: Hesse (1969) as, for summaries of other proposals. Popper (1959) holds that scientists should prefer highly falsifiable (improbable) theories: He tries to show that simpler theories are more falsifiable, also Quine (1966), in contrast, sees a virtue in theories that are highly probable, he argues for a general connexion between simplicity and high probability.

Both these proposals are global. They attempt to explain why simplicity should be part of the scientific method in a way that spans all scientific subject matters. No assumption about the details of any particular scientific problem serves as a premiss in Popper or Quine's arguments.

Newton and Leibniz thought that the justification of parsimony and simplicity flows from the hand of God: Popper and Quine try to justify these methodologically median of importance is without assuming anything substantive about the way the world is. In spite of these differences in approach, they have something in common. They assume that all users of parsimony and simplicity in the separate sciences can be encompassed in a single justifying argument. That recent developments in confirmation theory suggest that this assumption should be scrutinized. Good (1983) and Rosenkrantz (1977) has emphasized the role of auxiliary assumptions in mediating the connexion between hypotheses and observations. Whether a hypothesis is well supported by some observations, or whether one hypothesis is better supported than another by those observations, crucially depends on empirical background assumptions about the inference problem here. The same view applies to the idea of prior probability (or, prior plausibility). In of a single hypo-physical science if chosen as an alternative to another even though they are equally supported by current observations, this must be due to an empirical background assumption.

Principles of parsimony and simplicity mediate the epistemic connexion between hypotheses and observations. Perhaps these principles are able to do this because they are surrogates for an empirical background theory. It is not that there is one background theory presupposed by every appeal to parsimony; this has the quantifier order backwards. Rather, the suggestion is that each parsimony argument is justified only to each degree that it reflects an empirical background theory about the subjective matter. On this theory is brought out into the open, but the principle of parsimony is entirely dispensable (Sober, 1988).

This local approach to the principles of parsimony and simplicity resurrects the idea that they make sense only if the world is one way rather than another. It rejects the idea that these maxims are purely methodological. How defensible this point of view is, will depend on detailed case studies of scientific hypothesis evaluation and on further developments in the theory of scientific inference.

It is usually not found of one and the same that, an inference is a (perhaps very complex) act of thought by virtue of which act (1) I pass from a set of one or more propositions or statements to a proposition or statement and (2) it appears that the latter are true if the former is or are. This psychological characterization has occurred over a wider summation of literature under lesser than inessential variations. Desiring a better characterization of inference is natural. Yet attempts to do so by constructing a fuller psychological explanation fail to comprehend the grounds on which inference will be objectively valid-A point elaborately made by Gottlob Frége. Attempts to understand the nature of inference through the device of the representation of inference by formal-logical calculations or derivations better (1) leave us puzzled about the relation of formal-logical derivations to the informal inferences they are supposedly to represent or reconstruct, and (2) leaves us worried about the sense of such formal derivations. Are these derivations inference? Are not informal inferences needed in order to apply the rules governing the constructions of formal derivations (inferring that this operation is an application of that formal rule)? These are concerns cultivated by, for example, Wittgenstein.

Coming up with an adequate characterized inference, and even working out what would count as a very adequate characterization here is demandingly by no means nearly some resolved philosophical problem.

Traditionally, a proposition that is not a conditional, as with the affirmative and negative, modern opinion is wary of the distinction, since what appears categorical may vary with the choice of a primitive vocabulary and notation. Apparently categorical propositions may also turn out to be disguised conditionals: X is intelligent (categorical?) Equivalent, if X is given a range of tasks, she does them better than many people (conditional?). The problem is not merely one of classification, since deep metaphysical questions arise when facts that seems to be categorical and therefore solid, come to seem by contrast conditional, or purely hypothetical or potential.

Its condition of some classified necessity is so proven sufficient that if p is a necessary condition of q, then q cannot be true unless p; is true? If p is a sufficient condition, thus steering well is a necessary condition of driving in a satisfactory manner, but it is not sufficient, for one can steer well but drive badly for other reasons. Confusion may result if the distinction is not heeded. For example, the statement that A causes B may be interpreted to mean that A is itself a sufficient condition for B, or that it is only a necessary condition fort B, or perhaps a necessary parts of a total sufficient condition. Lists of conditions to be met for satisfying some administrative or legal requirement frequently attempt to give individually necessary and jointly sufficient sets of conditions.

What is more that if any proposition of the form if p then q. The condition hypothesized p. Is called the antecedent of the conditionals, and q, the consequent? Various kinds of conditional have been distinguished. Its weakest is that of material implication, merely telling that either not-p or q. stronger conditionals include elements of modality, corresponding to the thought that if p is truer then q must be true. Ordinary language is very flexible in its use of the conditional form, and there is controversy whether conditionals are better treated semantically, yielding differently finds of conditionals with different meanings, or pragmatically, in which case there should be one basic meaning with surface differences arising from other implicatures.

It follows from the definition of strict implication that a necessary proposition is strictly implied by any proposition, and that an impossible proposition strictly implies any proposition. If strict implication corresponds to q follows from p, then this means that a necessary proposition follows from anything at all, and anything at all follows from an impossible proposition. This is a problem if we wish to distinguish between valid and invalid arguments with necessary conclusions or impossible premises.

The Humean problem of induction is that if we would suppose that there is some property 'A' concerning and observational or an experimental situation, and that out of a large number of observed instances of 'A', some fraction m/n (possibly equal to 1) has also been instances of some logically independent property 'B'. Suppose further that the background proportionate circumstances not specified in these descriptions have been varied to a substantial degree and that there is no collateral information available concerning the frequency of 'B's' among 'A's' or concerning causal or nomologically connections between instances of 'A' and instances of 'B'.

In this situation, an enumerative or instantial induction inference would move rights from the premise, that m/n of observed 'A's' are 'B's' to the conclusion that approximately m/n of all 'A's' are 'B's'. (The usual probability qualification will be assumed to apply to the inference, rather than being part of the conclusion.) Here the class of 'A's' should be taken to include not only unobserved 'A's' and future 'A's', but also possible or hypothetical 'A's' (an alternative conclusion would concern the probability or likelihood of the adjacently observed 'A' being a 'B').

The traditional or Humean problem of induction, often referred to simply as the problem of induction, is the problem of whether and why inferences that fit this schema should be considered rationally acceptable or justified from an epistemic or cognitive standpoint, i.e., whether and why reasoning in this way is likely to lead to true claims about the world. Is there any sort of argument or rationale that can be offered for thinking that conclusions reached in this way are likely to be true in the corresponding premises is true - or even that their chances of truth are significantly enhanced?

Humes discussion of this issue deals explicitly only with cases where all observed 'A's' are 'B's' and his argument applies just as well to the more general case. His conclusion is entirely negative and sceptical: Inductive inferences are not rationally justified, but are instead the result of an essentially a-rational process, custom or habit. Hume (1711-76) challenges the proponent of induction to supply a cogent lignin of reasoning that leads from an inductive premise to the corresponding conclusion and offers an extremely influential argument in the form of a dilemma (a few times referred to as Humes fork), that either our actions are determined, in which case we are not responsible for them, or they are the result of random events, under which case we are also not responsible for them.

Such reasoning would, he argues, have to be either deductively demonstrative reasoning in the concerning relations of ideas or experimental, i.e., empirical, that reasoning concerning matters of fact or existence. It cannot be the former, because all demonstrative reasoning relies on the avoidance of contradiction, and it is not a contradiction to suppose that the course of nature may change, that an order that was observed in the past and not of its continuing against the future: But it cannot be, as the latter, since any empirical argument would appeal to the success of such reasoning about an experience, and the justifiability of generalizing from experience are precisely what is at issue-so that any such appeal would be question-begging. Hence, Hume concludes that there can be no such reasoning (1748).

An alternative version of the problem may be obtained by formulating it with reference to the so-called Principle of Induction, which says roughly that the future will resemble the past or, somewhat better, that unobserved cases will resemble observed cases. An inductive argument may be viewed as enthymematic, with this principle serving as a supposed premiss, in which case the issue is obviously how such a premiss can be justified. Hume's argument is then that no such justification is possible: The principle cannot be justified a prior because having possession of been true in experiences without obviously begging the question is not contradictory to have possession of been true in experiences without obviously begging the question.

The predominant recent responses to the problem of induction, at least in the analytic tradition, in effect accept the main conclusion of Humes argument, namely, that inductive inferences cannot be justified in the sense of showing that the conclusion of such an inference is likely to be true if the premise is true, and thus attempt to find another sort of justification for induction. Such responses fall into two main categories: (I) Pragmatic justifications or vindications of induction, mainly developed by Hans Reichenbach (1891-1953), and (ii) ordinary language justifications of induction, whose most important proponent is Frederick, Peter Strawson (1919- ). In contrast, some philosophers still attempt to reject Humes dilemma by arguing either (iii) That, contrary to appearances, induction can be inductively justified without vicious circularity, or (iv) that an anticipatory justification of induction is possible after all. In that:

(1) Reichenbachs view is that induction is best regarded, not as a form of inference, but rather as a method for arriving at posits regarding, i.e., the proportion of what remain besides such a speculative assertion would not claim asserted to be true, but is instead an intellectual wager analogous to a bet made by a gambler. Understood in this way, the inductive method says that one should posit that the observed proportion is, within some measure of an approximation, the true proportion and then continually correct that initial posit as new information comes in.

The gamblers bet is normally an appraised posit, i.e., he knows the chances or odds that the outcome on which he bets will actually occur. In contrast, the inductive bet is a blind posit: We do not know the chances that it will succeed or even that success is that it will succeed or even that success is possible. What we are gambling on when we make such a bet is the value of a certain proportion in the independent world, which Reichenbach construes as the limit of the observed proportion as the number of cases increases to infinity. Nevertheless, we have no way of knowing that there are even such a limit, and no way of knowing that the proportion of As are in addition of convergence, the end on some stable value than varying at random. If we cannot know that this limit exists, then we obviously cannot know that we have any definite chance of finding it.

What we can know, according to Reichenbach, is that if there is a truth of this sort to be found, the inductive method will eventually find it. That this is so is an analytic consequence of Reichenbachs account of what it is for such a limit to exist. The only way that the inductive method of making an initial posit and then refining it in light of new observations can fail eventually to arrive at the true proportion is if the series of observed proportions never converges on any stable value, which means that there is no truth to be found pertaining the proportion of an additionally make up of an induction if justified, not by showing that it will succeed or indeed, that it has any definite likelihood of success, but only by showing that it will succeed if success is possible. Reichenbachs claim is that no more than this can be established for any method, and hence that induction gives us our best chance for success, our best gamble in a situation where there is no alternative to gambling.

This pragmatic response to the problem of induction faces several serious problems. First, there are indefinitely many other methods for arriving at posits for which the same sort of defence can be given-methods that yield the same result as the inductive method over time but differ arbitrarily before long. Despite the efforts of others, it is unclear that there is any satisfactory way to exclude such alternatives, in order to avoid the result that any arbitrarily chosen short-term posit is just as reasonable as the inductive posit. Second, even if there is a truth of the requisite sort to be found, the inductive method is only guaranteed to find it or even to come within any specifiable distance of it in the indefinite long run. All the same, any actual application of inductive results always takes place in the presence to the future eventful states in making the relevance of the pragmatic justification to actual practice uncertainly. Third, and most important, it needs to be emphasized that Reichenbachs response to the problem simply accepts the claim of the Humean sceptic that an inductive premise never provides the slightest reason for thinking that the corresponding inductive conclusion is true. Reichenbach himself is quite candid on this point, but this does not alleviate the intuitive implausibility of saying that we have no more reason for thinking that our scientific and commonsense conclusions that result in the induction of it . . . is true than, to use Reichenbachs own analogy (1949), a blind man wandering in the mountains who feels an apparent trail with his stick has for thinking that following it will lead him to safety.

An approach to induction resembling Reichenbachs claiming in that those particular inductive conclusions are posits or conjectures, than the conclusions of cogent inferences, is offered by Popper. However, Poppers view is even more overtly sceptical: It amounts to saying that all that can ever be said in favours of the truth of an inductive claim is that the claim has been tested and not yet been shown to be false.

(2) The ordinary language response to the problem of induction has been advocated by many philosophers, none the less, Strawson claims that the question whether induction is justified or reasonable makes sense only if it tacitly involves the demand that inductive reasoning meet the standards appropriate to deductive reasoning, i.e., that the inductive conclusions are shown to follow deductively from the inductive assumption. Such a demand cannot, of course, be met, but only because it is illegitimate: Inductive and deductive reasons are simply fundamentally different kinds of reasoning, each possessing its own autonomous standards, and there is no reason to demand or expect that one of these kinds meet the standards of the other. Whereas, if induction is assessed by inductive standards, the only ones that are appropriate, then it is obviously justified.

The problem here is to understand to what this allegedly obvious justification of an induction amount. In his main discussion of the point (1952), Strawson claims that it is an analytic true statement that believing it a conclusion for which there is strong evidence is reasonable and an analytic truth that inductive evidence of the sort captured by the schema presented earlier constitutes strong evidence for the corresponding inductive conclusion, thus, apparently yielding the analytic conclusion that believing it a conclusion for which there is inductive evidence is reasonable. Nevertheless, he also admits, indeed insists, that the claim that inductive conclusions will be true in the future is contingent, empirical, and may turn out to be false (1952). Thus, the notion of reasonable belief and the correlative notion of strong evidence must apparently be understood in ways that have nothing to do with likelihood of truth, presumably by appeal to the standard of reasonableness and strength of evidence that are accepted by the community and are embodied in ordinary usage.

Understood in this way, Strawsons response to the problem of inductive reasoning does not speak to the central issue raised by Humean scepticism: The issue of whether the conclusions of inductive arguments are likely to be true. It amounts to saying merely that if we reason in this way; we can correctly call ourselves reasonable and our evidence strong, according to our accepted community standards. Nevertheless, to the undersealing of issue of whether following these standards is a good way to find the truth; the ordinary language response appears to have nothing to say.

(3) The main attempts to show that induction can be justified inductively have concentrated on showing that such as a defence can avoid circularity. Skyrms (1975) formulate, perhaps the clearest version of this general strategy. The basic idea is to distinguish different levels of inductive argument: A first level in which induction is applied to things other than arguments: A second level in which it is applied to arguments at the first level, arguing that they have been observed to succeed so far and hence are likely to succeed in general: A third level in which it is applied in the same way to arguments at the second level, and so on. Circularity is allegedly avoided by treating each of these levels as autonomous and justifying the argument at each level by appeal to an argument at the next level.

One problem with this sort of move is that even if circularity is avoided, the movement to Higher and Higher levels will clearly eventually fail simply for lack of evidence: A level will reach at which there have been enough successful inductive arguments to provide a basis for inductive justification at the next Higher level, and if this is so, then the whole series of justifications collapses. A more fundamental difficulty is that the epistemological significance of the distinction between levels is obscure. If the issue is whether reasoning in accord with the original schema offered above ever provides a good reason for thinking that the conclusion is likely to be true, then it still seems question-begging, even if not flatly circular, to answer this question by appeal to anther argument of the same form.

(4) The idea that induction can be justified on a pure priori basis is in one way the most natural response of all: It alone treats an inductive argument as an independently cogent piece of reasoning whose conclusion can be seen rationally to follow, although perhaps only with probability from its premise. Such an approach has, however, only rarely been advocated (Russell, 19132 and BonJour, 1986), and is widely thought to be clearly and demonstrably hopeless.

Many on the reasons for this pessimistic view depend on general epistemological theses about the possible or nature of anticipatory cognition. Thus if, as Quine alleges, there is no a prior justification of any kind, then obviously a prior justification for induction is ruled out. Or if, as more moderate empiricists have in claiming some preexistent knowledge should be analytic, then again a prevenient justification for induction seems to be precluded, since the claim that if an inductive premise is truer, then the conclusion is likely to be true does not fit the standard conceptions of analyticity. A consideration of these matters is beyond the scope of the present spoken exchange.

There are, however, two more specific and quite influential reasons for thinking that an early approach is impossible that can be briefly considered, first, there is the assumption, originating in Hume, but since adopted by very many of others, that a move forward in the defence of induction would have to involve turning induction into deduction, i.e., showing, per impossible, that the inductive conclusion follows deductively from the premise, so that it is a formal contradiction to accept the latter and deny the former. However, it is unclear why a prior approach need be committed to anything this strong. It would be enough if it could be argued that it is deductively unlikely that such a premise is true and corresponding conclusion false.

Second, Reichenbach defends his view that pragmatic justification is the best that is possible by pointing out that a completely chaotic world in which there is simply not true conclusion to be found as to the proportion of, in addition that occur of, but is neither impossible nor unlikely from a purely a prior standpoint, the suggestion being that therefore there can be no a prior reason for thinking that such a conclusion is true. Nevertheless, there is still a substring way in laying that a chaotic world is a prior neither impossible nor unlikely without any further evidence does not show that such a world is not a prior unlikely and a world containing such-and-such regularity might anticipatorily be somewhat likely in relation to an occurrence of a long running pattern of evidence in which a certain stable proportion of observed events -. An occurrence, it might be claimed, that would be highly unlikely in a chaotic world (BonJour, 1986).

Goodman's new riddle of induction purports that we suppose that before some specific time t (perhaps the year 2000) we observe a larger number of emeralds (property A) and find them all to be green (property B). We proceed to reason inductively and conclude that all emeralds are green Goodman points out, however, that we could have drawn a quite different conclusion from the same evidence. If we define the term stuff to mean green if examined before states of blueness were examined there after. Then all of our observed emeralds will also be gruing. A parallel inductive argument will yield the conclusion that all emeralds are gruing, and hence that all those examined after the year 2000 will be blue. Presumably the first of these concisions is genuinely supported by our observations and the second is not. Nevertheless, the problem is to say why this is so and to impose some further restriction upon inductive reasoning that will permit the first argument and exclude the second.

The obvious alternative suggestion is that stuff. Similar predicates do not correspond to genuine, purely qualitative properties in the way that green and blueness does, and that this is why inductive arguments involving them are unacceptable. Goodman, however, claims to be unable to make clear sense of this suggestion, pointing out that the relations of formal desirability are perfectly symmetrical: Stuff may be defined in terms if, green and blue, but green an equally well be defined in terms of stuff and green (blue if examined before t and green if examined after t).

The stuff that has been recognized from its complicated and most puzzling of named paradoxes that only demonstrate the importance of categorization, in that sometimes it is itemized as gruing, if examined of a presence to the future, before future time t and green, or not so examined and blue. Even though all emeralds in our evidence class stuff, we ought to must infer that all emeralds are gruing. For stuff is unpredictable, and cannot transmit credibility from known to unknown cases. Only projectable predicates are right for induction. Goodman considers entrenchment the key to projectibility having a long history of successful protection, stuff is entrenched, lacking such a history, stuff is not. A hypothesis is projectable, Goodman suggests, only if its predicates (or suitable related ones) are much better entrenched than its rivalrous past successes that do not assume future ones. Induction remains a risky business. The rationale for favouring entrenched predicates is pragmatic. Of the possible projections from our evidence class, the one that fits with past practices enables us to utilize our cognitive resources best. Its prospects of being true are worse than its competitors and its cognitive utility is greater.

So, to a better understanding of induction we should then linearize its term for which is most widely used for any process of reasoning that takes us from empirical premises to empirical conclusions supported by the premises, but not deductively entailed by them. Inductive arguments are therefore kinds of applicative arguments, in which something beyond the content of the premise is inferred as probable or supported by them. Induction is, however, commonly distinguished from arguments to theoretical explanations, which share this applicative character, by being confined to inferences in which he conclusion involves the same properties or relations as the premises. The central example is induction by simple enumeration, where from premises telling that Fa, Fb, Fc . . . where 'a', 'b', 'C's', are all of some kind 'G', it is inferred that 'G's' from outside the sample, such as future 'G's', will be 'F', or perhaps that all 'G's' are 'F'. In this, which and the other persons deceive them, children may infer that everyone is a deceiver: Different, but similar inferences of a property by some object to the same objects future possession of the same property, or from the constancy of some law-like pattern in events and states of affairs ti its future constancy. All objects we know of attract each other with a force inversely proportional to the square of the distance between them, so perhaps they all do so, and will always do so.

The rational basis of any inference was challenged by Hume, who believed that induction presupposed belief in the uniformity of nature, but that this belief has no defence in reason, and merely reflected a habit or custom of the mind. Hume was not therefore sceptical about the role of reason in either explaining it or justifying it. Trying to answer Hume and to show that there is something rationally compelling about the inference referred to as the problem of induction. It is widely recognized that any rational defence of induction will have to partition well-behaved properties for which the inference is plausible (often called projectable properties) from badly behaved ones, for which it is not. It is also recognized that actual inductive habits are more complex than those of similar enumeration, and that both common sense and science pay attention to such giving factors as variations within the sample giving us the evidence, the application of ancillary beliefs about the order of nature, and so on.

Nevertheless, the fundamental problem remains that and experience condition by application show us only events occurring within a very restricted part of a vast spatial and temporal order about which we then come to believe things.

Uncompounded by its belonging of a confirmation theory finding of the measure to which evidence supports a theory fully formalized confirmation theory would dictate the degree of confidence that a rational investigator might have in a theory, given some-body of evidence. The grandfather of confirmation theory is Gottfried Leibniz (1646-1718), who believed that a logically transparent language of science would be able to resolve all disputes. In the 20th century a fully formal confirmation theory was a main goal of the logical positivist, since without it the central concept of verification by empirical evidence itself remains distressingly unscientific. The principal developments were due to Rudolf Carnap (1891-1970), culminating in his Logical Foundations of Probability (1950). Carnaps idea was that the measure necessitated would be the proportion of logically possible states of affairs in which the theory and the evidence both hold, compared ti the number in which the evidence itself holds that the probability of a preposition, relative to some evidence, is a proportion of the range of possibilities under which the proposition is true, compared to the total range of possibilities left by the evidence. The difficulty with the theory lies in identifying sets of possibilities so that they admit of measurement. It therefore demands that we can put a measure on the range of possibilities consistent with theory and evidence, compared with the range consistent with the evidence alone.

Among the obstacles the enterprise meets, is the fact that while evidence covers only a finite range of data, the hypotheses of science may cover an infinite range. In addition, confirmation proves to vary with the language in which the science is couched, and the Carnapian programme has difficulty in separating genuinely confirming variety of evidence from less compelling repetition of the same experiment. Confirmation also proved to be susceptible to acute paradoxes. Finally, scientific judgment seems to depend on such intangible factors as the problems facing rival theories, and most workers have come to stress instead the historically situated scene of what would appear as a plausible distinction of a scientific knowledge at a given time.

Arose to the paradox of which when a set of apparent incontrovertible premises is given to unacceptable or contradictory conclusions. To solve a paradox will involve showing either that there is a hidden flaw in the premises or that the reasoning is erroneous, or that the apparently unacceptable conclusion can, in fact, be tolerated. Paradoxes are therefore important in philosophy, for until one is solved it shows that there is something about our reasoning and our concepts that we do not understand. What is more, and somewhat loosely, a paradox is a compelling argument from unacceptable premises to an unacceptable conclusion: More strictly speaking, a paradox is specified to be a sentence that is true if and only if it is false. A characterized objection lesson of it would be: The displayed sentence is false.

Seeing that this sentence is false if true is easy, and true if false, a paradox, in either of the senses distinguished, presents an important philosophical challenger. Epistemologists are especially concerned with various paradoxes having to do with knowledge and belief. In other words, for example, the Knower paradox is an argument that begins with apparently impeccable premises about the concepts of knowledge and inference and derives an explicit contradiction. The origin of the reasoning is the surprise examination paradox: A teacher announces that there will be a surprise examination next week. A clever student argues that this is impossible. The test cannot be on Friday, the last day of the week, because it would not be a surprise. We would know the day of the test on Thursday evening. This means we can also rule out Thursday. For after we learn that no test has been given by Wednesday, we would know the test is on Thursday or Friday -and would already know that it s not on Friday and would already know that it is not on Friday by the previous reasoning. The remaining days can be eliminated in the same manner.

This puzzle has over a dozen variants. The first was probably invented by the Swedish mathematician Lennard Ekbon in 1943. Although the first few commentators regarded the reverse elimination argument as cogent, every writer on the subject since 1950 agrees that the argument is unsound. The controversy has been over the proper diagnosis of the flaw.

Initial analyses of the subject's argument tried to lay the blame on a simple equivocation. Their failure led to more sophisticated diagnoses. The general format has been assimilation to better-known paradoxes. One tradition casts the surprise examination paradox as a self-referential problem, as fundamentally akin to the Liar, the paradox of the Knower, or Gödels incompleteness theorem. That in of itself says enough that Kaplan and Montague (1960) distilled the following self-referential paradox, the Knower. Consider the sentence: (S) The negation of this sentence is known (to be true).

Suppose that (S) is true. Then its negation is known and hence true. However, if its negation is true, then (S) must be false. Therefore (s) is false, or what is the name, the negation of (S) is true.

This paradox and its accompanying reasoning are strongly reminiscent of the Lair Paradox that (in one version) begins by considering a sentence this sentence is false and derives a contradiction. Versions of both arguments using axiomatic formulations of arithmetic and Gödel-numbers to achieve the effect of self-reference yields important meta-theorems about what can be expressed in such systems. Roughly these are to the effect that no predicates definable in the formalized arithmetic can have the properties we demand of truth (Tarskis Theorem) or of knowledge (Montague, 1963).

These meta-theorems still leave us; with the problem that if we suppose that we add of these formalized languages predicates intended to express the concept of knowledge (or truth) and inference - as one mighty does if logic of these concepts is desired. Then the sentence expressing the leading principles of the Knower Paradox will be true.

Explicitly, the assumption about knowledge and inferences are:

(1) If sentences A are known, then a.

(2) (1) is known?

(3) If ‘B’ is correctly inferred from ‘A’, and ‘A’ is known, then ‘B’ is known.

To give an absolutely explicit t derivation of the paradox by applying these principles to (S), we must add (contingent) assumptions to the effect that certain inferences have been done. Still, as we go through the argument of the Knower, these inferences are done. Even if we can somehow restrict such principles and construct a consistent formal logic of knowledge and inference, the paradoxical argument as expressed in the natural language still demands some explanation.

The usual proposals for dealing with the Liar often have their analogues for the Knower, e.g., that there is something wrong with a self-reference or that knowledge (or truth) is properly a predicate of propositions and not of sentences. The relies that show that some of these are not adequate are often parallel to those for the Liar paradox. In addition, one can try of what seems to be an adequate solution for the Surprise Examination Paradox, namely the observation that new knowledge can drive out knowledge, but this does not seem to work on the Knower (Anderson, 1983).

There are a number of paradoxes of the Liar family. The simplest example is the sentence this sentence is false, which must be false if it is true, and true if it is false. One suggestion is that the sentence fails to say anything, but sentences that fail to say anything are at least not true. In fact case, we consider to sentences this sentence is not true, which, if it fails to say anything is not true, and hence (this kind of reasoning is sometimes called the strengthened Liar). Other versions of the Liar introduce pairs of sentences, as in a slogan on the front of a T-shirt saying This sentence on the back of this T-shirt is false, and one on the back saying The sentence on the front of this T-shirt is true. It is clear that each sentence individually is well formed, and was it not for the other, might have said something true. So any attempt to dismiss the paradox by settling in that of the sentence involved are meaningless will face problems.

Even so, the two approaches that have some hope of adequately dealing with this paradox is hierarchy solutions and truth-value gap solutions. According to the first, knowledge is structured into levels. It is argued that there be one-careened notion expressed by the verb; knows, but rather a whole series of notions, of the knowable knows, and so on (perhaps into transfinite), stated ion terms of predicate expressing such ramified concepts and properly restricted, (1)-(3) lead to no contradictions. The main objections to this procedure are that the meaning of these levels has not been adequately explained and that the idea of such subscripts, even implicit, in a natural language is highly counterintuitive the truth-value gap solution takes sentences such as (S) to lack truth-value. They are neither true nor false, but they do not express propositions. These defeats a crucial step in the reasoning used in the derivation of the paradoxes. Kripler (1986) has developed this approach in connexion with the Liar and Asher and Kamp (1986) has worked out some details of a parallel solution to the Knower. The principal objection is that strengthened or super versions of the paradoxes tend to reappear when the solution itself is stated.

Since the paradoxical deduction uses only the properties (1)-(3) and since the argument is formally valid, any notion that satisfies these conditions will lead to a paradox. Thus, Grim (1988) notes that this may be read as is known by an omniscient God and concludes that there is no careened single notion of omniscience. Thomason (1980) observes that with some different conditions, analogous reasoning about belief can lead to paradoxical consequence.

Overall, it looks as if we should conclude that knowledge and truth are ultimately intrinsically stratified concepts. It would seem that we must simply accept the fact that these (and similar) concepts cannot be assigned of any-one fixed, finite or infinite. Still, the meaning of this idea certainly needs further clarification.

Its paradox arises when a set of apparently incontrovertible premises gives unacceptable or contradictory conclusions, to solve a paradox will involve showing either that there is a hidden flaw in the premises, or that the reasoning is erroneous, or that the apparently unacceptable conclusion can, in fact, be tolerated. Paradoxes are therefore important in philosophy, for until one is solved its shows that there is something about our reasoning and of concepts that we do not understand. Famous families of paradoxes include the semantic paradoxes and Zenos paradoxes. Art the beginning of the 20th century, paradox and other set-theoretical paradoxes led to the complete overhaul of the foundations of set theory, while the Sorites paradox has lead to the investigations of the semantics of vagueness and fuzzy logics.

It is, however, to what extent can analysis be informative? This is the question that gives a riser to what philosophers have traditionally called the paradox of analysis. Thus, consider the following proposition:

(1) To be an instance of knowledge is to be an instance of justified true belief not essentially grounded in any falsehood. (1) If true, illustrates an important type of philosophical analysis. For convenience of exposition, I will assume (1) is a correct analysis. The paradox arises from the fact that if the concept of justified true belief not been essentially grounded in any falsification is the analyzed of the concept of knowledge, it would seem that they are the same concept and hence that: (2) To be an instance of knowledge is to be as an instance of knowledge and would have to be the same propositions as (1). But then how can (1) be informative when (2) is not? This is what is called the first paradox of analysis. Classical writings on analysis suggest a second paradoxical analysis (Moore, 1942). (3) An analysis of the concept of being a brother is that to be a

Brother is to be a male sibling. If (3) is true, it would seem that the concept of being a brother would have to be the same concept as the concept of being a male sibling and tat: (4) An analysis of the concept of being a brother is that to be a brother is to be a brother would also have to be true and in fact, would have to be the same proposition as (3). Yet (3) is true and (4) is false.

Both these paradoxes rest upon the assumptions that analysis is a relation between concepts, than one involving entity of other sorts, such as linguistic expressions, and tat in a true analysis, analyzed and analysandum are the same concept. Both these assumptions are explicit in Moore, but some of Moores remarks hint at a solution to that of another statement of an analysis is a statement partly about the concept involved and partly about the verbal expressions used to express it. He says he thinks a solution of this sort is bound to be right, but fails to suggest one because he cannot see a way in which the analysis can be even partly about the expression (Moore, 1942).

Elsewhere, of such ways, as a solution to the second paradox, to explicating (3) as: (5) - An analysis is given by saying that the verbal expression '?' is a brother expresses the same concept as is expressed by the conjunction of the verbal expressions '?' is male when used to express the concept of being male and '?' is a sibling when used to express the concept of being a sibling? (Ackerman, 1990). An important point about (5) is as follows. Stripped of its philosophical jargon (analysis, concept, '?' is a . . . '), (5) seems to state the sort of information generally stated in a definition of the verbal expression brother in terms of the verbal expressions male and sibling, where this definition is designed to draw upon listeners antecedent understanding of the verbal expression male and sibling, and thus, to tell listeners what the verbal expression brother really means, instead of merely providing the information that two verbal expressions are synonymous without specifying the meaning of either one? Thus, its solution to the second paradox seems to make the sort of analysis that gives rise to this paradox is a matter of specifying the meaning of a verbal expression in terms of separate verbal expressions already understood and saying how the meanings of these separate, already-understood verbal expressions are combined. This corresponds to Moore's intuitive requirement that an analysis should both specify the constituent concepts of the analysandum and tell how they are combined, but is this all there is to philosophical analysis?

We must note that, in addition too there being two paradoxes of analysis, there is two types of analyses that are relevant here. (There are also other types of analysis, such as reformatory analysis, where the analyzed are intended to improve on and replace the analysandum. But since reformatory analysis involves no commitment to conceptual identity between analysandum and analysandum, reformatory analysis does not generate a paradox of analysis and so will not concern us here.) One way to recognize the difference between the two types of analysis concerning us here is to focus on the difference between the two paradoxes. This can be done by means of the Frége-inspired sense-individuation condition, which is the condition that two expressions have the same sense if and only if they can be interchangeably salva veritate whenever used in propositional attitude context. If the expressions for the analsands and the analysandum in (1) met this condition, (1) and (2) would not raise the first paradox, but the second paradox arises regardless of whether the expression for the analysandum and the analysandum meet this condition. The second paradox is a matter of the failure of such expressions to be interchangeable salva veritate in sentences involving such contexts as an analysis is given thereof. Thus, a solution (such as the one offered) that is aimed only at such contexts can solve the second paradox. This is clearly false for the first paradox, however, which will apply to all pairs of propositions expressed by sentences in which expressions for pairs of anal sands and analysandum raising the first paradox is interchangeable.

One approach to the first paradox is to argue that, despite the apparent epistemic in equivalence of (1) and (2), the concept of justified true belief not essentially grounded in any falsehood is still identical with the concept of knowledge (Sosa, 1983). Another approach is to argue that in the sort of analysis raising the first paradox, the analysandum and analysandum is concepts that are different but that bear a special epistemic relation to each other. Elsewhere, the development is such an approach and suggestion that this analysand-analysandum relation has the following facets.

(I) the analysandum and analysandum are necessarily coextensive, i.e., necessarily every instance of one is an instance of the other.

(ii) The analysandum and analysandum are knowable theoretical to be coextensive.

(iii) The analysandum is simpler than the anal sands a condition whose necessity is recognized in classical writings on analysis, such as, Langford, 1942.

(iv) The analysandum do not have the analysandum as a constituent.

Condition (iv) rules out circularity. But since many valuable quasi-analyses are partly circular, e.g., knowledge is justified true belief supported by known reasons not essentially grounded in any falsehood, it seems best to distinguish between full analysis, from that of (iv) is a necessary condition, and partial analysis, for which it is not.

These conditions, while necessary, are clearly insufficient. The basic problem is that they apply too many pairs of concepts that do not seem closely enough related epistemologically to count as analysandum and analysandum, such as the concept of being six and the concept of the fourth root of 1296. Accordingly, its solution upon what actually seems epistemologically distinctive about analyses of the sort under consideration, which is a certain way they can be justified. This is by the philosophical example-and-counterexample method, which is in a general term that goes as follows. 'J' investigates the analysis of 'K's' concept 'Q' (where 'K' can but need not be identical to 'J' by setting 'K' a series of armchair thought experiments, i.e., presenting 'K' with a series of simple described hypothetical test cases and asking 'K' questions of the form If such-and-such where the case would this count as a case of 'Q'? J then contrasts the descriptions of the cases to which; 'K' answers affirmatively with the description of the cases to which 'K' does not, and 'J' generalizes upon these descriptions to arrive at the concepts (if possible not including the analysandum) and their mode of combination that constitute the analysandum of 'K's' concept 'Q'. Since 'J' need not be identical with 'K', there is no requirement that K himself be able to perform this generalization, to recognize its result as correct, or even to understand the analysandum that is its result. This is reminiscent of Walton's observation that one can simply recognize a bird as a blue jay without realizing just what feature of the bird (beak, wing configurations, etc.) form the basis of this recognition. (The philosophical significance of this way of recognizing is discussed in Walton, 1972) 'K' answers the questions based solely on whether the described hypothetical cases just strike him as cases of 'Q'. 'J' observes certain strictures in formulating the cases and questions. He makes the cases as simple as possible, to minimize the possibility of confusion and to minimize the likelihood that 'K' will draw upon his philosophical theories (or quasi-philosophical, a rudimentary notion if he is unsophisticated philosophically) in answering the questions. For this conflicting result, the conflict should other things being equal be resolved in favours of the simpler case. 'J' makes the series of described cases wide-ranging and varied, with the aim of having it be a complete series, where a series is complete if and only if no case that is omitted in such that, if included, it would change the analysis arrived at. 'J' does not, of course, use as a test-case description anything complicated and general enough to express the analysandum. There is no requirement that the described hypothetical test cases be formulated only in terms of what can be observed. Moreover, using described hypothetical situations as test cases enables 'J' to frame the questions in such a way as to rule out extraneous background assumption to a degree, thus, even if 'K' correctly believes that all and only 'P's' are 'R's', the question of whether the concepts of 'P', 'R', or both enter the analysandum of his concept 'Q' can be investigated by asking him such questions as Suppose (even if it seems preposterous to you) that you were to find out that there was a 'P' that was not an 'R'. Would you still consider it a case of 'Q'?

Taking all this into account, the necessary conditions for this sort of analysandum-analysandum relations is as follows: If 'S' is the analysandum of 'Q', the proposition that necessarily all and only instances of S are instances of 'Q' can be justified by generalizing from intuition about the correct answers to questions of the sort indicated about a varied and wide-ranging series of simple described hypothetical situations. It so does occur of antinomy, when we are able to argue for, or demonstrate, both a proposition and its contradiction, roughly speaking, a contradiction of a proposition 'p' is one that can be expressed in form 'not-p', or, if 'p' can be expressed in the form 'not-q', then a contradiction is one that can be expressed in the form 'q'. Thus, e.g., if p is 2 + 1 = 4, then, 2 + 1 = 4 is the contradictory of 'p', for 2 + 1 = 4 can be expressed in the form not (2 + 1 = 4). If p is 2 + 1 = 4, then 2 + 1 = 4 is a contradictory of 'p', since 2 + 1 = 4 can be expressed in the form not (2 + 1 = 4). This is, mutually, but contradictory propositions can be expressed in the form, 'r', and 'not-r'. The Principle of Contradiction says that mutually contradictory propositions cannot both be true and cannot both be false. Thus, by this principle, since if p is true, not-p is false, no proposition p can be at once true and false (otherwise both 'p' and its contradictories would be false?). In particular, for any predicate 'p' and object '?', it cannot be that 'p'; is at once true of '?' and false of '?'? This is the classical formulation of the principle of contradiction, but it is nonetheless, that we cannot now fault either demonstrates. We would eventually hope to be able to solve the antinomy by managing, through careful thinking and analysis, eventually to fault either or both demonstrations.

The conjunction of a proposition and its negation, where the law of non-contradiction provides that no such conjunction can be true: not (p & not-p). The standard proof of the inconsistency of a set of propositions or sentences is to show that a contradiction may be derived from them.

In Hégélien and Marxist writing the term is used more widely, as a contradiction may be a pair of features that together produce an unstable tension in a political or social system: a 'contradiction' of capitalism might be the aérosol of expectations in the workers that the system cannot require. For Hegel the gap between this and genuine contradiction is not as wide as it is for other thinkers, given the equation between systems of thought and their historical embodiment.

A contradictarian approach to problems of ethics asks what solution could be agreed upon by contradicting parties, starting from certain idealized positions (for example, no ignorance, no inequalities of power enabling one party to force unjust solutions upon another, no malicious ambitions). The idea of thinking of civil society, with its different distribution of rights and obligations, as if it were established by a social contract, derives from the English philosopher and mathematician Thomas Hobbes and Jean-Jacques Rousseau (1712-78). The utility of such a model was attacked by the Scottish philosopher, historian and essayist David Hume (1711-76), who asks why, given that non-traditional-historical events of establishing a contract, took place. It is useful to allocate rights and duties as if it had; he also points out that the actual distribution of these things in a society owes too much to contingent circumstances to be derivable from any such model. Similar positions in general ethical theory, sometimes called contradictualism: The right thing to do so one that could be acknowledged in the achievements of opinion, feeling, or purpose are that they coincide or concur to exist or go together without conflict or incongruity, e.g., his conclusion agrees with the evidence, yet the agreeability is accorded to consonance, upon which is the further direction in hypothetical contract.

Somewhat loosely, a paradox arises when a set of apparent incontrovertible premises gives unacceptable or contradictory conclusions, to solve a paradox will involve showing either that there is a hidden flaw in the premises, or that the reasoning is erroneous, or that the apparent unacceptable conclusion can, in fact, be tolerated. Paradoxes are themselves important in philosophy, for until one is solved it shows that there is something that we do not understand. Such are the paradoxes as compelling arguments from unexceptionable premises to an unacceptable conclusion, and more strictly, a paradox is specified to be a sentence that is true if and only if it is false: For example of the latter would be: 'The displayed sentence is false.

It is easy to see that this sentence is false if true, and true if false. A paradox, in either of the senses distinguished, presents an important philosophical challenge. Epistemologists are especially concerned with various paradoxes having to do with knowledge and belief.

Moreover, paradoxes are as an easy source of antinomies, for example, Zeno gave some famously lets say, logical-non-mathematical arguments that might be interpreted as demonstrating that motion is impossible. But our eyes as it was, demonstrate motion (exhibit moving things) all the time. Where did Zeno go wrong? Where do our eyes go wrong? If we cannot readily answer at least one of these questions, then we are in antinomy. In the Critique of Pure Reason, Kant gave demonstrations of the same kind -in the Zeno example they were obviously not the same kind of both, e.g., that the world has a beginning in time and space, and that the world has no beginning in time or space. He argues that both demonstrations are at fault because they proceed on the basis of pure reason unconditioned by sense experience.

At this point, we display attributes to the theory of experience, as it is not possible to define in an illuminating way, however, we know what experiences are through acquaintances with some of our own, e.g., visual experiences of as afterimage, a feeling of physical nausea or a tactile experience of an abrasive surface (which might be caused by an actual surface -rough or smooth, or which might be part of a dream, or the product of a vivid sensory imagination). The essential feature of experience is it feels a certain way -that there is something that it is like to have it. We may refer to this feature of an experience as its character.

Another core feature of the sorts of experiences, with which this may be of a concern, is that they have representational content. (Unless otherwise indicated, experience will be reserved for their contentual representations.) The most obvious cases of experiences with content are sense experiences of the kind normally involved in perception. We may describe such experiences by mentioning their sensory modalities ad their contents, e.g., a gustatory experience (modality) of chocolate ice cream (content), but do so more commonly by means of perceptual verbs combined with noun phrases specifying their contents, as in Macbeth saw a dagger. This is, however, ambiguous between the perceptual claim There was a (material) dagger in the world that Macbeth perceived visually and Macbeth had a visual experience of a dagger (the reading with which we are concerned, as it is afforded by our imagination, or perhaps, experiencing mentally hallucinogenic imagery).

As in the case of other mental states and events with content, it is important to distinguish between the properties that and experience represents and the properties that it possesses. To talk of the representational properties of an experience is to say something about its content, not to attribute those properties to the experience itself. Like every other experience, a visual; experience of a sculpture-sculptured square, of which is a mental event, and it is therefore not itself, or finds to some irregularity or is it square, even though it represents those properties. It is, perhaps, fleeting, pleasant or unusual, even though it does not represent those properties. An experience may represent a property that it possesses, and it may even do so in virtue of a rapidly changing (complex) experience representing something as changing rapidly. However, this is the exception and not the rule.

Which properties can be [directly] represented in sense experience is subject to debate. Traditionalists include only properties whose presence could not be doubted by a subject having appropriate experiences, e.g., colour and shape in the case of visual experience, and apparent shape, surface texture, hardness, etc., in the case of tactile experience. This view is natural to anyone who has an egocentric, Cartesian perspective in epistemology, and who wishes for pure data in experiences to serve as logically certain foundations for knowledge, especially to the immediate objects of perceptual awareness in or of sense-data, such categorized of colour patches and shapes, which are usually supposed distinct from surfaces of physical objectivity. Qualities of sense-data are supposed to be distinct from physical qualities because their perception is more relative to conditions, more certain, and more immediate, and because sense-data is private and cannot appear other than they are they are objects that change in our perceptual field when conditions of perception change: Physical objects remain constant.

Others who do not think that this wish can be satisfied, and who are more impressed with the role of experience in providing animisms with ecologically significant information about the world around them, claim that sense experiences represent properties, characteristic and kinds that are much richer and much more wide-ranging than the traditional sensory qualities. We do not see only colours and shapes, but they tell us, but also Earth, water, men, women and fire: We do not smell only odours, but also food and filth. There is no space here to examine the factors relevantly responsible to their choice of situational alternatives. Yet, this suggests that character and content are not really distinct, and there is a close tie between them. For one thing, the relative complexity of the character of sense experience places limitations upon its possible content, e.g., a tactile experience of something touching ones left ear is just too simple to carry the same amount of content as typically convincing to an every day, visual experience. Moreover, the content of a sense experience of a given character depends on the normal causes of appropriately similar experiences, e.g., the sort of gustatory experience that we have when eating chocolate would be not represented as chocolate unless it was normally caused by chocolate. Granting a contingent ties between the character of an experience and its possible causal origins, once, again follows that its possible content is limited by its character.

Character and content are none the less irreducibly different, for the following reasons. (1) There are experiences that completely lack content, e.g., certain bodily pleasures. (2) Not every aspect of the character of an experience with content is relevant to that content, e.g., the unpleasantness of an aural experience of chalk squeaking on a board may have no representational significance. (3) Experiences in different modalities may overlap in content without a parallel overlap in character, e.g., visual and tactile experiences of circularity feel completely different. (4) The content of an experience with a given character may vary according to the background of the subject, e.g., a certain content singing bird only after the subject has learned something about birds.

According to the act/object analysis of experience (which is a special case of the act/object analysis of consciousness), every experience involves an object of experience even if it has no material object. Two main lines of argument may be offered in support of this view, one phenomenological and the other semantic.

In an outline, or projective view, the phenomenological argument is as follows. Whenever we have an experience, even if nothing beyond the experience answers to it, we seem to be presented with something through the experience (which is it diaphanous). The object of the experience is whatever is so presented to us-is that it is an individual thing, an event, or a state of affairs.

The semantic argument is that objects of experience are required in order to make sense of certain features of our talk about experience, including, in particular, the following. (1) Simple attributions of experience, e.g., Rod is experiencing an oddity that is not really square but in appearance it seems more than likely a square, this seems to be relational. (2) We appear to refer to objects of experience and to attribute properties to them, e.g., the after-image that John experienced was certainly odd. (3) We appear to quantify over objects of experience, e.g., Macbeth saw something that his wife did not see.

The act/object analysis comes to grips with several problems concerning the status of objects of experiences. Currently the most common view is that they are sense-data - private mental entities that actually posses the traditional sensory qualities represented by the experiences of which they are the objects. But the very idea of an essentially private entity is suspect. Moreover, since an experience may apparently represent something as having a determinable property, e.g., redness, without representing it as having any subordinate determinate property, e.g., any specific shade of red, a sense-datum may actually have a determinate property subordinate to it. Even more disturbing is that sense-data may have contradictory properties, since experiences can have contradictory contents. A case in point is the waterfall illusion: If you stare at a waterfall for a minute and then immediately fixate on a nearby rock, you are likely to have an experience of the rocks moving upward while it remains in the same place. The sense-data theorist must either deny that there are such experiences or admit contradictory objects.

These problems can be avoided by treating objects of experience as properties. This, however, fails to do justice to the appearances, for experience seems not to present us with properties embodied in individuals. The view that objects of experience is Meinongian objects accommodate this point. It is also attractive in as far as (1) it allows experiences to represent properties other than traditional sensory qualities, and (2) it allows for the identification of objects of experience and objects of perception in the case of experiences that constitute perception.

According to the act/object analysis of experience, every experience with content involves an object of experience to which the subject is related by an act of awareness (the event of experiencing that object). This is meant to apply not only to perceptions, which have material objects (whatever is perceived), but also to experiences like hallucinations and dream experiences, which do not. Such experiences none the less appear to represent something, and their objects are supposed to be whatever it is that they represent. Act/object theorists may differ on the nature of objects of experience, which have been treated as properties. Meinongian objects (which may not exist or have any form of being), and, more commonly private mental entities with sensory qualities. (The term sense-data is now usually applied to the latter, but has also been used as a general term for objects of sense experiences, as in the work of G.E. Moore) Act/object theorists may also differ on the relationship between objects of experience and objects of perception. In terms of perception (of which we are indirectly aware) are always distinct from objects of experience (of which we are directly aware). Meinongian, however, may treat objects of perception as existing objects of experience. But sense-datum theorists must either deny that there is such experiences or admit contradictory objects. Still, most philosophers will feel that the Meinongians acceptance of impossible objects is too high a prime rate for prices that don't pay for such benefits.

A general problem for the act/object analysis is that the question of whether two subjects are experiencing one and the same thing (as opposed to having exactly similar experiences) appears to have an answer only on the assumption that the experiences concerned are perceptions with material objects. But in terms of the act/object analysis the question must have an answer even when this condition is not satisfied. (The answer is always negative on the sense-datum theory; it could be positive on other versions of the act/object analysis, depending on the facts of the case.)

In view of the above problems, the case for the act/object analysis should be reassessed. The Phenomenological argument is not, on reflection, convincing, for it is easy enough to grant that any experience appears to present us with an object without accepting that it actually does. The semantic argument is more impressive, but is none the less answerable. The seemingly relational structure of attributions of experience is a challenge dealt with below in connexion with the adverbial theory. Apparent reference to and quantification over objects of experience can be handled by analyzing them as reference to experiences themselves and quantification over experiences tacitly typed according to content. Thus, the afterimage that John experienced was colour fully appealing becomes Johns afterimage experience was an experience of colour, and Macbeth saw something that his wife did not see becomes Macbeth had a visual experience that his wife did not have.

Pure cognitive attempts to avoid the problems facing the act/object analysis by reducing experiences to cognitive events or associated disposition, e.g., Susys experience of a rough surface beneath her hand might be identified with the event of her acquiring the belief that there is a rough surface beneath her hand, or, if she does not acquire this belief, with a disposition to acquire it that has somehow been blocked.

This position has attractions. It does full justice to the cognitive contents of experience, and to the important role of experience as a source of belief acquisition. It would also help clear the way for a naturalistic theory of mind, since there seems to be some prospect of a physicalist/functionalist account of belief and other intentional states. But pure cognitive is completely undermined by its failure to accommodate the fact that experiences have a felt character that cannot be reduced to their content, as aforementioned.

The adverbial theory is an attempt to undermine the act/object analysis by suggesting a semantic account of attributions of experience that does not require objects of experience. Unfortunately, the oddities of explicit adverbializations of such statements have driven off potential supporters of the theory. Furthermore, the theory remains largely undeveloped, and attempted refutations have traded on this. It may, however, be founded on sound basis intuitions, and there is reason to believe that an effective development of the theory (which is merely hinting at) is possible.

The relevant intuitions are (1) that when we say that someone is experiencing an A, or has an experience of an A, we are using this content-expression to specify the type of thing that the experience is especially apt to fit, (2) that doing this is a matter of saying something about the experience itself (and maybe about the normal causes of like experiences), and (3) that it is no-good of reasons to posit of its position to presuppose that of any involvements, is that its descriptions of an object in which the experience is. Thus the effective role of the content-expression in a statement of experience is to modify the verb it compliments, not to introduce a special type of object.

Perhaps, the most important criticism of the adverbial theory is the many property problem, according to which the theory does not have the resources to distinguish between, e.g.

(1) Frank has an experience of a brown triangle

And:

(2) Frank has an experience of brown and an experience of a triangle.

Which is entailed by (1) but does not entail it? The act/object analysis can easily accommodate the difference between (1) and (2) by claiming that the truth of (1) requires a single object of experience that is both brown and triangular, while that of the (2) allows for the possibility of two objects of experience, one brown and the other triangular, however, (1) is equivalent to:

(1*) Frank has an experience of something being both brown and triangular.

And (2) is equivalent to:

(2*) Frank has an experience of something being brown and an experience of something being triangular,

And the difference between these can be explained quite simply in terms of logical scope without invoking objects of experience. The Adverbialists may use this to answer the many-property problem by arguing that the phrase a brown triangle in (1) does the same work as the clause something being both brown and triangular in (1*). This is perfectly compatible with the view that it also has the adverbial function of modifying the verb has an experience of, for it specifies the experience more narrowly just by giving a necessary condition for the satisfaction of the experience (the condition being that there are something both brown and triangular before Frank).

A final position that should be mentioned is the state theory, according to which a sense experience of an ‘A’ is concurrently a non-relational state of the kind that the subject would be in when perceiving an A. Suitably qualified, this claim is no doubt true, but its significance is subject to debate. Here it is enough to remark that the claim is compatible with both pure cognitive and the adverbial theory, and that state theorists are probably best advised to adopt adverbials as a means of developing their intuitions.

Yet, clarifying sense-data, if taken literally, is that which is given by the senses. But in response to the question of what exactly is so given, sense-data theories posit private showings in the consciousness of the subject. In the case of vision this would be a kind of inner picture shown which it only indirectly represents aspects of the external world that has in and of itself a worldly representation. The view has been widely rejected as implying that we really only see extremely thin coloured pictures interposed between our minds eye and reality. Modern approaches to perception tend to reject any conception of the eye as a camera or lenses, simply responsible for producing private images, and stress the active life of the subject in and of the world, as the determinant of experience.

Nevertheless, the argument from illusion is of itself the usually intended directive to establish that certain familiar facts about illusion disprove the theory of perception called naivety or direct realism. There are, however, many different versions of the argument that must be distinguished carefully. Some of these distinctions centre on the content of the premises (the nature of the appeal to illusion); others centre on the interpretation of the conclusion (the kind of direct realism under attack). Let us set about by distinguishing the importantly different versions of direct realism which one might take to be vulnerable to familiar facts about the possibility of perceptual illusion.

A crude statement of direct realism might go as follows. In perception, we sometimes directly perceive physical objects and their properties; we do not always perceive physical objects by perceiving something else, e.g., a sense-datum. There are, however, difficulties with this formulation of the view, as for one thing a great many philosophers who are not direct realists would admit that it is a mistake to describe people as actually perceiving something other than a physical object. In particular, such philosophers might admit, we should never say that we perceive sense-data. To talk that way would be to suppose that we should model our understanding of our relationship to sense-data on our understanding of the ordinary use of perceptual verbs as they describe our relation to and of the physical world, and that is the last thing paradigm sense-datum theorists should want. At least, many of the philosophers who objected to direct realism would prefer to express in what they were of objecting too in terms of a technical (and philosophically controversial) concept such as acquaintance. Using such a notion, we could define direct realism this way: In veridical experience we are directly acquainted with parts, e.g., surfaces, or constituents of physical objects. Less cautious venison of the view might drop the reference to veridical experience and claim simply that in all experience we are directly acquainted with parts or constituents of physical objects. The expressions knowledge by acquaintance and knowledge by description, and the distinction they mark between knowing things and knowing about things, are generally associated with Bertrand Russell (1872-1970), that scientific philosophy required analyzing many objects of belief as logical constructions or logical fictions, and the programme of analysis that this inaugurated dominated the subsequent philosophy of logical atomism, and then of other philosophers, Russells The Analysis of Mind, the mind itself is treated in a fashion reminiscent of Hume, as no more than the collection of neutral perceptions or sense-data that make up the flux of conscious experience, and that looked at another way that also was to make up the external world (neutral monism), but An Inquiry into Meaning and Truth (1940) represents a more empirical approach to the problem. Yet, philosophers have perennially investigated this and related distinctions using varying terminology.

Distinction in our ways of knowing things, highlighted by Russell and forming a central element in his philosophy after the discovery of the theory of definite descriptions. A thing is known by acquaintance when there is direct experience of it. It is known by description if it can only be described as a thing with such-and-such properties. In everyday parlance, I might know my spouse and children by acquaintance, but know someone as the first person born at sea only by description. However, for a variety of reasons Russell shrinks the area of things that can be known by acquaintance until eventually only current experience, perhaps my own self, and certain universals or meanings qualify anything else is known only as the thing that has such-and-such qualities.

Because one can interpret the relation of acquaintance or awareness as one that is not epistemic, i.e., not a kind of propositional knowledge, it is important to distinguish the above aforementioned views read as ontological theses from a view one might call epistemological direct realism? In perception we are, on at least some occasions, non-inferentially justified in believing a proposition asserting the existence of a physical object. Since it is that these objects exist independently of any mind that might perceive them, and so it thereby rules out all forms of idealism and phenomenalism, which hold that there are no such independently existing objects. Its being to direct realism rules out those views defended under the cubic of critical naive realism, or representational realism, in which there is some non-physical intermediary -usually called a sense-datum or a sense impression -that must first be perceived or experienced in order to perceive the object that exists independently of this perception. Often the distinction between direct realism and other theories of perception is explained more fully in terms of what is immediately perceived, than immediately perceived. What relevance does illusion have for these two forms of direct realism?

The fundamental premise of the arguments is from illusion seems to be the theses that things can appear to be other than they are. Thus, for example, straight sticks when immerged in water looks bent, a penny when viewed from certain perspective appears as an illusory spatial elliptic circularity, when something that is yellow when place under red fluorescent light looks red. In all of these cases, one version of the argument goes; it is implausible to maintain that what we are directly acquainted with is the real nature of the object in question. Indeed, it is hard to see how we can be said to be aware of the really physical object at all. In the above illusions the things we were aware of actually were bent, elliptical and red, respectively. But, by hypothesis, the really physical objects lacked these properties. Thus, we were not aware of the substantial reality of been real as a physical objects or theory.

So far, if the argument is relevant to any of the direct realizes distinguished above, it seems relevant only to the claim that in all sense experience we are directly acquainted with parts or constituents of physical objects. After all, even if in illusion we are not acquainted with physical objects, but their surfaces, or their constituents, why should we conclude anything about the hidden nature of our relations to the physical world in veridical experience?

We are supposed to discover the answer to this question by noticing the similarities between illusory experience and veridical experience and by reflecting on what makes illusion possible at all. Illusion can occur because the nature of the illusory experience is determined, not just by the nature of events or sorted, conflicting affairs but the object perceived as itself the event in cause, but also by other conditions, both external and internal as becoming of an inner or as the outer experience. But all of our sensations are subject to these causal influences and it would be gratuitous and arbitrary to select from indefinitely of many and subtly different perceptual experiences some special ones those that get us in touch with the real nature of the physical world and its surrounding surfaces. Red fluorescent light affects the way things look, but so doe's sunlight. Water reflects light, but so does air. We have no unmediated access to the external world.

The Philosophy of science and scientific epistemology are not the only area where philosophers have lately urged the relevance of neuroscientific discoveries. Kathleen Akins argues that a traditional view of the senses underlies the variety of sophisticated naturalistic programs about intentionality. Current neuroscientific understanding of the mechanisms and coding strategies implemented by sensory receptors shows that this traditional view is mistaken. The traditional view holds that sensory systems are veridical in at least three ways. (1) Each signal in the system correlates along with diminutive ranging properties in the external (to the body) environment. (2) The structure in the relevant relations between the external properties the receptors are sensitive to is preserved in the structure of the relations between the resulting sensory states, and (3) the sensory system theory, is not properly a single theory, but any approach to a complicated or complex structure that abstract away from the particular physical, chemical or biological nature of its components and simply considers the structure they together administer the terms of the functional role of individual parts and their contribution to the functioning of the whole, without fabricated additions or embellishments, that this is an external event. Using recent neurobiological discoveries about response properties of thermal receptors in the skin as an illustration, are, presently concurring of some acceptable of sensory systems from which are narcissistic than veridical. All three traditional assumptions are violated. These neurobiological details and their philosophical implications open novel questions for the philosophy of perception and for the appropriate foundations for naturalistic projects about intentionality. Armed with the known neurophysiology of sensory receptors, for example, our philosophy of perception or of perceptual intentionality will no longer focus on the search for correlations between states of sensory systems and periodically detected external properties. This traditionally philosophical (and scientific) project rests upon a mistaken veridical view of the senses. Neurophysiologic constructs allow for the knowledge of sensory receptors to actively show that sensory experience does not serve the naturalist as well as a simple paradigm case of intentional relations between representation and the world. Once again, available scientific detail shows the naivety of some traditional philosophical projects.

Focussing on the anatomy and physiology of the pain transmission system, Valerie Hard castle (1997) urges a similar negative implication for a popular methodological assumption. Pain experiences have long been philosopher's favourite cases for analysis and theorizing about conscious experience generally. Nevertheless, every position about pain experiences has been defended recently: eliminativism, a variety of objectivists view, relational views, and subjectivist views. Why so little agreement, despite agreement that pain experience is the place to start an analysis or theory of consciousness? Hard castle urges two answers. First, philosophers tend to be uninformed about the neuronal complexity of our pain transmission systems, and build their analyses or theories on the outcome of a single component of a fractionally-compounded system. Second, even those who understand some of the underlying neurobiology of pain tends to advocate gate-control theories. But the best existing gate-control theories are vague about the neural mechanisms of the gates. Hardcastle instead proposes a dissociable dual system of pain transmission, consisting of a pain sensory system closely analogous in its neurobiological implementation to other sensory systems, and a descending pain inhibitory system. She argues that this dual system is consistent with recent neuroscientific discoveries and accounts for all the pain phenomena that have tempted philosophers toward particular (but limited) theories of pain experience. The neurobiological uniqueness of the pain inhibitory system, contrasted with the mechanisms of other sensory modalities, renders pain processing atypical. In particular, the pain inhibitory system dissociates pains sensation from stimulation of nociceptors (pain receptors). Hardcastle concludes from the neurobiological uniqueness of pain transmission that pain experiences are atypical conscious events, and hence not a good place to start theorizing about or analyzing the general type.

Developing and defending theories of content is a central topic in current philosophy of mind. A common desideratum in this debate is a theory of cognitive representation consistent with a physical or naturalistic ontology. Here, described are a few contributional functions responsible in the dynamic construction of Neurophilosophers have made to this literature.

When one perceives or remembers that he is out of coffee, his brain state possesses intentionality or aboutness. The percept or memory is about ones being out of coffee, and it represents one for being out of coffee. The representational state has content. Some psychosemantics seek to explain what it is for a representational state to be about something: to provide an account of how states and events can have specific representational content. Some physicalist psychosemantics seek to do these using resources of the physical sciences exclusively. Neuro-philosophers have contributed to two types of physicalist psychosemantics: the Functional Role approach and the Informational approach.

The nucleus of functional roles of semantics holds that a representation has its content in virtue of relations it bears to other representations. Its paradigm application is to concepts of truth-functional logic, like the conjunctive and disjunctive or, a physical event instantiates the function as justly the case that it maps two true inputs onto a single true output. Thus an expression bears the relations to others that give it the semantic content of and, proponents of functional role semantics propose similar analyses for the content of all representations (Form 1986). A physical event represents birds, for example, if it bears the right relations to events representing feathers and others representing beaks. By contrast, informational semantics associates content to a state depending upon the causal relations obtaining between the state and the object it represents. A physical state represents birds, for example, just in case an appropriate causal relation obtains between it and birds. At the heart of informational semantics is a causal account of information. Red spots on a face carry the information that one has measles because the red spots are caused by the measles virus. A common criticism of informational semantics holds that mere causal variations are insufficient for representation, since information (in the causal sense) is by definition, always veridical while representations can misrepresent. A popular solution to this challenge invokes a teleological analysis of function. A brain state represents X by virtue of having the function of carrying information about being caused by X (Dretske 1988). These two approaches do not exhaust the popular options for some psychosemantics, but are the ones to which Neuro philosophers have contributed.

Jerry Fodor and Ernest LePore raise an important challenge to Churchlands psychosemantics. Location in a state space alone seems insufficient to fix representational states endorsed by content. Churchland never explains why a point in a three-dimensional state space represents the Collor, as opposed to any other quality, object, or event that varies along three dimensions. Churchlands account achieves its explanatory power by the interpretation imposed on the dimensions. Fodor and LePore allege that Churchland never specifies how a dimension comes to represent, e.g., degree of saltiness, as opposed to yellow-blue wavelength opposition. One obvious answer appeals to the stimuli that form the external inputs to the neural network in question. Then, for example, the individuating conditions on neural representations of colours are that opponent processing neurons receive input from a specific class of photoreceptors. The latter in turn have electromagnetic radiation (of a specific portion of the visible spectrum) as their activating stimuli. Nonetheless, this appeal to exterior impulsions as the ultimate stimulus that included individual conditions for representational content and context, for which makes the resulting approaches of an interpretation implied by the versional information to semantics. If, not only, from which this approach is accordantly supported with other neurobiological inferences.

The neurobiological paradigm for informational semantics is the feature detector: One or more neurons that are (I) maximally responsive to a particular type of stimulus, and (ii) have the function of indicating the presence of that stimulus type. Examples of such stimulus-types for visual feature detectors include high-contrast edges, motion direction, and colours. A favourite feature detector among philosophers is the alleged fly detector in the frog. Lettvin et al. (1959) identified cells in the frog retina that responded maximally to small shapes moving across the visual field. The idea that this cell's activity functioned to detect flies rested upon knowledge of the frogs' diet. Using experimental techniques ranging from single-cell recording to sophisticated functional imaging, neuroscientists have recently discovered a host of neurons that are maximally responsive to a variety of stimuli. However, establishing condition (ii) on a feature detector is much more difficult. Even some paradigm examples have been called into question. David Hubel and Torsten Wiesels (1962) Nobel Prize adherents, who strove to establish the receptive fields of neurons in striate cortices were often interpreted as revealing cells manouevre with those that function continued of their detection, however, Lehky and Sejnowski (1988) have challenged this interpretation. They trained an artificial neural network to distinguish the three-dimensional shape and orientation of an object from its two-dimensional shading pattern. Their network incorporates many features of visual neurophysiology. Nodes in the trained network turned out to be maximally responsive to edge contrasts, but did not appear to have the function of edge detection.

Kathleen Akins (1996) offers a different Neuro philosophical challenge to informational semantics and its affiliated feature-detection view of sensory representation. We saw in the previous section how Akins argues that the physiology of thermo receptor violates three necessary conditions on veridical representation. From this fact she draws doubts about looking for feature detecting neurons to ground some psychosemantics generally, including thought contents. Human thoughts about flies, for example, are sensitive to numerical distinctions between particular flies and the particular locations they can occupy. But the ends of frog nutrition are well served without a representational system sensitive to such ontological refinements. Whether a fly seen now is numerically identical to one seen a moment ago, need not, and perhaps cannot, figure into the frogs feature detection repertoire. Akins critique casts doubt on whether details of sensory transduction will scale up to encompass of some adequately unified psychosemantics. It also raises new questions for human intentionality. How do we get from activity patterns in narcissistic sensory receptors, keyed not to objective environmental features but rather only to effects of the stimuli on the patch of tissue enervated, to the human ontology replete with enduring objects with stable configurations of properties and relations, types and their tokens (as the fly-thought example presented above reveals), and the rest? And how did the development of a stable and rich ontology confer survival advantages to human ancestors?

Consciousness has reemerged as a topic in philosophy of mind and the cognition and attitudinal values over the past three decades. Instead of ignoring it, many physicalists now seek to explain it (Dennett, 1991). Here we focus exclusively on ways those neuroscientific discoveries have impacted philosophical debates about the nature of consciousness and its relation to physical mechanisms. Thomas Nagel (1937 - ), argues that conscious experience is subjective, and thus permanently recalcitrant to objective scientific understanding. He invites us to ponder what it is like to be a bat and urges the intuition that no amount of physical-scientific knowledge (including neuroscientific) supplies a complete answer. Nagels work is centrally concerned with the nature of moral motivation and the possibility of as rational theory of moral and political commitment, and has been a major impetus of interests in realistic and Kantian approaches to these issues. The modern philosophy of mind has been his 'What is it Like to Be a Bat? , Arguing that there is an irreducible subjective aspect of experience that cannot be grasped by the objective methods of natural science, or by philosophies such as functionalism that confine themselves to those methods, as the intuition pump up has generated extensive philosophical discussion. At least two well-known replies make direct appeal to neurophysiology. John Biro suggests that part of the intuition pumped by Nagel, that bat experience is substantially different from human experience, presupposes systematic relations between physiology and phenomenology. Kathleen Akins (1993) delves deeper into existing knowledge of bat physiology and reports much that is pertinent to Nagels question. She argues that many of the questions about subjectivity that we still consider open hinge on questions that remain unanswered about neuroscientific details.

The more recent philosopher David Chalmers (1996) has argued that any possible brain-process account of consciousness will leave open an explanatory gap between the brain process and properties of the conscious experience. This is because no brain-process theory can answer the hard question: Why should that particular brain process give rise to conscious experience? We can always imagine (conceive of) a universe populated by creatures having those brain processes but completely lacking conscious experience. A theory of consciousness requires an explanation of how and why some brain process causes consciousness replete with all the features we commonly experience. The fact that the more difficult of questions remains unanswered implicates that we will probably never get to culminate of an explanation of consciousness, in that, at the level of neural compliance. Paul and Patricia Churchland have recently offered the following diagnosis and reply. Chalmers offer a conceptual argument, based on our ability to imagine creatures possessing brains like ours but wholly lacking in conscious experience. But the more one learns about how the brain produces conscious experience-and literature is beginning to emerge (e.g., Gazzaniga, 1995) - the harder it becomes to imagine a universe consisting of creatures with brain processes like ours but lacking consciousness. This is not just too bare assertions. The Churchlands appeal to some neurobiological detail. For example, Paul Churchland (1995) develops a neuroscientific account of consciousness based on recurrent connections between thalamic nuclei (particularly diffusely projecting nuclei like the intralaminar nuclei) and the cortex. Churchland argues that the thalamocortical recurrences accounts for the selective features of consciousness, for the effects of short-term memory on conscious experience, for vivid dreaming during REM. (rapid-eye movement) sleep and other core features of conscious experience. In other words, the Churchlands are claiming that when one learns about activity patterns in these recurrent circuits, one can't imagine or conceive of this activity occurring without these core features of conscious experience. (Other than just mouthing the words, I am now imagining activity in these circuits without selective attention/the effects of short-term memory/vivid dreaming . . .)

A second focus of sceptical arguments about a complete neuroscientific explanation of consciousness is sensory Qualia: the introspectable qualitative aspects of sensory experience, the features by which subjects discern similarities and differences among their experiences. The colours of visual sensations are a philosopher's favourite example. One famous puzzle about colour Qualia is the alleged conceivability of spectral inversions. Many philosophers claim that it is conceptually possible (if perhaps physically impossible) for two humans not to diverge apart of similarities, but such are the compatibles as forwarded by their differing negation to neurophysiology. While the colour that fires engines and tomatoes appear to have of only one subject, is the colour that grasses and frogs appear in having the other (and vice versa). A large amount of neurophysiologically informed philosophy has addressed this question. A related area where neurophilosophical considerations have emerged concerns the metaphysics of colours themselves (rather than Collor experiences). A longstanding philosophical dispute is whether colours are objective properties existing external to perceiver or rather identifiable as or dependent upon minds or nervous systems. Some recent work on this problem begins with characteristics of Collor experiences: For example those Collor similarity judgments produce Collor orderings that align on a circle. With this resource, one can seek mappings of phenomenology onto environmental or physiological regularities. Identifying colours with particular frequencies of electromagnetic radiation does not preserve the structure of the hue circle, whereas identifying colours with activity in opponent processing neurons does. Such a tidbit is not decisive for the Collor objectivist-subjectivist debate, but it does convey the type of neurophilosophical work being done on traditional metaphysical issues beyond the philosophy of mind.

We saw in the discussion of Hardcastle (1997) two sections above that Neurophilosophers have entered disputes about the nature and methodological import of pain experiences. Two decades earlier, Dan Dennett (1978) took up the question of whether it is possible to build a computer that feels pain. He compares and notes the strong move between Neurophysiologic discoveries and common sense intuitions about pain experience. He suspects that the incommensurability between scientific and common sense views is due to incoherence in the latter. His attitude is wait-and-see. But foreshadowing Churchlands reply to Chalmers, Dennett favours scientific investigations over conceivability-based philosophical arguments.

Neurological deficits have attracted philosophical interest. For thirty years philosophers have found implications for the unity of the self in experiments with commissurotomy patients. In carefully controlled experiments, commissurotomy patients display two dissociable seats of consciousness. Patricia Churchland scouts philosophical implications of a variety of neurological deficits. One deficit is blindsight. Some patients with lesions to primary visual cortex report being unable to see items in regions of their visual fields, yet perform far better than chance in forced guess trials about stimuli in those regions. A variety of scientific and philosophical interpretations have been offered. Ned Form (1988) worries that many of these conflate distinct notions of consciousness. He labels these notions phenomenal consciousness (P-consciousness) and access consciousness (A-consciousness). The former is that which, what it is like-ness of experience. The latter are the availability of representational content to self-initiated action and speech. Form argues that P-consciousness is not always representational whereas A-consciousness is. Dennett and Michael Tye are sceptical of non-representational analyses of consciousness in general. They provide accounts of blind sight that do not depend on Forms distinction.

Many other topics are worth neurophilosophical pursuit. We mentioned commissurotomy and the unity of consciousness and the self, which continues to generate discussion. Qualia beyond those of Collor and pain have begun to attract neurophilosophical attention has self-consciousness. The first issues to arise in the philosophy of neuroscience (before there was a recognized area) were the localization of cognitive functions to specific neural regions. Although the localization approach had dubious origins in the phrenology of Gall and Spurzheim, and was challenged severely by Flourens throughout the early nineteenth century, it reemerged in the study of aphasia by Bouillaud, Auburtin, Broca, and Wernicke. These neurologists made careful studies (where possible) of linguistic deficits in their aphasic patients followed by brain autophsys postmortem. Brocas initial study of twenty-two patients in the mid-nineteenth century confirmed that damage to the left cortical hemisphere was predominant, and that damage to the second and third frontal convolutions was necessary to produce speech production deficits. Although the anatomical coordinates Brocas postulates for the speech production centres do not correlate exactly with damage producing production deficits as both are in this area of frontal cortexes and speech production requires of some greater degree of composure, in at least, that still bears his name (Brocas area and Brocas aphasia). Less than two decades later Carl Wernicke published evidence for a second language Centre. This area is anatomically distinct from Brocas area, and damage to it produced a very different set of aphasic symptoms. The cortical area that still bears his name (Wernickes area) is located around the first and second convolutions in temporal cortex and the aphasia that bear his name (Wernickes aphasia) involves deficits in language comprehension. Wernickes method, like Brocas, was based on lesion studies: a careful evaluation of the behavioural deficits followed by post mortem examination to find the sites of tissue damage and atrophy. Lesion studies suggesting more precise localization of specific linguistic functions remain the groundwork of a strengthening foundation to which supports all while it remains in tack to this day in unarticulated research

Lesion studies have also produced evidence for the localization of other cognitive functions: for example, sensory processing and certain types of learning and memory. However, localization arguments for these other functions invariably include studies using animal models. With an animal model, one can perform careful behavioural measures in highly controlled settings then ablate specific areas of neural tissue (or use a variety of other techniques to Form or enhance activity in these areas) and remeasure performance on the same behavioural tests. But since we lack an animal model for (human) language production and comprehension, this additional evidence isn't available to the neurologist or neurologist. This fact makes the study of language a paradigm case for evaluating the logic of the lesion/deficit method of inferring functional localization. Philosopher Barbara Von Eckardt (1978) attempts to make explicitly the steps of reasoning involved in this common and historically important method. Her analysis begins with Robert Cummins early analysis of functional explanation, but she extends it into a notion of structurally adequate functional analysis. These analyses break down a complex capacity 'C' into its constituent capacities, C2, . . . Cn, where the constituent capacities are consistent with the underlying structural details of the system. For example, human speech production (complex capacity C) results from formulating a speech intention, then selecting appropriate linguistic representations to capture the content of the speech intention, then formulating the motor commands to produce the appropriate sounds, then communicating these motor commands to the appropriate motor pathways (constituent capacities c1, c2, . . . , Cn). A functional-localization hypothesis has the form: Brain structure S in an organism (type) O has constituent capacity ci, where ci is a function of some part of O. An example, Brains Brocas area (S) in humans (O) formulates motor commands to produce the appropriate sounds (one of the constituent capacities ci). Such hypotheses specify aspects of the structural realization of a functional-component model. They are part of the theory of the neural realization of the functional model.

Armed with these characterizations, Von Eckardt argues that inference to some functional-localization hypothesis proceeds in two steps. First, a functional deficit in a patient is hypothesized based on the abnormal behavior the patient exhibits. Second, localization of function in normal brains is inferred on the basis of the functional deficit hypothesis plus the evidence about the site of brain damage. The structurally-adequate functional analysis of the capacity connects the pathological behavior to the hypothesized functional deficit. This connexion suggests four adequacy conditions on a functional deficit hypothesis. First, the pathological behavior P (e.g., the speech deficits characteristic of Brocas aphasia) must result from failing to exercise some complex capacity C (human speech production). Second, there must be a structurally-adequate functional analysis of how people exercise capacity C that involves some constituent capacity ci (formulating motor commands to produce the appropriate sounds). Third, the operation of the steps described by the structurally-adequate functional analysis minus the operation of the component performing ci (Brocas area) must result in pathological behavior P. Fourth, there must not be a better available explanation for why the patient does P. Arguments to a functional deficit hypothesis on the basis of pathological behavior is thus an instance of argument to the best available explanation. When postulating a deficit in a normal functional component provides the best available explanation of the pathological data, we are justified in drawing the inference.

Von Eckardt applies this analysis to a neurological case study involving a controversial reinterpretation of anoxia. Her philosophical explication of this important neurological method reveals that most challenges to localization arguments of whether to argue only against the localization of a particular type of functional capacity or against generalizing from localization of function in one individual to all normal individuals. (She presents examples of each from the neurological literature.) Such challenges do not impugn the validity of standard arguments for functional localization from deficits. It does not follow that such arguments are unproblematic. But they face difficult factual and methodological problems, not logical ones. Furthermore, the analysis of these arguments as involving a type of functional analysis and inference to the best available explanation carries an important implication for the biological study of cognitive function. Functional analyses require functional theories, and structurally adequate functional analyses require checks imposed by the lower level sciences investigating the underlying physical mechanisms. Arguments to best available explanation are often hampered by a lack of theoretical imagination: the available explanations are often severely limited. We must seek theoretical inspiration from any level of theory and explanation. Hence making explicitly the logic of this common and historically important form of neurological explanation reveals the necessity of joint participation from all scientific levels, from cognitive psychology down to molecular neuroscience. Von Eckardt anticipated what came to be heralded as the co-evolutionary research methodology, which remains a centerpiece of Neurophilosophers to the present day.

Over the last two decades, evidence for localization of cognitive function has come increasingly from a new source: the development and refinement of neuroimaging techniques. The form of localization-of-function argument appears not to have changed from that employing lesion studies (as analyzed by Von Eckardt). Instead, these imaging technologies resolve some of the methodological problems that plague lesion studies. For example, researchers do not need to wait until the patient dies, and in the meantime probably acquires additional brain damage, to find the lesion sites. Two functional imaging techniques are prominent: Positron emission tomography, or PET, and functional magnetic resonance imaging, or MRI. Although these measure different biological markers of functional activity, both now have a resolution down too around one millimeter. As these techniques increase spatial and temporal resolution of functional markers and continue to be used with sophisticated behavioural methodologies, the possibility of localizing specific psychological functions to increasingly specific neural regions continues to grow



What we now know about the cellular and molecular mechanisms of neural conductance and transmission is spectacular. The same evaluation holds for all levels of explanation and theory about the mind/brain: maps, networks, systems, and behavior. This is a natural outcome of increasing scientific specialization. We develop the technology, the experimental techniques, and the theoretical frameworks within specific disciplines to push forward our understanding. Still, a crucial aspect of the total picture gets neglected: the relationships between the levels, the glue that binds knowledge of neuron activity to subcellular and molecular mechanisms, network activity patterns to the activity of and connectivity between single neurons, and behavioural network activity. This problem is especially glaring when we focus on the relationship between cognitive psychological theories, postulating information-bearing representations and processes operating over their contents, and the activity patterns in networks of neurons. Co-evolution between explanatory levels still seems more like a distant dream rather than an operative methodology.

It is here that some neuroscientists appeal to computational methods. If we examine the way that computational models function in more developed sciences (like physics), we find the resources of dynamical systems constantly employed. Global effects (such as large-scale meteorological patterns) are explained in terms of the interaction of local lower-level physical phenomena, but only by dynamical, nonlinear, and often chaotic sequences and combinations. Addressing the interlocking levels of theory and explanation in the mind/brain using computational resources that have worked to bridge levels in more mature sciences might yield comparable results. This methodology is necessarily interdisciplinary, drawing on resources and researchers from a variety of levels, including higher levels like experimental psychology, program-writing and connectionist artificial intelligence, and philosophy of science.

However, the use of computational methods in neuroscience is not new. Hodgkin, Huxley, and Katz incorporated values of voltage-dependent potassium conductance they had measured experimentally in the squid giant axon into an equation from physics describing the time evolution of a first-order kinetic process. This equation enabled them to calculate best-fit curves for modelled conductance versus time data that reproduced the S-shaped (sigmoidal) function suggested by their experimental data. Using equations borrowed from physics, Rall (1959) developed the cable model of dendrites. This theory provided an account of how the various inputs from across the dendritic tree interact temporally and spatially to determine the input-output properties of single neurons. It remains influential today, and has been incorporated into the genesis software for programming neurally realistic networks. More recently, David Sparks and his colleagues have shown that a vector-averaging model of activity in neurons of correctly predicts experimental results about the amplitude and direction of saccadic eye movements. Working with a more sophisticated mathematical model, Apostolos Georgopoulos and his colleagues have predicted direction and amplitude of hand and arm movements based on averaged activity of 224 cells in motor cortices. Their predictions have borne out under a variety of experimental tests. We mention these particular studies only because we are familiar with them. We could multiply examples of the fruitful interaction of computational and experimental methods in neuroscience easily by one-hundred-fold. Many of these extend back before computational neuroscience was a recognized research endeavour.

We've already seen one example, the vector transformation accounts, of neural representation and computation, under active development in cognitive neuroscience. Other approaches using cognitivist resources are also being pursued. Many of these projects draw upon cognitivist characterizations of the phenomena to be explained. Many exploit cognitivist experimental techniques and methodologies, but, yet, some even attempt to derive cognitivist explanations from cell-biological processes (e.g., Hawkins and Kandel 1984). As Stephen Kosslyn puts it, cognitive neuroscientists employ the information processing view of the mind characteristic of cognitivism without trying to separate it from theories of brain mechanisms. Such an endeavour calls for an interdisciplinary community willing to communicate the relevant portions of the mountain of detail gathered in individual disciplines with interested nonspecialists: not just people willing to confer with those working at related levels, but researchers trained in the methods and factual details of a variety of levels. This is a daunting requirement, but it does offer some hope for philosophers wishing to contribute to future neuroscience. Thinkers trained in both the synoptic vision afforded by philosophy and the factual and experimental basis of genuine graduate-level science would be ideally equipped for this task. Recognition of this potential niche has been slow among graduate programs in philosophy, but there is some hope that a few programs are taking steps to fill it.

No comments:

Post a Comment