https://en.wikipedia.org/wiki/Category:Concepts_in_epistemology
The distinction between absolute and relative terms was introduced by Peter Unger in his 1971 paper A Defense of Skepticism and differentiates between terms that, in their most literal sense, don't admit of degrees (absolute terms) and those that do (relative terms).[1] According to his account, the term "flat", for example, is an absolute term because a surface is either perfectly (or absolutely) flat or isn't flat at all. The terms "bumpy" or "curved", on the other hand, are relative terms because there is no such thing as "absolute bumpiness" or "absolute curvedness" (although in analytic geometry curvedness is quantified). A bumpy surface can always be made bumpier. A truly flat surface, however, can never be made flatter. Colloquially, he acknowledges, we do say things like "surface A is flatter than surface B", but this is just a shorter way of saying "surface A is closer to being flat than surface B". This paraphrasing, however, doesn't work for relative terms. Another important aspect of absolute terms, one that motivated this choice of terminology, is that they can always be modified by the term "absolutely". For example, it is quite natural to say "this surface is absolutely flat", but it would be very strange and barely even meaningful to say "this surface is absolutely bumpy".
https://en.wikipedia.org/wiki/Absolute_and_relative_terms
Part of a series on |
Epistemology |
---|
Core concepts |
Distinctions |
Schools of thought |
Topics and views |
Specialized domains of inquiry |
Notable epistemologists |
Related fields |
https://en.wikipedia.org/wiki/Absolute_and_relative_terms
Part of a series on |
Epistemology |
---|
Core concepts |
Distinctions |
Schools of thought |
Topics and views |
Specialized domains of inquiry |
Notable epistemologists |
Related fields |
A priori ("from the earlier") and a posteriori ("from the later") are Latin phrases used in philosophy to distinguish types of knowledge, justification, or argument by their reliance on empirical evidence or experience. A priori knowledge is independent from current experience (e.g., as part of a new study). Examples include mathematics,[i] tautologies, and deduction from pure reason.[ii] A posteriori knowledge depends on empirical evidence. Examples include most fields of science and aspects of personal knowledge.
The terms originate from the analytic methods found in Organon, a collection of works by Aristotle. Prior analytics (a priori) is about deductive logic, which comes from definitions and first principles. Posterior analytics (a posteriori) is about inductive logic, which comes from observational evidence.
Both terms appear in Euclid's Elements and were popularized by Immanuel Kant's Critique of Pure Reason, an influential work in the history of philosophy.[1] Both terms are primarily used as modifiers to the noun "knowledge" (i.e. "a priori knowledge"). A priori can be used to modify other nouns such as "truth". Philosophers may use apriority, apriorist, and aprioricity as nouns referring to the quality of being a priori.[2]
https://en.wikipedia.org/wiki/A_priori_and_a_posteriori
Justification (also called epistemic justification) is the property of belief that qualifies it as knowledge rather than mere opinion. Epistemology is the study of reasons that someone holds a rationally admissible belief (although the term is also sometimes applied to other propositional attitudes such as doubt).[1] Epistemologists are concerned with various epistemic features of belief, which include the ideas of warrant (a proper justification for holding a belief), knowledge, rationality, and probability, among others.
Debates surrounding epistemic justification often involve the structure of justification, including whether there are foundational justified beliefs or whether mere coherence is sufficient for a system of beliefs to qualify as justified. Another major subject of debate is the sources of justification, which might include perceptual experience (the evidence of the senses), reason, and authoritative testimony, among others.
Justification and knowledge
"Justification" involves the reasons why someone holds a belief that one should hold based on one's current evidence.[1] Justification is a property of beliefs insofar as they are held blamelessly. In other words, a justified belief is a belief that a person is entitled to hold.
Many philosophers from Plato onward have treated "justified true belief" as constituting knowledge. It is particularly associated with a theory discussed in his dialogues Meno and Theaetetus. While in fact Plato seems to disavow justified true belief as constituting knowledge at the end of Theaetetus, the claim that Plato unquestioningly accepted this view of knowledge stuck until the proposal of the Gettier problem.[1]
The subject of justification has played a major role in the value of knowledge as "justified true belief".[citation needed] Some contemporary epistemologists, such as Jonathan Kvanvig assert that justification isn't necessary in getting to the truth and avoiding errors. Kvanvig attempts to show that knowledge is no more valuable than true belief, and in the process dismissed the necessity of justification due to justification not being connected to the truth.[citation needed]
Conceptions of justification
William P. Alston identifies two conceptions of justification.[2]: 15–16 One conception is "deontological" justification, which holds that justification evaluates the obligation and responsibility of a person having only true beliefs. This conception implies, for instance, that a person who has made his best effort but is incapable of concluding the correct belief from his evidence is still justified. The deontological conception of justification corresponds to epistemic internalism. Another conception is "truth-conducive" justification, which holds that justification is based on having sufficient evidence or reasons that entails that the belief is at least likely to be true. The truth-conductive conception of justification corresponds to epistemic externalism.
Theories of justification
There are several different views as to what entails justification, mostly focusing on the question "How sure do we need to be that our beliefs correspond to the actual world?" Different theories of justification require different conditions before a belief can be considered justified. Theories of justification generally include other aspects of epistemology, such as defining knowledge.
Notable theories of justification include:
- Foundationalism – Basic beliefs justify other, non-basic beliefs.
- Epistemic coherentism – Beliefs are justified if they cohere with other beliefs a person holds, each belief is justified if it coheres with the overall system of beliefs.
- Infinitism – Beliefs are justified by infinite chains of reasons.
- Foundherentism – Both fallible foundations and coherence are components of justification—proposed by Susan Haack.
- Internalism and externalism – The believer must be able to justify a belief through internal knowledge (internalism), or outside sources of knowledge can be used to justify a belief (externalism).
- Reformed epistemology – Beliefs are warranted by proper cognitive function—proposed by Alvin Plantinga.
- Evidentialism – Beliefs depend solely on the evidence for them.
- Reliabilism – A belief is justified if it is the result of a reliable process.
- Infallibilism – Knowledge is incompatible with the possibility of being wrong.
- Fallibilism – Claims can be accepted even though they cannot be conclusively proven or justified.
- Non-justificationism – Knowledge is produced by attacking claims and refuting them instead of justifying them.
- Skepticism – Knowledge is impossible or undecidable.
Criticism of theories of justification
Robert Fogelin claims to detect a suspicious resemblance between the theories of justification and Agrippa's five modes leading to the suspension of belief. He concludes that the modern proponents have made no significant progress in responding to the ancient modes of Pyrrhonian skepticism.[3]
William P. Alston criticizes the very idea of a theory of justification. He claims: "There isn't any unique, epistemically crucial property of beliefs picked out by 'justified'. Epistemologists who suppose the contrary have been chasing a will-o'-the-wisp. What has really been happening is this. Different epistemologists have been emphasizing, concentrating on, "pushing" different epistemic desiderata, different features of belief that are positively valuable from the standpoint of the aims of cognition."[2]: 22
See also
References
- Robert J. Fogelin, Pyrrhonian Reflections on Knowledge and Justification, Oxford University Press, 1994, ISBN 978-0-19-508987-5
External links
Stanford Encyclopedia of Philosophy
- Stanford Encyclopedia of Philosophy entry on Foundationalist Theories of Epistemic Justification
- Stanford Encyclopedia of Philosophy entry on Epistemology, 2. What is Justification?
- Stanford Encyclopedia of Philosophy entry on Internalist vs. Externalist Conceptions of Epistemic Justification
- Stanford Encyclopedia of Philosophy entry on Coherentist Theories of Epistemic Justification
Internet Encyclopedia of Philosophy
- Internet Encyclopedia of Philosophy entry on Epistemic Justification
- Internet Encyclopedia of Philosophy entry on Epistemic Entitlement
- Internet Encyclopedia of Philosophy entry on Internalism and Externalism in Epistemology
- Internet Encyclopedia of Philosophy entry on Epistemic Consequentialism
- Internet Encyclopedia of Philosophy entry on Coherentism in Epistemology
- Internet Encyclopedia of Philosophy entry on Contextualism in Epistemology
- Internet Encyclopedia of Philosophy entry on Knowledge-First Theories of Justification
https://en.wikipedia.org/wiki/Justification_(epistemology)
Part of a series on |
Epistemology |
---|
Core concepts |
Distinctions |
Schools of thought |
Topics and views |
Specialized domains of inquiry |
Notable epistemologists |
Related fields |
Knowledge is a form of awareness or familiarity. It is often understood as awareness of facts or as practical skills, and may also mean familiarity with objects or situations. Knowledge of facts, also called propositional knowledge, is often defined as true belief that is distinct from opinion or guesswork by virtue of justification. While there is wide agreement among philosophers that propositional knowledge is a form of true belief, many controversies in philosophy focus on justification: whether it is needed at all, how to understand it, and whether something else besides it is needed. These controversies intensified due to a series of thought experiments by Edmund Gettier and have provoked various alternative definitions. Some of them deny that justification is necessary and suggest alternative criteria while others accept that justification is an essential aspect and formulate additional requirements.
Knowledge can be produced in many different ways. The most important source of empirical knowledge is perception, which is the usage of the senses. Many theorists also include introspection as a source of knowledge, not of external physical objects, but of one's own mental states. Other sources often discussed include memory, rational intuition, inference, and testimony. According to foundationalism, some of these sources are basic in the sense that they can justify beliefs without depending on other mental states. This claim is rejected by coherentists, who contend that a sufficient degree of coherence among all the mental states of the believer is necessary for knowledge. According to infinitism, an infinite chain of beliefs is needed.
Many different aspects of knowledge are investigated, and it plays a role in various disciplines. It is the primary subject of the field of epistemology, which studies what someone knows, how they come to know it, and what it means to know something. The problem of the value of knowledge concerns the question of why knowledge is more valuable than mere true belief. Philosophical skepticism is the thesis that humans lack any form of knowledge or that knowledge is impossible. Formal epistemology studies, among other things, the rules governing how knowledge and related states behave and in what relations they stand to each other. Science tries to acquire knowledge using the scientific method, which is based on repeatable experimentation, observation, and measurement. Many religions hold that humans should seek knowledge and that God or the divine is the source of knowledge.
Definitions
Numerous definitions of knowledge have been suggested.[1][2][3] Most definitions of knowledge in analytic philosophy recognize three fundamental types. "Knowledge-that", also called propositional knowledge, can be expressed using that-clauses as in "I know that Dave is at home".[4][5][6] "Knowledge-how" (know-how) expresses practical competence, as in "she knows how to swim". Finally, "knowledge by acquaintance" refers to a familiarity with the known object based on previous direct experience.[5][7][8] Analytical philosophers usually aim to identify the essential features of propositional knowledge in their definitions.[9] There is wide, though not universal, agreement among philosophers that knowledge involves a cognitive success or an epistemic contact with reality, like making a discovery, and that propositional knowledge is a form of true belief.[10][11]
Despite the agreement about these general characteristics of knowledge, many deep disagreements remain regarding its exact definition. These disagreements relate to the goals and methods within epistemology and other fields, or to differences concerning the standards of knowledge that people intend to uphold, for example, what degree of certainty is required. One approach is to focus on knowledge's most salient features in order to give a practically useful definition.[12][5] Another is to try to provide a theoretically precise definition by listing the conditions that are individually necessary and jointly sufficient. The term "analysis of knowledge" (or equivalently, "conception of knowledge" or "theory of knowledge") is often used for this approach.[1][13][14] It can be understood in analogy to how chemists analyze a sample by seeking a list of all the chemical elements composing it.[1][15][16] An example of this approach is characterizing knowledge as justified true belief (JTB), which is seen by many philosophers as the standard definition.[4][17] Others seek a common core among diverse forms of knowledge, for example, that they all involve some kind of awareness or that they all belong to a special type of successful performance.[11][18][19][20]
Methodological differences concern whether researchers base their inquiry on abstract and general intuitions or on concrete and specific cases, referred to as methodism and particularism, respectively.[21][22][23] Another source of disagreement is the role of ordinary language in one's inquiry: the weight given to how the term "knowledge" is used in everyday discourse.[6][14] According to Ludwig Wittgenstein, for example, there is no clear-cut definition of knowledge since it is just a cluster of concepts related through family resemblance.[24][25] Different conceptions of the standards of knowledge are also responsible for various disagreements. Some epistemologists, like René Descartes, hold that knowledge demands very high requirements, like certainty, and is therefore quite rare. Others see knowledge as a rather common phenomenon, prevalent in many everyday situations, without excessively high standards.[1][5][26]
In analytic philosophy, knowledge is usually understood as a mental state possessed by an individual person, but the term is sometimes used to refer to a characteristic of a group of people as group knowledge, social knowledge, or collective knowledge.[27][28] In a slightly different sense, it can also mean knowledge stored in documents, as in "knowledge housed in the library"[29][30] or the knowledge base of an expert system.[31][32] The English word knowledge is a broad term. It includes various meanings that some other languages distinguish using several words. For example, Latin uses the words cognitio and scientia for "knowledge" while Spanish uses the words conocer and saber for "to know".[11] In ancient Greek, four important terms for knowledge were used: epistēmē (unchanging theoretical knowledge), technē (expert technical knowledge), mētis (strategic knowledge), and gnōsis (personal intellectual knowledge).[20] Knowledge is often contrasted with ignorance, which is associated with a lack of understanding, education, and true beliefs.[33][34][35] Epistemology, also referred to as the theory of knowledge, is the philosophical discipline studying knowledge. It investigates topics like the nature of knowledge and justification, how knowledge arises, and what value it has. Further issues include the different types of knowledge and the extent to which the beliefs of most people amount to knowledge as well as the limits of what can be known.[36][37][11]
Justified true belief
Many philosophers define knowledge as justified true belief (JTB). This definition characterizes knowledge through three essential features: as (1) a belief that is (2) true and (3) justified.[4][17] In the dialogue Theaetetus by the ancient Greek philosopher Plato, Socrates pondered the distinction between knowledge and true belief but rejected the JTB definition of knowledge.[38][39] The most widely accepted feature is truth: one can believe something false but one cannot know something false.[5][6] A few ordinary language philosophers have raised doubts that knowledge is a form of belief based on everyday expressions like "I do not believe that; I know it".[4][5][40] Most theorists reject this distinction and explain such expressions through ambiguities of natural language.[4][5]
The main controversy surrounding the JTB definition concerns its third feature: justification.[1][4][41] This component is often included because of the impression that some true beliefs are not forms of knowledge. Specifically, this covers cases of superstition, lucky guesses, or erroneous reasoning. The corresponding beliefs may even be true but it seems there is more to knowledge than just being right about something.[4][5][14] The JTB definition solves this problem by identifying proper justification as the additional component needed, which is absent in the above-mentioned cases. Many philosophers have understood justification internalistically (internalism): a belief is justified if it is supported by another mental state of the person, such as a perceptual experience, a memory, or a second belief. This mental state has to constitute a sufficiently strong evidence or reason for the believed proposition. Some modern versions modify the JTB definition by using an externalist conception of justification instead. This means that justification depends not just on factors internal to the subject but also on external factors. According to reliabilist theories of justification, justified beliefs are produced by a reliable process. According to causal theories of justification, justification requires that the believed fact causes the belief. This is the case, for example, when a bird sits on a tree and a person forms a belief about this fact because they see the bird.[1][4][5]
Gettier problem and alternatives
The JTB definition came under severe criticism in the 20th century, when Edmund Gettier gave a series of counterexamples.[42] They purport to present concrete cases of justified true beliefs that fail to constitute knowledge. The reason for their failure is usually a form of epistemic luck: the justification is not relevant to the truth.[4][5][41] In a well-known example, there is a country road with many barn facades and only one real barn. The person driving is not aware of this, stops in front of the real barn by a lucky coincidence, and forms the belief that he is in front of a barn. It has been argued that this justified true belief does not constitute knowledge since the person would not have been able to tell the difference without the fortuitous accident.[43][44][45] So even though the belief is justified, it is a lucky coincidence that it is also true.[1]
The responses to these counterexamples have been diverse. According to some, they show that the JTB definition of knowledge is deeply flawed and that a radical reconceptualization of knowledge is necessary, often by denying justification a role.[1] This can happen, for example, by replacing justification with reliability or by understanding knowledge as the manifestation of cognitive virtues. Another approach is to define it in regard to the cognitive role it plays. For example, one role of knowledge is to provide reasons for thinking something or for doing something.[11] Various theorists are diametrically opposed to the radical reconceptualization and either deny that Gettier cases pose problems or they try to solve them by making smaller modifications to how justification is defined. Such approaches result in a minimal modification of the JTB definition.[1]
Between these two extremes, some philosophers have suggested various moderate departures. They agree that the JTB definition includes some correct claims: justified true belief is a necessary condition of knowledge. However, they disagree that it is a sufficient condition. They hold instead that an additional criterion, some feature X, is necessary for knowledge. For this reason, they are often referred to as JTB+X definitions of knowledge.[1][46] A closely related approach speaks not of justification but of warrant and defines warrant as justification together with whatever else is necessary to arrive at knowledge.[4][47]
Many candidates for the fourth feature have been suggested. In this regard, knowledge may be defined as justified true belief that does not depend on any false beliefs, that there are no defeaters[a] present, or that the person would not have the belief if it was false.[14][45] Such and similar definitions are successful at avoiding many of the original Gettier cases. However, they are often undermined by newly conceived counterexamples.[49] To avoid all possible cases, it may be necessary to find a criterion that excludes all forms of epistemic luck. It has been argued that such a criterion would set the required standards of knowledge very high: the belief has to be infallible to succeed in all cases.[5][50] This would mean that very few of our beliefs amount to knowledge, if any.[5][51][52] For example, Richard Kirkham suggests that our definition of knowledge requires that the evidence for the belief necessitates its truth.[53] There is still very little consensus in the academic discourse as to which of the proposed modifications or reconceptualizations is correct.[1][54][11]
Types
A common distinction among types of knowledge is between propositional knowledge, or knowledge-that, and non-propositional knowledge in the form of practical skills or acquaintance.[5][55][56] The distinctions between these major types are usually drawn based on the linguistic formulations used to express them.[1]
Propositional knowledge
Propositional knowledge, also referred to as descriptive knowledge, is a form of theoretical knowledge about facts, like knowing that "2 + 2 = 4". It is the paradigmatic type of knowledge in analytic philosophy.[4][5][6] Propositional knowledge is propositional in the sense that it involves a relation to a proposition. Since propositions are often expressed through that-clauses, it is also referred to as knowledge-that, as in "Akari knows that kangaroos hop".[5][6][8] In this case, Akari stands in the relation of knowing to the proposition "kangaroos hop". Closely related types of knowledge are know-wh, for example, knowing who is coming to dinner and knowing why they are coming.[5] These expressions are normally understood as types of propositional knowledge since they usually can be paraphrased using a that-clause.[5][6] It is usually held that the capacity for propositional knowledge is exclusive to relatively sophisticated creatures, such as humans. This is based on the claim that advanced intellectual capacities are required to believe a proposition that expresses what the world is like.[57]
Propositional knowledge is either occurrent or dispositional. Occurrent knowledge is knowledge that is active, for example, because a person is currently thinking about it. Dispositional knowledge, on the other hand, is stored in the back of a person's mind without being involved in cognitive processes at the moment. In this regard, it refers to the mere ability to access the relevant information. For example, a person may know for most of their life that cats have whiskers. This knowledge is dispositional most of the time. It becomes occurrent when the person actively thinks about the whiskers of cats. A similar classification is often discussed in relation to beliefs as the difference between occurrent and dispositional beliefs.[6][58][59]
Non-propositional knowledge
Non-propositional knowledge is knowledge in which no essential relation to a proposition is involved. The two most well-known forms are knowledge-how (know-how or procedural knowledge) and knowledge by acquaintance.[5][6][7] The term "know-how" refers to some form of practical ability or skill. It can be defined as having the corresponding competence.[5][57] Examples include knowing how to ride a bicycle or knowing how to swim. Some of the abilities responsible for know-how may also involve certain forms of knowledge-that, as in knowing how to prove a mathematical theorem, but this is not generally the case.[11] Some forms of practical knowledge do not require a highly developed mind, in contrast to propositional knowledge. In this regard, practical knowledge is more common in the animal kingdom. For example, an ant knows how to walk even though it presumably lacks a mind sufficiently developed enough to stand in a relation to the corresponding proposition by representing it.[57] Knowledge-how is closely related to tacit knowledge. Tacit knowledge is knowledge that cannot be fully articulated or explained, in contrast to explicit knowledge.[60][61]
Knowledge by acquaintance is familiarity with an individual that results from direct experiential contact with this individual.[5][6][8] This individual can a person or a regular object. On the linguistic level, it does not require a that-clause and can be expressed using a direct object. So when someone claims that they know Wladimir Klitschko personally, they are expressing that they had a certain kind of contact with him and not that they know a certain fact about him. This is usually understood to mean that it constitutes a relation to a concrete individual and not to a proposition. Knowledge by acquaintance plays a central role in Bertrand Russell's epistemology. He contrasts it with knowledge by description, which is a form of propositional knowledge not based on direct perceptual experience.[62][63][5] However, there is some controversy about whether it is possible to acquire knowledge by acquaintance in its pure non-propositional form. In this regard, some theorists, like Peter D. Klein, have suggested that it can be understood as one type of propositional knowledge that is only expressed in a grammatically different way.[4]
Other distinctions
A priori and a posteriori
The distinction between a priori and a posteriori knowledge came to prominence in Immanuel Kant's philosophy and is often discussed in the academic literature. To which category a knowledge attitude belongs depends on the role of experience in its formation and justification.[6][64][65] To know something a posteriori means to know it on the basis of experience.[66][67] For example, to know that it is currently raining or that the baby is crying belongs to a posteriori knowledge since it is based on some form of experience, like visual or auditory experience.[64] A priori knowledge, however, is possible without any experience to justify or support the known proposition.[65][68] Mathematical knowledge, such as that 2 + 2 = 4, is a paradigmatic case of a priori knowledge since no empirical investigation is necessary to confirm this fact.[67][68] The distinction between a posteriori and a priori knowledge is usually equated with the distinction between empirical and non-empirical knowledge.[67] This distinction pertains primarily to knowledge but it can also be applied to propositions or arguments. For example, an a priori proposition is a proposition that can be known independently of experience.[64]
The relevant experience in question is primarily identified with sensory experience. However, some non-sensory experiences, like memory and introspection, are often included as well. But certain conscious phenomena are excluded in this context. For example, the conscious phenomenon of a rational insight into the solution of a mathematical problem does not make the resulting knowledge a posteriori.[64][65] It is sometimes argued that, in a trivial sense, some form of experience is required even for a priori knowledge, the experience needed to learn the language in which the claim is expressed. For a priori knowledge, this is the only form of experience required. For this reason, knowing that "all bachelors are unmarried" is considered a form of a priori knowledge since, given an understanding of the terms "bachelor" and "unmarried", no further experience is necessary to know that it is true.[64][65]
One difficulty for a priori knowledge is to explain how it is possible. It is usually seen as unproblematic that one can come to know things through experience but it is not clear how knowledge is possible without experience. One of the earliest solutions to this problem is due to Plato, who argues that, in the context of geometry, the soul already possesses the knowledge and just needs to recollect or remember it to access it again.[68][69] A similar explanation is given by René Descartes, who holds that a priori knowledge exists as innate knowledge present in the mind of each human.[68] A different approach is to posit a special mental faculty responsible for this type of knowledge, often referred to as rational insight or rational intuition.[64]
Self-knowledge
In philosophy, "self-knowledge" usually refers to a person's knowledge of their own sensations, thoughts, beliefs, and other mental states. Many philosophers hold that it is a special type of knowledge since it is more direct than knowledge of the external world, which is mediated through the senses. Traditionally, it was often claimed that self-knowledge is indubitable. For example, when someone is in pain, they cannot be wrong about this fact. However, various contemporary theorists reject this position. A closely related issue is to explain how self-knowledge works. Some philosophers, like Russell, understand it as a form of knowledge by acquaintance while others, like John Locke, claim that there is an inner sense that works in analogy to how the external five senses work. According to a different perspective, self-knowledge is indirect in the sense that a person has to interpret their internal and external behavior in order to learn about their mental states, similar to how one can learn about the mental states of other people by interpreting their external behavior.[70][71][72]
In a slightly different sense, the term self-knowledge can also refer to the knowledge of the self as a persisting entity that has certain personality traits, preferences, physical attributes, relationships, goals, and social identities. This meaning is of particular interest to psychology and refers to a person's awareness of their own characteristics.[73][74][75] Self-knowledge is closely related to self-concept, the difference being that the self-concept also includes unrealistic aspects of how a person sees themselves. In this regard, self-knowledge is often measured by comparing a person's self-assessment of their character traits with how other people assess this person's traits.[74]
Situated knowledge
Situated knowledge is knowledge specific to a particular situation.[76][77] It is closely related to practical or tacit knowledge, which is learned and applied in specific circumstances. This especially concerns certain forms of acquiring knowledge, such as trial and error or learning from experience.[78] In this regard, situated knowledge usually lacks a more explicit structure and is not articulated in terms of universal ideas.[77] The term is often used in feminism and postmodernism to argue that many forms of knowledge are not absolute but depend on the concrete historical, cultural, and linguistic context.[76][77] Understood in this way, it is frequently used to argue against absolute or universal knowledge claims stated in the scientific discourse. Donna Haraway is a prominent defender of this position.[78][79] One of her arguments is based on the idea that perception is embodied and is not a universal "gaze from nowhere".[78]
Higher and lower knowledge
Many forms of eastern spirituality and religion distinguish between higher and lower knowledge. They are also referred to as para vidya and apara vidya in Hinduism or the two truths doctrine in Buddhism. Lower knowledge is based on the senses and the intellect. In this regard, all forms of empirical and objective knowledge belong to this category.[80][81] Most of the knowledge needed in one's everyday functioning is lower knowledge. It is about mundane or conventional things that are in tune with common sense. It includes the body of knowledge belonging to the empirical sciences.[80][82][83]
Higher knowledge, on the other hand, is understood as knowledge of God, the absolute, the true self, or the ultimate reality. It belongs neither to the external world of physical objects nor to the internal world of the experience of emotions and concepts. Many spiritual teachings emphasize the increased importance, or sometimes even exclusive importance, of higher knowledge in comparison to lower knowledge. This is usually based on the idea that achieving higher knowledge is one of the central steps on the spiritual path. In this regard, higher knowledge is seen as what frees the individual from ignorance, helps them realize God, or liberates them from the cycle of rebirth.[81][82] This is often combined with the view that lower knowledge is in some way based on a delusion: it belongs to the realm of mere appearances or Maya, while higher knowledge manages to view the reality underlying these appearances.[83] In the Buddhist tradition, the attainment of higher knowledge or ultimate truth is often associated with seeing the world from the perspective of sunyata, i.e. as a form of emptiness lacking inherent existence or intrinsic nature.[80][84][85]
Sources of knowledge
Sources of knowledge are ways how people come to know things. According to Andrea Kern, they can be understood as rational capacities that are exercised when a person acquires new knowledge.[86] Various sources of knowledge are discussed in the academic literature, often in terms of the mental faculties responsible. They include perception, introspection, memory, inference, and testimony. However, not everyone agrees that all of them actually lead to knowledge. Usually, perception or observation, i.e. using one of the senses, is identified as the most important source of empirical knowledge.[5][6][87] So knowing that the baby is sleeping constitutes observational knowledge if it was caused by a perception of the snoring baby. But this would not be the case if one learned about this fact through a telephone conversation with one's spouse. Direct realists explain observational knowledge by holding that perception constitutes a direct contact with the perceived object. Indirect realists, on the other hand, contend that this contact happens indirectly: people can only directly perceive sense data, which are then interpreted as representing external objects. This distinction affects whether the knowledge of external objects is direct or indirect and may thus have an impact on how certain perceptual knowledge is.[11]
Introspection is often seen in analogy to perception as a source of knowledge, not of external physical objects, but of internal mental states. Traditionally, various theorists have ascribed a special epistemic status to introspection by claiming that it is infallible or that there is no introspective difference between appearance and reality. However, this claim has been contested in the contemporary discourse. Critics argue that it may be possible, for example, to mistake an unpleasant itch for a pain or to confuse the experience of a slight ellipse for the experience of a circle.[11] Perceptual and introspective knowledge often act as a form of fundamental or basic knowledge. According to some empiricists, perceptual knowledge is the only source of basic knowledge and provides the foundation for all other knowledge.[5][6]
Memory is usually identified as another source of knowledge. It differs from perception and introspection in that it is not as independent or fundamental as they are since it depends on other previous experiences.[11][88] The faculty of memory retains knowledge acquired in the past and makes it accessible in the present, as when remembering a past event or a friend's phone number.[89][90] It is generally considered a reliable source of knowledge, but it can be deceptive at times nonetheless, either because the original experience was unreliable or because the memory degraded and does not accurately represent the original experience anymore.[11]
Knowledge based on perception, introspection, or memory may also give rise to inferential knowledge, which comes about when reasoning is applied to draw inferences from another known fact.[5][6][11] In this regard, the perceptual knowledge of a Czech stamp on a postcard may give rise to the inferential knowledge that one's friend is visiting the Czech Republic. According to rationalists, some forms of knowledge are completely independent of observation and introspection. They are needed to explain how certain a priori beliefs, like the mathematical belief that 2 + 2 = 4, constitute knowledge. Some theorists, like Robert Audi, hold that the faculty of pure reason or rational intuition is responsible in these cases since there seem to be no sensory perceptions that could justify such general and abstract knowledge.[88][91] However, difficulties in providing a clear account of pure reason or rational intuition have led various empirically minded epistemologists to doubt that they constitute independent sources of knowledge.[5][6][11] A closely related approach is to hold that this type of knowledge is innate. According to Plato's theory of recollection, for example, it is accessed through a special form of remembering.[5][6]
Testimony is often included as an additional source of knowledge. Unlike the other sources, it is not tied to one specific cognitive faculty. Instead, it is based on the idea that one person can come to know a fact because another person talks about this fact. Testimony can happen in numerous ways, like regular speech, a letter, the newspaper, or an online blog. The problem of testimony consists in clarifying under what circumstances and why it constitutes a source of knowledge. A popular response is that it depends on the reliability of the person pronouncing the testimony: only testimony from reliable sources can lead to knowledge.[11][92][93]
Structure of knowledge
The structure of knowledge is the way in which the mental states of a person need to be related to each other for knowledge to arise.[94][95] Most theorists hold that, among other things, an agent has to have good reasons for holding a belief if this belief is to amount to knowledge. So when challenged, the agent may justify their belief by referring to their reason for holding it. In many cases, this reason is itself a belief that may as well be challenged. So when the agent believes that Ford cars are cheaper than BMWs because they believe to have heard this from a reliable source, they may be challenged to justify why they believe that their source is reliable. Whatever support they present may also be challenged.[4][11][14] This threatens to lead to an infinite regress since the epistemic status at each step depends on the epistemic status of the previous step.[96][97] Theories of the structure of knowledge offer responses for how to solve this problem.[4][11][14]
The three most common theories are foundationalism, coherentism, and infinitism. Foundationalists and coherentists deny the existence of this infinite regress, in contrast to infinitists.[4][11][14] According to foundationalists, some basic reasons have their epistemic status independent of other reasons and thereby constitute the endpoint of the regress. Against this view, it has been argued that the concept of "basic reason" is contradictory: there should be a reason for why some reasons are basic and others are non-basic, in which case the basic reasons would depend on another reason after all and would therefore not be basic. An additional problem consists in finding plausible candidates for basic reasons.[4][11][14]
Coherentists and infinitists avoid these problems by denying the distinction between basic and non-basic reasons. Coherentists argue that there is only a finite number of reasons, which mutually support each other and thereby ensure each other's epistemic status.[4][11] Their critics contend that this constitutes the fallacy of circular reasoning.[98][99] For example, if belief b1 supports belief b2 and belief b2 supports belief b1, the agent has a reason for accepting one belief if they already have the other. However, their mutual support alone is not a good reason for newly accepting both beliefs at once. A closely related issue is that there can be various distinct sets of coherent beliefs and coherentists face the problem of explaining why someone should accept one coherent set rather than another.[4][11] For infinitists, in contrast to foundationalists and coherentists, there is an infinite number of reasons. This position faces the problem of explaining how human knowledge is possible at all, as it seems that the human mind is limited and cannot possess an infinite number of reasons.[4] In their traditional forms, both foundationalists and coherentists face the Gettier problem, i.e. that having a reason or justification for a true belief is not sufficient for knowledge in cases where cognitive luck is responsible for the success.[4][100]
Value of knowledge
Knowledge may be valuable either because it is useful or because it is good in itself. Knowledge can be useful by helping a person achieve their goals. An example of this form of instrumental value is knowledge of a disease, which can be beneficial because it enables a person to identify and treat the disease. Or to reach an important job interview, knowledge of where and when it takes place helps the person arrive there on time.[101][102][103] However, this does not imply that knowledge is always useful. In this regard, many true beliefs about trivial matters have neither positive nor negative value. This concerns, for example, knowing how many grains of sand are on a specific beach or memorizing phone numbers one never intends to call. In a few cases, knowledge may even have a negative value. For example, if a person's life depends on gathering the courage to jump over a ravine, then having a true belief about the involved dangers may hinder the person to do so.[102]
Besides having instrumental value, knowledge may also have intrinsic value. This means that certain forms of knowledge are good in themselves even if they do not provide any practical benefits. According to Duncan Pritchard, this applies to certain forms of knowledge associated with wisdom.[102][101] The value of knowledge is relevant to the field of education, specifically to the issue of choosing which knowledge should be passed on to the student.[101]
A more specific issue in epistemology concerns the question of whether or why knowledge is more valuable than mere true belief.[104][103] There is wide agreement that knowledge is good in some sense but the thesis that knowledge is better than true belief is controversial. An early discussion of this problem is found in Plato's Meno in relation to the claim that both knowledge and true belief can successfully guide action and, therefore, have apparently the same value. For example, it seems that mere true belief is as effective as knowledge when trying to find the way to Larissa.[103][104][105] According to Plato, knowledge is better because it is more stable.[103] A different approach is to hold that knowledge gets its additional value from justification. However, if the value in question is understood primarily as an instrumental value, it is not clear in what sense knowledge is better than mere true belief since they are usually equally useful.[104]
The problem of the value of knowledge is often discussed in relation to reliabilism and virtue epistemology.[104][103][106] Reliabilism can be defined as the thesis that knowledge is reliably-formed true belief. On this view, it is difficult to explain how a reliable belief-forming process adds additional value.[104] According to an analogy by Linda Zagzebski, a cup of coffee made by a reliable coffee machine has the same value as an equally good cup of coffee made by an unreliable coffee machine.[107] This difficulty in solving the value problem is sometimes used as an argument against reliabilism. Virtue epistemologists see knowledge as the manifestation of cognitive virtues and can thus argue that knowledge has additional value due to its association with virtue. However, not everyone agrees that knowledge actually has additional value over true belief. A similar view is defended by Jonathan Kvanvig, who argues that the main epistemic value resides not in knowledge but in understanding, which implies grasping how one's beliefs cohere with each other.[104][103][108]
Philosophical skepticism
Philosophical skepticism in its strongest form, also referred to as global skepticism, is the thesis that humans lack any form of knowledge or that knowledge is impossible. Very few philosophers have explicitly defended this position. However, it has been influential nonetheless, usually in a negative sense: many researchers see it as a serious challenge to any epistemological theory and often try to show how their preferred theory overcomes it.[4][5][11] For example, it is commonly accepted that perceptual experience constitutes a source of knowledge. However, according to the dream argument, this is not the case since dreaming provides unreliable information and since the agent could be dreaming without knowing it. Because of this inability to discriminate between dream and perception, it is argued that there is no perceptual knowledge.[4][5][6] A similar often cited thought experiment assumes that the agent is actually a brain in a vat that is just fed electrical stimuli. Such a brain would have the false impression of having a body and interacting with the external world. The basic argument is the same: since the agent is unable to tell the difference, they do not know that they have a body responsible for reliable perceptions.[11]
One issue revealed through these thought experiments is the problem of underdetermination: that the evidence available is not sufficient to make a rational decision between competing theories. And if two contrary hypotheses explain the appearances equally well, then the agent is not justified in believing one of those hypotheses rather than the other. Based on this premise, the general skeptic has to argue that this is true for all our knowledge, that there is always an alternative and very different explanation.[11] Another skeptic argument is based on the idea that human cognition is fallible and therefore lacks absolute certainty. More specific arguments target particular theories of knowledge, such as foundationalism or coherentism, and try to show that their concept of knowledge is deeply flawed.[4][11] An important argument against global skepticism is that it seems to contradict itself: the claim that there is no knowledge appears to constitute a knowledge-claim itself.[6] Other responses come from common sense philosophy and reject global skepticism based on the fact that it contradicts many plausible ordinary beliefs. It is then argued against skepticism by seeing common sense as more reliable than the abstract reasoning cited in favor of skepticism.[11][109]
Certain less radical forms of skepticism deny that knowledge exists within a specific area or discipline, sometimes referred to as local or selective skepticism.[5][6][11] It is often motivated by the idea that certain phenomena do not accurately represent their subject matter. They may thus lead to false impressions concerning its nature. External world skeptics hold that one can only know about one's own sensory impressions and experiences but not about the external world. This is based on the idea that beliefs about the external world are mediated through the senses. The senses are faulty at times and may thus show things that are not really there. This problem is avoided on the level of sensory impressions, which are given to the experiencer directly without an intermediary. In this sense, the person may be wrong about seeing a red Ferrari in the street (it might have been a Maserati or a mere light reflection) but they cannot be wrong about having a sensory impression of the color red.[5][6][11]
The inverse path is taken by some materialists, who accept the existence of the external physical world but deny the existence of the internal realm of mind and consciousness based on the difficulty of explaining how the two realms can exist together.[14] Other forms of local skepticism accept scientific knowledge but deny the possibility of moral knowledge, for example, because there is no reliable way to empirically measure whether a moral claim is true or false.[5]
The issue of the definition and standards of knowledge is central to the question of whether skepticism in its different forms is true. If very high standards are used, for example, that knowledge implies infallibility, then skepticism becomes more plausible. In this case, the skeptic only has to show that no belief is absolutely certain; that while the actual belief is true, it could have been false. However, the more these standards are weakened to how the term is used in everyday language, the less plausible skepticism becomes.[6][110][11] For example, such a position is defended in the pragmatist epistemology, which sees all beliefs and theories as fallible hypotheses and holds that they may need to be revised as new evidence is acquired.[111][112][113]
In various disciplines
Formal epistemology
Formal epistemology studies knowledge using formal tools, such as mathematics and logic.[114] An important issue in this field concerns the epistemic principles of knowledge. These are rules governing how knowledge and related states behave and in what relations they stand to each other. The transparency principle, also referred to as the luminosity of knowledge, is an often discussed principle. It states that knowing something implies the second-order knowledge that one knows it. This principle implies that if Heike knows that today is Monday, then she also knows that she knows that today is Monday.[11][115][116] Other commonly discussed principles are the conjunction principle, the closure principle, and the evidence transfer principle. For example, the conjunction principle states that having two justified beliefs in two separate propositions implies that the agent is also justified in believing the conjunction of these two propositions. In this regard, if Bob knows that dogs are animals and he also knows that cats are animals, then he knows the conjunction of these two propositions, i.e. he knows that dogs and cats are animals.[4]
Science
The scientific approach is usually regarded as an exemplary process for how to gain knowledge about empirical facts.[117][118] Scientific knowledge includes mundane knowledge about easily observable facts, for example, chemical knowledge that certain reactants become hot when mixed together. But it also encompasses knowledge of less tangible issues, like claims about the behavior of genes, neutrinos, and black holes.[119]
A key aspect of most forms of science is that they seek natural laws that explain empirical observations.[117][118] Scientific knowledge is discovered and tested using the scientific method. This method aims to arrive at reliable knowledge by formulating the problem in a clear manner and by ensuring that the evidence used to support or refute a specific theory is public, reliable, and replicable. This way, other researchers can repeat the experiments and observations in the initial study to confirm or disconfirm it.[120] The scientific method is often analyzed as a series of steps. According to some formulations, it begins with regular observation and data collection. Based on these insights, the scientists then try to find a hypothesis that explains the observations. The hypothesis is then tested using a controlled experiment to compare whether predictions based on the hypothesis match the actual results. As a last step, the results are interpreted and a conclusion is reached whether and to what degree the findings confirm or disconfirm the hypothesis.[121][122][123]
The progress of scientific knowledge is traditionally seen as a gradual and continuous process in which the existing body of knowledge is increased in each step. However, this view has been challenged by various philosophers of science, such as Thomas Kuhn, who holds that between phases of incremental progress, there are so-called scientific revolutions in which a paradigm shift occurs. According to this view, various fundamental assumptions are changed due to the paradigm shift, resulting in a radically new perspective on the body of scientific knowledge that is incommensurable with the previous outlook.[124][125]
Religion
Knowledge plays a central role in many religions. Knowledge claims about the existence of God or religious doctrines about how each one should live their lives are found in almost every culture.[126] However, such knowledge claims are often controversial and are commonly rejected by religious skeptics and atheists.[127] The epistemology of religion is the field of inquiry that investigates whether belief in God and in other religious doctrines is rational and amounts to knowledge.[126][128] One important view in this field is evidentialism. It states that belief in religious doctrines is justified if it is supported by sufficient evidence. Suggested examples of evidence for religious doctrines include religious experiences such as direct contact with the divine or inner testimony as when hearing God's voice.[126][128][129] However, evidentialists often reject that belief in religious doctrines amount to knowledge based on the claim that there is not sufficient evidence.[126][128] A famous saying in this regard is due to Bertrand Russell. When asked how he would justify his lack of belief in God when facing his judgment after death, he replied "Not enough evidence, God! Not enough evidence."[126]
However, religious teachings about the existence and nature of God are not always understood as knowledge claims by their defenders and some explicitly state that the proper attitude towards such doctrines is not knowledge but faith. This is often combined with the assumption that these doctrines are true but cannot be fully understood by reason or verified through rational inquiry. For this reason, it is claimed that one should accept them even though they do not amount to knowledge.[127] Such a view is reflected in a famous saying by Immanuel Kant where he claims that he "had to deny knowledge in order to make room for faith."[130]
Distinct religions often differ from each other concerning the doctrines they proclaim as well as their understanding of the role of knowledge in religious practice.[131][132] In both the Jewish and the Christian tradition, knowledge plays a role in the fall of man in which Adam and Eve were expelled from the Garden of Eden. Responsible for this fall was that they ignored God's command and ate from the tree of knowledge, which gave them the knowledge of good and evil. This is understood as a rebellion against God since this knowledge belongs to God and it is not for humans to decide what is right or wrong.[133][134][135] In the Christian literature, knowledge is seen as one of the seven gifts of the Holy Spirit.[136] In Islam, "the Knowing" (al-ʿAlīm) is one of the 99 names reflecting distinct attributes of God. The Qur'an asserts that knowledge comes from God and the acquisition of knowledge is encouraged in the teachings of Muhammad.[137][138]
In Buddhism, knowledge that leads to liberation is called vijjā. It contrasts with avijjā or ignorance, which is understood as the root of all suffering. This is often explained in relation to the claim that humans suffer because they crave things that are impermanent. The ignorance of the impermanent nature of things is seen as the factor responsible for this craving.[139][140][141] The central goal of Buddhist practice is to stop suffering. This aim is to be achieved by understanding and practicing the teaching known as the Four Noble Truths and thereby overcoming ignorance.[140][141] Knowledge plays a key role in the classical path of Hinduism known as jñāna yoga or "path of knowledge". Its aim is to achieve oneness with the divine by fostering an understanding of the self and its relation to Brahman or ultimate reality.[142][143]
Anthropology
The anthropology of knowledge is a multi-disciplinary field of inquiry.[144][145] It studies how knowledge is acquired, stored, retrieved, and communicated.[146] Special interest is given to how knowledge is reproduced and undergoes changes in relation to social and cultural circumstances.[144] In this context, the term knowledge is used in a very broad sense, roughly equivalent to terms like understanding and culture.[144][147] This means that the forms and reproduction of understanding are studied irrespective of their truth value. In epistemology, on the other hand, knowledge is usually restricted to forms of true belief. The main focus in anthropology is on empirical observations of how people ascribe truth values to meaning contents, like when affirming an assertion, even if these contents are false.[144] But it also includes practical components: knowledge is what is employed when interpreting and acting on the world and involves diverse phenomena, such as feelings, embodied skills, information, and concepts. It is used to understand and anticipate events in order to prepare and react accordingly.[147]
The reproduction of knowledge and its changes often happen through some form of communication.[144][146] This includes face-to-face discussions and online communications as well as seminars and rituals. An important role in this context falls to institutions, like university departments or scientific journals in the academic context.[144] A tradition or traditional knowledge may be defined as knowledge that has been reproduced within a society or geographic region over several generations. However, societies also respond to various external influences, such as other societies, whose understanding is often interpreted and incorporated in a modified form.[144][147][148]
Individuals belonging to the same social group usually understand things and organize knowledge in similar ways to one another. In this regard, social identities play a significant role: individuals who associate themselves with similar identities, like age-influenced, professional, religious, and ethnic identities, tend to embody similar forms of knowledge. Such identities concern both how an individual sees themselves, for example, in terms of the ideals they pursue, as well as how other people see them, such as the expectations they have toward the individual.[144][149]
Sociology
The sociology of knowledge is closely related to the anthropology of knowledge.[150][147][151] It is the subfield of sociology that investigates how thought and society are related to each other. In it, the term "knowledge" is understood in a very wide sense that encompasses many types of mental products. It includes philosophical and political ideas as well as religious and ideological doctrines and can also be applied to folklore, law, and technology. In this regard, knowledge is sometimes treated as a synonym for culture. The sociology of knowledge studies in what sociohistorical circumstances knowledge arises and on what existential conditions it depends. The investigated conditions can include physical, demographic, economic, and sociocultural factors. The sociology of knowledge differs from most forms of epistemology since its main focus is on the question of how ideas originate and what consequences they have rather than asking whether these ideas are coherent or correct. An example of a theory in this field is due to Karl Marx, who claims that the dominant ideology in a society is a product of and changes with the underlying socioeconomic conditions.[152][153][154] A related perspective is found in certain forms of decolonial scholarship that claim that colonial powers are responsible for the hegemony of western knowledge systems and seek a decolonization of knowledge to undermine this hegemony.[155][156]
See also
- Common knowledge
- Domain knowledge
- Epistemic modal logic
- General knowledge
- Inductive inference
- Inductive probability
- Intelligence
- Knowledge falsification
- Knowledge transfer
- Metaknowledge
- Omniscience
- Outline of human intelligence
- Outline of knowledge
References
Notes
- A defeater of a belief is evidence that this belief is false.[48]
Citations
Popper: Putting our ideas into words, or better, writing them down, makes an important difference. ... It is what I call 'knowledge in the objective sense'. Scientific knowledge belongs to it. It is this knowledge which is stored in our libraries rather than our heads. Magee: And you regard the knowledge stored in our libraries as more important than the knowledge stored in our heads. Popper: Much more important, from every point of view
It is a pervasive assumption in recent analytical philosophy that knowledge can be defined as a modality representing a rational agent's true and consistent beliefs. Such views are based on rationality assumptions. One is that knowledge can only consist of true propositions. This way of speaking is sharply at odds with the way we speak about knowledge, for example, in computing, where a so-called knowledge base can be a database, that is, a set of data that has been collected and is thought to consist of true propositions, even though, realistically speaking, many of them might later be shown to be false or untenable. ... The pragmatic account of knowledge starts with a knowledge system, meaning a working system with an agent having a database. ... The notion of a search can be a social one, in many instances. A group of agents can be engaged in the search, and some of them can know things that others do not know.
What the Gettier cases show is that this condition is insufficient to capture all the ways in which accidental truth is incompatible with knowledge.
- Dreyer, Jaco S. (21 April 2017). "Practical theology and the call for the decolonisation of higher education in South Africa: Reflections and proposals". HTS Teologiese Studies / Theological Studies. 73 (4): 1–7. doi:10.4102/hts.v73i4.4805.
Bibliography
- Zagzebski, Linda (1999). "What Is Knowledge?". In Greco, John; Sosa, Ernest (eds.). The Blackwell Guide to Epistemology. Malden, MA: Blackwell. pp. 92–116. doi:10.1002/9781405164863.ch3. ISBN 978-0-631-20290-5. OCLC 39269507. S2CID 158886670. Archived from the original on 2 June 2022. Retrieved 12 June 2022.
External links
- Knowledge at PhilPapers
- "Knowledge". Internet Encyclopedia of Philosophy.
- Zalta, Edward N. (ed.). "The Value of Knowledge". Stanford Encyclopedia of Philosophy.
- Zalta, Edward N. (ed.). "The Analysis of Knowledge". Stanford Encyclopedia of Philosophy.
- Zalta, Edward N. (ed.). "Knowledge by Acquaintance vs. Description". Stanford Encyclopedia of Philosophy.
- Knowledge at the Indiana Philosophy Ontology Project
https://en.wikipedia.org/wiki/Knowledge
Abstraction in its main sense is a conceptual process wherein general rules and concepts are derived from the usage and classification of specific examples, literal ("real" or "concrete") signifiers, first principles, or other methods.
"An abstraction" is the outcome of this process—a concept that acts as a common noun for all subordinate concepts and connects any related concepts as a group, field, or category.[1]
Conceptual abstractions may be formed by filtering the information content of a concept or an observable phenomenon, selecting only those aspects which are relevant for a particular purpose. For example, abstracting a leather soccer ball to the more general idea of a ball selects only the information on general ball attributes and behavior, excluding but not eliminating the other phenomenal and cognitive characteristics of that particular ball.[1] In a type–token distinction, a type (e.g., a 'ball') is more abstract than its tokens (e.g., 'that leather soccer ball').
Abstraction in its secondary use is a material process,[2] discussed in the themes below.
https://en.wikipedia.org/wiki/Abstraction
Aestheticism (also the Aesthetic movement) was an art movement in the late 19th century which privileged the aesthetic value of literature, music and the arts over their socio-political functions.[1][2] According to Aestheticism, art should be produced to be beautiful, rather than to serve a moral, allegorical, or other didactic purpose, a sentiment exemplified by the slogan "art for art's sake." Aestheticism originated in 1860s England with a radical group of artists and designers, including William Morris and Dante Gabriel Rossetti. It flourished in the 1870s and 1880s, gaining prominence and the support of notable writers such as Walter Pater and Oscar Wilde.
Aestheticism challenged the values of mainstream Victorian culture, as many Victorians believed that literature and art fulfilled important ethical roles.[3] Writing in The Guardian, Fiona McCarthy states that "the aesthetic movement stood in stark and sometimes shocking contrast to the crass materialism of Britain in the 19th century."[4]
Aestheticism was named by the critic Walter Hamilton in The Aesthetic Movement in England in 1882.[5] By the 1890s, decadence, a term with origins in common with aestheticism, was in use across Europe.[3]
https://en.wikipedia.org/wiki/Aestheticism
The argument from illusion is an argument for the existence of sense-data. It is posed as a criticism of direct realism.
https://en.wikipedia.org/wiki/Argument_from_illusion
The anthropic principle, also known as the "observation selection effect",[1] is the hypothesis, first proposed in 1957 by Robert Dicke, that the range of possible observations that we could make about the universe is limited by the fact that observations could only happen in a universe capable of developing intelligent life in the first place.[2] Proponents of the anthropic principle argue that it explains why this universe has the age and the fundamental physical constants necessary to accommodate conscious life, since if either had been different, we would not have been around to make observations. Anthropic reasoning is often used to deal with the notion that the universe seems to be finely tuned for the existence of life.[3]
There are many different formulations of the anthropic principle. Philosopher Nick Bostrom counts them at thirty, but the underlying principles can be divided into "weak" and "strong" forms, depending on the types of cosmological claims they entail. The weak anthropic principle (WAP), as defined by Brandon Carter, states that the universe's ostensible fine tuning is the result of selection bias (specifically survivorship bias). Most such arguments draw upon some notion of the multiverse for there to be a statistical population of universes to select from. However, a single vast universe is sufficient for most forms of the WAP that do not specifically deal with fine tuning. Carter distinguished the WAP from the strong anthropic principle (SAP), which considers the universe in some sense compelled to eventually have conscious and sapient life emerge within it.[4][5] A form of the latter known as the participatory anthropic principle, articulated by John Archibald Wheeler, suggests on the basis of quantum mechanics that the universe, as a condition of its existence, must be observed, so implying one or more observers. Stronger yet is the final anthropic principle (FAP), proposed by John D. Barrow and Frank Tipler, which views the universe's structure as expressible by bits of information in such a way that information processing is inevitable and eternal.[4]
https://en.wikipedia.org/wiki/Anthropic_principle
The analytic–synthetic distinction is a semantic distinction, used primarily in philosophy to distinguish between propositions (in particular, statements that are affirmative subject–predicate judgments) that are of two types: analytic propositions and synthetic propositions. Analytic propositions are true or not true solely by virtue of their meaning, whereas synthetic propositions' truth, if any, derives from how their meaning relates to the world.[1]
While the distinction was first proposed by Immanuel Kant, it was revised considerably over time, and different philosophers have used the terms in very different ways. Furthermore, some philosophers (starting with W.V.O. Quine) have questioned whether there is even a clear distinction to be made between propositions which are analytically true and propositions which are synthetically true.[2] Debates regarding the nature and usefulness of the distinction continue to this day in contemporary philosophy of language.[2]
Kant
Conceptual containment
The philosopher Immanuel Kant uses the terms "analytic" and "synthetic" to divide propositions into two types. Kant introduces the analytic–synthetic distinction in the Introduction to his Critique of Pure Reason (1781/1998, A6–7/B10–11). There, he restricts his attention to statements that are affirmative subject–predicate judgments and defines "analytic proposition" and "synthetic proposition" as follows:
- analytic proposition: a proposition whose predicate concept is contained in its subject concept
- synthetic proposition: a proposition whose predicate concept is not contained in its subject concept but related
Examples of analytic propositions, on Kant's definition, include:
- "All bachelors are unmarried."
- "All triangles have three sides."
Kant's own example is:
- "All bodies are extended," that is, occupy space. (A7/B11)
Each of these statements is an affirmative subject–predicate judgment, and, in each, the predicate concept is contained within the subject concept. The concept "bachelor" contains the concept "unmarried"; the concept "unmarried" is part of the definition of the concept "bachelor". Likewise, for "triangle" and "has three sides", and so on.
Examples of synthetic propositions, on Kant's definition, include:
- "All bachelors are alone."
- "All creatures with hearts have kidneys."
Kant's own example is:
- "All bodies are heavy," that is, they experience a gravitational force. (A7/B11)
As with the previous examples classified as analytic propositions, each of these new statements is an affirmative subject–predicate judgment. However, in none of these cases does the subject concept contain the predicate concept. The concept "bachelor" does not contain the concept "alone"; "alone" is not a part of the definition of "bachelor". The same is true for "creatures with hearts" and "have kidneys"; even if every creature with a heart also has kidneys, the concept "creature with a heart" does not contain the concept "has kidneys". So the philosophical issue is: What kind of statement is "Language is used to transmit meaning"?
Kant's version and the a priori / a posteriori distinction
In the Introduction to the Critique of Pure Reason, Kant contrasts his distinction between analytic and synthetic propositions with another distinction, the distinction between a priori and a posteriori propositions. He defines these terms as follows:
- a priori proposition: a proposition whose justification does not rely upon experience. Moreover, the proposition can be validated by experience, but is not grounded in experience. Therefore, it is logically necessary.
- a posteriori proposition: a proposition whose justification does rely upon experience. The proposition is validated by, and grounded in, experience. Therefore, it is logically contingent.
Examples of a priori propositions include:
- "All bachelors are unmarried."
- "7 + 5 = 12."
The justification of these propositions does not depend upon experience: one need not consult experience to determine whether all bachelors are unmarried, nor whether 7 + 5 = 12. (Of course, as Kant would grant, experience is required to understand the concepts "bachelor", "unmarried", "7", "+" and so forth. However, the a priori - a posteriori distinction as employed here by Kant refers not to the origins of the concepts but to the justification of the propositions. Once we have the concepts, experience is no longer necessary.)
Examples of a posteriori propositions include:
- "All bachelors are unhappy."
- "Tables exist."
Both of these propositions are a posteriori: any justification of them would require one's experience.
The analytic/synthetic distinction and the a priori - a posteriori distinction together yield four types of propositions:
- analytic a priori
- synthetic a priori
- analytic a posteriori
- synthetic a posteriori
Kant posits the third type as obviously self-contradictory. Ruling it out, he discusses only the remaining three types as components of his epistemological framework—each, for brevity's sake, becoming, respectively, "analytic", "synthetic a priori", and "empirical" or "a posteriori" propositions. This triad accounts for all propositions possible. Examples of analytic and a posteriori statements have already been given, for synthetic a priori propositions he gives those in mathematics and physics.
The ease of knowing analytic propositions
Part of Kant's argument in the Introduction to the Critique of Pure Reason involves arguing that there is no problem figuring out how knowledge of analytic propositions is possible. To know an analytic proposition, Kant argued, one need not consult experience. Instead, one needs merely to take the subject and "extract from it, in accordance with the principle of contradiction, the required predicate" (A7/B12). In analytic propositions, the predicate concept is contained in the subject concept. Thus, to know an analytic proposition is true, one need merely examine the concept of the subject. If one finds the predicate contained in the subject, the judgment is true.
Thus, for example, one need not consult experience to determine whether "All bachelors are unmarried" is true. One need merely examine the subject concept ("bachelors") and see if the predicate concept "unmarried" is contained in it. And in fact, it is: "unmarried" is part of the definition of "bachelor" and so is contained within it. Thus the proposition "All bachelors are unmarried" can be known to be true without consulting experience.
It follows from this, Kant argued, first: All analytic propositions are a priori; there are no a posteriori analytic propositions. It follows, second: There is no problem understanding how we can know analytic propositions; we can know them because we only need to consult our concepts in order to determine that they are true.
The possibility of metaphysics
After ruling out the possibility of analytic a posteriori propositions, and explaining how we can obtain knowledge of analytic a priori propositions, Kant also explains how we can obtain knowledge of synthetic a posteriori propositions. That leaves only the question of how knowledge of synthetic a priori propositions is possible. This question is exceedingly important, Kant maintains, because all scientific knowledge (for him Newtonian physics and mathematics) is made up of synthetic a priori propositions. If it is impossible to determine which synthetic a priori propositions are true, he argues, then metaphysics as a discipline is impossible. The remainder of the Critique of Pure Reason is devoted to examining whether and how knowledge of synthetic a priori propositions is possible.[3]
https://en.wikipedia.org/wiki/Analytic%E2%80%93synthetic_distinction
The analogy of the sun (or simile of the sun or metaphor of the sun) is found in the sixth book of The Republic (507b–509c), written by the Greek philosopher Plato as a dialogue between his brother Glaucon and Socrates, and narrated by the latter. Upon being urged by Glaucon to define goodness, a cautious Socrates professes himself incapable of doing so.[1]: 169 Instead he draws an analogy and offers to talk about "the child of goodness"[1]: 169 (Greek: "ἔκγονός τε τοῦ ἀγαθοῦ"). Socrates reveals this "child of goodness" to be the sun, proposing that just as the sun illuminates, bestowing the ability to see and be seen by the eye,[1]: 169 with its light, so the idea of goodness illumines the intelligible with truth. While the analogy sets forth both epistemological and ontological theories, it is debated whether these are most authentic to the teaching of Socrates or its later interpretations by Plato.
https://en.wikipedia.org/wiki/Analogy_of_the_sun
The analogy of the divided line (Greek: γραμμὴ δίχα τετμημένη, translit. grammē dicha tetmēmenē) is presented by the Greek philosopher Plato in the Republic (509d–511e). It is written as a dialogue between Glaucon and Socrates, in which the latter further elaborates upon the immediately preceding analogy of the sun at the former's request. Socrates asks Glaucon to not only envision this unequally bisected line but to imagine further bisecting each of the two segments. Socrates explains that the four resulting segments represent four separate 'affections' (παθήματα) of the psyche. The lower two sections are said to represent the visible while the higher two are said to represent the intelligible. These affections are described in succession as corresponding to increasing levels of reality and truth from conjecture (εἰκασία) to belief (πίστις) to thought (διάνοια) and finally to understanding (νόησις). Furthermore, this analogy not only elaborates a theory of the psyche but also presents metaphysical and epistemological views.
https://en.wikipedia.org/wiki/Analogy_of_the_divided_line
Ambiguity is the type of meaning in which a phrase, statement or resolution is not explicitly defined, making several interpretations plausible. A common aspect of ambiguity is uncertainty. It is thus an attribute of any idea or statement whose intended meaning cannot be definitively resolved, according to a rule or process with a finite number of steps. (The ambi- part of the term reflects an idea of "two," as in "two meanings.")
The concept of ambiguity is generally contrasted with vagueness. In ambiguity, specific and distinct interpretations are permitted (although some may not be immediately obvious), whereas with information that is vague, it is difficult to form any interpretation at the desired level of specificity.
Linguistic forms
Lexical ambiguity is contrasted with semantic ambiguity. The former represents a choice between a finite number of known and meaningful context-dependent interpretations. The latter represents a choice between any number of possible interpretations, none of which may have a standard agreed-upon meaning. This form of ambiguity is closely related to vagueness.
Ambiguity in human language is argued to reflect principles of efficient communication.[2][3] Languages that communicate efficiently will avoid sending information that is redundant with information provided in the context. This can be shown mathematically to result in a system which is ambiguous when context is neglected. In this way, ambiguity is viewed as a generally useful feature of a linguistic system.
Linguistic ambiguity can be a problem in law, because the interpretation of written documents and oral agreements is often of paramount importance.
Lexical ambiguity
The lexical ambiguity of a word or phrase pertains to its having more than one meaning in the language to which the word belongs.[4] "Meaning" here refers to whatever should be captured by a good dictionary. For instance, the word "bank" has several distinct lexical definitions, including "financial institution" and "edge of a river". Or consider "apothecary". One could say "I bought herbs from the apothecary". This could mean one actually spoke to the apothecary (pharmacist) or went to the apothecary (pharmacy).
The context in which an ambiguous word is used often makes it evident which of the meanings is intended. If, for instance, someone says "I buried $100 in the bank", most people would not think someone used a shovel to dig in the mud. However, some linguistic contexts do not provide sufficient information to disambiguate a used word.
Lexical ambiguity can be addressed by algorithmic methods that automatically associate the appropriate meaning with a word in context, a task referred to as word-sense disambiguation.
The use of multi-defined words requires the author or speaker to clarify their context, and sometimes elaborate on their specific intended meaning (in which case, a less ambiguous term should have been used). The goal of clear concise communication is that the receiver(s) have no misunderstanding about what was meant to be conveyed. An exception to this could include a politician whose "weasel words" and obfuscation are necessary to gain support from multiple constituents with mutually exclusive conflicting desires from their candidate of choice. Ambiguity is a powerful tool of political science.
More problematic are words whose senses express closely related concepts. "Good", for example, can mean "useful" or "functional" (That's a good hammer), "exemplary" (She's a good student), "pleasing" (This is good soup), "moral" (a good person versus the lesson to be learned from a story), "righteous", etc. "I have a good daughter" is not clear about which sense is intended. The various ways to apply prefixes and suffixes can also create ambiguity ("unlockable" can mean "capable of being unlocked" or "impossible to lock").
Semantic and syntactic ambiguity
Semantic ambiguity occurs when a word, phrase or sentence, taken out of context, has more than one interpretation. In "We saw her duck" (example due to Richard Nordquist), the words "her duck" can refer either
- to the person's bird (the noun "duck", modified by the possessive pronoun "her"), or
- to a motion she made (the verb "duck", the subject of which is the objective pronoun "her", object of the verb "saw").[5]
Syntactic ambiguity arises when a sentence can have two (or more) different meanings because of the structure of the sentence—its syntax. This is often due to a modifying expression, such as a prepositional phrase, the application of which is unclear. "He ate the cookies on the couch", for example, could mean that he ate those cookies that were on the couch (as opposed to those that were on the table), or it could mean that he was sitting on the couch when he ate the cookies. "To get in, you will need an entrance fee of $10 or your voucher and your drivers' license." This could mean that you need EITHER ten dollars OR BOTH your voucher and your license. Or it could mean that you need your license AND you need EITHER ten dollars OR a voucher. Only rewriting the sentence, or placing appropriate punctuation can resolve a syntactic ambiguity.[5] For the notion of, and theoretic results about, syntactic ambiguity in artificial, formal languages (such as computer programming languages), see Ambiguous grammar.
Usually, semantic and syntactic ambiguity go hand in hand. The sentence "We saw her duck" is also syntactically ambiguous. Conversely, a sentence like "He ate the cookies on the couch" is also semantically ambiguous. Rarely, but occasionally, the different parsings of a syntactically ambiguous phrase result in the same meaning. For example, the command "Cook, cook!" can be parsed as "Cook (noun used as vocative), cook (imperative verb form)!", but also as "Cook (imperative verb form), cook (noun used as vocative)!". It is more common that a syntactically unambiguous phrase has a semantic ambiguity; for example, the lexical ambiguity in "Your boss is a funny man" is purely semantic, leading to the response "Funny ha-ha or funny peculiar?"
Spoken language can contain many more types of ambiguities which are called phonological ambiguities, where there is more than one way to compose a set of sounds into words. For example, "ice cream" and "I scream". Such ambiguity is generally resolved according to the context. A mishearing of such, based on incorrectly resolved ambiguity, is called a mondegreen.
https://en.wikipedia.org/wiki/Ambiguity
Always-already is a philosophical term regarding the perception of phenomena by the mind of an observer. The features of a phenomenon that seem to precede any perception of it are said to be "always already" present.
https://en.wikipedia.org/wiki/Always_already
The Allegory of the Cave, or Plato's Cave, is an allegory presented by the Greek philosopher Plato in his work Republic (514a–520a) to compare "the effect of education (παιδεία) and the lack of it on our nature". It is written as a dialogue between Plato's brother Glaucon and his mentor Socrates, narrated by the latter. The allegory is presented after the analogy of the sun (508b–509c) and the analogy of the divided line (509d–511e).
In the allegory "The Cave", Plato describes a group of people who have lived chained to the wall of a cave all their lives, facing a blank wall. The people watch shadows projected on the wall from objects passing in front of a fire behind them and give names to these shadows. The shadows are the prisoners' reality, but are not accurate representations of the real world. The shadows represent the fragment of reality that we can normally perceive through our senses, while the objects under the sun represent the true forms of objects that we can only perceive through reason. Three higher levels exist: the natural sciences; mathematics, geometry, and deductive logic; and the theory of forms.
Socrates explains how the philosopher is like a prisoner who is freed from the cave and comes to understand that the shadows on the wall are actually not the direct source of the images seen. A philosopher aims to understand and perceive the higher levels of reality. However, the other inmates of the cave do not even desire to leave their prison, for they know no better life.[1]
Socrates remarks that this allegory can be paired with previous writings, namely the analogy of the sun and the analogy of the divided line.
Summary
Imprisonment in the cave
Plato begins by having Socrates ask Glaucon to imagine a cave where people have been imprisoned from childhood, but not from birth. These prisoners are chained so that their legs and necks are fixed, forcing them to gaze at the wall in front of them and not to look around at the cave, each other, or themselves (514a–b).[2] Behind the prisoners is a fire, and between the fire and the prisoners is a raised walkway with a low wall, behind which people walk carrying objects or puppets "of men and other living things" (514b).[2]
The people walk behind the wall so their bodies do not cast shadows for the prisoners to see, but the objects they carry do ("just as puppet showmen have screens in front of them at which they work their puppets") (514a).[2] The prisoners cannot see any of what is happening behind them; they are only able to see the shadows cast upon the cave wall in front of them. The sounds of the people talking echo off the walls; the prisoners believe these sounds come from the shadows (514c).[2]
Socrates suggests that the shadows are reality for the prisoners because they have never seen anything else; they do not realize that what they see are shadows of objects in front of a fire, much less that these objects are inspired by real things outside the cave which they do not see (514b–515a).[2]
Departure from the cave
Socrates then supposes that the prisoners are released.[3]: 199 A freed prisoner would look around and see the fire. The light would hurt his eyes and make it difficult for him to see the objects casting the shadows. If he were told that what he is seeing is real instead of the other version of reality he sees on the wall, he would not believe it. In his pain, Socrates continues, the freed prisoner would turn away and run back to what he is accustomed to (that is, the shadows of the carried objects). The light "... would hurt his eyes, and he would escape by turning away to the things which he was able to look at, and these he would believe to be clearer than what was being shown to him."[2]
Socrates continues: "Suppose... that someone should drag him... by force, up the rough ascent, the steep way up, and never stop until he could drag him out into the light of the sun."[2] The prisoner would be angry and in pain, and this would only worsen when the radiant light of the sun overwhelms his eyes and blinds him.[2]
"Slowly, his eyes adjust to the light of the sun. First he can see only shadows. Gradually he can see the reflections of people and things in water and then later see the people and things themselves. Eventually, he is able to look at the stars and moon at night until finally he can look upon the sun itself (516a)."[2] Only after he can look straight at the sun "is he able to reason about it" and what it is (516b).[2] (See also Plato's analogy of the sun, which occurs near the end of The Republic, Book VI.)[4][5]
Return to the cave
Socrates continues, saying that the free prisoner would think that the world outside the cave was superior to the world he experienced in the cave and attempt to share this with the prisoners remaining in the cave attempting to bring them onto the journey he had just endured; "he would bless himself for the change, and pity [the other prisoners]" and would want to bring his fellow cave dwellers out of the cave and into the sunlight (516c).[2]
The returning prisoner, whose eyes have become accustomed to the sunlight, would be blind when he re-entered the cave, just as he was when he was first exposed to the sun (516e).[2] The prisoners who remained, according to the dialogue, would infer from the returning man's blindness that the journey out of the cave had harmed him and that they should not undertake a similar journey. Socrates concludes that the prisoners, if they were able, would therefore reach out and kill anyone who attempted to drag them out of the cave (517a).[2]
Themes in the allegory appearing elsewhere in Plato's work
The allegory is related to Plato's theory of Forms, according to which the "Forms" (or "Ideas"), and not the material world known to us through sensation, possess the highest and most fundamental kind of reality. Knowledge of the Forms constitutes real knowledge or what Socrates considers "the Good".[6] Socrates informs Glaucon that the most excellent people must follow the highest of all studies, which is to behold the Good. Those who have ascended to this highest level, however, must not remain there but must return to the cave and dwell with the prisoners, sharing in their labors and honors.
Plato's Phaedo contains similar imagery to that of the allegory of the cave; a philosopher recognizes that before philosophy, his soul was "a veritable prisoner fast bound within his body... and that instead of investigating reality of itself and in itself is compelled to peer through the bars of a prison."[7]
Scholarly discussion
Scholars debate the possible interpretations of the allegory of the cave, either looking at it from an epistemological standpoint—one based on the study of how Plato believes we come to know things—or through a political (politeia) lens.[8] Much of the scholarship on the allegory falls between these two perspectives, with some completely independent of either. The epistemological view and the political view, fathered by Richard Lewis Nettleship and A. S. Ferguson, respectively, tend to be discussed most frequently.[8]
Nettleship interprets the allegory of the cave as representative of our innate intellectual incapacity, in order to contrast our lesser understanding with that of the philosopher, as well as an allegory about people who are unable or unwilling to seek truth and wisdom.[9][8] Ferguson, on the other hand, bases his interpretation of the allegory on the claim that the cave is an allegory of human nature and that it symbolizes the opposition between the philosopher and the corruption of the prevailing political condition.[1]
Cleavages have emerged within these respective camps of thought, however. Much of the modern scholarly debate surrounding the allegory has emerged from Martin Heidegger's exploration of the allegory, and philosophy as a whole, through the lens of human freedom in his book The Essence of Human Freedom: An Introduction to Philosophy and The Essence of Truth: On Plato's Cave Allegory and Theaetetus.[10] In response, Hannah Arendt, an advocate of the political interpretation of the allegory, suggests that through the allegory, Plato "wanted to apply his own theory of ideas to politics".[11] Conversely, Heidegger argues that the essence of truth is a way of being and not an object.[12] Arendt criticised Heidegger's interpretation of the allegory, writing that "Heidegger ... is off base in using the cave simile to interpret and 'criticize' Plato's theory of ideas".[11]
Various scholars also debate the possibility of a connection between the work in the allegory and the cave and the work done by Plato considering the analogy of the divided line and the analogy of the sun. The divided line is a theory presented to us in Plato's work the Republic. This is displayed through a dialogue given between Socrates and Glaucon in which they explore the possibility of a visible and intelligible world, with the visible world consisting of items such as shadows and reflections (displayed as AB) then elevating to the physical item itself (displayed as BC) while the intelligible world consists of mathematical reasoning (displayed by CD) and philosophical understanding (displayed by DE).[3]
Many see this as an explanation for the way in which the prisoner in the allegory of the cave goes through the journey, first in the visible world with shadows such as those on the wall,[3] then the realization of the physical with the understanding of concepts such as the tree being separate from its shadow. It enters the intelligible world as the prisoner looks at the sun.[13]
The Analogy of the Sun refers to the moment in book six in which Socrates, after being urged by Glaucon to define goodness, proposes instead an analogy through a "child of goodness". Socrates reveals this "child of goodness" to be the sun, proposing that just as the sun illuminates, bestowing the ability to see and be seen by the eye[15]: 169 with its light, so the idea of goodness illumines the intelligible with truth, leading some scholars to believe this forms a connection of the sun and the intelligible world within the realm of the allegory of the cave.
Influence
The themes and imagery of Plato's cave have appeared throughout Western thought and culture. Some examples include:
- Francis Bacon used the term "Idols of the Cave" to refer to errors of reason arising from the idiosyncratic biases and preoccupations of individuals.
- Thomas Browne in his 1658 discourse Urn Burial stated: "A Dialogue between two Infants in the womb concerning the state of this world, might handsomely illustrate our ignorance of the next, whereof methinks we yet discourse in Platoes denne, and are but Embryon Philosophers".
- Evolutionary biologist Jeremy Griffith's book A Species In Denial includes the chapter "Deciphering Plato's Cave Allegory".[16]
- The films The Conformist, The Matrix, Cube, Dark City, The Truman Show, Us and City of Ember model Plato's allegory of the cave, as does the TV series 1899.[17]
- The 2013 movie After the Dark has a segment where Mr. Zimit likens James' life to the Allegory of the Cave.
- The Cave by José Saramago culminates in the discovery of Plato's Cave underneath the center, "an immense complex fusing the functions of an office tower, a shopping mall and a condominium".[18]
- Emma Donoghue acknowledges the influence of Plato's allegory of the cave on her novel Room.[19]
- Ray Bradbury's novel Fahrenheit 451 explores the themes of reality and perception also explored in Plato's allegory of the cave and Bradbury references Plato's work in the novel.[20][21]
- José Carlos Somoza's novel The Athenian Murders is presented as a murder mystery but features many references to Plato's philosophy including the allegory of the cave.[22]
- Novelist James Reich argues Nicholas Ray's film Rebel Without a Cause, starring James Dean, Natalie Wood, and Sal Mineo as John "Plato" Crawford is influenced by and enacts aspects of the allegory of the cave.[23]
- In an episode of the television show Legion, titled "Chapter 16", the narrator uses Plato's Cave to explain "the most alarming delusion of all", narcissism.
- H. G. Wells' short novel The Country of the Blind has a similar "Return to the Cave" situation when a man accidentally discovers a village of blind people and wherein he tries to explain how he can "see", only to be ridiculed.
- Daniel F. Galouye's post-apocalyptic novel Dark Universe describes the Survivors, who live underground in total darkness, using echolocation to navigate. Another race of people evolve, who are able to see using infrared.
- C. S. Lewis' novels The Silver Chair and The Last Battle both reference the ideas and imagery of the Cave. In the former in Chapter 12, the Witch dismisses the idea of a greater reality outside the bounds of her Underworld. In The Last Battle most of the characters learn that the Narnia which they have known is but a "shadow" of the true Narnia. Lord Digory says in Chapter 15, "It's all in Plato, all in Plato".
- In season 1, episode 2 of the 2015 Catalan television series Merlí, titled "Plato", a high school philosophy teacher demonstrates the allegory using household objects for a non-verbal, agoraphobic student, and makes a promise to him that "I'll get you out of the cave".
- In the 2016 season 1, episode 1 of The Path, titled "What the Fire Throws", a cult leader uses the allegory in a sermon to inspire the members to follow him "up out of the world of shadows ... into the light".
See also
- Allegorical interpretations of Plato
- Anekantavada
- Archetype
- Brain in a vat
- Experience machine
- Flatland
- The Form of the Good
- Intelligibility (philosophy)
- Nous – Noumenon
- Phaneron
- Plato's Republic in popular culture
- Simulation hypothesis
- Holographic principle
- Blind men and an elephant, a rough equivalent in Eastern Philosophy
- Maya (illusion)
References
- "Plato's Cave: Rebel Without a Cause and Platonic Allegory – OUTSIDER ACADEMY". Retrieved 2017-06-25.[permanent dead link]
Further reading
The following is a list of supplementary scholarly literature on the allegory of the cave that includes articles from epistemological, political, alternative, and independent viewpoints on the allegory:
- Kim, A. (2004). "Shades of Truth: Phenomenological Perspectives on the Allegory of the Cave". Idealistic Studies. 34 (1): 1–24. doi:10.5840/idstudies200434118. INIST:16811501.
- Zamosc, Gabriel (2017). "The Political Significance of Plato's Allegory of the Cave". Ideas y Valores. 66 (165). doi:10.15446/ideasyvalores.v66n165.55201. ProQuest 1994433580.
- Mitta, Dimitra (1 January 2003). "Reading Platonic Myths from a Ritualistic Point of View: Gyges' Ring and the Cave Allegory". Kernos (16): 133–141. doi:10.4000/kernos.815.
- McNeill, William (5 January 2003). "The Essence of Human Freedom: An Introduction to Philosophy and The Essence of Truth: On Plato's Cave Allegory and Theaetetus". Notre Dame Philosophical Reviews.
- Eckert, Maureen (2012). "Cinematic Spelunking Inside Plato's Cave". Glipmse Journal. 9: 42–49.
- Tsabar, Boaz (1 January 2014). "'Poverty and Resourcefulness': On the Formative Significance of Eros in Educational Practice". Studies in Philosophy and Education. 33 (1): 75–87. doi:10.1007/s11217-013-9364-5. S2CID 144408538.
- Malcolm, J. (May 1981). "The Cave Revisited". The Classical Quarterly. 31 (1): 60–68. doi:10.1017/S0009838800021078. S2CID 170697508.
- Murphy, N. R. (April 1932). "The 'Simile Of Light' In Plato's Republic". The Classical Quarterly. 26 (2): 93–102. doi:10.1017/S0009838800002366. S2CID 170223655.
External links
- Allegory of the cave at PhilPapers
- Animated interpretation of Plato's Allegory of the Cave
- Ted-ed: Plato's Allegory of the Cave
- Plato: The Republic, Book VII at Project Gutenberg
- Plato: The Allegory of the Cave, from The Republic at University of Washington – Faculty
- Plato: Book VII of The Republic, Allegory of the Cave at Shippensburg University
https://en.wikipedia.org/wiki/Allegory_of_the_cave
In philosophy and psychology, an alief is an automatic or habitual belief-like attitude, particularly one that is in tension with a person's explicit beliefs.[1]
For example, a person standing on a transparent balcony may believe that they are safe, but alieve that they are in danger. A person watching a sad movie may believe that the characters are completely fictional, but their aliefs may lead them to cry nonetheless. A person who is hesitant to eat fudge that has been formed into the shape of feces, or who exhibits reluctance in drinking from a sterilized bedpan may believe that the substances are safe to eat and drink, but may alieve that they are not.
The term alief was introduced by Tamar Gendler, a professor of philosophy and cognitive science at Yale University, in a pair of influential articles published in 2008.[2] Since the publication of these original articles, the notion of alief has been utilized by Gendler and others — including Paul Bloom[3] and Daniel Dennett[4] — to explain a range of psychological phenomena in addition to those listed above, including the pleasure of stories,[3] the persistence of positive illusions,[4] certain religious beliefs,[5] and certain psychiatric disturbances, such as phobias and obsessive–compulsive disorder.[4]
https://en.wikipedia.org/wiki/Alief_(mental_state)
The basic premise of the concept of mundane reason is that the standard assumptions about reality that people typically make as they go about day to day, including the very fact that they experience their reality as perfectly natural, are actually the result of social, cultural, and historical processes that make a particular perception of the world readily available. It is the reasoning about the world, self, and others which presupposes the world and its relationship to the observer; according to Steven Shapin (Shapin 1994:31), it is a set of presuppositions about the subject, the object, and the nature of their relations.[1]
https://en.wikipedia.org/wiki/Mundane_reason
Moral rationalism, also called ethical rationalism, is a view in meta-ethics (specifically the epistemology of ethics) according to which moral principles are knowable a priori, by reason alone.[1] Some prominent figures in the history of philosophy who have defended moral rationalism are Plato and Immanuel Kant. Perhaps the most prominent figure in the history of philosophy who has rejected moral rationalism is David Hume. Recent philosophers who have defended moral rationalism include Richard Hare, Christine Korsgaard, Alan Gewirth, and Michael Smith.
Moral rationalism is similar to the rationalist version of ethical intuitionism; however, they are distinct views. Moral rationalism is neutral on whether basic moral beliefs are known via inference or not. A moral rationalist who believes that some moral beliefs are justified non-inferentially is a rationalist ethical intuitionist. So, rationalist ethical intuitionism implies moral rationalism, but the reverse does not hold.
https://en.wikipedia.org/wiki/Moral_rationalism
In semantics, semiotics, philosophy of language, metaphysics, and metasemantics, meaning "is a relationship between two sorts of things: signs and the kinds of things they intend, express, or signify".[1]
The types of meanings vary according to the types of the thing that is being represented. Namely:
- There are the things, which might have meaning;
- There are things that are also signs of other things, and so, are always meaningful (i.e., natural signs of the physical world and ideas within the mind);
- There are things that are necessarily meaningful such as words and nonverbal symbols.
The major contemporary positions of meaning come under the following partial definitions of meaning:
- Psychological theories, involving notions of thought, intention, or understanding;
- Logical theories, involving notions such as intension, cognitive content, or sense, along with extension, reference, or denotation;
- Message, content, information, or communication;
- Truth conditions;
- Usage, and the instructions for usage; and
- Measurement, computation, or operation.
https://en.wikipedia.org/wiki/Meaning_(philosophy)
Metafiction is a form of fiction that emphasises its own narrative structure in a way that continually reminds the audience that they are reading or viewing a fictional work. Metafiction is self-conscious about language, literary form, and story-telling, and works of metafiction directly or indirectly draw attention to their status as artifacts.[1] Metafiction is frequently used as a form of parody or a tool to undermine literary conventions and explore the relationship between literature and reality, life, and art.[2]
Although metafiction is most commonly associated with postmodern literature that developed in the mid-20th century, its use can be traced back to much earlier works of fiction, such as The Canterbury Tales (Geoffrey Chaucer, 1387), Don Quixote (Miguel de Cervantes, 1605), "Chymical Wedding of Christian Rosenkreutz" (Johann Valentin Andreae, 1617) The Life and Opinions of Tristram Shandy, Gentleman (Laurence Sterne, 1759), Sartor Resartus (Thomas Carlyle, 1833–34),[3] and Vanity Fair (William Makepeace Thackeray, 1847).
Metafiction became particularly prominent in the 1960s, with works such as Lost in the Funhouse by John Barth, Pale Fire by Vladimir Nabokov, "The Babysitter" and "The Magic Poker" by Robert Coover, Slaughterhouse-Five by Kurt Vonnegut,[4] The French Lieutenant's Woman by John Fowles, The Crying of Lot 49 by Thomas Pynchon, and Willie Master's Lonesome Wife by William H. Gass.
Since the 1980s, contemporary Latino literature has an abundance of self-reflexive, metafictional works, including novels and short stories by Junot Díaz (The Brief Wondrous Life of Oscar Wao),[5] Sandra Cisneros (Caramelo),[6] Salvador Plascencia (The People of Paper),[7] Carmen Maria Machado (Her Body),[8] Rita Indiana (Tentacle),[9] and Valeria Luiselli (Lost Children Archive).[7]
https://en.wikipedia.org/wiki/Metafiction
Mimesis (/mɪˈmiːsɪs, mə-, maɪ-, -əs/;[1] Ancient Greek: μίμησις, mīmēsis) is a term used in literary criticism and philosophy that carries a wide range of meanings, including imitatio, imitation, nonsensuous similarity, receptivity, representation, mimicry, the act of expression, the act of resembling, and the presentation of the self.[2]
The original Ancient Greek term mīmēsis (μίμησις) derives from mīmeisthai (μιμεῖσθαι, 'to imitate'), itself coming from mimos (μῖμος, 'imitator, actor'). In ancient Greece, mīmēsis was an idea that governed the creation of works of art, in particular, with correspondence to the physical world understood as a model for beauty, truth, and the good. Plato contrasted mimesis, or imitation, with diegesis, or narrative. After Plato, the meaning of mimesis eventually shifted toward a specifically literary function in ancient Greek society.[3]
One of the best-known modern studies of mimesis—understood in literature as a form of realism—is Erich Auerbach's Mimesis: The Representation of Reality in Western Literature, which opens with a comparison between the way the world is represented in Homer's Odyssey and the way it appears in the Bible.[4]
In addition to Plato and Auerbach, mimesis has been theorised by thinkers as diverse as Aristotle,[5] Philip Sidney, Samuel Taylor Coleridge, Adam Smith, Gabriel Tarde, Sigmund Freud, Walter Benjamin,[6] Theodor Adorno,[7] Paul Ricœur, Luce Irigaray, Jacques Derrida, René Girard, Nikolas Kompridis, Philippe Lacoue-Labarthe, Michael Taussig,[8] Merlin Donald, Homi Bhabha, Roberto Calasso, and Nidesh Lawtoo. During the nineteenth century, the racial politics of imitation towards African Americans influenced the term mimesis and its evolution.[9]
https://en.wikipedia.org/wiki/Mimesis
The burden of proof (Latin: onus probandi, shortened from Onus probandi incumbit ei qui dicit, non ei qui negat) is the obligation on a party in a dispute to provide sufficient warrant for its position.
https://en.wikipedia.org/wiki/Burden_of_proof_(philosophy)
In contemporary philosophy, a brute fact is a fact that cannot be explained in terms of a deeper, more "fundamental" fact.[1] There are two main ways to explain something: say what "brought it about", or describe it at a more "fundamental" level.[citation needed] For example, a cat displayed on a computer screen can be explained, more "fundamentally", in terms of certain voltages in bits of metal in the screen, which in turn can be explained, more "fundamentally", in terms of certain subatomic particles moving in a certain manner. If one were to keep explaining the world in this way and reach a point at which no more "deeper" explanations can be given, then one would have found some facts which are brute or inexplicable, in the sense that we cannot give them an ontological explanation. As it might be put, there may exist some things that just are.
To reject the existence of brute facts is to think that everything can be explained ("Everything can be explained" is sometimes called the principle of sufficient reason).
https://en.wikipedia.org/wiki/Brute_fact
The principle of sufficient reason states that everything must have a reason or a cause. The principle was articulated and made prominent by Gottfried Wilhelm Leibniz, with many antecedents, and was further used and developed by Arthur Schopenhauer and Sir William Hamilton, 9th Baronet.
https://en.wikipedia.org/wiki/Principle_of_sufficient_reason
In metaphysics, ontology is the philosophical study of being, as well as related concepts such as existence, becoming, and reality.
Ontology addresses questions like how entities are grouped into categories and which of these entities exist on the most fundamental level. Ontologists often try to determine what the categories or highest kinds are and how they form a system of categories that encompasses the classification of all entities. Commonly proposed categories include substances, properties, relations, states of affairs, and events. These categories are characterized by fundamental ontological concepts, including particularity and universality, abstractness and concreteness, or possibility and necessity. Of special interest is the concept of ontological dependence, which determines whether the entities of a category exist on the most fundamental level. Disagreements within ontology are often about whether entities belonging to a certain category exist and, if so, how they are related to other entities.[1]
When used as a countable noun, the words ontology and ontologies refer not to the science of being but to theories within the science of being. Ontological theories can be divided into various types according to their theoretical commitments. Monocategorical ontologies hold that there is only one basic category, but polycategorical ontologies rejected this view. Hierarchical ontologies assert that some entities exist on a more fundamental level and that other entities depend on them. Flat ontologies, on the other hand, deny such a privileged status to any entity.
Etymology
The compound word ontology ('study of being') combines
- onto- (Greek: ὄν, on;[note 1] GEN. ὄντος, ontos, 'being' or 'that which is') and
- -logia (-λογία, 'logical discourse').[2][3]
While the etymology is Greek, the oldest extant records of the word itself is a Neo-Latin form ontologia, which appeared
- in 1606 in the Ogdoas Scholastica by Jacob Lorhard (Lorhardus), and
- in 1613 in the Lexicon philosophicum by Rudolf Göckel (Goclenius).
The first occurrence in English of ontology, as recorded by the Oxford English Dictionary,[4] came in 1664 through Archelogia philosophica nova... by Gideon Harvey.[5] The word was first used, in its Latin form, by philosophers, and based on the Latin roots (and in turn on the Greek ones).
https://en.wikipedia.org/wiki/Ontology
A binary opposition (also binary system) is a pair of related terms or concepts that are opposite in meaning. Binary opposition is the system of language and/or thought by which two theoretical opposites are strictly defined and set off against one another.[1] It is the contrast between two mutually exclusive terms, such as on and off, up and down, left and right.[2] Binary opposition is an important concept of structuralism, which sees such distinctions as fundamental to all language and thought.[2] In structuralism, a binary opposition is seen as a fundamental organizer of human philosophy, culture, and language.
Binary opposition originated in Saussurean structuralist theory.[3] According to Ferdinand de Saussure, the binary opposition is the means by which the units of language have value or meaning; each unit is defined in reciprocal determination with another term, as in binary code. It is not a contradictory relation but a structural, complementary one.[3] Saussure demonstrated that a sign's meaning is derived from its context (syntagmatic dimension) and the group (paradigm) to which it belongs.[4] An example of this is that one cannot conceive of 'good' if we do not understand 'evil'.[5]
Typically, one of the two opposites assumes a role of dominance over the other. The categorization of binary oppositions is "often value-laden and ethnocentric", with an illusory order and superficial meaning.[6] Furthermore, Pieter Fourie discovered that binary oppositions have a deeper or second level of binaries that help to reinforce meaning. As an example, the concepts hero and villain involve secondary binaries: good/bad, handsome/ugly, liked/disliked, and so on.[7]
https://en.wikipedia.org/wiki/Binary_opposition
In sociology, anthropology, archaeology, history, philosophy, and linguistics, structuralism is a general theory of culture and methodology that implies that elements of human culture must be understood by way of their relationship to a broader system.[1] It works to uncover the structures that underlie all the things that humans do, think, perceive, and feel.
Alternatively, as summarized by philosopher Simon Blackburn, structuralism is:[2]
[T]he belief that phenomena of human life are not intelligible except through their interrelations. These relations constitute a structure, and behind local variations in the surface phenomena there are constant laws of abstract structure.
Structuralism in Europe developed in the early 20th century, mainly in France and the Russian Empire, in the structural linguistics of Ferdinand de Saussure and the subsequent Prague,[3] Moscow,[3] and Copenhagen schools of linguistics. As an intellectual movement, structuralism became the heir to existentialism.[4] After World War II, an array of scholars in the humanities borrowed Saussure's concepts for use in their respective fields. French anthropologist Claude Lévi-Strauss was arguably the first such scholar, sparking a widespread interest in structuralism.[2]
The structuralist mode of reasoning has since been applied in a range of fields, including anthropology, sociology, psychology, literary criticism, economics, and architecture. Along with Lévi-Strauss, the most prominent thinkers associated with structuralism include linguist Roman Jakobson and psychoanalyst Jacques Lacan.
By the late 1960s, many of structuralism's basic tenets came under attack from a new wave of predominantly French intellectuals/philosophers such as historian Michel Foucault, Jacques Derrida, Marxist philosopher Louis Althusser, and literary critic Roland Barthes.[3] Though elements of their work necessarily relate to structuralism and are informed by it, these theorists eventually came to be referred to as post-structuralists. Many proponents of structuralism, such as Lacan, continue to influence continental philosophy and many of the fundamental assumptions of some of structuralism's post-structuralist critics are a continuation of structuralist thinking.[5]
https://en.wikipedia.org/wiki/Structuralism
A Basic Limiting Principle (B.L.P.) is a general principle that limits our explanations metaphysically or epistemologically, and which normally goes unquestioned or even unnoticed in our everyday or scientific thinking. The term was introduced by the philosopher C. D. Broad in his 1949 paper "The Relevance of Psychical research to Philosophy":
"There are certain limiting principles which we unhesitatingly take for granted as the framework within which all our practical activities and our scientific theories are confined. Some of these seem to be self-evident. Others are so overwhelmingly supported by all the empirical facts which fall within the range of ordinary experience and the scientific elaborations of it (including under this heading orthodox psychology) that it hardly enters our heads to question them. Let us call these Basic Limiting Principles."[1]
Broad offers nine examples of B.L.P.s, including the principle that there can be no backward causation, that there can be no action at a distance, and that one cannot perceive physical events or material things directly, unmediated by sensations.
Notes
- Broad (1949)
References
- C. D. Broad. 1949: "The Relevance of Psychical Research to Philosophy" (Philosophy 24, pp 291–309)
- Metaphysical principles
- Concepts in epistemology
- Razors (philosophy)
- Philosophy of science
- Metaphysics stubs
- Epistemology stubs
- Philosophy of science stubs
https://en.wikipedia.org/wiki/Basic_limiting_principle
Basic beliefs (also commonly called foundational beliefs or core beliefs) are, under the epistemological view called foundationalism, the axioms of a belief system.[example needed]
Categories of beliefs
Foundationalism holds that all beliefs must be justified in order to be known. Beliefs therefore fall into two categories:
- Beliefs that are properly basic, in that they do not depend upon justification of other beliefs, but on something outside the realm of belief (a "non-doxastic justification")
- Beliefs that derive from one or more basic beliefs, and therefore depend on the basic beliefs for their validity
Description
Within this basic framework of foundationalism exist a number of views regarding which types of beliefs qualify as properly basic; that is, what sorts of beliefs can be justifiably held without the justification of other beliefs.
In classical foundationalism, beliefs are held to be properly basic if they are either self-evident axioms, or evident to the senses (empiricism).[1] However Anthony Kenny and others have argued that this is a self-refuting idea.[2]
- In modern foundationalism, beliefs are held to be properly basic if they were either self-evident axiom or incorrigible.[3] One such axiom is René Descartes's axiom, Cogito ergo sum ("I think, therefore I am"). Incorrigible (lit. uncorrectable) beliefs are those one can believe without possibly being proven wrong. Notably, the evidence of the senses is not seen as properly basic because, Descartes argued, all our sensory experience could be an illusion.
- In what Keith Lehrer has called "fallible foundationalism",[4] also known as "moderate foundationalism", the division between inferential and non-inferential belief is retained, but the requirement of incorrigibility is dropped. This, it is claimed, allows the senses to resume their traditional role as the basis of non-inferential belief despite their fallibility.[5]
- In reformed epistemology, beliefs are held to be properly basic if they are reasonable and consistent with a sensible world view.
Anti-foundationalism rejects foundationalism and denies there is some fundamental belief or principle which is the basic ground or foundation of inquiry and knowledge.[6]
See also
- Doxastic logic
- First principle
- Mental model
- Mental representation
- Mindset
- Paradigm
- Set (psychology)
- Schema (psychology)
- Worldview
Notes and references
- J. Childers/G. Hentzi, The Columbia Dictionary of Modern Literary and Cultural Criticism (1995) p. 100
| |||
| |||
https://en.wikipedia.org/wiki/Basic_belief
Doxastic logic is a type of logic concerned with reasoning about beliefs.
The term doxastic derives from the Ancient Greek δόξα (doxa, "opinion, belief"), from which the English term doxa ("popular opinion or belief") is also borrowed. Typically, a doxastic logic uses the notation to mean "It is believed that is the case", and the set denotes a set of beliefs. In doxastic logic, belief is treated as a modal operator.
There is complete parallelism between a person who believes propositions and a formal system that derives propositions. Using doxastic logic, one can express the epistemic counterpart of Gödel's incompleteness theorem of metalogic, as well as Löb's theorem, and other metalogical results in terms of belief.[1]
https://en.wikipedia.org/wiki/Doxastic_logic
Nomothetic and idiographic are terms used by Neo-Kantian philosopher Wilhelm Windelband to describe two distinct approaches to knowledge, each one corresponding to a different intellectual tendency, and each one corresponding to a different branch of academia. To say that Windelband supported that last dichotomy is a consequent misunderstanding of his own thought. For him, any branch of science and any discipline can be handled by both methods as they offer two integrating points of view.[1]
- Nomothetic is based on what Kant described as a tendency to generalize, and is typical for the natural sciences. It describes the effort to derive laws that explain types or categories of objective phenomena, in general.
- Idiographic is based on what Kant described as a tendency to specify, and is typical for the humanities. It describes the effort to understand the meaning of contingent, unique, and often cultural or subjective phenomena.
https://en.wikipedia.org/wiki/Nomothetic_and_idiographic
A closed concept is a concept where all the necessary and sufficient conditions required to include something within the concept can be listed. For example, the concept of a triangle is closed because it is a three-sided polygon, and only a three-sided polygon, is a triangle. All the conditions required to call something a triangle can be, and are, are listed.
Its opposite is an "open concept".[1]
https://en.wikipedia.org/wiki/Closed_concept
In logic and mathematics, necessity and sufficiency are terms used to describe a conditional or implicational relationship between two statements. For example, in the conditional statement: "If P then Q", Q is necessary for P, because the truth of Q is guaranteed by the truth of P (equivalently, it is impossible to have P without Q).[1] Similarly, P is sufficient for Q, because P being true always implies that Q is true, but P not being true does not always imply that Q is not true.[2]
In general, a necessary condition is one that must be present in order for another condition to occur, while a sufficient condition is one that produces the said condition.[3] The assertion that a statement is a "necessary and sufficient" condition of another means that the former statement is true if and only if the latter is true. That is, the two statements must be either simultaneously true, or simultaneously false.[4][5][6]
In ordinary English (also natural language) "necessary" and "sufficient" indicate relations between conditions or states of affairs, not statements. For example, being a male is a necessary condition for being a brother, but it is not sufficient—while being a male sibling is a necessary and sufficient condition for being a brother. Any conditional statement consists of at least one sufficient condition and at least one necessary condition.
https://en.wikipedia.org/wiki/Necessity_and_sufficiency
Continuum fallacy
The continuum fallacy (also known as the fallacy of the beard,[9][10] line-drawing fallacy, or decision-point fallacy[11]) is an informal fallacy related to the sorites paradox. Both fallacies cause one to erroneously reject a vague claim simply because it is not as precise as one would like it to be. Vagueness alone does not necessarily imply invalidity. The fallacy is the argument that two states or conditions cannot be considered distinct (or do not exist at all) because between them there exists a continuum of states.
Strictly, the sorites paradox refers to situations where there are many discrete states (classically between 1 and 1,000,000 grains of sand, hence 1,000,000 possible states), while the continuum fallacy refers to situations where there is (or appears to be) a continuum of states, such as temperature. Whether any continua exist in the physical world is the classic question of atomism, and while both Newtonian physics and quantum physics model the world as continuous, there are some proposals in quantum gravity, such as loop quantum gravity, that suggest that notions of continuous length do not apply at the Planck length, and thus what appear to be continua may simply be as-yet undistinguishable discrete states.
For the purpose of the continuum fallacy, one assumes that there is in fact a continuum, though this is generally a minor distinction: in general, any argument against the sorites paradox can also be used against the continuum fallacy. One argument against the fallacy is based on the simple counterexample: there do exist bald people and people who are not bald. Another argument is that for each degree of change in states, the degree of the condition changes slightly, and these slight changes build up to shift the state from one category to another. For example, perhaps the addition of a grain of rice causes the total group of rice to be "slightly more" of a heap, and enough slight changes will certify the group's heap status – see fuzzy logic.
Proposed resolutions
Denying the existence of heaps
One may object to the first premise by denying that 1,000,000 grains of sand makes a heap. But 1,000,000 is just an arbitrary large number, and the argument will apply with any such number. So the response must deny outright that there are such things as heaps. Peter Unger defends this solution.[12]
Setting a fixed boundary
A common first response to the paradox is to term any set of grains that has more than a certain number of grains in it a heap. If one were to define the "fixed boundary" at 10,000 grains then one would claim that for fewer than 10,000, it is not a heap; for 10,000 or more, then it is a heap.[13]
Collins argues that such solutions are unsatisfactory as there seems little significance to the difference between 9,999 grains and 10,000 grains. The boundary, wherever it may be set, remains arbitrary, and so its precision is misleading. It is objectionable on both philosophical and linguistic grounds: the former on account of its arbitrariness and the latter on the ground that it is simply not how natural language is used.[14]
Unknowable boundaries (or epistemicism)
Timothy Williamson[15][16][17] and Roy Sorensen[18] claim that there are fixed boundaries but that they are necessarily unknowable.
Supervaluationism
Supervaluationism is a method for dealing with irreferential singular terms and vagueness. It allows one to retain the usual tautological laws even when dealing with undefined truth values.[19][20][21][22] As an example of a proposition about an irreferential singular term, consider the sentence "Pegasus likes licorice". Since the name "Pegasus" fails to refer, no truth value can be assigned to the sentence; there is nothing in the myth that would justify any such assignment. However, there are some statements about "Pegasus" which have definite truth values nevertheless, such as "Pegasus likes licorice or Pegasus doesn't like licorice". This sentence is an instance of the tautology "", i.e. the valid schema " or not-". According to supervaluationism, it should be true regardless of whether or not its components have a truth value.
By admitting sentences without defined truth values, supervaluationism avoids adjacent cases such that n grains of sand is a heap of sand, but n-1 grains is not; for example, "1,000 grains of sand is a heap" may be considered a border case having no defined truth value. Nevertheless, supervaluationism is able to handle a sentence like "1,000 grains of sand is a heap, or 1,000 grains of sand is not a heap" as a tautology, i.e. to assign it the value true.[citation needed]
Mathematical explanation
Let be a classical valuation defined on every atomic sentence of the language , and let be the number of distinct atomic sentences in . Then for every sentence , at most distinct classical valuations can exist. A supervaluation is a function from sentences to truth values such that, a sentence is super-true (i.e. ) if and only if for every classical valuation ; likewise for super-false. Otherwise, is undefined—i.e. exactly when there are two classical valuations and such that and .
For example, let be the formal translation of "Pegasus likes licorice". Then there are exactly two classical valuations and on , viz. and . So is neither super-true nor super-false. However, the tautology is evaluated to by every classical valuation; it is hence super-true. Similarly, the formalization of the above heap proposition is neither super-true nor super-false, but is super-true.
Truth gaps, gluts, and multi-valued logics
Another method is to use a multi-valued logic. In this context, the problem is with the principle of bivalence: the sand is either a heap or is not a heap, without any shades of gray. Instead of two logical states, heap and not-heap, a three value system can be used, for example heap, indeterminate and not-heap. A response to this proposed solution is that three valued systems do not truly resolve the paradox as there is still a dividing line between heap and indeterminate and also between indeterminate and not-heap. The third truth-value can be understood either as a truth-value gap or as a truth-value glut.[23]
Alternatively, fuzzy logic offers a continuous spectrum of logical states represented in the unit interval of real numbers [0,1]—it is a many-valued logic with infinitely-many truth-values, and thus the sand transitions gradually from "definitely heap" to "definitely not heap", with shades in the intermediate region. Fuzzy hedges are used to divide the continuum into regions corresponding to classes like definitely heap, mostly heap, partly heap, slightly heap, and not heap.[24][25] Though the problem remains of where these borders occur; e.g. at what number of grains sand starts being 'definitely' a heap.
Hysteresis
Another method, introduced by Raffman,[26] is to use hysteresis, that is, knowledge of what the collection of sand started as. Equivalent amounts of sand may be termed heaps or not based on how they got there. If a large heap (indisputably described as a heap) is diminished slowly, it preserves its "heap status" to a point, even as the actual amount of sand is reduced to a smaller number of grains. For example, 500 grains is a pile and 1,000 grains is a heap. There will be an overlap for these states. So if one is reducing it from a heap to a pile, it is a heap going down until 750. At that point, one would stop calling it a heap and start calling it a pile. But if one replaces one grain, it would not instantly turn back into a heap. When going up it would remain a pile until 900 grains. The numbers picked are arbitrary; the point is, that the same amount can be either a heap or a pile depending on what it was before the change. A common use of hysteresis would be the thermostat for air conditioning: the AC is set at 77 °F and it then cools the air to just below 77 °F, but does not activate again instantly when the air warms to 77.001 °F—it waits until almost 78 °F, to prevent instant change of state over and over again.[27]
Group consensus
One can establish the meaning of the word "heap" by appealing to consensus. Williamson, in his epistemic solution to the paradox, assumes that the meaning of vague terms must be determined by group usage.[28] The consensus method typically claims that a collection of grains is as much a "heap" as the proportion of people in a group who believe it to be so. In other words, the probability that any collection is considered a heap is the expected value of the distribution of the group's opinion.
A group may decide that:
- One grain of sand on its own is not a heap.
- A large collection of grains of sand is a heap.
Between the two extremes, individual members of the group may disagree with each other over whether any particular collection can be labelled a "heap". The collection can then not be definitively claimed to be a "heap" or "not a heap". This can be considered an appeal to descriptive linguistics rather than prescriptive linguistics, as it resolves the issue of definition based on how the population uses natural language. Indeed, if a precise prescriptive definition of "heap" is available then the group consensus will always be unanimous and the paradox does not occur.
Resolutions in utility theory
"X more or equally red than Y" modelled as quasitransitive relation ≈ : indistinguishable, > : clearly more red | ||||||
---|---|---|---|---|---|---|
Y X
|
f10 | e20 | d30 | c40 | b50 | a60 |
f10 | ≈ | ≈ | > | > | > | > |
e20 | ≈ | ≈ | ≈ | > | > | > |
d30 |
|
≈ | ≈ | ≈ | > | > |
c40 |
|
|
≈ | ≈ | ≈ | > |
b50 |
|
|
|
≈ | ≈ | ≈ |
a60 |
|
|
|
|
≈ | ≈ |
In the economics field of utility theory, the sorites paradox arises when a person's preferences patterns are investigated. As an example by Robert Duncan Luce, it is easy to find a person, say Peggy, who prefers in her coffee 3 grams (that is, 1 cube) of sugar to 15 grams (5 cubes), however, she will usually be indifferent between 3.00 and 3.03 grams, as well as between 3.03 and 3.06 grams, and so on, as well as finally between 14.97 and 15.00 grams.[29]
Two measures were taken by economists to avoid the sorites paradox in such a setting.
- Comparative, rather than positive, forms of properties are used. The above example deliberately does not make a statement like "Peggy likes a cup of coffee with 3 grams of sugar", or "Peggy does not like a cup of coffee with 15 grams of sugar". Instead, it states "Peggy likes a cup of coffee with 3 grams of sugar more than one with 15 grams of sugar".[33]
- Economists distinguish preference ("Peggy likes ... more than ...") from indifference ("Peggy likes ... as much as ... "), and do not consider the latter relation to be transitive.[35] In the above example, abbreviating "a cup of coffee with x grams of sugar" by "cx", and "Peggy is indifferent between cx and cy" as "cx ≈ cy", the facts c3.00 ≈ c3.03 and c3.03 ≈ c3.06 and ... and c14.97 ≈ c15.00 do not imply c3.00 ≈ c15.00.
Several kinds of relations were introduced to describe preference and indifference without running into the sorites paradox. Luce defined semi-orders and investigated their mathematical properties;[29] Amartya Sen performed a similar task for quasitransitive relations.[36] Abbreviating "Peggy likes cx more than cy" as "cx > cy", and abbreviating "cx > cy or cx ≈ cy" by "cx ≥ cy", it is reasonable that the relation ">" is a semi-order while ≥ is quasitransitive. Conversely, from a given semi-order > the indifference relation ≈ can be reconstructed by defining cx ≈ cy if neither cx > cy nor cy > cx. Similarly, from a given quasitransitive relation ≥ the indifference relation ≈ can be reconstructed by defining cx ≈ cy if both cx ≥ cy and cy ≥ cx. These reconstructed ≈ relations are usually not transitive.
The table to the right shows how the above color example can be modelled as a quasi-transitive relation ≥. Color differences overdone for readability. A color X is said to be more or equally red than a color Y if the table cell in row X and column Y is not empty. In that case, if it holds a "≈", then X and Y look indistinguishably equal, and if it holds a ">", then X looks clearly more red than Y. The relation ≥ is the disjoint union of the symmetric relation ≈ and the transitive relation >. Using the transitivity of >, the knowledge of both f10 > d30 and d30 > b50 allows one to infer that f10 > b50. However, since ≥ is not transitive, a "paradoxical" inference like "d30 ≥ e20 and e20 ≥ f10, hence d30 ≥ f10" is no longer possible. For the same reason, e.g. "d30 ≈ e20 and e20 ≈ f10, hence d30 ≈ f10" is no longer a valid inference. Similarly, to resolve the original heap variation of the paradox with this approach, the relation "X grains are more a heap than Y grains" could be considered quasitransitive rather than transitive.
Geometrical shape definition
The concept of a heap can be defined as a geometric shape in space, such as a pyramid, composed of randomly arranged sand particles. The existence of a heap can be said to cease when there are no longer sufficient sand particles to maintain the defined geometric shape. This philosophical resolution influenced by mathematical principles was proposed by Alex Dragomir, a Romanian philosopher, in 2023.[citation needed]
See also
References
- Sen, Amartya (1969). "Quasi-transitivity, rational choice and collective decisions". The Review of Economic Studies. 36 (3): 381–393. doi:10.2307/2296434. JSTOR 2296434. Zbl 0181.47302.
Bibliography
- Barnes, J. (1982). "Medicine, Experience and Logic". In Barnes, J.; Brunschwig, J.; Burnyeat, M. F.; Schofield, M. (eds.). Science and Speculation. Cambridge: Cambridge University Press.
- Black, Max (1970). Margins of Precision. Ithaca, NY: Cornell University Press. ISBN 978-0-8014-0602-7.
- Burns, Linda Claire (1991). Vagueness: An Investigation into Natural Languages and the Sorites Paradox. Dordrecht: Kluwer Academic Publishers. ISBN 978-0-7923-1489-9.
- Burnyeat, Myles (1982). "15. Gods and heaps". In Schofield, M.; Nussbaum, M. C. (eds.). Language and Logos. Cambridge: Cambridge University Press. pp. 315–.
- Collins, Rory (2018). "On the Borders of Vagueness and the Vagueness of Borders" (PDF). Vassar College Journal of Philosophy. 5: 30–44. Retrieved 21 June 2018.
- Dzhafarov, Ehtibar N.; Dzhafarov, Damir D. (2013). "The Sorites Paradox: A Behavioral Approach" (PDF). In Rudolph, Lee (ed.). Qualitative Mathematics for the Social Sciences: Mathematical Models for Research on Cultural Dynamics. pp. 105–136. doi:10.4324/9780203100806. ISBN 9780415444828.
- Gerla (2001). Fuzzy logic: Mathematical Tools for Approximate Reasoning. Dordrecht, Netherlands: Kluwer Academic Publishers. ISBN 978-0-7923-6941-7.
- Ludwig, Kirk; Ray, Greg (2002). "Vagueness and the Sorites Paradox". Philosophical Perspectives. 16: 419–461. JSTOR 3840922.
- Nouwen, Rick; Rooij, Robert van; Sauerland, Uli; Schmitz, Hans-Christian (2009). International Workshop on Vagueness in Communication (ViC; held as part of ESSLLI). LNAI. Vol. 6517. Springer. ISBN 978-3-642-18445-1.
- Oms, Sergi; Zardini, Elia, eds. (2019). The Sorites Paradox. Cambridge University Press.
- Sainsbury, R. M. (2009). Paradoxes (3rd ed.). Cambridge University Press.; Sect.3
External links
- Zalta, Edward N. (ed.). "Sorites Paradox". Stanford Encyclopedia of Philosophy. by Dominic Hyde.
- Sandra LaFave: Open and Closed Concepts and the Continuum Fallacy
https://en.wikipedia.org/wiki/Sorites_paradox#Continuum_fallacy
Causality (also called causation, or cause and effect) is influence by which one event, process, state, or object (a cause) contributes to the production of another event, process, state, or object (an effect) where the cause is partly responsible for the effect, and the effect is partly dependent on the cause. In general, a process has many causes,[1] which are also said to be causal factors for it, and all lie in its past. An effect can in turn be a cause of, or causal factor for, many other effects, which all lie in its future. Some writers have held that causality is metaphysically prior to notions of time and space.[2][3][4]
Causality is an abstraction that indicates how the world progresses.[5] As such a basic concept, it is more apt as an explanation of other concepts of progression than as something to be explained by others more basic. The concept is like those of agency and efficacy. For this reason, a leap of intuition may be needed to grasp it.[6][7] Accordingly, causality is implicit in the logic and structure of ordinary language,[8] as well as explicit in the language of scientific causal notation.
In English studies of Aristotelian philosophy, the word "cause" is used as a specialized technical term, the translation of Aristotle's term αἰτία, by which Aristotle meant "explanation" or "answer to a 'why' question". Aristotle categorized the four types of answers as material, formal, efficient, and final "causes". In this case, the "cause" is the explanans for the explanandum, and failure to recognize that different kinds of "cause" are being considered can lead to futile debate. Of Aristotle's four explanatory modes, the one nearest to the concerns of the present article is the "efficient" one.
David Hume, as part of his opposition to rationalism, argued that pure reason alone cannot prove the reality of efficient causality; instead, he appealed to custom and mental habit, observing that all human knowledge derives solely from experience.
The topic of causality remains a staple in contemporary philosophy.
https://en.wikipedia.org/wiki/Causality
Categorization is the ability and activity of recognizing shared features or similarities between the elements of the experience of the world (such as objects, events, or ideas), organizing and classifying experience by associating them to a more abstract group (that is, a category, class, or type),[1][2] on the basis of their traits, features, similarities or other criteria that are universal to the group. Categorization is considered one of the most fundamental cognitive abilities, and as such it is studied particularly by psychology and cognitive linguistics.
Categorization is sometimes considered synonymous with classification (cf., Classification synonyms). Categorization and classification allow humans to organize things, objects, and ideas that exist around them and simplify their understanding of the world.[3] Categorization is something that humans and other organisms do: "doing the right thing with the right kind of thing." The activity of categorizing things can be nonverbal or verbal. For humans, both concrete objects and abstract ideas are recognized, differentiated, and understood through categorization. Objects are usually categorized for some adaptive or pragmatic purposes.
Categorization is grounded in the features that distinguish the category's members from nonmembers. Categorization is important in learning, prediction, inference, decision making, language, and many forms of organisms' interaction with their environments.
Overview
Categories are distinct collections of concrete or abstract instances (category members) that are considered equivalent by the cognitive system. Using category knowledge requires one to access mental representations that define the core features of category members (cognitive psychologists refer to these category-specific mental representations as concepts).[4][5]
To categorization theorists, the categorization of objects is often considered using taxonomies with three hierarchical levels of abstraction.[6] For example, a plant could be identified at a high level of abstraction by simply labeling it a flower, a medium level of abstraction by specifying that the flower is a rose, or a low level of abstraction by further specifying this particular rose as a dog rose. Categories in a taxonomy are related to one another via class inclusion, with the highest level of abstraction being the most inclusive and the lowest level of abstraction being the least inclusive.[6] The three levels of abstraction are as follows:
- Superordinate level, Genus (e.g., Flower) - The highest and most inclusive level of abstraction. Exhibits the highest degree of generality and the lowest degree of within-category similarity.[7]
- Basic Level, Species (e.g., Rose) - The middle level of abstraction. Rosch and colleagues (1976) suggest the basic level to be the most cognitively efficient.[6] Basic level categories exhibit high within-category similarities and high between-category dissimilarities. Furthermore, the basic level is the most inclusive level at which category exemplars share a generalized identifiable shape.[6] Adults most-often use basic level object names, and children learn basic object names first.[6]
- Subordinate level (e.g., Dog Rose) - The lowest level of abstraction. Exhibits the highest degree of specificity and within-category similarity.[7]
https://en.wikipedia.org/wiki/Categorization
In philosophy, objectivity is the concept of truth independent from individual subjectivity (bias caused by one's perception, emotions, or imagination). A proposition is considered to have objective truth when its truth conditions are met without bias caused by the mind of a sentient being. Scientific objectivity refers to the ability to judge without partiality or external influence. Objectivity in the moral framework calls for moral codes to be assessed based on the well-being of the people in the society that follow it.[1] Moral objectivity also calls for moral codes to be compared to one another through a set of universal facts and not through subjectivity.[1]
Objectivity of knowledge
Plato considered geometry to be a condition of idealism concerned with universal truth.[clarification needed] In Republic, Socrates opposes the sophist Thrasymachus's relativistic account of justice, and argues that justice is mathematical in its conceptual structure, and that ethics was therefore a precise and objective enterprise with impartial standards for truth and correctness, like geometry.[2] The rigorous mathematical treatment Plato gave to moral concepts set the tone for the western tradition of moral objectivism that came after him.[citation needed] His contrasting between objectivity and opinion became the basis for philosophies intent on resolving the questions of reality, truth, and existence. He saw opinions as belonging to the shifting sphere of sensibilities, as opposed to a fixed, eternal and knowable incorporeality. Where Plato distinguished between how we know things and their ontological status, subjectivism such as George Berkeley's depends on perception.[3] In Platonic terms, a criticism of subjectivism is that it is difficult to distinguish between knowledge, opinions, and subjective knowledge.[4]
Platonic idealism is a form of metaphysical objectivism, holding that the ideas exist independently from the individual. Berkeley's empirical idealism, on the other hand, holds that things only exist as they are perceived. Both approaches boast an attempt at objectivity. Plato's definition of objectivity can be found in his epistemology, which is based on mathematics, and his metaphysics, where knowledge of the ontological status of objects and ideas is resistant to change.[3]
In opposition to philosopher René Descartes' method of personal deduction[clarification needed], natural philosopher Isaac Newton applied the relatively objective scientific method to look for evidence before forming a hypothesis.[5] Partially in response to Kant's rationalism, logician Gottlob Frege applied objectivity to his epistemological and metaphysical philosophies. If reality exists independently of consciousness, then it would logically include a plurality of indescribable forms. Objectivity requires a definition of truth formed by propositions with truth value. An attempt of forming an objective construct incorporates ontological commitments to the reality of objects.[6]
The importance of perception in evaluating and understanding objective reality is debated in the observer effect of quantum mechanics. Direct or naïve realists rely on perception as key in observing objective reality, while instrumentalists hold that observations are useful in predicting objective reality. The concepts that encompass these ideas are important in the philosophy of science. Philosophies of mind explore whether objectivity relies on perceptual constancy.[7]
Objectivity in ethics
Ethical subjectivism
The term "ethical subjectivism" covers two distinct theories in ethics. According to cognitive versions of ethical subjectivism, the truth of moral statements depends upon people's values, attitudes, feelings, or beliefs. Some forms of cognitivist ethical subjectivism can be counted as forms of realism, others are forms of anti-realism.[8] David Hume is a foundational figure for cognitive ethical subjectivism. On a standard interpretation of his theory, a trait of character counts as a moral virtue when it evokes a sentiment of approbation in a sympathetic, informed, and rational human observer.[9] Similarly, Roderick Firth's ideal observer theory held that right acts are those that an impartial, rational observer would approve of.[10] William James, another ethical subjectivist, held that an end is good (to or for a person) just in the case it is desired by that person (see also ethical egoism). According to non-cognitive versions of ethical subjectivism, such as emotivism, prescriptivism, and expressivism, ethical statements cannot be true or false, at all: rather, they are expressions of personal feelings or commands.[11] For example, on A. J. Ayer's emotivism, the statement, "Murder is wrong" is equivalent in meaning to the emotive, "Murder, Boo!"[12]
Ethical objectivism
According to the ethical objectivist, the truth or falsehood of typical moral judgments does not depend upon the beliefs or feelings of any person or group of persons. This view holds that moral propositions are analogous to propositions about chemistry, biology, or history, in so much as they are true despite what anyone believes, hopes, wishes, or feels. When they fail to describe this mind-independent moral reality, they are false—no matter what anyone believes, hopes, wishes, or feels.
There are many versions of ethical objectivism, including various religious views of morality, Platonistic intuitionism, Kantianism, utilitarianism, and certain forms of ethical egoism and contractualism. Note that Platonists define ethical objectivism in an even more narrow way, so that it requires the existence of intrinsic value. Consequently, they reject the idea that contractualists or egoists could be ethical objectivists. Objectivism, in turn, places primacy on the origin of the frame of reference—and, as such, considers any arbitrary frame of reference ultimately a form of ethical subjectivism by a transitive property, even when the frame incidentally coincides with reality and can be used for measurements.
Moral objectivism and relativism
Moral objectivism is the view that what is right or wrong does not depend on what anyone thinks is right or wrong.[1] Moral objectivism depends on how the moral code affects the well-being of the people of the society. Moral objectivism allows for moral codes to be compared to each other through a set of universal facts than mores of a society. Nicholas Reschar defines mores as customs within every society (e.g., what women can wear) and states that moral codes cannot be compared to one's personal moral compass.[1] An example is the categorical imperative of Immanuel Kant which says: "Act only according to that maxim [i.e., rule] whereby you can at the same time will that it become a universal law." John Stuart Mill was a consequential thinker and therefore proposed utilitarianism which asserts that in any situation, the right thing to do is whatever is likely to produce the most happiness overall. Moral relativism is the view where an actor's moral codes are locally derived from their culture.[13] The rules within moral codes are equal to each other and are only deemed "right" or "wrong" within their specific moral codes.[13] Relativism is opposite to Universalism because there is not a single moral code for every agent to follow.[13] Relativism differs from Nihilism because it validates every moral code that exists whereas nihilism does not.[13] When it comes to relativism, Russian philosopher and writer, Fyodor Dostoevsky, coined the phrase "If God doesn't exist, everything is permissible". That phrase was his view of the consequences for rejecting theism as a basis of ethics. American anthropologist Ruth Benedict argued that there is no single objective morality and that moral codes necessarily vary by culture.[14]
Objectivity in history
History as a discipline has wrestled with notions of objectivity from its very beginning. While its object of study is commonly thought to be the past, the only thing historians have to work with are different versions of stories based on individual perceptions of reality and memory.
Several history streams developed to devise ways to solve this dilemma: Historians like Leopold von Ranke (19th century) have advocated for the use of extensive evidence –especially archived physical paper documents– to recover the bygone past, claiming that, as opposed to people's memories, objects remain stable in what they say about the era they witnessed, and therefore represent a better insight into objective reality.[15] In the 20th century, the Annales School emphasized the importance of shifting focus away from the perspectives of influential men –usually politicians around whose actions narratives of the past were shaped–, and putting it on the voices of ordinary people.[16] Postcolonial streams of history challenge the colonial-postcolonial dichotomy and critique Eurocentric academia practices, such as the demand for historians from colonized regions to anchor their local narratives to events happening in the territories of their colonizers to earn credibility.[17] All the streams explained above try to uncover whose voice is more or less truth-bearing and how historians can stitch together versions of it to best explain what "actually happened."
The anthropologist Michel-Rolph Trouillot developed the concepts of historicity 1 and 2 to explain the difference between the materiality of socio-historical processes (H1) and the narratives that are told about the materiality of socio-historical processes (H2).[18] This distinction hints that H1 would be understood as the factual reality that elapses and is captured with the concept of "objective truth", and that H2 is the collection of subjectivities that humanity has stitched together to grasp the past. Debates about positivism, relativism, and postmodernism are relevant to evaluating these concepts' importance and the distinction between them.
Ethical considerations
In his book "Silencing the past", Trouillot wrote about the power dynamics at play in history-making, outlining four possible moments in which historical silences can be created: (1) making of sources (who gets to know how to write, or to have possessions that are later examined as historical evidence), (2) making of archives (what documents are deemed important to save and which are not, how to classify materials, and how to order them within physical or digital archives), (3) making of narratives (which accounts of history are consulted, which voices are given credibility), and (4) the making of history (the retrospective construction of what The Past is).[19]
Because history (official, public, familial, personal) informs current perceptions and how we make sense of the present, whose voice gets to be included in it –and how– has direct consequences in material socio-historical processes. Thinking of current historical narratives as impartial depictions of the totality of events unfolded in the past by labeling them as "objective" risks sealing historical understanding. Acknowledging that history is never objective and always incomplete has a meaningful opportunity to support social justice efforts. Under said notion, voices that have been silenced are placed on an equal footing to the grand and popular narratives of the world, appreciated for their unique insight of reality through their subjective lens.
See also
References
- Trouillot, Michel-Rolph. (1995). Silencing the past : power and the production of history. Boston, Mass. :Beacon Press,
Further reading
- Bachelard, Gaston. La formation de l'esprit scientifique: contribution à une psychanalyse de la connaissance. Paris: Vrin, 2004. ISBN 2-7116-1150-7.
- Castillejo, David. The Formation of Modern Objectivity. Madrid: Ediciones de Arte y Bibliofilia, 1982.
- Gaukroger, Stephen. (2012). Objectivity. Oxford University Press.
- Kuhn, Thomas S. The Structure of Scientific Revolutions. Chicago: University of Chicago Press, 1996, 3rd ed. ISBN 0-226-45808-3.
- Megill, Allan. Rethinking Objectivity. London: Duke UP, 1994.
- Nagel, Ernest. The Structure of Science. New York: Brace and World, 1961.
- Nagel, Thomas. The View from Nowhere. Oxford: Oxford UP, 1986
- Nozick, Robert. Invariances: the structure of the objective world. Cambridge: Harvard UP, 2001.
- Popper, Karl. R. Objective Knowledge: An Evolutionary Approach. Oxford University Press, 1972. ISBN 0-19-875024-2.
- Rescher, Nicholas. Objectivity: the obligations of impersonal reason. Notre Dame: Notre Dame Press, 1977.
- Rorty, Richard. Objectivity, Relativism, and Truth. Cambridge: Cambridge University Press, 1991
- Rousset, Bernard. La théorie kantienne de l'objectivité, Paris: Vrin, 1967.
- Scheffler, Israel. Science and Subjectivity. Hackett, 1982. Voices of Wisdom; a multicultural philosophy reader. Kessler
External links
https://en.wikipedia.org/wiki/Objectivity_(philosophy)
An object is a philosophical term often used in contrast to the term subject. A subject is an observer and an object is a thing observed. For modern philosophers like Descartes, consciousness is a state of cognition that includes the subject—which can never be doubted as only it can be the one who doubts—and some object(s) that may be considered as not having real or full existence or value independent of the subject who observes it. Metaphysical frameworks also differ in whether they consider objects existing independently of their properties and, if so, in what way.[1]
The pragmatist Charles S. Peirce defines the broad notion of an object as anything that we can think or talk about.[2] In a general sense it is any entity: the pyramids, gods,[3] Socrates,[3] Alpha Centauri, the number seven, a disbelief in predestination or the fear of cats. In a strict sense it refers to any definite being.
A related notion is objecthood. Objecthood is the state of being an object. One approach to defining it is in terms of objects' properties and relations. Descriptions of all bodies, minds, and persons must be in terms of their properties and relations. The philosophical question of the nature of objecthood concerns how objects are related to their properties and relations. For example, it seems that the only way to describe an apple is by describing its properties and how it is related to other things. Its properties may include its redness, its size, and its composition, while its relations may include "on the table", "in the room" and "being bigger than other apples".
The notion of an object must address two problems: the change problems and the problems of substances. Two leading theories about objecthood are substance theory, wherein substances (objects) are distinct from their properties, and bundle theory, wherein objects are no more than bundles of their properties.
Etymology
In English the word object is derived from the Latin objectus (p.p. of obicere) with the meaning "to throw, or put before or against", from ob- and jacere, "to throw".[4] As such it is a root for several important words used to derive meaning, such as objectify (to materialize), objective (a future reference), and objectivism (a philosophical doctrine that knowledge is based on objective reality).
Terms and usage
Broadly construed, the word object names a maximally general category, whose members are eligible for being referred to, quantified over and thought of. Terms similar to the broad notion of object include thing, being, entity, item, existent, term, unit, and individual.[3]
In ordinary language, one is inclined to call only a material object "object".[3] In certain contexts, it may be socially inappropriate to apply the word object to animate beings, especially to human beings, while the words entity and being are more acceptable.
Some authors use object in contrast to property; that is to say, an object is an entity that is not a property. Objects differ from properties in that objects cannot be referred to by predicates. Such usage may exclude abstracts objects from counting as objects. Terms similar to such usage of object include substance, individual, and particular. [3]
The word object can be used in contrast to subject as well. There are two definitions. The first definition holds that an object is an entity that fails to experience and that is not conscious. The second definition holds that an object is an entity experienced. The second definition differs from the first one in that the second definition allows for a subject to be an object at the same time.[3]
Change
An attribute of an object is called a property if it can be experienced (e.g. its color, size, weight, smell, taste, and location). Objects manifest themselves through their properties. These manifestations seem to change in a regular and unified way, suggesting that something underlies the properties. The change problem asks what that underlying thing is. According to substance theory, the answer is a substance, that which stands for the change.
Problem of substance
Because substances are only experienced through their properties a substance itself is never directly experienced. The problem of substance asks on what basis can one conclude the existence of a substance that cannot be seen or scientifically verified. According to David Hume's bundle theory, the answer is none; thus an object is merely its properties.
In the Mūlamadhyamakakārikā Nagarjuna seizes the dichotomy between objects as collections of properties or as separate from those properties to demonstrate that both assertions fall apart under analysis. By uncovering this paradox he then provides a solution (pratītyasamutpāda – "dependent origination") that lies at the very root of Buddhist praxis. Although Pratītyasamutpāda is normally limited to caused objects, Nagarjuna extends his argument to objects in general by differentiating two distinct ideas – dependent designation and dependent origination. He proposes that all objects are dependent upon designation, and therefore any discussion regarding the nature of objects can only be made in light of the context. The validity of objects can only be established within those conventions that assert them.[5]
Facts
Bertrand Russell updated the classical terminology with one more term, the fact;[6] "Everything that there is in the world I call a fact." Facts, objects, are opposed to beliefs, which are "subjective" and may be errors on the part of the subject, the knower who is their source and who is certain of himself and little else. All doubt implies the possibility of error and therefore admits the distinction between subjectivity and objectivity. The knower is limited in ability to tell fact from belief, false from true objects and engages in reality testing, an activity that will result in more or less certainty regarding the reality of the object. According to Russell,[7] "we need a description of the fact which would make a given belief true" where "Truth is a property of beliefs." Knowledge is "true beliefs".[8]
Applications
Value theory
Value theory concerns the value of objects. When it concerns economic value, it generally deals with physical objects. However, when concerning philosophic or ethic value, an object may be both a physical object and an abstract object (e.g. an action).[citation needed]
Physics
Limiting discussions of objecthood to the realm of physical objects may simplify them. However, defining physical objects in terms of fundamental particles (e.g. quarks) leaves open the question of what is the nature of a fundamental particle and thus asks what categories of being can be used to explain physical objects.[citation needed]
Semantics
Symbols represent objects; how they do so, the map–territory relation, is the basic problem of semantics.[9]
See also
- Abstract object theory
- Abstraction
- Category theory
- Continuous predicate
- Concept
- Hypostatic abstraction
- Hypostasis (philosophy and religion)
- Noumenon and phenomenon
- Objectivity (philosophy)
- Observer (physics)
- Sign relation
- Ship of Theseus
- Object-oriented ontology
- Subject (grammar)
- Subject (philosophy)
- Nonexistent object
References
- Dąmbska, Izydora (2016). "Symbols". Poznan Studies in the Philosophy of the Sciences & the Humanities. 105: 201–209 – via Humanities Source.
Sources
- Russell, Bertrand (1948). Human Knowledge Its Scope and Limits. New York: Simon and Schuster.
External links
- Bradley Retter & Andrew M. Bailey (2017). Object. Stanford Encyclopedia of Philosophy. Metaphysics Research Lab, Stanford University.
- Gideon Rosen (2022). Abstract objects. Stanford Encyclopedia of Philosophy. Metaphysics Research Lab, Stanford University.
- Colin Smith. "Even More Abstract Objects". Crazy Objects And Their Affect On Reality.
https://en.wikipedia.org/wiki/Object_(philosophy)
Part of a series on |
Science |
---|
|
|
|
A paradigm shift is a fundamental change in the basic concepts and experimental practices of a scientific discipline. It is a concept in the philosophy of science that was introduced and brought into the common lexicon by the American physicist and philosopher Thomas Kuhn. Even though Kuhn restricted the use of the term to the natural sciences, the concept of a paradigm shift has also been used in numerous non-scientific contexts to describe a profound change in a fundamental model or perception of events.
Kuhn presented his notion of a paradigm shift in his influential book The Structure of Scientific Revolutions (1962).
Kuhn contrasts paradigm shifts, which characterize a Scientific Revolution, to the activity of normal science, which he describes as scientific work done within a prevailing framework or paradigm. Paradigm shifts arise when the dominant paradigm under which normal science operates is rendered incompatible with new phenomena, facilitating the adoption of a new theory or paradigm.[1]
As one commentator summarizes:
Kuhn acknowledges having used the term "paradigm" in two different meanings. In the first one, "paradigm" designates what the members of a certain scientific community have in common, that is to say, the whole of techniques, patents and values shared by the members of the community. In the second sense, the paradigm is a single element of a whole, say for instance Newton’s Principia, which, acting as a common model or an example... stands for the explicit rules and thus defines a coherent tradition of investigation. Thus the question is for Kuhn to investigate by means of the paradigm what makes possible the constitution of what he calls "normal science". That is to say, the science which can decide if a certain problem will be considered scientific or not. Normal science does not mean at all a science guided by a coherent system of rules, on the contrary, the rules can be derived from the paradigms, but the paradigms can guide the investigation also in the absence of rules. This is precisely the second meaning of the term "paradigm", which Kuhn considered the most new and profound, though it is in truth the oldest.[2]
https://en.wikipedia.org/wiki/Paradigm_shift
The Composition of Causes was a set of philosophical laws advanced by John Stuart Mill in his watershed essay A System of Logic. These laws outlined Mill's view of the epistemological components of emergentism, a school of philosophical laws that posited a decidedly opportunistic approach to the classic dilemma of causation nullification.
Mill was determined to prove that the intrinsic properties of all things relied on three primary tenets, which he called the Composition of Causes. These were:
1. The Cause of Inherent Efficiency, a methodological understanding of deterministic forces engaged in the perpetual axes of the soul, as it pertained to its own self-awareness.
2. The so-called Sixth Cause, a conceptual notion embodied by the system of inter-related segments of social and elemental vitra. This was a hotly debated matter in early 17th-century philosophical circles, especially in the halls of the Reichtaven in Meins, where the spirit of Geudl still lingered.
3. The Cause of Multitude, an evolutionary step taken from Hemmlich's Plurality of a Dysfunctional Enterprise, detailing the necessary linkage between both sets of perception-based self-awareness.
Furthermore, the Composition of Causes elevated Mill's standing in ontological circles, lauded by his contemporaries for applying a conceptual vision of an often-argued discipline.
External links
https://en.wikipedia.org/wiki/Composition_of_Causes
In philosophy, condition of possibility (German: Bedingungen der Möglichkeit) is a concept made popular by the German philosopher Immanuel Kant, and is an important part of his philosophy.
A condition of possibility is a necessary framework for the possible appearance of a given list of entities. It is often used in contrast to the unilateral causality concept, or even to the notion of interaction[disambiguation needed]. For example, consider a cube made by an artisan. All cubes are three-dimensional. If an object is three-dimensional, then it is an extended object. But extension is an impossibility without space. Therefore, space is a condition of possibility because it is a necessary condition for the existence of cubes to be possible. Note, however, that space did not cause the cube, but that the artisan did, and that the cube and space are distinct entities, so space is not part of the definition of cube.
Gilles Deleuze presented it as a dichotomy in contradistinction to the classical phenomenon/noumenon dichotomy.[citation needed] From Plato to Descartes, what was presented by the senses was deemed illusory and denigrated[clarification needed]. It was believed that the perceptions ought to be overcome to grasp the thing-in-itself, the essential essence, also known as Plato's allegory of the cave. With Kant comes a transition in philosophy from this dichotomy to the dichotomy of the apparition[clarification needed]/conditions-of-appearance[clarification needed]. There is no longer any higher essence behind the apparition[clarification needed]. It is what it is, a brute fact, and what one must now examine is the conditions that are necessary for its appearance. Immanuel Kant does just this in the Transcendental Aesthetic, when he examines the necessary conditions for the synthetic a priori cognition of mathematics. But Kant was a transitional thinker[clarification needed], so he still maintains the phenomenon/noumenon dichotomy, but what he did achieve was to render Noumena as unknowable and irrelevant.[1]
Foucault would come to adapt it in a historical sense through the concept of "episteme":
what I am attempting to bring to light is the epistemological field, the épistémè in which knowledge, envisaged apart from all criteria having reference to its rational value or to its objective forms, grounds its positivity and thereby manifests a history which is not that of its growing perfection, but rather that of its conditions of possibility; in this account, what should appear are those configurations within the space of knowledge which have given rise to the diverse forms of empirical science. Such an enterprise is not so much a history, in the traditional meaning of that word, as an ‘archaeology’.[2]
References
- Foucault, Michel (2002), The Order of Things, p. xxiii-xxiv
- Foucault, Michel (2002) [1966], Les Mots et Les Choses [The Order of Things], Routledge.
https://en.wikipedia.org/wiki/Condition_of_possibility
In science and history, consilience (also convergence of evidence or concordance of evidence) is the principle that evidence from independent, unrelated sources can "converge" on strong conclusions. That is, when multiple sources of evidence are in agreement, the conclusion can be very strong even when none of the individual sources of evidence is significantly so on its own. Most established scientific knowledge is supported by a convergence of evidence: if not, the evidence is comparatively weak, and there will probably not be a strong scientific consensus.
The principle is based on unity of knowledge; measuring the same result by several different methods should lead to the same answer. For example, it should not matter whether one measures distances within the Giza pyramid complex by laser rangefinding, by satellite imaging, or with a meter stick – in all three cases, the answer should be approximately the same. For the same reason, different dating methods in geochronology should concur, a result in chemistry should not contradict a result in geology, etc.
The word consilience was originally coined as the phrase "consilience of inductions" by William Whewell (consilience refers to a "jumping together" of knowledge).[1][2] The word comes from Latin com- "together" and -siliens "jumping" (as in resilience).[3]
Description
Consilience requires the use of independent methods of measurement, meaning that the methods have few shared characteristics. That is, the mechanism by which the measurement is made is different; each method is dependent on an unrelated natural phenomenon. For example, the accuracy of laser rangefinding measurements is based on the scientific understanding of lasers, while satellite pictures and meter sticks rely on different phenomena. Because the methods are independent, when one of several methods is in error, it is very unlikely to be in error in the same way as any of the other methods, and a difference between the measurements will be observed.[note 1] If the scientific understanding of the properties of lasers were inaccurate, then the laser measurement would be inaccurate but the others would not.
As a result, when several different methods agree, this is strong evidence that none of the methods are in error and the conclusion is correct. This is because of a greatly reduced likelihood of errors: for a consensus estimate from multiple measurements to be wrong, the errors would have to be similar for all samples and all methods of measurement, which is extremely unlikely. Random errors will tend to cancel out as more measurements are made, due to regression to the mean; systematic errors will be detected by differences between the measurements (and will also tend to cancel out since the direction of the error will still be random). This is how scientific theories reach high confidence – over time, they build up a large degree of evidence which converges on the same conclusion.[note 2]
When results from different strong methods do appear to conflict, this is treated as a serious problem to be reconciled. For example, in the 19th century, the Sun appeared to be no more than 20 million years old, but the Earth appeared to be no less than 300 million years (resolved by the discovery of nuclear fusion and radioactivity, and the theory of quantum mechanics);[4] or current attempts to resolve theoretical differences between quantum mechanics and general relativity.[5]
Significance
Because of consilience, the strength of evidence for any particular conclusion is related to how many independent methods are supporting the conclusion, as well as how different these methods are. Those techniques with the fewest (or no) shared characteristics provide the strongest consilience and result in the strongest conclusions. This also means that confidence is usually strongest when considering evidence from different fields because the techniques are usually very different.
For example, the theory of evolution is supported by a convergence of evidence from genetics, molecular biology, paleontology, geology, biogeography, comparative anatomy, comparative physiology, and many other fields.[6] In fact, the evidence within each of these fields is itself a convergence providing evidence for the theory. (As a result, to disprove evolution, most or all of these independent lines of evidence would have to be found to be in error.[2]) The strength of the evidence, considered together as a whole, results in the strong scientific consensus that the theory is correct.[6] In a similar way, evidence about the history of the universe is drawn from astronomy, astrophysics, planetary geology, and physics.[2]
Finding similar conclusions from multiple independent methods is also evidence for the reliability of the methods themselves, because consilience eliminates the possibility of all potential errors that do not affect all the methods equally. This is also used for the validation of new techniques through comparison with the consilient ones. If only partial consilience is observed, this allows for the detection of errors in methodology; any weaknesses in one technique can be compensated for by the strengths of the others. Alternatively, if using more than one or two techniques for every experiment is infeasible, some of the benefits of consilience may still be obtained if it is well-established that these techniques usually give the same result.
Consilience is important across all of science, including the social sciences,[7] and is often used as an argument for scientific realism by philosophers of science. Each branch of science studies a subset of reality that depends on factors studied in other branches. Atomic physics underlies the workings of chemistry, which studies emergent properties that in turn are the basis of biology. Psychology is not separate from the study of properties emergent from the interaction of neurons and synapses. Sociology, economics, and anthropology are each, in turn, studies of properties emergent from the interaction of countless individual humans. The concept that all the different areas of research are studying one real, existing universe is an apparent explanation of why scientific knowledge determined in one field of inquiry has often helped in understanding other fields.
Deviations
Consilience does not forbid deviations: in fact, since not all experiments are perfect, some deviations from established knowledge are expected. However, when the convergence is strong enough, then new evidence inconsistent with the previous conclusion is not usually enough to outweigh that convergence. Without an equally strong convergence on the new result, the weight of evidence will still favor the established result. This means that the new evidence is most likely to be wrong.
Science denialism (for example, AIDS denialism) is often based on a misunderstanding of this property of consilience. A denier may promote small gaps not yet accounted for by the consilient evidence, or small amounts of evidence contradicting a conclusion without accounting for the pre-existing strength resulting from consilience. More generally, to insist that all evidence converge precisely with no deviations would be naïve falsificationism,[8] equivalent to considering a single contrary result to falsify a theory when another explanation, such as equipment malfunction or misinterpretation of results, is much more likely.[8][note 3]
In history
Historical evidence also converges in an analogous way. For example: if five ancient historians, none of whom knew each other, all claim that Julius Caesar seized power in Rome in 49 BCE, this is strong evidence in favor of that event occurring even if each individual historian is only partially reliable. By contrast, if the same historian had made the same claim five times in five different places (and no other types of evidence were available), the claim is much weaker because it originates from a single source. The evidence from the ancient historians could also converge with evidence from other fields, such as archaeology: for example, evidence that many senators fled Rome at the time, that the battles of Caesar’s civil war occurred, and so forth.
Consilience has also been discussed in reference to Holocaust denial.
"We [have now discussed] eighteen proofs all converging on one conclusion...the deniers shift the burden of proof to historians by demanding that each piece of evidence, independently and without corroboration between them, prove the Holocaust. Yet no historian has ever claimed that one piece of evidence proves the Holocaust. We must examine the collective whole."[2]
That is, individually the evidence may underdetermine the conclusion, but together they overdetermine it. A similar way to state this is that to ask for one particular piece of evidence in favor of a conclusion is a flawed question.[6][9]
Outside the sciences
In addition to the sciences, consilience can be important to the arts, ethics and religion. Both artists and scientists have identified the importance of biology in the process of artistic innovation.[1]
History of the concept
Consilience has its roots in the ancient Greek concept of an intrinsic orderliness that governs our cosmos, inherently comprehensible by logical process, a vision at odds with mystical views in many cultures that surrounded the Hellenes. The rational view was recovered during the high Middle Ages, separated from theology during the Renaissance and found its apogee in the Age of Enlightenment.[1]
Whewell's definition was that:[10]
The Consilience of Inductions takes place when an Induction, obtained from one class of facts, coincides with an Induction obtained from another different class. Thus Consilience is a test of the truth of the Theory in which it occurs.
More recent descriptions include:
"Where there is a convergence of evidence, where the same explanation is implied, there is increased confidence in the explanation. Where there is divergence, then either the explanation is at fault or one or more of the sources of information is in error or requires reinterpretation."[11]
"Proof is derived through a convergence of evidence from numerous lines of inquiry--multiple, independent inductions, all of which point to an unmistakable conclusion."[6]
Edward O. Wilson
Although the concept of consilience in Whewell's sense was widely discussed by philosophers of science, the term was unfamiliar to the broader public until the end of the 20th century, when it was revived in Consilience: The Unity of Knowledge, a 1998 book by the author and biologist E. O. Wilson, as an attempt to bridge the cultural gap between the sciences and the humanities that was the subject of C. P. Snow's The Two Cultures and the Scientific Revolution (1959).[1] Wilson believed that "the humanities, ranging from philosophy and history to moral reasoning, comparative religion, and interpretation of the arts, will draw closer to the sciences and partly fuse with them" with the result that science and the scientific method, from within this fusion, would not only explain the physical phenomenon but also provide moral guidance and be the ultimate source of all truths.[12]
Wilson held that with the rise of the modern sciences, the sense of unity gradually was lost in the increasing fragmentation and specialization of knowledge in the last two centuries. He asserted that the sciences, humanities, and arts have a common goal: to give a purpose to understand the details, to lend to all inquirers "a conviction, far deeper than a mere working proposition, that the world is orderly and can be explained by a small number of natural laws." An important point made by Wilson is that hereditary human nature and evolution itself profoundly effect the evolution of culture, in essence, a sociobiological concept. Wilson's concept is a much broader notion of consilience than that of Whewell, who was merely pointing out that generalizations invented to account for one set of phenomena often account for others as well.[1]
A parallel view lies in the term universology, which literally means "the science of the universe." Universology was first promoted for the study of the interconnecting principles and truths of all domains of knowledge by Stephen Pearl Andrews, a 19th-century utopian futurist and anarchist.[1]
See also
- Appeal to authority – Fallacious form of argument
- Appeal to tradition – Logical fallacy in which a thesis is deemed correct on the basis of tradition
- Philosophy of science#Coherentism – Study of foundations, methods, and implications of science
- Equifinality – Principle that in open systems a given end state can be reached by many potential means
- Scientific method – Interplay between observation, experiment and theory in science
- Syncretism – Assimilation of two or more originally discrete religious traditions
- Tree of knowledge system – Map of history from Big Bang to present
- Unified Science
Notes
- More generally, anything which results in a false positive or false negative.
References
- Horgan, John. "Science Should Not Try to Absorb Religion and Other Ways of Knowing". Scientific American. Retrieved 1 July 2021.
External links
https://en.wikipedia.org/wiki/Consilience
In epistemology, criteria of truth (or tests of truth) are standards and rules used to judge the accuracy of statements and claims. They are tools of verification, and as in the problem of the criterion, the reliability of these tools is disputed. Understanding a philosophy's criteria of truth is fundamental to a clear evaluation of that philosophy. This necessity is driven by the varying, and conflicting, claims of different philosophies. The rules of logic have no ability to distinguish truth on their own. An individual must determine what standards distinguish truth from falsehood. Not all criteria are equally valid. Some standards are sufficient, while others are questionable.[1]
The criteria listed represent those most commonly used by scholars and the general public.[2]
Authority
The opinions of those with significant experience, highly trained or possessing an advanced degree are often considered a form of proof. Their knowledge and familiarity within a given field or area of knowledge command respect and allow their statements to be criteria of truth. A person may not simply declare themselves an authority, but rather must be properly qualified. Despite the wide respect given to expert testimony, it is not an infallible criterion. For example, multiple authorities may conflict in their claims and conclusions.[3]
Coherence
Coherence refers to a consistent and overarching explanation for all facts. To be coherent, all pertinent facts must be arranged in a consistent and cohesive fashion as an integrated whole. The theory that most effectively reconciles all facts in this fashion may be considered most likely to be true. Coherence is the most potentially effective test of truth because it most adequately addresses all elements. The main limitation lies not in the standard, but in the human inability to acquire all facts of an experience. Only an omniscient mind could be aware of all of the relevant information. A scholar must accept this limitation and accept as true the most coherent explanation for the available facts. Coherence is difficult to dispute as a criterion of truth, since arguing against coherence is validating incoherence, which is inherently illogical.[4]
Consensus gentium
Some view opinions held by all people to be valid criteria of truth. According to consensus gentium, the universal consent of all mankind (all humans holding a distinct belief), proves it is true. There is some value in the criterion if it means innate truth, such as the laws of logic and mathematics. If it merely means agreement, as in a unanimous vote, its value is questionable. For example, general assent once held that the sun revolved about the flat earth.[5]
Consistency (mere)
Mere consistency is when correct statements do not contradict, but are not necessarily related. Accordingly, an individual is consistent if he does not contradict himself. It is inadequate as a criterion because it treats facts in an isolated fashion without true cohesion and integration; nevertheless it remains a necessary condition for the truth of any argument, owing to the law of noncontradiction. The value of a proof largely lies in its ability to reconcile individual facts into a coherent whole.[6]
Consistency (strict)
Strict consistency is when claims are connected in such a fashion that one statement follows from another. Formal logic and mathematical rules are examples of rigorous consistency. An example would be: if all As are Bs and all Bs are Cs, then all As are Cs. While this standard is of high value, it is limited. For example, the premises are a priori (or self-apparent), requiring another test of truth to employ this criterion. Additionally, strict consistency may produce results lacking coherence and completeness. While a philosophical system may demonstrate rigorous consistency with the facts it considers, all facts must be taken into consideration for an adequate criterion of truth, regardless of their detriment to any given system.[6]
Correspondence
Correspondence is quite simply when a claim corresponds with its object. For example, the claim that the White House is in Washington, D.C. is true, if the White House is actually located in Washington. Correspondence is held by many philosophers to be the most valid of the criteria of truth. An idea that corresponds to its object is indeed true, but determining if the correspondence is perfect requires additional tests of truth. This indicates that correspondence is a perfectly valid definition of truth, but is not of itself a valid criterion of truth. An additional test beyond this "definition" is required to determine the precise degree of similarity between what is posited and what exists in objective reality.[7] Establishing correspondence between what is posited and what exists is fraught with its own difficulties, see Map–territory relation.
Custom
Most people consciously or unknowingly employ custom as a criterion of truth, based on the assumption that doing what is customary will prevent error. It is particularly applied in the determination of moral truth and reflected in the statement "when in Rome, do as the Romans do". People stick closely to the principle of custom when they use common vernacular, wear common fashions and so forth; essentially, when they do what is popular. Custom is not considered a serious, or valid, test of truth. For example, public opinion polls do not determine truth.[8]
Emotions
Many people allow feelings to determine judgment, often in the face of contrary evidence or without even attempting to collect evidence and facts. They are implicitly accepting emotions as a criterion of truth. Most people will admit that feelings are not an adequate test for truth. For example, a seasoned businessman will put aside his emotions and search for the best available facts when making an investment. Similarly, scholars are trained to put aside such subjective judgments when evaluating knowledge.[9] Emotions are real, however, and thus must be considered within any social scientific system of coherence.
Instinct
The existence of distinct instincts has long been debated. Proponents of instinct argue that we eat because of hunger, drink because of thirst, and so forth. Some have even argued for the existence of God based on this criterion, arguing that the object of every instinct has a referent in reality. The counterpoint of hunger is food; for thirst it is liquid; for the sex drive it is a mate. Instincts are not accepted as a reliable test because they are most often indistinct, variant and difficult to define. Additionally, universal instincts are so few that they offer little to the greater body of philosophy as a criterion.[10]
Intuition
Intuition is an assumed truth with an unknown, or possibly unexamined, source. It is a judgment that is not dependent on a rational examination of the facts. It is usually experienced as a sudden sensation and/or rush of thoughts that feel "right". Many persons experience intuitive epiphanies which later prove to be true. Scholars have sometimes come upon valid theories and proofs while daydreaming or otherwise mentally occupied with something bearing no apparent relationship to the truth they seek to reveal. Intuition is at best a source for truths, rather than a criterion with which to evaluate them. Intuitive knowledge requires testing by means of other criteria of truth in order to confirm its accuracy.[11]
Majority rule
Majority rule is a statistical method of accepting assertions and proposals. In democratic systems, majority rule is used to determine group decisions, particularly those relating to personal morality and social behavior. Some systems divided into several oppositional factions may depend on mere plurality. While majority rule may make for a good democratic system, it is a poor determinant of truth, subject to the criticisms of the broad version of consensus gentium.[5]
Naïve realism
Naïve realism posits that only that which is directly observable by the human senses is true. First-hand observation determines the truth or falsity of a given statement. Naïve Realism is an insufficient criterion of truth. A host of natural phenomena are demonstrably true, but not observable by the unaided sense. For example, Naïve Realism would deny the existence of sounds beyond the range of human hearing and the existence of x-rays. Similarly, there are a number of sense experiments which show a disconnect between the perceived sensation and the reality of its cause.[12]
Pragmatic
If an idea works then it must be true, to the Pragmatist. The consequences of applying a concept reveal its truth value upon examination of the results. The full meaning of an idea is self-apparent in its application. For example, the therapeutic value and effect of penicillin in relation to infections is proven in its administration. Although pragmatism is considered a valuable criterion, it must be used with caution and reservation, due to its potential for false positives. For example, a doctor may prescribe a patient medication for an illness, but it could later turn out that a placebo is equally effective. Thus, untrue concepts could appear to be working contrary to the purpose of the pragmatic test. However, it has validity as a test, particularly in the form William Ernest Hocking called "negative pragmatism". In essence, it states that ideas that do not work cannot possibly be true, though ideas which do work may or may not be true.[13]
Revelation
The principal distinction between intuition and revelation is that revelation has an assumed source: God (or another higher power). Revelation may be defined as truth emanating from God. Many religions fundamentally rely on revelation as a test of truth. This criterion is subject to the same criticisms as intuition. It may be a valid reference of truth for an individual, but it is inadequate for providing a coherent proof of the knowledge to others.[14]
Time
Time is a criterion commonly appealed to in debate, often referred to as "the test of time". This criterion posits that over time erroneous beliefs and logical errors will be revealed, while if the belief is true, the mere passage of time cannot adversely affect its validity. Time is an inadequate test for truth, since it is subject to similar flaws as custom and tradition (which are simply specific variations of the time factor). Many demonstrably false beliefs have endured for centuries and even millennia (e.g. vitalism). It is commonly rejected as a valid criterion. For example, most people will not convert to another faith simply because the other religion is centuries (or even millennia) older than their current beliefs.[15]
Tradition
Tradition, closely related to custom, is the standard stating that which is held for generations is true. Those accepting tradition argue that ideas gaining the loyalty multiple generations possesses a measure of credibility. Tradition possesses many of the same failings as custom. It is possible for falsehoods to be passed down from generation to generation, since tradition generally emphasizes repetition over critical evaluation.[8]
See also
Footnotes
- Sahakian & Sahakian, Ideas, pp. 4–5.
https://en.wikipedia.org/wiki/Criteria_of_truth
A pattern is a regularity in the world, in human-made design,[1] or in abstract ideas. As such, the elements of a pattern repeat in a predictable manner. A geometric pattern is a kind of pattern formed of geometric shapes and typically repeated like a wallpaper design.
Any of the senses may directly observe patterns. Conversely, abstract patterns in science, mathematics, or language may be observable only by analysis. Direct observation in practice means seeing visual patterns, which are widespread in nature and in art. Visual patterns in nature are often chaotic, rarely exactly repeating, and often involve fractals. Natural patterns include spirals, meanders, waves, foams, tilings, cracks, and those created by symmetries of rotation and reflection. Patterns have an underlying mathematical structure;[2]: 6 indeed, mathematics can be seen as the search for regularities, and the output of any function is a mathematical pattern. Similarly in the sciences, theories explain and predict regularities in the world.
In art and architecture, decorations or visual motifs may be combined and repeated to form patterns designed to have a chosen effect on the viewer. In computer science, a software design pattern is a known solution to a class of problems in programming. In fashion, the pattern is a template used to create any number of similar garments.
In many areas of the decorative arts, from ceramics and textiles to wallpaper, "pattern" is used for an ornamental design that is manufactured, perhaps for many different shapes of object.
Nature
Nature provides examples of many kinds of pattern, including symmetries, trees and other structures with a fractal dimension, spirals, meanders, waves, foams, tilings, cracks and stripes.[3]
Symmetry
Symmetry is widespread in living things. Animals that move usually have bilateral or mirror symmetry as this favours movement.[2]: 48–49 Plants often have radial or rotational symmetry, as do many flowers, as well as animals which are largely static as adults, such as sea anemones. Fivefold symmetry is found in the echinoderms, including starfish, sea urchins, and sea lilies.[2]: 64–65
Among non-living things, snowflakes have striking sixfold symmetry: each flake is unique, its structure recording the varying conditions during its crystallisation similarly on each of its six arms.[2]: 52 Crystals have a highly specific set of possible crystal symmetries; they can be cubic or octahedral, but cannot have fivefold symmetry (unlike quasicrystals).[2]: 82–84
Spirals
Spiral patterns are found in the body plans of animals including molluscs such as the nautilus, and in the phyllotaxis of many plants, both of leaves spiralling around stems, and in the multiple spirals found in flowerheads such as the sunflower and fruit structures like the pineapple.[4]
Chaos, turbulence, meanders and complexity
Chaos theory predicts that while the laws of physics are deterministic, there are events and patterns in nature that never exactly repeat because extremely small differences in starting conditions can lead to widely differing outcomes.[5] The patterns in nature tend to be static due to dissipation on the emergence process, but when there is interplay between injection of energy and dissipation there can arise a complex dynamic.[6] Many natural patterns are shaped by this complexity, including vortex streets,[7] other effects of turbulent flow such as meanders in rivers.[8] or nonlinear interaction of the system [9]
Waves, dunes
Waves are disturbances that carry energy as they move. Mechanical waves propagate through a medium – air or water, making it oscillate as they pass by.[10] Wind waves are surface waves that create the chaotic patterns of the sea. As they pass over sand, such waves create patterns of ripples; similarly, as the wind passes over sand, it creates patterns of dunes.[11]
Bubbles, foam
Foams obey Plateau's laws, which require films to be smooth and continuous, and to have a constant average curvature. Foam and bubble patterns occur widely in nature, for example in radiolarians, sponge spicules, and the skeletons of silicoflagellates and sea urchins.[12][13]
Cracks
Cracks form in materials to relieve stress: with 120 degree joints in elastic materials, but at 90 degrees in inelastic materials. Thus the pattern of cracks indicates whether the material is elastic or not. Cracking patterns are widespread in nature, for example in rocks, mud, tree bark and the glazes of old paintings and ceramics.[14]
Spots, stripes
Alan Turing,[15] and later the mathematical biologist James D. Murray[16] and other scientists, described a mechanism that spontaneously creates spotted or striped patterns, for example in the skin of mammals or the plumage of birds: a reaction–diffusion system involving two counter-acting chemical mechanisms, one that activates and one that inhibits a development, such as of dark pigment in the skin.[17] These spatiotemporal patterns slowly drift, the animals' appearance changing imperceptibly as Turing predicted.
Art and architecture
Tilings
In visual art, pattern consists in regularity which in some way "organizes surfaces or structures in a consistent, regular manner." At its simplest, a pattern in art may be a geometric or other repeating shape in a painting, drawing, tapestry, ceramic tiling or carpet, but a pattern need not necessarily repeat exactly as long as it provides some form or organizing "skeleton" in the artwork.[18] In mathematics, a tessellation is the tiling of a plane using one or more geometric shapes (which mathematicians call tiles), with no overlaps and no gaps.[19]
Zentangles
The concept and process of Zentangles, a blend of meditative Zen practice with the purposeful drawing of repetitive patterns or artistic tangles has been trademarked by Rick Roberts and Maria Thomas.[20] The process, using patterns such as cross hatching, dots, curves and other mark making, on small pieces of paper or tiles which can then be put together to form mosaic clusters, or shaded or coloured in, can, like the doodle, be used as a therapeutic device to help to relieve stress and anxiety in children and adults.[21] [22] Zentangles comprising relevant or irrelevant shapes can be drawn within the outline of an animal, human or object to provide texture and interest. [1]
In architecture
In architecture, motifs are repeated in various ways to form patterns. Most simply, structures such as windows can be repeated horizontally and vertically (see leading picture). Architects can use and repeat decorative and structural elements such as columns, pediments, and lintels.[23] Repetitions need not be identical; for example, temples in South India have a roughly pyramidal form, where elements of the pattern repeat in a fractal-like way at different sizes.[24]
See also: pattern book.
Science and mathematics
Mathematics is sometimes called the "Science of Pattern", in the sense of rules that can be applied wherever needed.[25] For example, any sequence of numbers that may be modeled by a mathematical function can be considered a pattern. Mathematics can be taught as a collection of patterns.[26]
Fractals
Some mathematical rule-patterns can be visualised, and among these are those that explain patterns in nature including the mathematics of symmetry, waves, meanders, and fractals. Fractals are mathematical patterns that are scale invariant. This means that the shape of the pattern does not depend on how closely you look at it. Self-similarity is found in fractals. Examples of natural fractals are coast lines and tree shapes, which repeat their shape regardless of what magnification you view at. While self-similar patterns can appear indefinitely complex, the rules needed to describe or produce their formation can be simple (e.g. Lindenmayer systems describing tree shapes).[27]
In pattern theory, devised by Ulf Grenander, mathematicians attempt to describe the world in terms of patterns. The goal is to lay out the world in a more computationally friendly manner.[28]
In the broadest sense, any regularity that can be explained by a scientific theory is a pattern. As in mathematics, science can be taught as a set of patterns.[29]
Computer science
In computer science, a software design pattern, in the sense of a template, is a general solution to a problem in programming. A design pattern provides a reusable architectural outline that may speed the development of many computer programs.[30]
Fashion
In fashion, the pattern is a template, a technical two-dimensional tool used to create any number of identical garments. It can be considered as a means of translating from the drawing to the real garment.[31]
See also
- Archetype
- Cellular automata
- Form constant
- Fractal
- Pattern (sewing)
- Pattern coin
- Pattern matching
- Pattern recognition
- Pattern (casting)
- Patterns in nature
- Pedagogical patterns
References
- "An Artist Centric Marketplace for Fashion Sketch Templates, Croquis & More". Illustrator Stuff. Retrieved 7 January 2018.
Bibliography
In nature
- Adam, John A. Mathematics in Nature: Modeling Patterns in the Natural World. Princeton, 2006.
- Ball, Philip The Self-made Tapestry: Pattern Formation in Nature. Oxford, 2001.
- Edmaier, Bernhard Patterns of the Earth. Phaidon Press, 2007.
- Haeckel, Ernst Art Forms of Nature. Dover, 1974.
- Stevens, Peter S. Patterns in Nature. Penguin, 1974.
- Stewart, Ian. What Shape is a Snowflake? Magical Numbers in Nature. Weidenfeld & Nicolson, 2001.
- Thompson, D'Arcy W. On Growth and Form. 1942 2nd ed. (1st ed., 1917). ISBN 0-486-67135-6
In art and architecture
- Alexander, C. A Pattern Language: Towns, Buildings, Construction. Oxford, 1977.
- de Baeck, P. Patterns. Booqs, 2009.
- Garcia, M. The Patterns of Architecture. Wiley, 2009.
- Kiely, O. Pattern. Conran Octopus, 2010.
- Pritchard, S. V&A Pattern: The Fifties. V&A Publishing, 2009.
In science and mathematics
- Adam, J. A. Mathematics in Nature: Modeling Patterns in the Natural World. Princeton, 2006.
- Resnik, M. D. Mathematics as a Science of Patterns. Oxford, 1999.
In computing
- Gamma, E., Helm, R., Johnson, R., Vlissides, J. Design Patterns. Addison-Wesley, 1994.
- Bishop, C. M. Pattern Recognition and Machine Learning. Springer, 2007.
https://en.wikipedia.org/wiki/Pattern
The Peripatetic axiom is: "Nothing is in the intellect that was not first in the senses" (Latin: "Nihil est in intellectu quod non sit prius in sensu"). It is found in Thomas Aquinas's De veritate, q. 2 a. 3 arg. 19.[1]
Aquinas adopted this principle from the Peripatetic school of Greek philosophy, established by Aristotle.[where?] Aquinas argued that the existence of God could be proved by reasoning from sense data.[2] He used a variation on the Aristotelian notion of the "active intellect" ("intellectus agens")[3] which he interpreted as the ability to abstract universal meanings from particular empirical data.[4]
https://en.wikipedia.org/wiki/Peripatetic_axiom
Perspicacity (also called perspicaciousness) is a penetrating discernment (from the Latin perspicācitās, meaning throughsightedness, discrimination)—a clarity of vision or intellect which provides a deep understanding and insight.[1] It takes the concept of wisdom deeper in the sense that it denotes a keenness of sense and intelligence applied to insight. It has been described as a deeper level of internalization.[2] Another definition refers to it as the "ability to recognize subtle differences between similar objects or ideas".[3]
The artist René Magritte illustrated the quality in his 1936 painting Perspicacity. The picture shows an artist at work who studies his subject intently: it is an egg. But the painting which he is creating is not of an egg; it is an adult bird in flight.[4]
Perspicacity is also used to indicate practical wisdom in the areas of politics and finance. [5][6] Being perspicacious about other people, rather than having false illusions, is a sign of good mental health.[7] The quality is needed in psychotherapists who engage in person-to-person dialogue and counselling of the mentally ill.[8]
Perspicacity is different from acuity, which also describes a keen insight, since it does not include physical abilities such as sight or hearing.[3]
In an article dated October 7, 1966, the journal Science discussed NASA scientist-astronaut program recruitment efforts:
To quote an Academy brochure, the quality most needed by a scientist-astronaut is "perspicacity." He must, the brochure says, be able to quickly pick out, from among the thousands of things he sees, those that are significant, and to synthesize observations and develop and test working hypotheses.[9]
Concept
In 17th-century Europe, René Descartes devised systematic rules for clear thinking in his work Regulæ ad directionem ingenii (Rules for the direction of natural intelligence). In Descartes' scheme, intelligence consisted of two faculties: perspicacity, which provided an understanding or intuition of distinct detail; and sagacity, which enabled reasoning about the details in order to make deductions. Rule 9 was De Perspicacitate Intuitionis (On the Perspicacity of Intuition).[10] He summarised the rule as
Oportet ingenii aciem ad res minimas et maxime faciles totam convertere, atque in illis diutius immorari, donec assuescamus veritatem distincte et perspicue intueri.
We should totally focus the vision of the natural intelligence on the smallest and easiest things, and we should dwell on them for a long time, so long, until we have become accustomed to intuiting the truth distinctly and perspicuously.
In his study of the elements of wisdom, the modern psychometrician Robert Sternberg identified perspicacity as one of its six components or dimensions; the other five being reasoning, sagacity, learning, judgement and the expeditious use of information.[11] In his analysis, the perspicacious individual is someone who
...has intuition; can offer solutions that are on the side of right and truth; is able to see through things — read between the lines; has the ability to understand and interpret his or her environment.
— Robert J. Sternberg, Wisdom: its nature, origins, and development
See also
- Neuro-linguistic programming
- Organizational politics
- Stress management
- Time management
- Personality
- Temperament
References
{{citation}}
: |author=
has generic name (help)
- Robert J. Sternberg (1990), Wisdom, Cambridge University Press, pp. 146, 157, ISBN 978-0-521-36718-9
https://en.wikipedia.org/wiki/Perspicacity
In philosophy, a point of view is a specific attitude or manner through which a person thinks about something.[1][2] This figurative usage of the expression dates back to 1760.[3] In this meaning, the usage is synonymous with one of the meanings of the term perspective[4][5] (also epistemic perspective).[6]
The concept of the "point of view" is highly multifunctional and ambiguous. Many things may be judged from certain personal, traditional or moral points of view (as in "beauty is in the eye of the beholder"). Our knowledge about reality is often relative to a certain point of view.[4]
Vázquez Campos and Manuel Liz Gutierrez suggested to analyse the concept of "point of view" using two approaches: one based on the concept of "propositional attitudes", the other on the concepts of "location" and "access".[7]
https://en.wikipedia.org/wiki/Point_of_view_(philosophy)
In philosophy, practical reason is the use of reason to decide how to act. It contrasts with theoretical reason, often called speculative reason, the use of reason to decide what to follow. For example, agents use practical reason to decide whether to build a telescope, but theoretical reason to decide which of two theories of light and optics is the best.
https://en.wikipedia.org/wiki/Practical_reason
The preface paradox, or the paradox of the preface,[1] was introduced by David Makinson in 1965. Similar to the lottery paradox, it presents an argument according to which it can be rational to accept mutually incompatible beliefs. While the preface paradox nullifies a claim contrary to one's belief, it is opposite to Moore's paradox which asserts a claim contrary to one's belief.
https://en.wikipedia.org/wiki/Preface_paradox
In epistemology, a presupposition relates to a belief system, or Weltanschauung, that is required for the argument to make sense. A variety of Christian apologetics, called presuppositional apologetics, argues that the existence or non-existence of God is the basic presupposition of all human thought, and that all people arrive at a worldview which is ultimately determined by the theology they presuppose. Evidence and arguments are only developed after the fact in an attempt to justify the theological assumptions already made. According to this view, it is impossible to demonstrate the existence of God unless one presupposes that God exists, with the stance that modern science relies on methodological naturalism, and thus is incapable of discovering the supernatural. It thereby fashions a Procrustean bed which rejects any observation which would disprove the naturalistic assumption. Apologetics argue that the resulting worldview is inconsistent with itself and therefore irrational (for example, via the Argument from morality or via the Transcendental argument for the existence of God).
https://en.wikipedia.org/wiki/Presupposition_(philosophy)
The primary–secondary quality distinction is a conceptual distinction in epistemology and metaphysics, concerning the nature of reality. It is most explicitly articulated by John Locke in his Essay concerning Human Understanding, but earlier thinkers such as Galileo and Descartes made similar distinctions.
Primary qualities are thought to be properties of objects that are independent of any observer, such as solidity, extension, motion, number and figure. These characteristics convey facts. They exist in the thing itself, can be determined with certainty, and do not rely on subjective judgments. For example, if an object is spherical, no one can reasonably argue that it is triangular. Primary qualities as mentioned earlier, exist outside of the observer. They inhere to an object in such a way that if the object was changed, e.g. divided (if the object is divisible; a sphere is not, since dividing a sphere would result in two non-spheres), the primary qualities would remain. When dividing a divisible object, “solidity, extension, figure, and mobility” [1] would not be altered because the primary qualities are built into the object itself. Another key component of primary qualities is that they create ideas in our minds through experience; they represent the actual object. Because of this, primary qualities such as size, weight, solidity, motion, and so forth can all be measured in some form.[2] Using an apple as an example, the shape and size can actually be measured and produce the idea in our minds of what the object is. A clear distinction to make is that qualities do not exist in the mind, rather they produce ideas in our minds and exist within the objects. In the case of primary qualities, they exist inside the actual body/substance and create an idea in our mind that resembles the object.
Secondary qualities are thought to be properties that produce sensations in observers, such as color, taste, smell, and sound. They can be described as the effect things have on certain people. Secondary qualities use the power of reflection in order to be perceived by our minds. These qualities “would ordinarily be said to be only a power in rather than a quality of the object”.[3] They are sensible qualities that produce different ideas in our mind from the actual object. Going back to the example of the aforementioned apple, something such as the redness of the apple does not produce an image of the object itself, but rather the idea of red. Secondary qualities are used to classify similar ideas produced by an object. That is why when we see something “red” it is only “red” in our minds because they produce the same idea as another object. So, going back to the color of the apple, it produces an idea of red, which we classify and identify with other red ideas. Again, secondary qualities do not exist inside the mind; they are simply the powers that allow us to sense a certain object and thus ‘reflect’ and classify similar ideas.[4]
According to the theory, primary qualities are measurable aspects of physical reality; secondary qualities are subjective.
https://en.wikipedia.org/wiki/Primary%E2%80%93secondary_quality_distinction
A propositional attitude is a mental state held by an agent or organism toward a proposition.
In philosophy, propositional attitudes can be considered to be neurally-realized causally efficacious content-bearing internal states.[1]
Linguistically, propositional attitudes are denoted by a verb (e.g. "believed") governing an embedded "that" clause, for example, 'Sally believed that she had won'.
Propositional attitudes are often assumed to be the fundamental units of thought and their contents, being propositions, are true or false from the perspective of the person. An agent can have different propositional attitudes toward the same proposition (e.g., "S believes that her ice-cream is cold," and "S fears that her ice-cream is cold").
Propositional attitudes have directions of fit: some are meant to reflect the world, others to influence it.
One topic of central concern is the relation between the modalities of assertion and belief, perhaps with intention thrown in for good measure. For example, we frequently find ourselves faced with the question of whether or not a person's assertions conform to his or her beliefs. Discrepancies here can occur for many reasons, but when the departure of assertion from belief is intentional, we usually call that a lie.
Other comparisons of multiple modalities that frequently arise are the relationships between belief and knowledge and the discrepancies that occur among observations, expectations, and intentions. Deviations of observations from expectations are commonly perceived as surprises, phenomena that call for explanations to reduce the shock of amazement.
https://en.wikipedia.org/wiki/Propositional_attitude
A proof is sufficient evidence or a sufficient argument for the truth of a proposition.[1][2][3][4]
The concept applies in a variety of disciplines,[5] with both the nature of the evidence or justification and the criteria for sufficiency being area-dependent. In the area of oral and written communication such as conversation, dialog, rhetoric, etc., a proof is a persuasive perlocutionary speech act, which demonstrates the truth of a proposition.[6] In any area of mathematics defined by its assumptions or axioms, a proof is an argument establishing a theorem of that area via accepted rules of inference starting from those axioms and from other previously established theorems.[7] The subject of logic, in particular proof theory, formalizes and studies the notion of formal proof.[8] In some areas of epistemology and theology, the notion of justification plays approximately the role of proof,[9] while in jurisprudence the corresponding term is evidence,[10] with "burden of proof" as a concept common to both philosophy and law.
In most disciplines, evidence is required to prove something. Evidence is drawn from the experience of the world around us, with science obtaining its evidence from nature,[11] law obtaining its evidence from witnesses and forensic investigation,[12] and so on. A notable exception is mathematics, whose proofs are drawn from a mathematical world begun with axioms and further developed and enriched by theorems proved earlier.
Exactly what evidence is sufficient to prove something is also strongly area-dependent, usually with no absolute threshold of sufficiency at which evidence becomes proof.[13][14] In law, the same evidence that may convince one jury may not persuade another. Formal proof provides the main exception, where the criteria for proofhood are ironclad and it is impermissible to defend any step in the reasoning as "obvious" (except for the necessary ability of the one proving and the one being proven to, to correctly identify any symbol used in the proof.);[15] for a well-formed formula to qualify as part of a formal proof, it must be the result of applying a rule of the deductive apparatus of some formal system to the previous well-formed formulae in the proof sequence.[16]
Proofs have been presented since antiquity. Aristotle used the observation that patterns of nature never display the machine-like uniformity of determinism as proof that chance is an inherent part of nature.[17] On the other hand, Thomas Aquinas used the observation of the existence of rich patterns in nature as proof that nature is not ruled by chance.[18]
Proofs need not be verbal. Before Copernicus, people took the apparent motion of the Sun across the sky as proof that the Sun went round the Earth.[19] Suitably incriminating evidence left at the scene of a crime may serve as proof of the identity of the perpetrator. Conversely, a verbal entity need not assert a proposition to constitute a proof of that proposition. For example, a signature constitutes direct proof of authorship; less directly, handwriting analysis may be submitted as proof of authorship of a document.[20] Privileged information in a document can serve as proof that the document's author had access to that information; such access might in turn establish the location of the author at certain time, which might then provide the author with an alibi.
Proof vs evidence
18th-century Scottish philosopher David Hume built on Aristotle's separation of belief from knowledge,[21] recognizing that one can be said to "know" something only if one has firsthand experience with it, in a strict sense proof, while one can infer that something is true and therefore "believe" it without knowing, via evidence or supposition. This speaks to one way of separating proof from evidence:
If one cannot find their chocolate bar, and sees chocolate on their napping roommate's face, this evidence can cause one to believe their roommate ate the chocolate bar. But they do not know their roommate ate it. It may turn out that the roommate put the candy away when straightening up, but was thus inspired to go eat their own chocolate. Only if one directly experiences proof of the roommate eating it, perhaps by walking in on them doing so, does one know the roommate did it.
In an absolute sense, one can be argued not to "know" anything, except for the existence of one's own thoughts, as 17th-century philosopher John Locke pointed out.[22] Even earlier, Descartes addressed when saying cogito, ergo sum (I think, therefore I am). While Descartes was attempting to "prove" logically that the world exists, his legacy in doing so is to have shown that one cannot have such proof, because all of one's perceptions could be false (such as under the evil demon or simulated reality hypotheses). But one at least has proof of one's own thoughts existing, and strong evidence that the world exists, enough to be considered "proof" by practical standards, though always indirect and impossible to objectively confirm.
See also
- Mathematical proof
- Proof theory
- Proof of concept
- Provability logic
- Evidence, information which tends to determine or demonstrate the truth of a proposition
- Proof procedure
- Proof complexity
- Standard of proof
References
https://en.wikipedia.org/wiki/Proof_(truth)
In the field of epistemology, the problem of the criterion is an issue regarding the starting point of knowledge. This is a separate and more fundamental issue than the regress argument found in discussions on justification of knowledge.[1]
In Western philosophy the earliest surviving documentation of the problem of the criterion is in the works of the Pyrrhonist philosopher Sextus Empiricus.[1] In Outlines of Pyrrhonism Sextus Empiricus demonstrated that no criterion of truth had been established, contrary to the position of dogmatists such as the Stoics and their doctrine of katalepsis.[2] In this Sextus was repeating or building upon earlier Pyrrhonist arguments about the problem of the criterion, as Pyrrho, the founder of Pyrrhonism, had declared that "neither our sense-perceptions nor our doxai (views, theories, beliefs) tell us the truth or lie.[3]
American philosopher Roderick Chisholm in his Theory of Knowledge details the problem of the criterion with two sets of questions:
- What do we know? or What is the extent of our knowledge?
- How do we know? or What is the criterion for deciding whether we have knowledge in any particular case?
An answer to either set of questions will allow us to devise a means of answering the other. Answering the former question set first is called particularism, whereas answering the latter set first is called methodism. A third solution is skepticism, which proclaims that since one cannot have an answer to the first set of questions without first answering the second set, and one cannot hope to answer the second set of questions without first knowing the answers to the first set, we are, therefore, unable to answer either. This has the result of us being unable to justify any of our beliefs.[citation needed]
Particularist theories organize things already known and attempt to use these particulars of knowledge to find a method of how we know, thus answering the second question set. Methodist theories propose an answer to question set two and proceed to use this to establish what we, in fact, know. Classical empiricism embraces the methodist approach.[citation needed]
See also
References
- Beckwith, Christopher I. (2015). Greek Buddha: Pyrrho's Encounter with Early Buddhism in Central Asia (PDF). Princeton University Press. pp. 22–23. ISBN 9781400866328.
https://en.wikipedia.org/wiki/Problem_of_the_criterion
The problem of other minds is a philosophical problem traditionally stated as the following epistemological question: Given that I can only observe the behavior of others, how can I know that others have minds?[1] The problem is that knowledge of other minds is always indirect. The problem of other minds does not negatively impact social interactions due to people having a "theory of mind" - the ability to spontaneously infer the mental states of others - supported by innate mirror neurons,[2] a theory of mind mechanism,[3] or a tacit theory.[4] There has also been an increase in evidence that behavior results from cognition which in turn requires consciousness and the brain.
It is a problem of the philosophical idea known as solipsism: the notion that for any person only one's own mind is known to exist. The problem of other minds maintains that no matter how sophisticated someone's behavior is, that does not reasonably guarantee the same presence of thought occurring in the self.[5] However, it is often disregarded by most philosophers as outdated. Behavior is recognized to occur due to a number of processes within the brain quelling much of the debate on this problem.
Phenomenology studies the subjective experience of human life resulting from consciousness.
See also
References
- Thornton, Stephen. "Solipsism and the Problem of Other Minds". Internet Encyclopedia of Philosophy. ISSN 2161-0002. Retrieved 2021-06-02.
Further reading
- Godfrey-Smith, Peter (2016). Other minds: the octopus, the sea, and the deep origins of consciousness. New York: Farrar, Straus, and Giroux. ISBN 9780374227760. OCLC 957696590.
- Inami, Masahiro (August 2001). "The problem of other minds in the Buddhist epistemological tradition". Journal of Indian Philosophy. 29 (4): 465–483. doi:10.1023/A:1013151011789. JSTOR 23496930. S2CID 169112249.
- Avramides, Anita (2001). Other minds. The problems of philosophy. London; New York: Routledge. doi:10.4324/9780203870174. ISBN 0415033365. OCLC 44613007. S2CID 240744968.
- Dennett, Daniel C. (2017) [1978]. Brainstorms: philosophical essays on mind and psychology. A Bradford book (40th anniversary ed.). Cambridge, MA: MIT Press. doi:10.7551/mitpress/11146.001.0001. ISBN 9780262534383. OCLC 988171852.
- Buford, Thomas O., ed. (1970). Essays on other minds. Urbana, IL: University of Illinois Press. ISBN 0252001230. OCLC 133061.
- Plantinga, Alvin (1967). God and other minds: a study of the rational justification of belief in God. Contemporary philosophy. Ithaca, NY: Cornell University Press. OCLC 383698.
- Wisdom, John (1966) [1952]. Other minds (2nd ed.). Oxford: Blackwell. OCLC 1076880.
External links
- Problem of other minds at PhilPapers
- Zalta, Edward N. (ed.). "Other minds". Stanford Encyclopedia of Philosophy.
- Problem of other minds at the Indiana Philosophy Ontology Project
- "Problem of other minds". Internet Encyclopedia of Philosophy.
https://en.wikipedia.org/wiki/Problem_of_other_minds
First formulated by David Hume, the problem of induction questions our reasons for believing that the future will resemble the past, or more broadly it questions predictions about unobserved things based on previous observations. This inference from the observed to the unobserved is known as "inductive inferences", and Hume, while acknowledging that everyone does and must make such inferences, argued that there is no non-circular way to justify them, thereby undermining one of the Enlightenment pillars of rationality.[1]
While David Hume is credited with raising the issue in Western analytic philosophy in the 18th century, the Pyrrhonist school of Hellenistic philosophy and the Cārvāka school of ancient Indian philosophy had expressed skepticism about inductive justification long prior to that.
The traditional inductivist view is that all claimed empirical laws, either in everyday life or through the scientific method, can be justified through some form of reasoning. The problem is that many philosophers tried to find such a justification but their proposals were not accepted by others. Identifying the inductivist view as the scientific view, C. D. Broad once said that induction is "the glory of science and the scandal of philosophy".[2] In contrast, Karl Popper's critical rationalism claimed that inductive justifications are never used in science and proposed instead that science is based on the procedure of conjecturing hypotheses, deductively calculating consequences, and then empirically attempting to falsify them.
https://en.wikipedia.org/wiki/Problem_of_induction
The private language argument argues that a language understandable by only a single individual is incoherent, and was introduced by Ludwig Wittgenstein in his later work, especially in the Philosophical Investigations.[1] The argument was central to philosophical discussion in the second half of the 20th century.
https://en.wikipedia.org/wiki/Private_language_argument
A principle is a proposition or value that is a guide for behavior or evaluation. In law, it is a rule that has to be or usually is to be followed. It can be desirably followed, or it can be an inevitable consequence of something, such as the laws observed in nature or the way that a system is constructed. The principles of such a system are understood by its users as the essential characteristics of the system, or reflecting system's designed purpose, and the effective operation or use of which would be impossible if any one of the principles was to be ignored.[2] A system may be explicitly based on and implemented from a document of principles as was done in IBM's 360/370 Principles of Operation.
Examples of principles are, entropy in a number of fields, least action in physics, those in descriptive comprehensive and fundamental law: doctrines or assumptions forming normative rules of conduct, separation of church and state in statecraft, the central dogma of molecular biology, fairness in ethics, etc.
In common English, it is a substantive and collective term referring to rule governance, the absence of which, being "unprincipled", is considered a character defect. It may also be used to declare that a reality has diverged from some ideal or norm as when something is said to be true only "in principle" but not in fact.
As law
As moral law
A principle represents values that orient and rule the conduct of persons in a particular society. To "act on principle" is to act in accordance with one's moral ideals.[4] Principles are absorbed in childhood through a process of socialization. There is a presumption of liberty of individuals that is restrained. Exemplary principles include First, do no harm, the golden rule and the doctrine of the mean.
As a juridic law
It represents a set of values that inspire the written norms that organize the life of a society submitting to the powers of an authority, generally the State. The law establishes a legal obligation, in a coercive way; it therefore acts as principle conditioning of the action that limits the liberty of the individuals. See, for examples, the territorial principle, homestead principle, and precautionary principle.
As scientific law
Archimedes principle, relating buoyancy to the weight of displaced water, is an early example of a law in science. Another early one developed by Malthus is the population principle, now called the Malthusian principle.[5] Freud also wrote on principles, especially the reality principle necessary to keep the id and pleasure principle in check. Biologists use the principle of priority and principle of Binomial nomenclature for precision in naming species. There are many principles observed in physics, notably in cosmology which observes the mediocrity principle, the anthropic principle, the principle of relativity and the cosmological principle. Other well-known principles include the uncertainty principle in quantum mechanics and the pigeonhole principle and superposition principle in mathematics.
As axiom or logical fundament
Principle of sufficient reason
The principle states that every event has a rational explanation.[6] The principle has a variety of expressions, all of which are perhaps best summarized by the following:
- For every entity x, if x exists, then there is a sufficient explanation for why x exists.
- For every event e, if e occurs, then there is a sufficient explanation for why e occurs.
- For every proposition p, if p is true, then there is a sufficient explanation for why p is true.
However, one realizes that in every sentence there is a direct relation between the predicate and the subject. To say that "the Earth is round", corresponds to a direct relation between the subject and the predicate.
Principle of non-contradiction
According to Aristotle, “It is impossible for the same thing to belong and not to belong at the same time to the same thing and in the same respect.”[7] For example, it is not possible that in exactly the same moment and place, it rains and does not rain.[8]
Principle of excluded middle
The principle of the excluding third or "principium tertium exclusum" is a principle of the traditional logic formulated canonically by Leibniz as: either A is B or A isn't B. It is read the following way: either P is true, or its denial ¬P is.[9] It is also known as "tertium non datur" ('A third (thing) is not'). Classically it is considered to be one of the most important fundamental principles or laws of thought (along with the principles of identity, non-contradiction and sufficient reason).
See also
References
- Whitehead, Alfred North (2005). Principia mathematica, by Alfred North Whitehead ... and Bertrand Russell.
External links
- The dictionary definition of principle at Wiktionary
- Actus Essendi and the Habit of the First Principle in Thomas Aquinas (New York: Einsiedler Press, 2019).
https://en.wikipedia.org/wiki/Principle
Doubt is a mental state in which the mind remains suspended between two or more contradictory propositions, unable to be certain of any of them.[1][better source needed] Doubt on an emotional level is indecision between belief and disbelief. It may involve uncertainty, distrust or lack of conviction on certain facts, actions, motives, or decisions. Doubt can result in delaying or rejecting relevant action out of concern for mistakes or missed opportunities.
https://en.wikipedia.org/wiki/Doubt
Distinction, the fundamental philosophical abstraction, involves the recognition of difference.[1]
In classical philosophy, there were various ways in which things could be distinguished. The merely logical or virtual distinction, such as the difference between concavity and convexity, involves the mental apprehension of two definitions, but which cannot be realized outside the mind, as any concave line would be a convex line considered from another perspective. A real distinction involves a level of ontological separation, as when squirrels are distinguished from llamas (for no squirrel is a llama, and no llama is a squirrel).[2] A real distinction is thus different than a merely conceptual one, in that in a real distinction, one of the terms can be realized in reality without the other being realized.
Later developments include Duns Scotus's formal distinction, which developed in part out of the recognition in previous authors that there need to be an intermediary between logical and real distinctions.[3]
Some relevant distinctions to the history of Western philosophy include:
Distinctions in contemporary thought
Analytic–synthetic distinction
While there are anticipation of this distinction prior to Kant in the British Empiricists (and even further in Scholastic thought), it was Kant who introduced the terminology. The distinction concerns the relation of a subject to its predicate: analytic claims are those in which the subject contains the predicate, as in "All bodies are extended." Synthetic claims bring two concepts together, as in "All events are caused." The distinction was recently called into question by W.V.O. Quine, in his paper "Two Dogmas of Empiricism."
A priori and a posteriori
The origins of the distinction are less clear, and it concerns the origins of knowledge. A posteriori knowledge arises from, or is caused by, experience. A priori knowledge may come temporally after experience, but its certainty is not derivable from the experience itself. Saul Kripke was the first major thinker to propose that there are analytic a posteriori knowledge claims.
Notable distinctions in historical authors
Aristotle
Aristotle makes the distinction between actuality and potentiality.[4] Actuality is a realization of the way a thing could be, while potency refers simply to the way a thing could be. There are two levels to each: matter itself can be anything, and becomes something actually by causes, making it something which then has the ability to be in a certain way, and that ability can then be realized. The matter of an ax can be an ax, then is made into an ax. The ax thereby is able to cut, and reaches a new form of actuality in actually cutting.
Aquinas
The major distinction Aquinas makes is that of essence and existence. It is a distinction already in Avicenna, but Aquinas maps the distinction onto the actuality/potentiality distinction of Aristotle, such that the essence of a thing is in potency to the existence of a thing, which is that thing's actuality.[5]
Kant
In Kant, the distinction between appearance and thing-in-itself is foundational to his entire philosophical project.[6] The distinction separates the way a thing appears to us on the one hand, and the way a thing really is.
See also
- Use-mention distinction – Difference between using a word and mentioning it
- Type-token distinction – Distinguishing objects and classes of objects
References
- Colin, Marshall (2013). "Kant's One Self and the Appearance/Thing-in-itself Distinction". Kant-Studien. 104 (4). ISSN 0022-8877.
https://en.wikipedia.org/wiki/Distinction_(philosophy)
Direct experience or immediate experience generally denotes experience gained through immediate sense perception. Many philosophical systems hold that knowledge or skills gained through direct experience cannot be fully put into words.
See also
References
- Henri Bergson (2012). Time and Free Will: An Essay on the Immediate Data of Consciousness. Courier Corporation. ISBN 978-0-486-11971-7.
https://en.wikipedia.org/wiki/Direct_experience
Declarative knowledge is an awareness of facts that can be expressed using declarative sentences, like knowing that Princess Diana died in 1997. It is also called theoretical knowledge, descriptive knowledge, propositional knowledge, and knowledge-that. It is not restricted to one specific use or purpose and can be stored in books or on computers.
Epistemology is the main discipline studying declarative knowledge. Among other things, it investigates the essential components of declarative knowledge. According to a traditionally influential view, it has three components: it is a belief that is true and justified. As a belief, it is a subjective commitment to the accuracy of the believed claim while truth is an objective aspect. To be justified, a belief has to be rational by being based on good reasons. This means that mere guesses do not amount to knowledge even if they are true. In contemporary epistemology, various additional or alternative components have been suggested, for example, that no defeating evidence is present, that the belief was caused by a reliable cognitive process, and that the belief is infallible.
Different types of declarative knowledge can be distinguished based on the source of knowledge, the type of claim that is known, and how certain the knowledge is. A central distinction is between a posteriori knowledge, which arises from experience, and a priori knowledge, which is grounded in pure rational reflection. Other classifications include domain-specific knowledge and general knowledge, knowledge of facts, concepts, and principles as well as explicit and implicit knowledge.
Declarative knowledge is often contrasted with practical knowledge and knowledge by acquaintance. Practical knowledge consists of skills, like knowing how to ride a horse. It is a form of non-intellectual knowledge since it does not need to involve true beliefs. Knowledge by acquaintance is a familiarity with something based on first-hand experience, like knowing the taste of chocolate. This familiarity can be present even if the person does not possess any factual information about the object. Some theorists also contrast declarative knowledge with conditional knowledge, prescriptive knowledge, structural knowledge, case knowledge, and strategic knowledge.
Declarative knowledge is required for various activities, such as labeling phenomena as well as describing and explaining them. It can guide the processes of problem-solving and decision-making. In many cases, its value is based on its usefulness in achieving one's goals. However, its usefulness is not always obvious and not all instances of declarative knowledge are valuable. A lot of knowledge taught at school is declarative knowledge. It can be learned through rote memorization of individual facts but in many cases, it is advantageous to foster a deeper understanding that integrates the new information into wider structures and connects it to pre-existing knowledge. Sources of declarative knowledge are perception, introspection, memory, reasoning, and testimony.
Definition and semantic field
Declarative knowledge is an awareness or understanding of facts. It can be expressed through spoken and written language using declarative sentences and can thus be acquired through verbal communication.[1] Examples of declarative knowledge are knowing "that Princess Diana died in 1997" or "that Goethe was 83 when he finished writing Faust".[2] Declarative knowledge involves mental representations in the form of concepts, ideas, theories, and general rules. Through these representations, the person stands in a relationship to a particular aspect of reality by depicting what it is like. Declarative knowledge tends to be context-independent: it is not tied to any specific use and may be employed for many different tasks.[3][4][5] It includes a wide range of phenomena and encompasses both specific knowledge of individual facts, for example, that the atomic mass of gold is 196.97 u, as well as general laws, for example, that the color of leaves of some trees changes in autumn.[6] Due to its verbal nature, declarative knowledge can be stored in media like books and harddisks. It may also be processed using computers and plays a key role in various forms of artificial intelligence, for example, in the knowledge base of expert systems.[7]
Terms like theoretical knowledge, descriptive knowledge, propositional knowledge, and knowledge-that are used as synonyms of declarative knowledge and express its different aspects. Theoretical knowledge is knowledge of what is the case, in the past, present, or future independent of a practical outlook concerning how to achieve a specific goal. Descriptive knowledge is knowledge that involves descriptions of actual or speculative objects, events, or concepts. Propositional knowledge asserts that a certain proposition or claim about the world is true. This is often expressed using a that-clause, as in "knowing that kangaroos hop" or "knowing that 2 + 2 = 4". For this reason, it is also referred to as knowledge-that.[8] Declarative knowledge contrasts with non-declarative knowledge, which does not concern the explicit comprehension of factual information regarding the world. In this regard, practical knowledge in the form of skills and knowledge by acquaintance as a type of experiential familiarity are not forms of declarative knowledge.[9][10][11] The main discipline investigating declarative knowledge is called epistemology. It tries to determine its nature, how it arises, what value it has, and what its limits are.[12][13][14]
Components
A central issue in epistemology is to determine the components or essential features of declarative knowledge. This field of inquiry is called the analysis of knowledge. It aims to provide the conditions that are individually necessary and jointly sufficient for a state to amount to declarative knowledge. In this regard, it is similar to how a chemist breaks down a sample by identifying all the chemical elements composing it.[15][16][17]
A traditionally influential view states that declarative knowledge has three essential features: it is (1) a belief that is (2) true and (3) justified.[18][19][20] This position is referred to as the justified-true-belief conception of knowledge and is often seen as the standard view.[21][22] This view faced significant criticism following a series of counterexamples given by Edmund Gettier in the latter half of the 20th century. In response, various alternative theories of the components of declarative knowledge have been suggested. Some see justified true belief as a necessary condition that is not sufficient by itself and discuss additional components that are needed. Another response is to deny that justification is needed and seek a different component to replace it.[23][24][25] Some theorists, like Timothy Williamson, reject the idea that declarative knowledge can be deconstructed into various constituent parts. They argue instead that it is a fundamental and unanalyzable epistemological state.[26]
Belief
One commonly accepted component of knowledge is belief. In this sense, whoever knows that whales are animals automatically also believes that whales are animals. A belief is a mental state that affirms that something is the case. As an attitude toward a proposition, it belongs to the subjective side of knowledge. Some theorists, like Luis Villoro, distinguish between weak and strong beliefs. Having a weak belief implies that the person merely presumes that something is the case. They guess that the claim is probably correct while acknowledging at the same time that they might very well be mistaken about it. This contrasts with strong belief, which implies a substantial commitment to the believed claim. It involves certainty in the form of being sure about it. For declarative knowledge, this stronger sense of belief is relevant.[27]
A few epistemologists, like Katalin Farkas, claim that, at least in some cases, knowledge is not a form of belief but a different type of mental state. One argument for this position is based on statements like "I don't believe it, I know it", which may be used to express that the person is very certain and has good reason to affirm this claim. However, this argument is not generally accepted since knowing something does not imply that the person disbelieves the claim. A different explanation is to hold that this statement is a linguistic tool to emphasize that the person is well-informed. In this regard, it only denies that a weak belief exists without rejecting that a stronger form of belief is involved.[28]
Truth
Beliefs are either true or false depending on whether they accurately represent reality. Truth is usually seen as one of the essential components of knowledge. This means that it is impossible to know a claim that is false. For example, it is possible to believe that Hillary Clinton won the 2016 US Presidential election but nobody can know it because this event did not occur. That a proposition is true does not imply that it is common knowledge, that an irrefutable proof exists, or that someone is thinking about it. Instead, it only means that it presents things as they are. For example, when flipping a coin, it may be true that it will land heads even if it is not possible to predict this with certainty. Truth is an objective factor of knowledge that goes beyond the psychological sphere of belief since it usually depends on what the world outside the person's mind is like.[29][30][31]
Some epistemologists hold that there are at least some forms of knowledge that do not require truth. For example, Joseph Thomas Tolliver argues that certain mental states amount to knowledge only because of the causes and effects they have even though they do not represent anything and are therefore neither true nor false.[31][32] A different outlook is found in the field of the anthropology of knowledge, which studies how knowledge is acquired, stored, retrieved, and communicated. In this discipline, knowledge is often understood in a very wide sense that is roughly equivalent to understanding and culture. In this regard, the main interest is usually about how people ascribe truth values to meaning-contents, like when affirming an assertion, independent of whether this assertion is true or false.[33][34][35] Despite these positions, it is widely accepted in epistemology that truth is an essential component of declarative knowledge.[29]
Justification
In epistemology, justification means that a proposition is supported by evidence or that a person has good reasons for believing it. This implies some form of appraisal in relation to an evaluative standard of rationality.[36][37] For example, a person who just checked their bank account and saw that their balance is 500 dollars has a good reason to believe that they have 500 dollars in their bank account.[38] However, justification by itself does not imply that a belief is true. For example, if someone reads the time from their clock they may form a justified belief about the current time even if the clock stopped a while ago and shows a false time now.[39] If a person has a justified belief then they are often able to articulate what this belief is and to provide arguments stating the reasons supporting it. However, this ability to articulate one's reasons is not an essential requirement of justification.[37]
Justification is usually included as a component of knowledge to exclude lucky guesses. For example, a compulsive gambler flipping a coin may be certain that it will land heads this time without a good reason for this belief. In this case, the belief does not amount to knowledge even if it turns out that it was true. This observation can be easily explained by including justification as an essential component: the gambler's belief does not amount to knowledge because it lacks justification. In this regard, mere true opinion is not enough to establish knowledge. A central issue in epistemology concerns the standards of justification, i.e., what conditions have to be fulfilled for a belief to be justified. Internalists understand justification as a purely subjective component, akin to belief. They claim that a belief is justified if it stands in the right relation to other mental states of the believer. For example, perceptual experiences can justify beliefs about the perceived object. This contrasts with externalists, who claim that justification involves objective factors that are external to the person's mind. Such factors can include causal relations with the object of the belief or that reliable cognitive processes are responsible for the formation of the belief.[40][41]
A closely related issue concerns the question of how the different mental states have to be related to each other to be justified. For example, one belief may be supported by another belief. However, it is questionable whether this is sufficient for justification if the second belief is itself not justified. For example, a person may believe that Ford cars are cheaper than BMWs because they heard this from a friend. However, this belief may not be justified if there is no good reason to think that the friend is a reliable source of information. This can lead to an infinite regress since whatever reason is provided for the friend's reliability may itself lack justification. Three popular responses to this problem are foundationalism, coherentism, and infinitism. According to foundationalists, some reasons are foundational and do not depend on other reasons for their justification. Coherentists also reject the idea that an infinite chain of reasons is needed and argue that different beliefs can mutually support each other without one being more basic than the others. Infinitists, on the other hand, accept the idea that an infinite chain of reasons is required.[42]
Many debates concerning the nature of declarative knowledge focus on the role of justification, specifically whether it is needed at all and what else might be needed to complement it. Influential in this regard was a series of thought experiments by Edmund Gettier. They present concrete cases of justified true beliefs that fail to amount to knowledge. The reason for their failure is a type of epistemic luck. This means that the justification is not relevant to whether the belief is true. In one thought experiment, Smith and Jones apply for a job and before officially declaring the result, the company president tells Smith that Jones will get the job. Smith saw that Jones has 10 coins in his pocket so he comes to form the justified belief that the successful candidate has 10 coins in his pocket. In the end, it turns out that Smith gets the job after all. By lucky coincidence, Smith also has 10 coins in his pocket. Gettier claims that, because of this coincidence, Smith's belief that the successful candidate has 10 coins in his pocket does not amount to knowledge even though it is justified and true because the justification is not relevant to the truth.[43][44]
Others
In response to Gettier's thought experiments, various further components of declarative knowledge have been suggested. Some of them are intended as additional elements besides belief, truth, and justification while others are understood as replacements for justification.[45][46][47]
According to defeasibility theory, an additional condition besides having evidence in favor of the belief is that no defeating evidence is present. Defeating evidence of a belief is evidence that undermines the justification of the belief. For example, if a person looks outside the window and sees a rainbow then this impression justifies their belief that there is a rainbow. However, if the person just ate a psychedelic drug then this is defeating evidence since it undermines the reliability of their experiences. Defeasibility theorists claim that, in this case, the belief does not amount to knowledge because defeating evidence is present. As an additional component of knowledge, they require that the person has no defeating evidence of the belief.[48][49][50] Some theorists demand the stronger requirement that there is no true proposition that would defeat the belief, independent of whether the person is aware of this proposition or not.[51] A closely related theory holds that beliefs can only amount to knowledge if they are not inferred from a falsehood.[52]
A different theory is based on the idea that knowledge states should be responsive to what the world is like. One suggested component in this regard is that the belief is safe or sensitive. This means that the person has the belief because it is true but that they would not hold the belief if it was false. In this regard, the person's belief tracks the state of the world.[53]
Some theories do not try to provide additional requirements but instead propose replacing justification with alternative components. For example, according to some forms of reliabilism, a true belief amounts to knowledge if it was formed through a reliable cognitive process. A cognitive process is reliable if it produces mostly true beliefs in actual situations and would also do so in counterfactual situations. [47][54][55] Examples of reliable processes are perception and reasoning.[56] A consequence of reliabilism is that knowledge is not restricted to humans since reliable belief-formation processes may also be present in other animals, like dogs, apes, or rats, even if they do not possess justification for their beliefs.[47] Virtue epistemology is a closely related approach that understands knowledge as the manifestation of epistemic virtues. It agrees with regular forms of reliabilism that knowledge is not a matter of luck but puts additional emphasis on the evaluative aspect of knowledge and the underlying skills responsible for it.[57][58][59]
According to causal theories of knowledge, a necessary element of knowing a fact is that this fact somehow caused the knowledge of it. This is the case, for example, if a belief about the color of a house is based on a perceptual experience, which causally connects the house to the belief. This causal connection does not have to be direct and can be mediated through different steps like activating memories and drawing inferences.[60][47]
In many cases, the goal of suggesting additional components is to avoid cases of epistemic luck. In this regard, some theorists have argued that the additional component would have to ensure that the belief is true. This approach is reflected in the idea that knowledge implies a form of certainty. But it sets the standards of knowledge very high and may require that a belief has to be infallible to amount to knowledge. This means that the justification ensures that the belief is true. For example, Richard Kirkham argues that the justification required for knowledge must be based on self-evident premises that deductively entail the held belief. Such a position leads to a form of skepticism about knowledge since the great majority of regular beliefs do not live up to these requirements. It would imply that people know very little and that most who claim to know a certain fact are mistaken. However, a more common view among epistemologists is that knowledge does not require infallibility and that many knowledge claims in everyday life are true.[61]
Types
Declarative knowledge arises in different forms. It is possible to distinguish between them based on the type of content of what is known. For example, empirical knowledge is knowledge of observable facts while conceptual knowledge is an understanding of general categorizations and theories as well as the relations between them.[62][63][64] Other examples are ethical, religious, scientific, mathematical, and logical knowledge as well as self-knowledge. A different distinction focuses on the mode of how something is known. On a causal level, different sources of knowledge correspond to different types of declarative knowledge. Examples are knowledge through perception, introspection, memory, reasoning, and testimony.[62][65][66] On a logical level, forms of knowledge can be distinguished based on how a knowledge claim is supported by its premises. This classification corresponds to the different forms of logical reasoning, such as deductive and inductive reasoning.[62][67][68] A closely related categorization focuses on the strength of the source of the justification and distinguishes between probabilistic and apodictic knowledge while the distinction between a priori and a posteriori knowledge focuses on the type of the source. These different classifications overlap with each other at various points. For example, a priori knowledge is closely connected to apodictic, conceptual, deductive, and logical knowledge. A posteriori knowledge, on the other hand, is associated with probabilistic, empirical, inductive, and scientific knowledge. Self-knowledge may be identified with introspective knowledge.[62][69]
The distinction between a priori and a posteriori knowledge is determined by the role of experience and matches the distinction between empirical and non-empirical knowledge. A posteriori knowledge is knowledge from experience. This means that experience, like regular perception, is responsible for its formation and justification. Knowing that the door of one's house is green is one example of a posteriori knowledge since some form of sensory observation is required. For a priori knowledge, on the other hand, no experience is required. It is based on pure rational reflection and can neither be verified nor falsified through experience. Examples are knowing that 7 + 5 = 12 or that whatever is red everywhere is not blue everywhere.[70] In this context, experience means primarily sensory observation but can also include related processes, like introspection and memory. However, it does not include all conscious phenomena. For example, having a rational insight into the solution of a mathematical problem does not mean that the resulting knowledge is a posteriori. And knowing that 7 + 5 = 12 is a priori knowledge even though some form of consciousness is involved in learning what symbols like "7" and "+" mean and in becoming aware of the associated concepts.[71][72][69]
One classification distinguishes between knowledge of facts, concepts, and principles. Knowledge of facts pertains to the association of concrete information, for example, that the red color on a traffic light means stop or that Christopher Columbus sailed in 1492 from Spain to America. Knowledge of concepts applies to more abstract and general ideas that group together many individual phenomena. For example, knowledge of the concept of jogging implies knowing how it differs from walking and running as well as being able to apply this concept to concrete cases. Knowledge of principles is an awareness of general patterns of cause and effect, including rules of thumb. It is a form of understanding how things work and being aware of the explanation of why something happened the way it did. Examples are that if there is lightning then there will be thunder or if a person robs a bank then they may go to jail.[73][74] Similar classifications distinguish between declarative knowledge of persons, events, principles, maxims, and norms.[75][76][77]
Declarative knowledge is traditionally identified with explicit knowledge and contrasted with tacit or implicit knowledge. Explicit knowledge is knowledge of which the person is aware and which can be articulated. Implicit knowledge, on the other hand, is a form of embodied knowledge that the person cannot articulate. The traditional association of declarative knowledge with explicit knowledge is not always accepted in the contemporary literature and some theorists have argued that there are forms of implicit declarative knowledge. A putative example is a person who has learned a concept and is now able to correctly classify objects according to this concept even though they are not able to provide a verbal rationale for their decision.[78][79][80]
A further distinction is between domain-specific and general knowledge. Domain-specific knowledge applies to a narrow subject or a particular task but is useless outside this focus. General knowledge, on the other hand, concerns wide topics or has general applications. For example, declarative knowledge of the rules of grammar belongs to general knowledge while having memorized the lines of the poem The Raven is domain-specific knowledge. This distinction is based on a continuum of cases that are more or less general without a clear-cut line between the types.[6][81] According to Paul Kurtz, there are six types of descriptive knowledge: knowledge of available means, of consequences, of particular facts, of general causal laws, of established values, and of fundamental needs.[82] Another classification distinguishes between structural knowledge and perceptual knowledge.[83]
Contrast with other forms of knowledge
Declarative knowledge is often contrasted with other types of knowledge. A common classification in epistemology distinguishes it from practical knowledge and knowledge by acquaintance. All of them can be expressed with the verb "to know" but their differences are reflected in the grammatical structures used to articulate them. Declarative knowledge is usually expressed with a that-clause, as in "Ann knows that koalas sleep most of the time". For practical knowledge, a how-clause is used instead, for example, "Dave knows how to read the time on a clock". Knowledge by acquaintance can be articulated using a direct object without a preposition, as in "Emily knows Obama personally".[84]
Practical knowledge consists of skills. Knowing how to ride a horse or how to play the guitar are forms of practical knowledge. The terms "procedural knowledge" and "knowledge-how" are often used as synonyms.[11][85][86] It differs from declarative knowledge in various aspects. It is usually imprecise and cannot be proven by deducing it from premises. It is non-propositional and, for the most part, cannot be taught in abstract without concrete exercise. In this regard, it is a form of non-intellectual knowledge.[87][10] It is tied to a specific goal and its value lies not in being true, but rather in how effective it is to accomplish its goal.[88] Practical knowledge can be present without any beliefs and may even involve false beliefs. For example, a ball player may know how to catch a ball despite falsely believing that their eyes continuously track the ball while, in truth, their eyes perform a series of abrupt movements that anticipate the ball's trajectory rather than following it.[89] Another difference is that declarative knowledge is commonly only ascribed to animals with highly developed minds, like humans. Practical knowledge, on the other hand, is more prevalent in the animal kingdom. For example, ants know how to walk through the kitchen despite presumably lacking the mental capacity for the declarative knowledge that they are walking through the kitchen.[90]
Declarative knowledge is also distinguished from knowledge by acquaintance, which is also known as objectual knowledge, and knowledge-of. Knowledge by acquaintance is a form of familiarity or direct awareness that a person has with another person, a thing, or a place. For example, a person who has tasted the flavor of chocolate knows chocolate in this sense, just like a person who visited Lake Taupō knows Lake Taupō. Knowledge by acquaintance does not imply that the person can provide factual information about the object. It is a form of non-inferential knowledge that depends on first-hand experience. For example, a person who has never left their home country may acquire a lot of declarative knowledge about other countries by reading books without any knowledge by acquaintance.[86][91][92] Knowledge by acquaintance plays a central role in the epistemology of Bertrand Russell. He holds that it is more basic than other forms of knowledge since to understand a proposition, one has to be acquainted with its constituents. According to Russell, knowledge by acquaintance covers a wide range of phenomena, such as thoughts, feelings, desires, memory, introspection, and sense data. It can happen in relation to particular things and universals. Knowledge of physical objects, on the other hand, belongs to declarative knowledge, which he calls knowledge by description. It also has a central role to play since it extends the realm of knowledge to things that lie beyond the personal sphere of experience.[93]
Some theorists, like Anita Woolfolk et. al., distinguish declarative knowledge and procedural knowledge from conditional knowledge. According to this view, conditional knowledge is about knowing when and why to use declarative and procedural knowledge. For many issues, like solving math problems and learning a foreign language, it is not sufficient to know facts and general procedures if the person does not know under which situations to use them. To master a language, for example, it is not enough to acquire declarative knowledge of different verb forms if one lacks conditional knowledge of when it is appropriate to use them. Some theorists understand conditional knowledge as one type of declarative knowledge and not as a distinct category.[94]
A further distinction is between declarative or descriptive knowledge in contrast to prescriptive knowledge. Descriptive knowledge represents what the world is like. It describes and classifies what phenomena are there and in what relations they stand toward each other. It is interested in what is true independently of what people want. Prescriptive knowledge is not about what things actually are like but what they should be like. This concerns specifically the question of what purposes people should follow and how they should act. It guides action by showing what people should do to fulfill their needs and desires. In this regard, it has a more subjective component since it depends on what people want. Some theorists equate prescriptive knowledge with procedural knowledge but others distinguish them based on the claim that prescriptive knowledge is about what should be done while procedural knowledge is about how to do it.[95] Other classifications contrast declarative knowledge with structural knowledge, meta knowledge, heuristic knowledge, control knowledge, case knowledge, and strategic knowledge.[96][76][77]
Some theorists argue that one type of knowledge is more fundamental than others. For example, Robert E. Haskell claims that declarative knowledge is the basic form of knowledge since it constitutes a general framework of understanding and thereby is a precondition for acquiring other forms of knowledge.[97] However, this position is not generally accepted and philosophers like Gilbert Ryle defend the opposing thesis that declarative knowledge presupposes procedural knowledge.[98][99]
Value
Declarative knowledge plays a central role in human understanding of the world. It underlies activities such as labeling phenomena, describing them, explaining them, and communicating with others about them.[100] The value of declarative knowledge depends in part on its usefulness in helping people achieve their objectives. For example, to treat a disease, knowledge of its symptoms and possible cures is beneficial. Or if a person has applied for a new job then knowing where and when the interview takes place is important.[101][102][103] Due to its context-independence, declarative knowledge can be used for a great variety of tasks and because of its compact nature, it can be easily stored and retrieved.[4][3] Declarative knowledge can be useful for procedural knowledge, for example, by knowing the list of steps needed to execute a skill. It also has a key role in understanding and solving problems and can guide the process of decision-making.[104][105][106] A related issue in the field of epistemology concerns the question of whether declarative knowledge is more valuable than true belief, since, for most purposes, true belief seems to be as useful as knowledge to achieve one's goals.[103][107][108]
Declarative knowledge is primarily desired in cases where it is immediately useful.[97] But not all forms of knowledge are useful. For example, indiscriminately memorizing phone numbers found in a foreign phone book is unlikely to result in useful declarative knowledge.[102] However, it is often difficult to assess the value of knowledge if one does not foresee a situation where it would be useful. In this regard, it can happen that the value of apparently useless knowledge is only discovered much later. For example, Maxwell's equations linking magnetism to electricity were considered useless at the time of discovery until experimental scientists discovered how to detect electromagnetic waves.[97] Occasionally, knowledge may have a negative value, for example, when it hinders someone to do what would be needed because their knowledge of associated dangers paralyzes them.[102]
Learning
The value of knowledge is specifically relevant in the field of education to make it possible to decide which of the vast amount of knowledge should become part of the curriculum to be passed on to students.[101] Many types of learning at school involve the acquisition of declarative knowledge.[100] One form of declarative knowledge learning is so-called rote learning. It is a memorization technique in which the claim to be learned is repeated again and again until it is fully memorized. Other forms of declarative knowledge learning focus more on developing an understanding of the subject. This means that the learner should not only be able to repeat the claim but also to explain, describe, and summarize it. For declarative knowledge to be useful, it is often advantageous if it is embedded in a meaningful structure. For example, learning about new concepts and ideas involves developing an understanding of how they are related to each other and to what is already known.[104]
According to Ellen Gagné, learning declarative knowledge happens in four steps. In the first step, the learner comes into contact with the material to be learned and apprehends it. Next, they translate this information into propositions. Following that, the learner's memory triggers and activates related propositions. As the last step, new connections are established and inferences are drawn.[104] A similar process is described by John V. Dempsey, who emphasizes that the new information must be organized, subdivided, and linked to existing knowledge. He distinguishes between learning that involves recalling information in contrast to learning that only requires being able to recognize certain patterns.[109] A related theory is defended by Anthony J. Rhem, who holds that the process of learning declarative knowledge involves organizing new information into groups and drawing relations between these groups as well as connecting the new information to pre-existing knowledge.[110]
Some theorists, like Robert Gagné and Leslie Briggs, distinguish between different types of declarative knowledge learning based on the cognitive processes involved: learning of labels and names, of facts and lists, and of organized discourse. Learning labels and names requires forming a mental connection between two elements. Examples include memorizing foreign vocabulary and learning the capital city of each state. Learning facts involves relationships between concepts, for example, that "Ann Richards was the governor of Texas in 1991". This process is usually easier if the person is not dealing with isolated facts but possesses a network of information into which the new fact is integrated. The case for learning lists is similar since it involves the association of many items. Learning organized discourse encompasses not discrete facts or items but a wider comprehension of the meaning present in an extensive body of information.[104][109][110]
Various sources of declarative knowledge are discussed in epistemology. They include perception, introspection, memory, reasoning, and testimony.[65][66][62] Perception is usually understood as the main source of empirical knowledge. It is based on the senses, like seeing that it is raining when looking out the window.[111][112][113] Introspection is similar to perception but provides knowledge of the internal sphere and not of external objects.[114] An example is directing one's attention to a pain in one's toe to assess whether it has intensified.[115] Memory differs from perception and introspection in that it does not produce new knowledge but merely stores and retrieves pre-existing knowledge. As such, it depends on other sources.[66][116][117] It is similar to reasoning in this regard, which starts from a known fact and arrives at new knowledge by drawing inferences from it. Empiricists hold that this is the only way how reason can arrive at knowledge while rationalists contend that certain claims can be known by pure reason independent of additional sources.[113][118][119] Testimony is different from the other sources since it does not have its own cognitive faculty. Rather, it is grounded in the notion that people can acquire knowledge through communication with others, for example, by speaking to someone or by reading a newspaper.[120][121][122] Some religious philosophers include religious experiences (through the so-called sensus divinitatis) as a source of knowledge of the divine. However, such claims are controversial.[66][123]
References
Citations
- Plantinga 2000, pp. 182–183.
Sources
- Allwood, Carl Martin (17 October 2013). "Anthropology of Knowledge". The Encyclopedia of Cross-Cultural Psychology. John Wiley & Sons, Inc. pp. 69–72. doi:10.1002/9781118339893.wbeccp025. ISBN 978-1-118-33989-3. Archived from the original on 26 September 2022. Retrieved 16 April 2023.
- Alter, Torin; Nagasawa, Yujin (1 April 2015). Consciousness in the Physical World: Perspectives on Russellian Monism. Oxford University Press. pp. 93–94. ISBN 978-0-19-992736-4. Archived from the original on 1 April 2023. Retrieved 16 April 2023.
- Audi, Robert (20 June 2005). Epistemology: A Contemporary Introduction to the Theory of Knowledge. Routledge. p. 315. ISBN 978-1-134-79181-1. Archived from the original on 9 April 2023. Retrieved 16 April 2023.
- Audi, Robert (2002). "The Sources of Knowledge". The Oxford Handbook of Epistemology. Oxford University Press. pp. 71–94. ISBN 978-0-19-513005-8. Archived from the original on 12 June 2022. Retrieved 12 June 2022.
- Baehr, Jason S. "A Priori and A Posteriori". Internet Encyclopedia of Philosophy. 1. An Initial Characterization. Archived from the original on 7 August 2021. Retrieved 17 September 2022.
- Baehr, Jason S. "Virtue Epistemology". Internet Encyclopedia of Philosophy. introduction; 2. Virtue Reliabilism. Archived from the original on 1 April 2023. Retrieved 31 March 2023.
- Barber, Alex; Stainton, Robert J. (6 April 2010). Concise Encyclopedia of Philosophy of Language and Linguistics. Elsevier. p. 11. ISBN 978-0-08-096501-7. Archived from the original on 6 April 2023. Retrieved 16 April 2023.
- Barth, Fredrik (February 2002). "An Anthropology of Knowledge". Current Anthropology. 43 (1): 1–18. doi:10.1086/324131. hdl:1956/4191. ISSN 0011-3204.
- Battaly, Heather (4 September 2018). The Routledge Handbook of Virtue Epistemology. Routledge. p. 772. ISBN 978-1-317-49528-4. Archived from the original on 1 April 2023. Retrieved 16 April 2023.
- Becker, Kelly (13 May 2013). Epistemology Modalized. Routledge. p. 12. ISBN 978-1-136-78632-7. Archived from the original on 31 March 2023. Retrieved 16 April 2023.
- Bengson, John; Moffett, Marc A. (6 January 2012). Knowing How: Essays on Knowledge, Mind, and Action. Oxford University Press. p. 328. ISBN 978-0-19-045283-4. Archived from the original on 4 April 2023. Retrieved 16 April 2023.
- Bernecker, Sven; Pritchard, Duncan (19 January 2011). The Routledge Companion to Epistemology. Routledge. p. 266. ISBN 978-1-136-88200-5. Archived from the original on 31 March 2023. Retrieved 16 April 2023.
- Bishop, M. J.; Boling, Elizabeth; Elen, Jan; Svihla, Vanessa (21 September 2020). Handbook of Research in Educational Communications and Technology: Learning Design. Springer Nature. p. 74. ISBN 978-3-030-36119-8. Archived from the original on 17 April 2023. Retrieved 17 April 2023.
We follow Markauskaite and Goodyear (2017) and Dohn (2017) in distinguishing between three knowledge types. Declarative knowledge is knowledge expressible in propositional statements, often also termed propositional knowledge or know-that (e.g., "President Donald Trump was inaugurated on 20.01.2017," "Karl Marx wrote Capital,' "Force = Mass x Acceleration'). Procedural knowledge is often called practical knowledge, skill, or know-how (e.g., riding a bicycle, performing surgery, carrying out a logical deduction). Relational knowledge, comprising experiential and contextual knowledge, sometimes just called experiential knowledge, knowledge-by-acquaintance, or know-of (e.g., knowing what kangaroo tastes like, what red looks like, and how local, cultural norms delimit what it is appropriate to say in a given context).
- Blaauw, Martijn (31 March 2020). Epistemology A-Z. Edinburgh University Press. p. 49. ISBN 978-0-7486-8082-5. Archived from the original on 4 April 2023. Retrieved 16 April 2023.
- Black, C. (1 April 1971). "Knowledge without belief". Analysis. 31 (5): 152–158. doi:10.1093/analys/31.5.152.
- Borges, Rodrigo; Almeida, Claudio de; Klein, Peter D. (1 December 2017). Explaining Knowledge: New Essays on the Gettier Problem. Oxford University Press. p. 180. ISBN 978-0-19-103682-8. Archived from the original on 22 April 2023. Retrieved 16 April 2023.
- Broadbent, Alex (5 February 2016). Philosophy for Graduate Students: Metaphysics and Epistemology. Routledge. p. 128. ISBN 978-1-317-39714-4. Archived from the original on 7 April 2023. Retrieved 16 April 2023.
- Bronkhorst, Hugo; Roorda, Gerrit; Suhre, Cor; Goedhart, Martin (December 2020). "Logical Reasoning in Formal and Everyday Reasoning Tasks". International Journal of Science and Mathematics Education. 18 (8): 1673–6. Bibcode:2020IJSME..18.1673B. doi:10.1007/s10763-019-10039-8. S2CID 254541202.
- Burgin, Mark (2016). Theory of Knowledge: Structures and Processes. Kackensack, NJ: World Scientific. p. 48. ISBN 9789814522670.
- Burstein, Frada; Holsapple, Clyde W. (22 January 2008). Handbook on Decision Support Systems 1: Basic Themes. Springer Science & Business Media. pp. 44–45. ISBN 978-3-540-48713-5. Archived from the original on 30 March 2023. Retrieved 16 April 2023.
- Cameron, Ross (2018). "Infinite Regress Arguments". Stanford Encyclopedia of Philosophy. Metaphysics Research Lab, Stanford University. Archived from the original on 2 January 2020. Retrieved 12 June 2022.
- Campbell, Joseph Keim; O'Rourke, Michael; Silverstein, Harry S. (21 May 2010). Knowledge and Skepticism. MIT Press. p. 10. ISBN 978-0-262-01408-3. Archived from the original on 4 April 2023. Retrieved 16 April 2023.
- Carter, J. Adam; Gordon, Emma C.; Jarvis, Benjamin W. (2017). Knowledge First: Approaches in Epistemology and Mind. Oxford University Press. p. 114. ISBN 978-0-19-871631-0. Archived from the original on 2023-03-30. Retrieved 2023-04-16.
- Cassirer, H. W. (11 March 2021). A Commentary of Kant's Critique of Judgment. Routledge. p. 208. ISBN 978-1-317-20272-1. Archived from the original on 6 April 2023. Retrieved 16 April 2023.
- Chen, Fang; Terken, Jacques (18 August 2022). Automotive Interaction Design: From Theory to Practice. Springer Nature. p. 49. ISBN 978-981-19-3448-3. Archived from the original on 15 April 2023. Retrieved 16 April 2023.
- Chiu, Chi-Yue; Hong, Ying-yi (16 December 2013). Social Psychology of Culture. Psychology Press. p. 102. ISBN 978-1-317-71018-9. Archived from the original on 1 April 2023. Retrieved 16 April 2023.
- Christensen, Peter Holdt (2003). Knowledge Management: Perspectives and Pitfalls. Copenhagen Business School Press DK. p. 29. ISBN 978-87-630-0119-9. Archived from the original on 2023-04-01. Retrieved 2023-04-16.
- Cohen, Emma (2010). "Anthropology of knowledge". The Journal of the Royal Anthropological Institute. 16: S193–S202. doi:10.1111/j.1467-9655.2010.01617.x. hdl:11858/00-001M-0000-0012-9B72-7. JSTOR 40606072. Archived from the original on 2022-09-26. Retrieved 2023-04-16.
- Colman, Andrew M. (1 January 2009a). "declarative knowledge". A Dictionary of Psychology. Oxford University Press. ISBN 978-0-19-953406-7. Archived from the original on 30 March 2023. Retrieved 16 April 2023.
- Colman, Andrew M. (1 January 2009b). "non-declarative knowledge". A Dictionary of Psychology. Oxford University Press. ISBN 978-0-19-953406-7. Archived from the original on 30 March 2023. Retrieved 16 April 2023.
- Cornelis, Gustaaf C.; Smets, Sonja; Bendegem, Jean-Paul van (11 November 2013). Metadebates on Science: The Blue Book of "Einstein Meets Magritte". Springer Science & Business Media. p. 37. ISBN 978-94-017-2245-2. Archived from the original on 22 April 2023. Retrieved 16 April 2023.
- Craig, Edward, ed. (1996). "Knowledge, defeasibility theory of". Routledge Encyclopedia of Philosophy. Routledge. ISBN 9780415073103. Archived from the original on 2023-03-31. Retrieved 2023-04-16.
- Crumley, Jack S. (30 July 2009). An Introduction to Epistemology - Second Edition. Broadview Press. p. 117. ISBN 978-1-4604-0116-3. Archived from the original on 31 March 2023. Retrieved 16 April 2023.
- Degenhardt, M. A. B. (13 August 2019). Education and the Value of Knowledge. Routledge. p. 1-6. ISBN 978-1-000-62799-2. Archived from the original on 5 April 2023. Retrieved 16 April 2023.
- Dempsey, John V. (1993). Interactive Instruction and Feedback. Educational Technology. pp. 80–81. ISBN 978-0-87778-260-5. Archived from the original on 2023-04-22. Retrieved 2023-04-16.
- Dunlap, Linda L. (2004). What All Children Need: Theory and Application. University Press of America. pp. 144–145. ISBN 978-0-7618-2925-6. Archived from the original on 2023-04-07. Retrieved 2023-04-16.
- Earley, P. Christopher; Ang, Soon (2003). Cultural Intelligence: Individual Interactions Across Cultures. Stanford University Press. p. 109. ISBN 978-0-8047-4312-9. Archived from the original on 2023-04-08. Retrieved 2023-04-16.
- Evans, Ian; Smith, Nicholas D. (25 April 2013). Knowledge. John Wiley & Sons. pp. 32–33. ISBN 978-0-7456-6141-4. Archived from the original on 31 March 2023. Retrieved 16 April 2023.
- Farkas, Katalin (February 2015). "Belief May Not Be a Necessary Condition for Knowledge". Erkenntnis. 80 (1): 185–200. doi:10.1007/s10670-014-9620-2. S2CID 254470269.
- Finlay, Janet (29 October 2020). An Introduction To Artificial Intelligence. CRC Press. p. 13. ISBN 978-1-000-15403-0. Archived from the original on 15 April 2023. Retrieved 16 April 2023.
- Fischer, Frank (4 June 2019). Politics, Values, And Public Policy: The Problem Of Methodology. Routledge. p. 66. ISBN 978-1-000-30762-7. Archived from the original on 1 April 2023. Retrieved 16 April 2023.
- Flick, Uwe (10 December 2013). The SAGE Handbook of Qualitative Data Analysis. SAGE. p. 123. ISBN 9781446296691. Archived from the original on 10 February 2023. Retrieved 16 April 2023.
- Foshay, Wellesley R.; Silber, Kenneth H. (19 November 2009). Handbook of Improving Performance in the Workplace, Instructional Design and Training Delivery. John Wiley & Sons. pp. 14–15. ISBN 978-0-470-52506-7. Archived from the original on 1 April 2023. Retrieved 16 April 2023.
- Foxall, Gordon (16 October 2017). Context and Cognition in Consumer Psychology: How Perception and Emotion Guide Action. Routledge. p. 75. ISBN 978-1-317-67738-3. Archived from the original on 1 April 2023. Retrieved 16 April 2023.
- Freitas, Sara de; Jameson, Jill (5 April 2012). The e-Learning Reader. Bloomsbury Publishing. p. 189. ISBN 978-1-4411-7679-0. Archived from the original on 15 April 2023. Retrieved 16 April 2023.
- Gardiner, J. M. (29 September 2001). "Episodic memory and autonoetic consciousness: a first-person approach". Philosophical Transactions of the Royal Society of London. Series B, Biological Sciences. 356 (1413): 1351–1361. doi:10.1098/rstb.2001.0955. ISSN 0962-8436. PMC 1088519. PMID 11571027.
- Gaskins, Irene West (3 May 2005). Success with Struggling Readers: The Benchmark School Approach. Guilford Press. p. 51. ISBN 978-1-59385-169-9. Archived from the original on 1 April 2023. Retrieved 16 April 2023.
- Goldberg, Elkhonon (16 February 2006). The Wisdom Paradox: How Your Mind Can Grow Stronger As Your Brain Grows Older. Penguin. pp. 121–122. ISBN 978-1-4406-2695-1. Archived from the original on 6 April 2023. Retrieved 16 April 2023.
- Goldman, Alvin I. (1992). Liaisons: Philosophy Meets the Cognitive and Social Sciences. MIT Press. pp. 105–106. ISBN 978-0-262-07135-2. Archived from the original on 2023-03-31. Retrieved 2023-04-16.
- Green, Christopher R. "Epistemology of Testimony". Internet Encyclopedia of Philosophy. introduction. Archived from the original on 7 March 2022. Retrieved 8 June 2022.
- Gupta, Anil (2021). "Definitions: 1.1 Real and nominal definitions". Stanford Encyclopedia of Philosophy. Metaphysics Research Lab, Stanford University. Archived from the original on 1 May 2022. Retrieved 28 May 2022.
- Hacker, P. M. S. (1 July 2013). The Intellectual Powers: A Study of Human Nature. John Wiley & Sons. p. 211. ISBN 978-1-118-60906-4. Archived from the original on 30 March 2023. Retrieved 16 April 2023.
- Hamilton, Christopher (2003). Understanding Philosophy for AS Level. Nelson Thornes. p. 23. ISBN 978-0-7487-6560-7. Archived from the original on 2023-04-04. Retrieved 2023-04-16.
- HarperCollins staff. "knowledge". The American Heritage Dictionary. HarperCollins. Archived from the original on 29 November 2022. Retrieved 25 October 2022.
- Hasan, Ali; Fumerton, Richard (2020). "Knowledge by Acquaintance vs. Description". The Stanford Encyclopedia of Philosophy. Metaphysics Research Lab, Stanford University. introduction. Archived from the original on 31 May 2022. Retrieved 31 March 2023.
- Haskell, Robert E. (2001). Transfer of Learning: Cognition and Instruction. Academic Press. p. 101-3. ISBN 978-0-12-330595-4. Archived from the original on 2023-04-15. Retrieved 2023-04-16.
- Haymes, Brian; Özdalga, Elisabeth (3 January 2016). Concept Of The Knowledge Of God. Springer. pp. 26–28. ISBN 978-1-349-19066-9. Archived from the original on 1 April 2023. Retrieved 16 April 2023.
- Hetherington, Stephen (2022). "Gettier Problems". Internet Encyclopedia of Philosophy. Archived from the original on 19 February 2009. Retrieved 28 May 2022.
- Hetherington, Stephen (2023). "Knowledge". Internet Encyclopedia of Philosophy. Archived from the original on 2 June 2022. Retrieved 31 March 2023.
- Hetherington, Stephen (1 September 2016). Knowledge and the Gettier Problem. Cambridge University Press. p. 219. ISBN 978-1-316-75729-1. Archived from the original on 30 March 2023. Retrieved 16 April 2023.
- Heydorn, Wendy; Jesudason, Susan (18 July 2013). Decoding Theory of Knowledge for the IB Diploma. Cambridge University Press. p. 10. ISBN 978-1-107-62842-7. Archived from the original on 1 April 2023. Retrieved 16 April 2023.
- Hohwy, Jakob (November 2013). The Predictive Mind. Oxford University Press. p. 245. ISBN 978-0-19-968273-7. Archived from the original on 2023-04-06. Retrieved 2023-04-16.
- Holyoak, Keith; Morrison, Robert (2005). The Cambridge Handbook of Thinking and Reasoning. Cambridge: Cambridge University Press. p. 371. ISBN 0521824176.
- Ichikawa, Jonathan Jenkins; Steup, Matthias (2018). "The Analysis of Knowledge". The Stanford Encyclopedia of Philosophy. Metaphysics Research Lab, Stanford University. Archived from the original on 2 May 2022. Retrieved 29 March 2023.
- Jankowski, Natasha A.; Marshall, David W. (11 January 2016). Partners in Advancing Student Learning: Degree Qualifications Profile and Tuning: New Directions for Institutional Research, Number 165. John Wiley & Sons. p. 70. ISBN 978-1-119-24064-8. Archived from the original on 1 April 2023. Retrieved 16 April 2023.
- Kent, Allen; Williams, James G. (18 November 1993). Encyclopedia of Microcomputers: Volume 13 - Optical Disks to Production Scheduling. CRC Press. p. 295. ISBN 978-0-8247-2711-6. Archived from the original on 22 April 2023. Retrieved 16 April 2023.
- Kikoski, Catherine Kano; Kikoski, John F. (2004). The Inquiring Organization: Tacit Knowledge, Conversation, and Knowledge Creation : Skills for 21st-century Organizations. Greenwood Publishing Group. pp. 62, 65–66. ISBN 978-1-56720-490-2. Archived from the original on 2023-04-06. Retrieved 2023-04-16.
- Kirkham, Richard L. (1984). "Does the Gettier Problem Rest on a Mistake?". Mind. XCIII (372): 503, 512–3. doi:10.1093/mind/XCIII.372.501.
- Klauer, Bernd; Manstetten, Reiner; Petersen, Thomas; Schiller, Johannes (1 September 2016). Sustainability and the Art of Long-Term Thinking. Taylor & Francis. pp. 105–106. ISBN 978-1-134-98618-7. Archived from the original on 1 April 2023. Retrieved 16 April 2023.
- Klein, Peter D. (1998). "Knowledge, concept of". In Craig, Edward (ed.). Routledge Encyclopedia of Philosophy. London; New York: Routledge. doi:10.4324/9780415249126-P031-1. ISBN 978-0-415-25069-6. OCLC 38096851. Archived from the original on 13 June 2022. Retrieved 13 June 2022.
- Kleinman, Paul (18 September 2013). Philosophy 101: From Plato and Socrates to Ethics and Metaphysics, an Essential Primer on the History of Thought. Simon and Schuster. p. 258. ISBN 978-1-4405-6768-1. Archived from the original on 7 April 2023. Retrieved 16 April 2023.
- Kornblith, Hilary (2008). "1 Knowledge Needs No Justification". In Smith, Quentin (ed.). Epistemology: New Essays. Oxford University Press. pp. 5–6. ISBN 9780191718472. Archived from the original on 2023-03-30. Retrieved 2023-04-16.
- Lalanda, Philippe; McCann, Julie A.; Diaconescu, Ada (13 May 2013). Autonomic Computing: Principles, Design and Implementation. Springer Science & Business Media. pp. 187–188. ISBN 978-1-4471-5007-7. Archived from the original on 7 April 2023. Retrieved 16 April 2023.
- Lee, James (25 August 2017). A Metaphysician's User Guide: The Epistemology of Metaphysics. Syracuse University. pp. 6–9. Archived from the original on 31 March 2023. Retrieved 16 April 2023.
- Lehrer, Keith (15 October 2015). "1. The Analysis of Knowledge". Theory of Knowledge. Routledge. ISBN 978-1-135-19609-7. Archived from the original on 2 June 2022. Retrieved 12 June 2022.
- Leonard, Nick (2021). "Epistemological Problems of Testimony". Stanford Encyclopedia of Philosophy. Metaphysics Research Lab, Stanford University. introduction. Archived from the original on 10 July 2022. Retrieved 8 June 2022.
- Leondes, Cornelius T. (26 September 2001). Expert Systems: The Technology of Knowledge Management and Decision Making for the 21st Century. Elsevier. p. 804. ISBN 978-0-08-053145-8. Archived from the original on 7 April 2023. Retrieved 16 April 2023.
- Leung, Yee (29 November 2019). International Encyclopedia of Human Geography. Elsevier. p. 210. ISBN 978-0-08-102296-2. Archived from the original on 25 April 2023. Retrieved 16 April 2023.
- Lilley, Simon; Lightfoot, Geoffrey; Amaral, Paulo (2004). Representing Organization: Knowledge, Management, and the Information Age. Oxford University Press. pp. 162–163. ISBN 978-0-19-877541-6. Archived from the original on 2023-03-10. Retrieved 2023-04-17.
In its more modern forms epistemology has taken the analysis of meaning and the status of claims to knowledge as its quarry. Consequently, writers such as Bertrand Arthur William Russell (also known as the third Earl Russell, 1872-1970), George Edward Moore (1873-1958), and Ludwig Joseph Johann Wittgenstein (1889-1951) have attempted to delineate three kinds of knowledge: 1. Knowledge that, or 'factual knowledge' ... 2. Knowledge how, or 'practical knowledge' ... 3. Knowledge of people, places, and things, or 'knowledge by acquaintance'
- Maedche, Alexander; Brocke, Jan vom; Hevner, Alan (22 May 2017). Designing the Digital Transformation: 12th International Conference, DESRIST 2017, Karlsruhe, Germany, May 30 – June 1, 2017, Proceedings. Springer. p. 403. ISBN 978-3-319-59144-5. Archived from the original on 7 April 2023. Retrieved 16 April 2023.
- Magee, Bryan; Popper, Karl R. (1971). "Conversation with Karl Popper". In Magee, Bryan (ed.). Modern British philosophy. New York: St. Martin's Press. pp. 74–75. ISBN 978-0-19-283047-0. OCLC 314039.
Popper: Putting our ideas into words, or better, writing them down, makes an important difference. ... It is what I call 'knowledge in the objective sense'. Scientific knowledge belongs to it. It is this knowledge which is stored in our libraries rather than our heads. Magee: And you regard the knowledge stored in our libraries as more important than the knowledge stored in our heads. Popper: Much more important, from every point of view
- Martinich, A. P.; Stroll, Ayrum. "Epistemology - Knowledge and certainty". www.britannica.com. Rationalism and empiricism. Archived from the original on 5 April 2023. Retrieved 5 April 2023.
- Merriënboer, Jeroen J. G. van (1997). Training Complex Cognitive Skills: A Four-Component Instructional Design Model for Technical Training. Educational Technology. p. 32. ISBN 978-0-87778-298-8. Archived from the original on 2023-04-01. Retrieved 2023-04-16.
- Miah, Sajahan (30 May 2006). Russell's Theory of Perception. Continuum. pp. 19–20. ISBN 978-1-84714-284-9. Archived from the original on 1 April 2023. Retrieved 16 April 2023.
- Morrison, Associate Professor of English Robert (18 April 2005). The Cambridge Handbook of Thinking and Reasoning. Cambridge University Press. p. 371. ISBN 978-0-521-82417-0. Archived from the original on 30 March 2023. Retrieved 16 April 2023.
- Moser, Paul K. "A posteriori". Routledge Encyclopedia of Philosophy. Routledge. Archived from the original on 20 September 2022. Retrieved 18 September 2022.
- Moser, Paul K. (27 October 2005). The Oxford Handbook of Epistemology. Oxford University Press. p. 3. ISBN 978-0-19-020818-9. Archived from the original on 8 March 2023. Retrieved 16 April 2023.
- Murphy, P. Karen; Alexander, Patricia A. (7 October 2005). Understanding How Students Learn: A Guide for Instructional Leaders. Corwin Press. pp. 38–39. ISBN 978-1-4833-6347-9. Archived from the original on 4 April 2023. Retrieved 16 April 2023.
- Nguyen, Van Tham; Nguyen, Ngoc Thanh; Tran, Trong Hieu (30 December 2022). Knowledge Integration Methods for Probabilistic Knowledge-based Systems. CRC Press. pp. 33–34. ISBN 978-1-000-80999-2. Archived from the original on 1 April 2023. Retrieved 16 April 2023.
- O’Brien, Daniel. "The Epistemology of Perception". Internet Encyclopedia of Philosophy. Archived from the original on 8 May 2009. Retrieved 25 October 2022.
- Olsson, Erik J (December 2011). "The Value of Knowledge: The Value of Knowledge". Philosophy Compass. 6 (12): 874–883. doi:10.1111/j.1747-9991.2011.00425.x. S2CID 143034920.
- Pavese, Carlotta (2022). "Knowledge How". The Stanford Encyclopedia of Philosophy. Metaphysics Research Lab, Stanford University. Archived from the original on 27 March 2023. Retrieved 31 March 2023.
- Peels, Rik (2023). Ignorance: A Philosophical Study. Oxford University Press. p. 28. ISBN 978-0-19-765451-4. Archived from the original on 2023-04-01. Retrieved 2023-04-16.
- Plantinga, Alvin (27 January 2000). Warranted Christian Belief. Oxford University Press. pp. 182–183. ISBN 978-0-19-803024-9. Archived from the original on 4 April 2023. Retrieved 16 April 2023.
- Plato (2002). Five Dialogues. Indianapolis, IN: Hackett Pub. Co. pp. 89–90, 97b–98a. ISBN 978-0-87220-633-5.
- Poston, Ted. "Internalism and Externalism in Epistemology". Internet Encyclopedia of Philosophy. introduction. Archived from the original on 31 March 2023. Retrieved 30 March 2023.
- Price, Kay M.; Nelson, Karna L. (1 January 2013). Planning Effective Instruction: Diversity Responsive Methods and Management. Cengage Learning. p. 4. ISBN 978-1-285-49991-8. Archived from the original on 1 April 2023. Retrieved 16 April 2023.
- Pritchard, Duncan; Turri, John; Carter, J. Adam (2022). "The Value of Knowledge". The Stanford Encyclopedia of Philosophy. Metaphysics Research Lab, Stanford University. introduction. Archived from the original on 10 July 2022. Retrieved 19 September 2022.
- Pritchard, Duncan (18 May 2023). What is this thing called Knowledge?. Taylor & Francis. ISBN 978-1-000-87480-8. Archived from the original on 31 March 2023. Retrieved 16 April 2023.
- Reber, Arthur S.; Allen, Rhianon (2022). The Cognitive Unconscious: The First Half Century. Oxford University Press. p. 281. ISBN 978-0-19-750157-3. Archived from the original on 2023-04-04. Retrieved 2023-04-16.
- Reif, Frederick (2008). Applying Cognitive Science to Education: Thinking and Learning in Scientific and Other Complex Domains. MIT Press. ISBN 978-0-262-18263-8. Archived from the original on 2023-04-15. Retrieved 2023-04-16.
- Rhem, Anthony J. (21 November 2005). UML for Developing Knowledge Management Systems. CRC Press. p. 42-3. ISBN 978-1-135-48553-5. Archived from the original on 7 April 2023. Retrieved 16 April 2023.
- Russell, Bruce (2020). "A Priori Justification and Knowledge". The Stanford Encyclopedia of Philosophy. Metaphysics Research Lab, Stanford University. Archived from the original on 12 August 2021. Retrieved 18 September 2022.
- Sadegh-Zadeh, Kazem (28 September 2011). Handbook of Analytic Philosophy of Medicine. Springer Science & Business Media. pp. 450–1, 470, 475. ISBN 978-94-007-2260-6. Archived from the original on 30 March 2023. Retrieved 16 April 2023.
- Schelling, Birte (2 May 2013). Knowledge - Genetic Foundations and Epistemic Coherence. Walter de Gruyter. pp. 55–56. ISBN 978-3-11-032266-8. Archived from the original on 31 March 2023. Retrieved 16 April 2023.
- Scott, Peter; Gallacher, Jim; Parry, Gareth (2017). New Languages and Landscapes of Higher Education. Oxford University Press. p. 97. ISBN 978-0-19-878708-2. Archived from the original on 2023-04-01. Retrieved 2023-04-16.
- Seel, Norbert M. (5 October 2011). Encyclopedia of the Sciences of Learning. Springer Science & Business Media. p. 1001. ISBN 978-1-4419-1427-9. Archived from the original on 30 March 2023. Retrieved 16 April 2023.
- Smith, Patricia L.; Ragan, Tillman J. (7 December 2004). Instructional Design. John Wiley & Sons. pp. 152–154. ISBN 978-0-471-39353-5. Archived from the original on 4 April 2023. Retrieved 16 April 2023.
- Soled, Suzanne Wegener (1995). Assessment, Testing, and Evaluation in Teacher Education. Greenwood Publishing Group. p. 49. ISBN 978-1-56750-153-7. Archived from the original on 2023-04-06. Retrieved 2023-04-16.
- Steup, Matthias; Neta, Ram (2020). "Epistemology". Stanford Encyclopedia of Philosophy. Metaphysics Research Lab, Stanford University. Archived from the original on 21 July 2020. Retrieved 22 May 2022.
- Stillings, Neil A.; Chase, Christopher H.; Weisler, Steven E.; Feinstein, Mark H.; Rissland, Edwina L. (1995). Cognitive Science: An Introduction. MIT Press. p. 370. ISBN 978-0-262-69175-8. Archived from the original on 2023-04-22. Retrieved 2023-04-16.
- Strube, G.; Wender, K. F. (1 October 1993). The Cognitive Psychology of Knowledge. Elsevier. p. 354. ISBN 978-0-08-086755-7. Archived from the original on 7 April 2023. Retrieved 16 April 2023.
- Sudduth, Michael. "Defeaters in Epistemology". Internet Encyclopedia of Philosophy. Archived from the original on 2 June 2022. Retrieved 30 March 2023.
- Tokuhama-Espinosa, Tracey (2011). Mind, Brain, and Education Science: A Comprehensive Guide to the New Brain-Based Teaching. New York: W. W. Norton & Company. p. 255. ISBN 9780393706079.
- Tolliver, Joseph Thomas (May 1989). "Knowledge without truth". Philosophical Studies. 56 (1): 29–51. doi:10.1007/bf00646208. S2CID 170980707.
- Truncellito, David A. "Epistemology". Internet Encyclopedia of Philosophy. Archived from the original on 13 January 2022. Retrieved 8 March 2023.
- Turri, John; Alfano, Mark; Greco, John (2021). "Virtue Epistemology". The Stanford Encyclopedia of Philosophy. Metaphysics Research Lab, Stanford University. 5. Knowledge. Archived from the original on 2023-03-26. Retrieved 2023-04-16.
- Villoro, Luis (1998). Belief, personal, and propositional knowledge. Amsterdam: Rodopi. ISBN 9042007427.
- Walton, Douglas N. (January 2005). "Pragmatic and idealized models of knowledge and ignorance". American Philosophical Quarterly. 42 (1): 59–69 (59, 64). JSTOR 20010182.
It is a pervasive assumption in recent analytical philosophy that knowledge can be defined as a modality representing a rational agent's true and consistent beliefs. Such views are based on rationality assumptions. One is that knowledge can only consist of true propositions. This way of speaking is sharply at odds with the way we speak about knowledge, for example, in computing, where a so-called knowledge base can be a database, that is, a set of data that has been collected and is thought to consist of true propositions, even though, realistically speaking, many of them might later be shown to be false or untenable. ... The pragmatic account of knowledge starts with a knowledge system, meaning a working system with an agent having a database. ... The notion of a search can be a social one, in many instances. A group of agents can be engaged in the search, and some of them can know things that others do not know.
- Watson, Jamie Carlin. "Justification, Epistemic". Internet Encyclopedia of Philosophy. Introduction. Archived from the original on 14 March 2023. Retrieved 30 March 2023.
- Woolfolk, Anita; Margetts, Kay (25 July 2012). Educational Psychology Australian Edition. Pearson Higher Education AU. p. 251. ISBN 978-1-4425-5145-9. Archived from the original on 6 April 2023. Retrieved 16 April 2023.
- Woolfolk, Anita E.; Hughes, Malcolm; Walkup, Vivienne (2008). Psychology in Education. Pearson Longman. p. 307. ISBN 978-1-4058-3541-1. Archived from the original on 2023-04-08. Retrieved 2023-04-16.
- Yamamoto, Sakae (4 July 2016). Human Interface and the Management of Information: Information, Design and Interaction: 18th International Conference, HCI International 2016 Toronto, Canada, July 17-22, 2016, Proceedings, Part I. Springer. p. 61. ISBN 978-3-319-40349-6. Archived from the original on 1 April 2023. Retrieved 16 April 2023.
- Zagzebski, Linda (1999). "What Is Knowledge?". In Greco, John; Sosa, Ernest (eds.). The Blackwell Guide to Epistemology. Malden, MA: Blackwell. pp. 92–116. doi:10.1002/9781405164863.ch3. ISBN 978-0-631-20290-5. OCLC 39269507. S2CID 158886670. Archived from the original on 2 June 2022. Retrieved 12 June 2022.
https://en.wikipedia.org/wiki/Declarative_knowledge
In philosophy of mind, qualia (/ˈkwɑːliə/ or /ˈkweɪliə/; singular form: quale) are defined as instances of subjective, conscious experience. The term qualia derives from the Latin neuter plural form (qualia) of the Latin adjective quālis (Latin pronunciation: [ˈkʷaːlɪs]) meaning "of what sort" or "of what kind" in a specific instance, such as "what it is like to taste a specific apple — this particular apple now".
Examples of qualia include the perceived sensation of pain of a headache, the taste of wine, and the redness of an evening sky. As qualitative characteristics of sensation, qualia stand in contrast to propositional attitudes,[1] where the focus is on beliefs about experience rather than what it is directly like to be experiencing.
Philosopher and cognitive scientist Daniel Dennett suggested that qualia was "an unfamiliar term for something that could not be more familiar to each of us: the ways things seem to us".[2]
Much of the debate over the importance of qualia hinges on the definition of the term, and various philosophers emphasize or deny the existence of certain features of qualia. Consequently, the nature and existence of qualia under various definitions remain controversial. While some philosophers of mind, like Daniel Dennett, argue that qualia do not exist and are incompatible with neuroscience and naturalism,[3] some neuroscientists and neurologists, like Gerald Edelman, Antonio Damasio, Vilayanur Ramachandran, Giulio Tononi, Christof Koch, and Rodolfo Llinás, state that qualia exist and that the desire by some philosophers to disregard qualia is based on an erroneous interpretation of what constitutes science.[4]
Definitions
Many definitions of qualia have been proposed. One of the simpler, broader definitions is: "The 'what it is like' character of mental states. The way it feels to have mental states such as pain, seeing red, smelling a rose, etc."[5]
C.S. Peirce introduced the term quale in philosophy in 1866[6][7] C.I. Lewis (1929)[7] was the first to use the term "qualia" in its generally agreed upon modern sense.
There are recognizable qualitative characters of the given, which may be repeated in different experiences, and are thus a sort of universals; I call these "qualia." But although such qualia are universals, in the sense of being recognized from one to another experience, they must be distinguished from the properties of objects. Confusion of these two is characteristic of many historical conceptions, as well as of current essence-theories. The quale is directly intuited, given, and is not the subject of any possible error because it is purely subjective.[7]
Frank Jackson later defined qualia as "...certain features of the bodily sensations especially, but also of certain perceptual experiences, which no amount of purely physical information includes".[8]: 273
Daniel Dennett identifies four properties that are commonly ascribed to qualia.[2] According to these, qualia are:
- ineffable – they cannot be communicated, or apprehended by any means other than direct experience.
- intrinsic – they are non-relational properties, which do not change depending on the experience's relation to other things.
- private – all interpersonal comparisons of qualia are systematically impossible.
- directly or immediately apprehensible by consciousness – to experience a quale is to know one experiences a quale, and to know all there is to know about that quale.
If qualia of this sort exist, then a normally sighted person who sees red would be unable to describe the experience of this perception in such a way that a listener who has never experienced color will be able to know everything there is to know about that experience. Though it is possible to make an analogy, such as "red looks hot", or to provide a description of the conditions under which the experience occurs, such as "it's the color you see when light of 700-nm wavelength is directed at you", supporters of this definition of qualia contend that such descriptions cannot provide a complete description of the experience.[citation needed]
Another way of defining qualia is as "raw feels". A raw feel is a perception in and of itself, considered entirely in isolation from any effect it might have on behavior and behavioral disposition. In contrast, a cooked feel is that perception seen in terms of its effects. For example, the perception of the taste of wine is an ineffable, raw feel, while the behavioral reaction one has to the warmth or bitterness caused by that taste of wine would be a cooked feel. Cooked feels are not qualia.[citation needed]
Saul Kripke argues that one key consequence of the claim that such things as raw feels can be meaningfully discussed – that qualia exist[clarification needed] – is that it leads to the logical possibility of two entities exhibiting identical behavior in all ways despite one of them entirely lacking qualia.[9] While few claim that such an entity, called a philosophical zombie, actually exists, the possibility is raised as a refutation of physicalism.[10][further explanation needed]
Arguably, the idea of hedonistic utilitarianism, where the ethical value of things is determined from the amount of subjective pleasure or pain they cause, is dependent on the existence of qualia.[11]
Arguments for the existence of qualia
Since by definition one cannot fully convey qualia verbally, one also cannot demonstrate them directly in an argument; so a more nuanced approach is needed. Arguments for qualia generally come in the form of thought experiments designed to lead one to the conclusion that qualia exist.[12]
"What's it like to be?" argument
Thomas Nagel's paper "What is it like to be a bat?",[13] although it does not use the word "qualia", is often cited in debates about qualia. Nagel argues that consciousness has an essentially subjective character, a what-it-is-like aspect. He states that "an organism has conscious mental states if and only if there is something that it is like to be that organism – something it is like for the organism."[13] Nagel suggests that this subjective aspect may never be sufficiently accounted for by the objective methods of reductionistic science. He claims that "if we acknowledge that a physical theory of mind must account for the subjective character of experience, we must admit that no presently available conception gives us a clue about how this could be done."[14]: 450 Furthermore, "it seems unlikely that any physical theory of mind can be contemplated until more thought has been given to the general problem of subjective and objective."[14]: 450
Inverted spectrum argument
The inverted spectrum thought experiment, originally developed by John Locke,[15] invites us to imagine that we wake up one morning and find that for some unknown reason all the colors in the world have been inverted, i.e. swapped to the hue on the opposite side of a color wheel. Furthermore, we discover that no physical changes have occurred in our brains or bodies that would explain this phenomenon. Supporters of the existence of qualia argue that since we can imagine this happening without contradiction, it follows that we are imagining a change in a property that determines the way things look to us, but that has no physical basis.[16][17] In more detail:
- Metaphysical identity holds of necessity.[clarification needed]
- If something is possibly false, it is not necessary.
- It is conceivable that qualia could have a different relationship to physical brain-states.
- If it is conceivable, then it is possible.
- Since it is possible for qualia to have a different relationship with physical brain-states, they cannot be identical to brain states (by 1).
- Therefore, qualia are non-physical.
The argument thus claims that if we find the inverted spectrum plausible, we must admit that qualia exist (and are non-physical). Some philosophers find it absurd that armchair theorizing can prove something to exist, and the detailed argument does involve a lot of assumptions about conceivability and possibility, which are open to criticism. Perhaps it is not possible for a given brain state to produce anything other than a given quale in our universe, and that is all that matters.
The idea that an inverted spectrum would be undetectable in practice is also open to criticism on more scientific grounds (see main article).[16][17] There is an actual experiment – albeit somewhat obscure – that parallels the inverted spectrum argument. George M. Stratton, professor of psychology at the University of California, Berkeley, performed an experiment in which he wore special prism glasses that caused the external world to appear upside down.[18] After a few days of continually wearing the glasses, he adapted and the external world appeared upright to him. When he removed the glasses, his perception of the external world again returned to the "normal" perceptual state. If this argument provides indicia that qualia exist, it does not necessarily follow that they must be non-physical, because that distinction should be considered a separate epistemological issue.
Zombie argument
An argument holds that it is not inconceivable for a set of people to have qualia, while physical duplicates of that set, called "philosophical zombies", do not. These "zombies" would demonstrate outward behavior, including utterances, exactly the same as normal humans (who are assumed to have subjective phenomenology), without subjective phenomenology. For there to be a valid distinction between "normal humans" and philosophical zombies there must be no specific part or parts of the brain that directly give rise to qualia: The zombie/normal-human distinction can only be valid if subjective consciousness is causally separate from the physical brain.[citation needed]
Are zombies possible? They're not just possible, they're actual. We're all zombies: Nobody is conscious. — D.C. Dennett (1992)[19][page needed]
Explanatory gap argument
Joseph Levine's paper Conceivability, Identity, and the Explanatory Gap takes up where the criticisms of conceivability arguments (such as the inverted spectrum argument and the zombie argument) leave off. Levine agrees that conceivability is a flawed means of establishing metaphysical realities, but points out that even if we come to the metaphysical conclusion that qualia are physical, there is still an explanatory problem.
While I think this materialist response is right in the end, it does not suffice to put the mind-body problem to rest. Even if conceivability considerations do not establish that the mind is in fact distinct from the body, or that mental properties are metaphysically irreducible to physical properties, still they do demonstrate that we lack an explanation of the mental in terms of the physical.
However, such an epistemological or explanatory problem might indicate an underlying metaphysical issue – the non-physicality of qualia, even if not proven by conceivability arguments, is far from ruled out.
In the end, we are right back where we started. The explanatory gap argument doesn't demonstrate a gap in nature, but a gap in our understanding of nature. Of course a plausible explanation for there being a gap in our understanding of nature is that there is a genuine gap in nature. But so long as we have countervailing reasons for doubting the latter, we have to look elsewhere for an explanation of the former.[20]
Knowledge argument
F.C. Jackson offers what he calls the "knowledge argument" for qualia.[8] One example runs as follows:
Mary the color scientist knows all the physical facts about color, including every physical fact about the experience of color in other people, from the behavior a particular color is likely to elicit to the specific sequence of neurological firings that register that a color has been seen. However, she has been confined from birth to a room that is black and white, and is only allowed to observe the outside world through a black and white monitor. When she is allowed to leave the room, it must be admitted that she learns something about the color red the first time she sees it – specifically, she learns what it is like to see that color.
This thought experiment has two purposes. First, it is intended to show that qualia exist. If we accept the thought experiment, we believe that Mary gains something after she leaves the room – that she acquires knowledge of a particular thing that she did not possess before. That knowledge, Jackson argues, is knowledge of the quale that corresponds to the experience of seeing red, and it must thus be conceded that qualia are real properties, since there is a difference between a person who has access to a particular quale and one who does not.
The second purpose of this argument is to refute the physicalist account of the mind. Specifically, the knowledge argument is an attack on the physicalist claim about the completeness of physical truths. The challenge posed to physicalism by the knowledge argument runs as follows:
- Before her release, Mary was in possession of all the physical information about color experiences of other people.
- After her release, Mary learns something about the color experiences of other people.
Therefore, - Before her release, Mary was not in possession of all the
information about other people's color experiences, even though she was
in possession of all the physical information.
Therefore, - There are truths about other people's color experience that are not physical.
Therefore, - Physicalism is false.
At first Jackson argued that qualia are epiphenomenal: not causally efficacious with respect to the physical world. Jackson does not give a positive justification for this claim – rather, he seems to assert it simply because it defends qualia against the classic problems of dualism. Our[who?] natural assumption would be that qualia must be causally efficacious in the physical world, but some would ask how we[who?] could argue for their existence if they did not affect our brains. If qualia are non-physical properties (which they must be in order to constitute an argument against physicalism), some[who?] question how they could have a causal effect on the physical world. By redefining qualia as epiphenomenal, Jackson attempts to protect them from the demand of playing a causal role.
Later, however, Jackson rejected epiphenomenalism. This, he argues, is because when Mary first sees red, she says "wow", so it must be Mary's qualia that cause her to say "wow". This contradicts epiphenomenalism. Since the Mary's room thought experiment seems to create this contradiction, there must be something wrong with it.[clarification needed] This[ambiguous] is often referred to as the "there must be a reply" reply.[citation needed]
Analytical philosophers who are critics of qualia
Daniel Dennett
In Consciousness Explained (1991)[19] and "Quining Qualia" (1988),[21] Daniel Dennett argues against qualia by claiming that the[which?] above definition breaks down if one tries to practically apply it. In a series of thought experiments, which he calls "intuition pumps", he brings qualia into the world of neurosurgery, clinical psychology, and psychological experimentation. He argues that, once the concept of qualia is so imported, we can either make no use of it, or the questions introduced by it are unanswerable precisely because of the special properties defined for qualia.[citation needed]
In Dennett's updated version of the inverted spectrum thought experiment, "alternative neurosurgery", you again awake to find that your qualia have been inverted – grass appears red, the sky appears orange, etc. According to the original account, you should be immediately aware that something has gone horribly wrong. Dennett argues, however, that it is impossible to know whether the diabolical neurosurgeons have indeed inverted your qualia (by tampering with your optic nerve, say), or have simply inverted your connection to memories of past qualia. Since both operations would produce the same result, you would have no means on your own to tell which operation has actually been conducted, and you are thus in the odd position of not knowing whether there has been a change in your "immediately apprehensible" qualia.[citation needed]
Dennett's argues that for qualia to be taken seriously as a component of experience – for them to make sense as a discrete concept – it must be possible to show that
- it is possible to know that a change in qualia has occurred, as opposed to a change in something else; or that
- there is a difference between having a change in qualia and not having one.
Dennett attempts to show that we cannot satisfy (a) either through introspection or through observation, and that qualia's very definition undermines its chances of satisfying (b).
Supporters of qualia could point out that in order for you to notice a change in qualia, you must compare your current qualia with your memories of past qualia. Arguably, such a comparison would involve immediate apprehension of your current qualia and your memories of past qualia, but not the past qualia themselves. Furthermore, modern functional brain imaging has increasingly suggested that the memory of an experience is processed in similar ways and in similar zones of the brain as those originally involved in the original perception.[22] This may mean that there would be asymmetry in outcomes between altering the mechanism of perception of qualia and altering their memories. If the diabolical neurosurgery altered the immediate perception of qualia, you might not even notice the inversion directly, since the brain zones which re-process the memories would themselves invert the qualia remembered. On the other hand, alteration of the qualia memories themselves would be processed without inversion, and thus you would perceive them as an inversion. Thus, you might know immediately if memory of your qualia had been altered, but might not know if immediate qualia were inverted or whether the diabolical neurosurgeons had done a sham procedure.[original research?]
Dennett responds to the "Mary the color scientist" thought experiment by arguing that Mary would not, in fact, learn something new if she stepped out of her black and white room to see the color red. Dennett asserts that if she already truly knew "everything about color", that knowledge would include a deep understanding of why and how human neurology causes us to sense the "quale" of color. Mary would therefore already know exactly what to expect of seeing red, before ever leaving the room. Dennett argues that the misleading aspect of the story is that Mary is supposed to not merely be knowledgeable about color but to actually know all the physical facts about it, which would be a knowledge so deep that it exceeds what can be imagined, and twists our intuitions.
If Mary really does know everything physical there is to know about the experience of color, then this effectively grants her almost omniscient powers of knowledge. Using this, she will be able to deduce her own reaction, and figure out exactly what the experience of seeing red will feel like.
Dennett finds that many people find it difficult to see this, so he uses the case of RoboMary to further illustrate what it would be like for Mary to possess such a vast knowledge of the physical workings of the human brain and color vision. RoboMary is an intelligent robot who has, instead of the ordinary color camera-eyes, a software lock such that she is only able to perceive black and white and shades in-between.[citation needed]
RoboMary can examine the computer brain of similar non-color-locked robots when they look at a red tomato, and see exactly how they react and what kinds of impulses occur. RoboMary can also construct a simulation of her own brain, unlock the simulation's color-lock and, with reference to the other robots, simulate exactly how this simulation of herself reacts to seeing a red tomato. RoboMary naturally has control over all of her internal states except for the color-lock. With the knowledge of her simulation's internal states upon seeing a red tomato, RoboMary can put her own internal states directly into the states they would be in upon seeing a red tomato. In this way, without ever seeing a red tomato through her cameras, she will know exactly what it is like to see a red tomato.[citation needed]
Dennett uses this example as attempt to show us that Mary's all-encompassing physical knowledge makes her own internal states as transparent as those of a robot or computer, and it is almost straightforward for her to figure out exactly how it feels to see red.[citation needed]
Perhaps Mary's failure to learn exactly what seeing red feels like is simply a failure of language, or a failure of our ability to describe experiences. An alien race with a different method of communication or description might be perfectly able to teach their version of Mary exactly how seeing the color red would feel. Perhaps it is simply a uniquely human failing to communicate first-person experiences from a third-person perspective. Dennett suggests that the description might even be possible using English. He uses a simpler version of the Mary thought experiment to show how this might work. What if Mary was in a room without triangles and was prevented from seeing or making any triangles? An English-language description of just a few words would be sufficient for her to imagine what it is like to see a triangle – she can simply and directly visualize a triangle in her mind. Similarly, Dennett proposes, it is perfectly, logically possible that the quale of what it is like to see red could eventually be described in an English-language description of millions or billions of words.[citation needed]
In "Are we explaining consciousness yet?" (2001),[23] Dennett approves of an account of qualia defined as the deep, rich collection of individual neural responses that are too fine-grained for language to capture. For instance, a person might have an alarming reaction to yellow because of a yellow car that hit her previously, and someone else might have a nostalgic reaction to a comfort food. These effects are too individual-specific to be captured by English words. "If one dubs this inevitable residue qualia, then qualia are guaranteed to exist, but they are just more of the same, dispositional properties that have not yet been entered in the catalog [...]."[23]
Paul Churchland
According to Paul Churchland, Mary might be considered to be like a feral child. Feral children have suffered extreme isolation during childhood. Technically when Mary leaves the room, she would not have the ability to see or know what the color red is. A brain has to learn and develop how to see colors. Patterns need to form in the V4 section of the visual cortex. These patterns are formed from exposure to wavelengths of light. This exposure is needed during the early stages of brain development. In Mary's case, the identifications and categorizations of color will only be in respect to representations of black and white.[24]
Gary Drescher
In his book Good and Real (2006),[25] Gary Drescher compares qualia with "gensyms" (generated symbols) in Common Lisp. These are objects that Lisp treats as having no properties or components and which can only be identified as equal or not equal to other objects. Drescher explains, "we have no introspective access to whatever internal properties make the red gensym recognizably distinct from the green [...] even though we know the sensation when we experience it."[25] Under this interpretation of qualia, Drescher responds to the Mary thought experiment by noting that "knowing about red-related cognitive structures and the dispositions they engender – even if that knowledge were implausibly detailed and exhaustive – would not necessarily give someone who lacks prior color-experience the slightest clue whether the card now being shown is of the color called red." This does not, however, imply that our experience of red is non-mechanical; "on the contrary, gensyms are a routine feature of computer-programming languages".[14]: 82
David Lewis
D.K. Lewis introduced a hypothesis about types of knowledge and their transmission in qualia cases. Lewis agrees that Mary cannot learn what red looks like through her monochrome physicalist studies. But he proposes that this does not matter. Learning transmits information, but experiencing qualia does not transmit information; instead it communicates abilities. When Mary sees red, she does not get any new information. She gains new abilities – now she can remember what red looks like, imagine what other red things might look like and recognize further instances of redness.
Lewis states that Jackson's thought experiment uses the phenomenal information hypothesis – that is, that the new knowledge that Mary gains upon seeing red is phenomenal information. Lewis then proposes a different ability hypothesis that differentiates between two types of knowledge: knowledge "that" (information) and knowledge "how" (abilities). Normally the two are entangled; ordinary learning is also an experience of the subject concerned, and people both learn information (for instance, that Freud was a psychologist) and gain ability (to recognize images of Freud). However, in the thought experiment, Mary can only use ordinary learning to gain know-that knowledge. She is prevented from using experience to gain the know-how knowledge that would allow her to remember, imagine and recognize the color red.
We have the intuition that Mary has been deprived of some vital data to do with the experience of redness. It is also uncontroversial that some things cannot be learned inside the room; for example, we do not expect Mary to learn how to ski within the room. Lewis has articulated that information and ability are potentially different things. In this way, physicalism is still compatible with the conclusion that Mary gains new knowledge. It is also useful for considering other instances of qualia; "being a bat" is an ability, so it is know-how knowledge.[26]
Marvin Minsky
Artificial intelligence researcher Marvin Minsky thinks the problems posed by qualia are essentially issues of complexity, or rather of mistaking complexity for simplicity.
Now, a philosophical dualist might then complain: "You've described how hurting affects your mind – but you still can't express how hurting feels." This, I maintain, is a huge mistake – that attempt to reify "feeling" as an independent entity, with an essence that's indescribable. As I see it, feelings are not strange alien things. It is precisely those cognitive changes themselves that constitute what "hurting" is – and this also includes all those clumsy attempts to represent and summarize those changes. The big mistake comes from looking for some single, simple, "essence" of hurting, rather than recognizing that this is the word we use for complex rearrangement of our disposition of resources.[27]
Michael Tye
Michael Tye believes there are no qualia, no "veils of perception" between us and the referents of our thought. He describes our experience of an object in the world as "transparent". By this he means that no matter what private understandings and/or misunderstandings we may have of some public entity, it is still there before us in reality. The idea that qualia intervene between ourselves and their origins he regards as "a massive error"; as he says, "it is just not credible that visual experiences are systematically misleading in this way";[14]: 46 "the only objects of which you are aware are the external ones making up the scene before your eyes";[14]: 47 there are "no such things as the qualities of experiences" for "they are qualities of external surfaces (and volumes and films) if they are qualities of anything."[14]: 48 He believes we can take our experiences at face value since there is no fear of losing contact with the realness of public objects.
In Tye's thought there is no question of qualia without information being contained within them; it is always "an awareness that", always "representational". He characterizes the perception of children as a misperception of referents that are undoubtedly as present for them as they are for grown-ups. As he puts it, they may not know that "the house is dilapidated", but there is no doubt about their seeing the house. After-images are dismissed as presenting no problem for the Transparency theory because, as he puts it, after-images being illusory, there is nothing that one sees.
Tye proposes that phenomenal experience has five basic elements, for which he has coined the acronym PANIC – Poised, Abstract, Nonconceptual, Intentional Content.[14]: 63 It is "Poised" in the sense that the phenomenal experience is always presented to the understanding, whether or not the agent is able to apply a concept to it. Tye adds that the experience is "maplike" in that, in most cases, it reaches through to the distribution[clarification needed] of shapes, edges, volumes, etc. in the world – you may not be reading the "map" but, as with an actual map there is a reliable match with what it is mapping. It is "Abstract" because it is still an open question in a particular case whether you are in touch with a concrete object (someone may feel a pain in a "left leg" when that leg has actually been amputated). It is "Nonconceptual" because a phenomenon can exist although one does not have the concept by which to recognize it. Nevertheless, it is "Intentional" in the sense that it represents something, again whether or not the particular observer is taking advantage of that fact; this is why Tye calls his theory "representationalism". This last makes it plain that Tye believes that he has retained a direct contact with what produces the phenomena and is therefore not hampered by any trace of a "veil of perception".[28]
Roger Scruton
Roger Scruton, while sceptical that neurobiology can tell us much about consciousness, believes qualia is an incoherent concept, and that Wittgenstein's private language argument effectively disproves it. Scruton writes,
The belief that these essentially private features of mental states exist, and that they form the introspectible essence of whatever possesses them, is grounded in a confusion, one that Wittgenstein tried to sweep away in his arguments against the possibility of a private language. When you judge that I am in pain, it is on the basis of my circumstances and behavior, and you could be wrong. When I ascribe a pain to myself, I don’t use any such evidence. I don’t find out that I am in pain by observation, nor can I be wrong. But that is not because there is some other fact about my pain, accessible only to me, which I consult in order to establish what I am feeling. For if there were this inner private quality, I could misperceive it; I could get it wrong, and I would have to find out whether I am in pain. To describe my inner state, I would also have to invent a language, intelligible only to me – and that, Wittgenstein plausibly argues, is impossible. The conclusion to draw is that I ascribe pain to myself not on the basis of some inner quale but on no basis at all.
In his book On Human Nature, Scruton poses a potential line of criticism to this, which is that while Wittgenstein's private language argument does disprove the concept of reference to qualia, or the idea that we can talk even to ourselves of their nature, it does not disprove their existence altogether. Scruton believes that this is a valid criticism, and this is why he stops short of actually saying that qualia do not exist, and instead merely suggests that we should abandon them as a concept. However, he quotes Wittgenstein in response: "Whereof one cannot speak, thereof one must be silent."[29]
Analytical philosophers who are proponents of qualia
David Chalmers
David Chalmers formulated the hard problem of consciousness, which raised the issue of qualia to a new level of importance and acceptance in the field[clarification needed].[30] In Chalmers (1995)[31] he also argued for what he called "the principle of organizational invariance": If a system such as one of appropriately configured computer hardware reproduces the functional organization of the brain, it will also reproduce the qualia associated with the brain.
E. J. Lowe
E. J. Lowe denies that indirect realism (in which we have access only to sensory features internal to the brain) necessarily implies a Cartesian dualism. He agrees with Bertrand Russell that our "retinal images" – that is, the distributions across[clarification needed] our retinas – are connected to[vague] "patterns of neural activity in the cortex".[32] He defends a version of the causal theory of perception in which a causal path can be traced between the external object and the perception of it. He is careful to deny that we do any inferring from the sensory field[clarification needed]; he believes this allows us to found an access to knowledge on that causal connection. In a later work he moves closer to the non-epistemic argument in that he postulates "a wholly non-conceptual component of perceptual experience",[32] but he refrains from analyzing the relation between the perceptual and the "non-conceptual". More recently he drew attention to the problems that hallucination raises for the direct realist and to their disinclination to enter the discussion on the topic.[33]
J. B. Maund
This section may need to be rewritten to comply with Wikipedia's quality standards. (September 2014) |
John Barry Maund, an Australian philosopher of perception, argues that qualia can be described on two levels, a fact that he refers to as "dual coding".[34]
If asked what we see on a television screen there are two varieties of answer that we might give. Consider the example of a "Movitype" screen, often used for advertisements and announcements in public places. A Movitype screen consists of a matrix – or "raster" as the neuroscientists prefer to call it (from the Latin rastrum, a "rake"; think of the lines on a TV screen as "raked" across) – that is made up of an array of tiny light-sources. A computer can excite these lights so as to give the impression of letters passing from right to left, or even to show moving pictures. In describing what we see on such a screen, we could adopt the everyday public language and say "I saw some sentences, followed by a picture of a 7-Up can." Although that is a perfectly adequate way of describing the sight, nevertheless, there is a scientific way of describing it which bears no relation whatsoever to this commonsense description. One could ask an electronics engineer to provide us with a computer print-out, staged across the seconds that we were watching, of the point-states of the raster of lights. This would no doubt be a long and complex document, with the state of each tiny light-source given its place in the sequence. Although such a list would give a comprehensive and point-by-point-detailed description of the state of the screen, nowhere would it contain mention of "English sentences" or "a 7-Up can".
This illustrates that there are two ways to describe such a screen, (1) the "commonsense" one which refers to publicly recognizable objects, and (2) an accurate point-by-point account of the actual state of the field that makes no mention of what any passer-by would or would not make of it. This second description would be non-epistemic from the common sense point of view, since no objects are mentioned in the print-out, but perfectly acceptable from the engineer's point of view. Note that, if one carries this analysis across to human sensing and perceiving, this rules out Dennett's claim that all qualiaphiles must regard qualia as "ineffable", for at this second level they are in principle quite "effable" – indeed, it is not ruled out that some neurophysiologist of the future might be able to describe the neural detail of qualia at this level.
Maund extended his argument with reference to color.[35] Color he sees as a dispositional property, not an objective one. Colors are "virtual properties"; is as if things possessed them. Although the naïve view attributes them to objects, they are intrinsic, non-relational, inner experiences. This allows for the facts of difference[clarification needed] between person and person, and also leaves aside the claim that external objects are colored.
Moreland Perkins
In his book Sensing the World,[36] Moreland Perkins argues that qualia need not be identified with their objective sources: a smell, for instance, bears no direct resemblance to the molecular shape that gives rise to it, nor is a toothache actually in the tooth. Like Hobbes he views the process of sensing as complete in itself; as he puts it, it is not like "kicking a football" where an external object is required – it is more like "kicking a kick". This explanation evades the Homunculus Objection, as adhered to, for example, by Gilbert Ryle. Ryle was unable to entertain this possibility, protesting that "in effect it explained the having of sensations as the not having of sensations."[37] However, A.J. Ayer called this objection "very weak" as it betrayed an inability to detach the notion of eyes, or indeed any sensory organ, from the neural sensory experience.[38]
Howard Robinson and William Robinson
Howard Robinson specialized in philosophy of mind. He argued against explanations of sensory experience that reduce them to physical origins. He never regarded the theory of sense-data[clarification needed] as refuted, but set out to refute persuasive objections to it. The version of the theory of sense-data he defends takes what is before consciousness in perception to be qualia as mental presentations that are causally linked to external entities, but which are not physical in themselves.[needs copy edit] He is therefore a dualist: one who takes both matter and mind to have real and metaphysically distinct natures. One of his articles takes the physicalist to task for ignoring the fact that sensory experience can be entirely free of representational character. He cites phosphenes as a stubborn example (phosphenes are flashes of light that result either from sudden pressure in the brain – as induced, for example, by intense coughing, or through direct physical pressure on the retina), and points out that it is counter-intuitive to argue that these are not visual experiences on a par with open-eye seeing.
William Robinson (no relation) takes a similar view in his book, Understanding Phenomenal Consciousness.[39] He is unusual as a dualist in calling for research programs that investigate the relation of qualia to the brain. The problem is so stubborn, he says, that too many philosophers would prefer "to explain it away", but he would rather have it explained and does not see why the effort should not be made. However, he does not expect a straightforward scientific reduction of phenomenal experience to neural architecture; he regards this as a forlorn hope. The "Qualitative Event Realism" that Robinson espouses sees phenomenal consciousness as non-material events that are caused by brain events but not identical with them.
He refuses to set aside the vividness – and commonness – of mental images, both visual and aural. In this he opposes Daniel Dennett, who has difficulty crediting such experience in others. He is similar to Moreland Perkins in keeping his investigation wide enough to apply to all the senses.
Edmond Wright
Edmond Wright is a philosopher who considers the inter-subjective aspect of perception.[40] From John Locke onwards it had been typical to frame perception problems in terms of a single subject S looking at a single entity E with a property p. However, if we begin with the facts of the differences in sensory registration from person to person, coupled with the differences in the criteria we have learned for distinguishing what we together call "the same" things, then a problem arises of how two persons align their differences on these two levels so that they can still get a practical overlap on parts of the real about them – and, in particular, update each other about them.[needs copy edit]
Wright was struck by the hearing difference between himself and his son: he discovered that his son could hear sounds up to nearly 20 kHz while his range only reached to 14 kHz or so. This implies that a difference in qualia could emerge in human action (for example, the son could warn the father of a high-pitched escape of a dangerous gas kept under pressure, the sound-waves of which would be producing no qualia evidence at all for the father). The relevance for language thus becomes critical, for an informative statement can best be understood as an updating of a perception – and this may involve a radical re-selection from the qualia fields viewed as non-epistemic, even perhaps of the presumed singularity of "the" referent, a fortiori if that "referent" is the self.[needs copy edit] He distinguishes his view from that of Revonsuo[clarification needed], who too readily makes his "virtual space" "egocentric".[importance?]
Wright emphasizes what he asserts is a core feature of communication, that, in order for an updating to be set up and made possible[clarification needed], both speaker and hearer have to behave as if they have identified "the same singular thing", which, he notes, partakes of the structure of a joke or a story[how?].[40] Wright says that this systematic ambiguity[clarification needed] seems to opponents of qualia to be a sign of fallacy in the argument (as ambiguity is in pure logic) whereas, on the contrary, it is sign – in talk about "what" is perceived – of something those speaking to each other have to learn to take advantage of. He argues that an important feature of human communication is the degree and character of the faith[clarification needed] maintained by the participants in the dialogue, something that has priority over such virtues of language as "sincerity", "truth", and "objectivity". Indeed, he considers that to prioritize them over faith is to move into superstition.
Erwin Schrödinger
Erwin Schrödinger, a theoretical physicist and one of the leading pioneers of quantum mechanics, also published in the areas of colorimetry and color perception. In several of his philosophical writings, he defends the notion that qualia are not physical.
The sensation of colour cannot be accounted for by the physicist's objective picture of light-waves. Could the physiologist account for it, if he had fuller knowledge than he has of the processes in the retina and the nervous processes set up by them in the optical nerve bundles and in the brain? I do not think so.[41]: 154
He continues on to remark that subjective experiences do not form a one-to-one correspondence with stimuli. For example, light of wavelengths in the neighborhood of 590 nm produces the sensation of yellow, whereas exactly the same sensation is produced by mixing red light, with wavelength 760 nm, with green light, at 535 nm. From this he concludes that there is no "numerical connection with these physical, objective characteristics of the waves" and the sensations they produce.
Schrödinger concludes with a proposal of how it is that we might arrive at the mistaken belief that a satisfactory theoretical account of qualitative experience has been – or might ever be – achieved:
Scientific theories serve to facilitate the survey of our observations and experimental findings. Every scientist knows how difficult it is to remember a moderately extended group of facts, before at least some primitive theoretical picture about them has been shaped. It is therefore small wonder, and by no means to be blamed on the authors of original papers or of text-books, that after a reasonably coherent theory has been formed, they do not describe the bare facts they have found or wish to convey to the reader, but clothe them in the terminology of that theory or theories. This procedure, while very useful for our remembering the fact in a well-ordered pattern, tends to obliterate the distinction between the actual observations and the theory arisen from them. And since the former always are of some sensual quality, theories are easily thought to account for sensual qualities; which, of course, they never do.[41]: 163–164
Neuroscientists who state that qualia exist
Gerald Edelman
Neuroscientist and Nobel laureate in Physiology / Medicine Gerald Edelman, in his book Bright Air, Brilliant Fire, argues:[42]
One alternative that definitely does not seem feasible is to ignore completely the reality of qualia, formulating a theory of consciousness that aims by its descriptions alone to convey to a hypothetical “qualia-free” observer what it is to feel warmth, see green, and so on. In other words, this is an attempt to propose a theory based on a kind of God's-eye view of consciousness. But no scientific theory of whatever kind can be presented without already assuming that observers have sensation as well as perception. To assume otherwise is to indulge the errors of theories that attempt syntactical formulations mapped onto objectivist interpretations – theories that ignore embodiment as a source of meaning (see the Postscript). There is no qualia-free scientific observer.
Antonio Damasio
Neurologist Antonio Damasio, in his book The Feeling Of What Happens, states:[43]
Qualia are the simple sensory qualities to be found in the blueness of the sky or the tone of sound produced by a cello, and the fundamental components of the images in the movie metaphor are thus made of qualia.
Damasio also argues that consciousness is subjective and is different from behavior:[43]: 307–309
The resistance found in some scientific quarters to the use of subjective observations is a revisitation of an old argument between behaviorists, who believed that only behaviors, not mental experiences, could be studied objectively, and cognitivists, who believed that studying only behavior did not do justice to human complexity. The mind and its consciousness are first and foremost private phenomena, much as they offer many public signs of their existence to the interested observer. The conscious mind and its constituent properties are real entities, not illusions, and they must be investigated as the personal, private, subjective experiences that they are. The idea that subjective experiences are not scientifically accessible is nonsense.
Subjective entities require, as do objective ones, that enough observers undertake rigorous observations according to the same experimental design; and they require that those observations be checked for consistency across observers and that they yield some form of measurement. Moreover, knowledge gathered from subjective observations, e.g., introspective insights, can inspire objective experiments, and, no less importantly, subjective experiences can be explained in terms of the available scientific knowledge. The idea that the nature of subjective experiences can be grasped effectively by the study of their behavioral correlates is wrong. Although both mind and behavior are biological phenomena, mind is mind and behavior is behavior.
Mind and behavior can be correlated, and the correlation will become closer as science progresses, but in their respective specifications, mind and behavior are different. This is why, in all likelihood, I will never know your thoughts unless you tell me, and you will never know mine until I tell you.
Damasio also addresses qualia in his book Self Comes to Mind.[44][importance?]
Rodolfo Llinás
Neurologist Rodolfo Llinás states in his book I of the Vortex that from a strictly neurological perspective, qualia exist and are important to the organism's survival. He argues that qualia were important for the evolution of the nervous system of organisms, including simple organisms such as insects:[45]: 201–221
There are today two similar beliefs concerning the nature of qualia. The first is that qualia represent an epiphenomenon that is not necessary for the acquisition of consciousness. Second and somewhat related is the belief that while being the basis for consciousness, qualia appeared only in the highest life forms, suggesting that qualia represent a recently evolved central function that is present in only the more advanced brains. This view relegates the more lowly animals, for example ants, to a realm characterized by the absence of subjective experiences of any kind. It implies that these animals are wired with sets of automatic, reflexively organized circuits that provide for survival by maintaining a successful, albeit purely reactive interaction with the ongoing external world. Although primitive creatures such as ants and cockroaches may be wildly successful, for all practical purposes they are biological automatons.
[…] To me, these views lack a proper evolutionary perspective, which is perhaps why qualia are given so little overall emphasis in the study of brain function. We clearly understand that the functional architecture of the brain is a product of the slow tumblings of evolution and that brain function implements what natural selection has found to be the most beneficial in terms of species survivability. What is not often understood is how deeply related qualia truly are to the evolutionary, functional structure of the brain. […]
One cannot operate without qualia; they are properties of mind of monumental importance.
Llinás argues that qualia are ancient and necessary for an organism's survival and a product of neuronal oscillation. He gives the evidence of anesthesia of the brain and subsequent stimulation of limbs to demonstrate that qualia can be "turned off" by changing only the variable of neuronal oscillation (local brain electrical activity), while all other connections remain intact. He argues for an oscillatory-electrical origin of qualia, or important aspects of them.[45]: 202–207
Vilayanur Ramachandran
Vilayanur S. Ramachandran and William Hirstein[46] proposed three laws of qualia (with a fourth later added), which are "functional criteria that need to be fulfilled in order for certain neural events to be associated with qualia" by philosophers of the mind:
- Qualia are irrevocable and indubitable. You don't say 'maybe it is red but I can visualize it as green if I want to'. An explicit neural representation of red is created that invariably and automatically 'reports' this to higher brain centres.
- Once the representation is created, what can be done with it is open-ended. You have the luxury of choice, e.g., if you have the percept of an apple you can use it to tempt Adam, to keep the doctor away, bake a pie, or just to eat. Even though the representation at the input level is immutable and automatic, the output is potentially infinite. This isn't true for, say, a spinal reflex arc where the output is also inevitable and automatic. Indeed, a paraplegic can even have an erection and ejaculate without an orgasm.
- Short-term memory. The input invariably creates a representation that persists in short-term memory – long enough to allow time for choice of output. Without this component, again, you get just a reflex arc.
- Attention. Qualia and attention are closely linked. You need attention to fulfill criterion number two; to choose. A study of circuits involved in attention, therefore, will shed much light on the riddle of qualia.[47]
These authors approach qualia from an empirical perspective and not as a logical or philosophical problem. They wonder how qualia evolved. They consider a skeptical point of view in which, since the objective scientific description of the world is complete without qualia, it is nonsense to ask the question of why they evolved or what qualia are for. But they rule out such an option.
Based on the parsimony principle of Occam's razor, one could accept epiphenomenalism and deny qualia since they are not necessary for a description of the functioning of the brain. However, they argue that Occam's razor is not useful for scientific discovery.[46] For example, the discovery of relativity in physics was not the product of accepting Occam's razor but rather of rejecting it and asking the question of whether a deeper generalization, not required by the currently available data, was true and would allow for unexpected predictions. Most scientific discoveries arise, these authors argue, from ontologically promiscuous conjectures that do not come from current data.
The authors then point out that skepticism might be justified in the philosophical field, but that science is the wrong place for such skepticism. Such skeptical questions might include asking if "your red is not my green" or if we can be logically certain that we are not dreaming. Science, these authors assert, deals with what is probably true, beyond reasonable doubt, not with what can be known with complete and absolute certainty. The authors say that most neuroscientists and even most psychologists dispute the very existence of the “problem” of qualia[46]
Roger Orpwood
Roger Orpwood, an engineer who studies neural mechanisms, proposed a neurobiological model that gives rise to qualia and ultimately, consciousness. Advancements in cognitive and computational neuroscience necessitate study of the mind and qualia from a scientific perspective. Orpwood does not deny the existence of qualia, nor does he debate its physical or non-physical existence. He suggests that qualia are created through the neurobiological mechanism of re-entrant feedback in cortical systems.[48][49]
Orpwood first addresses the issue of information. One unsolved aspect of qualia is the concept of the fundamental information[clarification needed] involved in creating the experience. He does not take a position on the metaphysics of the information underlying the experience of qualia, nor does he state what information actually is, but he does say that information is of two types: the information structure and information message. Information structures are defined by the physical vehicles and structural, biological patterns that encode information. That encoded information is the information message: a source that describes what that information is. The neural mechanism or network receives input information structures, completes a designated instructional task (firing of the neuron or network), and outputs a modified information structure to downstream regions. The information message is the purpose and meaning of the information structure and causally exists as a result of that particular information structure. Modification of the information structure changes the meaning of the information message, but the message itself cannot be directly altered.
Local cortical networks can receive feedback from their own output information structures. This form of local feedback continuously cycles part of the network's output structures as its next input information structure. Since the output structure must represent the information message derived from the input structure, each consecutive cycle that is fed-back represents the output structure the network just generated. As the network of mechanisms cannot recognize the information message, but only the input information structure, the network is unaware that it is representing its own previous outputs. The neural mechanisms are merely completing their instructional tasks and outputting any recognizable[clarification needed] information structures. Orpwood proposes that these local networks come into an attractor state that consistently outputs exactly the same information structure as the input structure. Instead of only representing the information message derived from the input structure, the network will now represent its own output and thereby its own information message. As the input structures are fed-back, the network identifies the previous information structure as being a previous representation of the information message. As Orpwood writes:
Once an attractor state has been established, the output [of a network] is a representation of its own identity to the network.[49]: 4
Orpwood explains the neurobiological manifestation of qualia as representations of the network's own output structures, by which it represents its own information message. This is particular to networks of pyramidal neurons. Although computational neuroscience still has much to investigate regarding pyramidal neurons, their complex circuitry is unusual. The complexity of pyramidal neuron networks in a species is directly related to the increase in that species's functional capabilities.[50] When human pyramidal networks are compared with other primate species and with species with less intricate behavioral and social interactions, the complexity of these neural networks drastically decline.[needs copy edit] These networks are also more complex[compared to?] in frontal brain regions. These regions are often associated with[weasel words] conscious assessment and modification of one's immediate environment; often referred to as executive functions.
One needs sensory input to gain information from the environment. Perception of that input is necessary in order to navigate and modify interactions with the environment. This suggests that frontal regions containing more complex pyramidal networks are associated with an increased perceptive capacity.[how?] As perception is necessary for conscious thought to occur,[citation needed] and since qualia derive from[clarification needed] consciousness of some perception, qualia may be specific to the functional capacity of[clarification needed] pyramidal networks.[non sequitur] For this reason Orpwood believes that the mechanisms of re-entrant feedback may not only create qualia, but also be the foundation to consciousness.
Other issues
Indeterminacy
A criticism similar to Nietzsche's criticism of Kant's "thing in itself" applies also to qualia: Qualia are unobservable in others and unquantifiable in us. We cannot possibly be sure, when discussing individual qualia, that we are even discussing the same phenomena. Thus, any discussion of them is of indeterminate value, as descriptions of qualia are necessarily of indeterminate accuracy.[citation needed]
Qualia, like "things in themselves", have no publicly demonstrable properties. This, along with the impossibility of being sure that we are communicating about the same qualia, makes them of indeterminate value and definition in any philosophy in which proof relies upon precise definition.[citation needed]
On the other hand, qualia could be considered akin to Kantian phenomena since they are held to be seemings of appearances. Revonsuo[clarification needed], however, considers that, within neurophysiological inquiry, a definition[vague] at the level of the fields[clarification needed] may become possible (just as we can define a television picture at the level of liquid crystal pixels).[non sequitur]
Causal efficacy
Whether qualia or consciousness can play any causal role in the physical world remains an open question. Epiphenomenalism acknowledges the existence of qualia while denying them any causal power. The position has been criticized by a number of philosophers,[a] if only because our own consciousness seem to be causally active.[54] In order to avoid epiphenomenalism, one who believes that qualia are nonphysical would need to embrace something like interactionist dualism; or perhaps emergentism, the claim that there are as yet unknown causal relations between the mental and physical. This in turn would imply that qualia can be detected by an external agency[clarification needed] through their causal powers.
Epistemological issues
Examples of qualia might include "the pain of a headache, the taste of wine, or the redness of an evening sky". But such examples prejudge a central debate on qualia.[citation needed] Suppose someone wants to know the nature of the liquid crystal pixels on a television screen—those tiny elements that make up the picture. It would not suffice as an answer to say that they are the "redness of an evening sky" as it appears on the screen; this would ignore their real character. Relying on a list like the one above assumes that we must tie sensations to both the notion of given objects in the world (the "head", "wine", "an evening sky") and to the properties with which we characterize the experiences themselves ("redness", for example).
Nor is it satisfactory to print a little red square as at the top of this article, for, since each person has a slightly different registration of the light-rays,[55] it confusingly suggests that we all have the same response.[how?] Imagine in a television shop seeing "a red square" on twenty screens at once, each slightly different – something of vital importance would be overlooked if a single example were to be taken as defining them all.[further explanation needed]
Yet it has been argued whether or not identification with the external object should still be the core of a correct approach to sensation, for there are many who state the definition thus because they regard the link with external reality as crucial.[needs copy edit] If sensations are defined as "raw feels", this threatens the reliability of knowledge because if one sees them as neurophysiological happenings in the brain, it is difficult to understand how they could have any connection to entities, whether in the body or the external world.
John McDowell, for example, declared that to countenance qualia as a "bare presence" prevents us ever gaining a certain ground for our knowledge.[56] The issue is epistemological: it would appear that access to knowledge is blocked if one allows the existence of qualia as fields in which only virtual constructs are before[clarification needed] the mind. His reason is that it[ambiguous] puts the entities about which we require knowledge behind a "veil of perception", an occult field of "appearance" which leaves us ignorant of the reality presumed to be beyond it. He is convinced that such uncertainty propels[needs copy edit] into the dangerous regions of relativism and solipsism. These constitute an ethical argument against qualia being something going on in the brain,[clarification needed] and these implications are probably largely responsible for the fact that in the 20th century it was regarded as not only freakish, but also dangerously misguided to uphold the notion of sensations as going on inside the head.[citation needed] The argument was usually strengthened with mockery at the very idea of "redness" being in the brain:[57] – "How can there be red neurons in the brain?"
Viewing sensations as "raw feels" implies that initially they have not yet – to carry on the metaphor – been "cooked", that is, unified into "things" and "persons", which is something the mind does after the sensation has responded to the blank input, that response driven by motivation, that is, initially by pain and pleasure, and subsequently, when memories have been implanted, by desire and fear.[needs copy edit] Such a "raw-feel" state has been called "non-epistemic". In support of this view, the theorists[who?] cite a range of empirical facts, for example:
- There are brain-damaged persons, known as "agnosics" (literally "not-knowing") who still have vivid visual sensations but are quite unable to identify any entity before them, including parts of their own body.
- There is also the similar predicament of persons, formerly blind, who are given sight for the first time. And consider what it is a newborn baby experiences.
German physicist Hermann von Helmholtz proposed an experiment to demonstrate the non-epistemic nature of qualia: His instructions were to stand in front of a familiar landscape, turn your back on it, bend down and look at the landscape between your legs – you will find it difficult in the upside-down view to recognize what you found familiar before.[58]
These examples suggest the reality of a "bare presence" – that is, knowledgeless sensation that is no more than evidence. Supporters of the non-epistemic theory thus regard sensations as only data in the sense that they are "given" (Latin datum, "given") and fundamentally involuntary, which is a good reason for not regarding them as basically mental.[why?] In the 20th century they were called "sense-data" by the proponents of qualia, but this led to the confusion that they carried with them reliable proofs of objective causal origins. For instance, one supporter of qualia was happy to speak of the redness and bulginess of a cricket ball as a typical "sense-datum",[59] though not all of them were happy to define qualia by their relation to external entities.[60] The modern argument centers on how we learn under the regime of motivation[definition needed] to interpret sensory evidence in terms of "things", "persons", and "selves" through a continuing process of feedback.
The definition of qualia inevitably brings with it philosophical and neurophysiological presuppositions. The question of what qualia can be raises profound issues in the philosophy of mind, since some materialists want to deny their existence altogether: on the other hand, if they are accepted, they cannot be easily accounted for as they raise the difficult problem of consciousness. There are committed dualists such as Richard L. Amoroso or John Hagelin who believe that the mental and the material are two distinct aspects of physical reality like the distinction between the classical and quantum regimes.[61] In contrast, there are direct realists for whom qualia are unscientific as there appears to be no way of making them fit within the modern scientific picture; and there are committed proselytizers for a final truth who reject them as forcing knowledge out of reach.
See also
- Binding problem – Unanswered question in the study of consciousness
- Blockhead (thought experiment) – philosophical thought experiment about a computer which has been programmed with all possible sentences in a natural language, so as to be able to pass the Turing test despite lacking intelligence
- Chinese room – Thought experiment on artificial intelligence by John Searle
- Explanatory gap – Inability to describe conscious experiences in solely physical or structural terms
- Feeling – Conscious subjective experience of emotion
- Form constant – Recurringly observed geometric pattern
- Further facts – Philosophy idea
- Hard problem of consciousness – Philosophical concept, first stated by David Chalmers in 1995
- Ideasthesia – Phenomenon in which concepts evoke sensory experiences
- Knowledge argument – Thought experiment
- Mind–body problem – Open question in philosophy of how abstract minds interact with physical bodies
- Open individualism – Philosophical concept
- Philosophical zombie – Thought experiment in philosophy
- Process philosophy – Philosophical approach
- Self-awareness – Capacity for introspection and individuation as a subject
- Self-reference – Sentence, idea or formula that refers to itself
- Subjectivity – Philosophical concept, related to consciousness, agency, personhood, reality, and truth
- Synesthesia – Neurological condition involving the crossing of senses
- Vertiginous question – philosophical question: “of all the subjects of experience out there, why is this one—the one corresponding to the human being referred to as me—the one whose experiences are live?”
Notes
References
- Dennett, D. (2015). "Why and how does consciousness seem the way it seems?". In Metzinger, T.; Windt, J. (eds.). Open mind. Mind Group. pp. 387–398.
- Edelman, G.; Gally, J.; Baars, B. (2011). "Biology of consciousness". Frontiers In Psychology. 2 (4): 1–6.
- Edelman, G. (1992). Bright air, brilliant fire. BasicBooks.
- Edelman, G. (2003). "Naturalizing consciousness: A theoretical framework". Proceedings of the National Academy of Sciences. 100 (9): 5520–5524.
- Koch, C. (2019). The feeling of life itself. The MIT Press.
- Llinás, R. (2003). I of the Vortex. MIT Press. pp. 202–207.
- Oizumi, M.; Albantakis, L.; Tononi, G. (2014). "From the phenomenology to the mechanisms of consciousness: Integrated information theory 3.0". PLOS Computational Biology. 10. e1003588
- Overgaard, M.; Mogensen, J.; Kirkeby-Hinrup, A., eds. (2021). Beyond neural correlates of consciousness. Routledge Taylor & Francis.
- Ramachandran, V.; Hirstein, W. (1997). "Three laws of qualia. What neurology tells us about the biological functions of consciousness, qualia and the self". Journal of consciousness studies. 4 (5–6): 429–458.
- Tononi, G.; Boly, M.; Massimini, M.; Koch, C. (2016). "Integrated information theory: From consciousness to its physical substrate". Nature Reviews Neuroscience. 17: 450–461.
- "Qualia". The Stanford Encyclopedia of Philosophy. Metaphysics Research Lab, Stanford University. 2018.
- Shepherd, J. (2018). Consciousness and Moral Status. Routledge, Taylor & Francis group.
- Slater, A.M., ed. (1999). Perceptual Development: Visual, auditory and speech perception in infancy. East Sussex: Psychology Press. ISBN 978-0863778513 – via Google Books.
- Tye, M. (1995). Ten Problems of Consciousness: A representational theory of the phenomenal mind. Cambridge, MA: MIT Press.
free copy available from
"qualia on consc.net".
- Maund, J.B. (2003). Perception. Chesham: Acumen Pub. Ltd.
- Wright, Edmond (2005-11-16). "Narrative, Perception, Language, and Faith". Basingstoke: Palgrave-Macmillan. Retrieved 2010-12-03.
also available from
"Three laws of qualia" (PDF). imprint.co.uk.
- Orpwood, Roger D. (June 2010). "Perceptual qualia and local network behavior in the cerebral cortex". Journal of Integrative Neuroscience. 09 (2): 123–152. doi:10.1142/s021963521000241x. PMID 20589951.
- Georgiev, Danko D. (2020). "Inner privacy of conscious experiences and quantum information". BioSystems. 187: 104051. arXiv:2001.00909. doi:10.1016/j.biosystems.2019.104051. PMID 31629783. S2CID 204813557.
- Amoroso, Richard L. (2010). Complementarity of Mind & Body: Realizing the dream of Descartes, Einstein, & Eccles. New York: Nova Science Publishers.
Other references
- various authors (2007-03-19). Chalmers, David (ed.). "Online collection of papers on qualia". Archived from the original on 2007-03-19 – via Archive.org.
- Lormand, Eric. "Qualia! (Now showing at a theatre near you)" (response to D. Dennett). University of Michigan.
- Dennett, Daniel. "Quining qualia". Cognitive Studies. Tufts University.
- Ramachandran, V.S.; Hirstein, W. "Three laws of qualia" (PDF). imprint.co.uk. (biological perspective)
- Gregory, Richard. "Brainy mind (qualia and the sensation of time)".
- Field Guide to the Philosophy of Mind
- Alter, Torin. "The Knowledge Argument". Field Guide to the Philosophy of Mind (article). University of Rome. Archived from the original on 2012-12-03. Retrieved 2007-01-22.
- Robinson, William. "Qualia realism". Field Guide to the Philosophy of Mind (article). University of Rome. Archived from the original on 2012-12-03. Retrieved 2007-01-22.
- Stanford Encyclopedia of Philosophy
- Byrne, Alex (2020). Inverted qualia. Stanford Encyclopedia of Philosophy (article). Stanford U.
- Tye, Michael (2021). Qualia. Stanford Encyclopedia of Philosophy (article). Stanford U.
- Nida-Rümelin, Martine (2021). Qualia: The knowledge argument. Stanford Encyclopedia of Philosophy (article). Stanford U.
Further reading
- Mroczko-Wąsowicz, A.; Nikolić, D. (2014). "Semantic mechanisms may be responsible for developing synesthesia" (PDF). Frontiers in Human Neuroscience. 8: 509. doi:10.3389/fnhum.2014.00509. PMC 4137691. PMID 25191239.
External links
- "Qualia". Internet Encyclopedia of Philosophy (article). University of Tennessee. ISSN 2161-0002.
- Qualia. Stanford Encyclopedia of Philosophy (article). Metaphysics Research Lab, Stanford University. 2021.
- Qualia
- Abstraction
- Cognition
- Concepts in epistemology
- Concepts in metaphilosophy
- Concepts in metaphysics
- Concepts in the philosophy of mind
- Concepts in the philosophy of science
- Consciousness studies
- Mental processes
- Metaphysics of mind
- Ontology
- Perception
- Subjective experience
- Theory of mind
- Thought
- Unsolved problems in neuroscience
https://en.wikipedia.org/wiki/Qualia
Reason is the capacity of consciously applying logic by drawing conclusions from new or existing information, with the aim of seeking the truth.[1][2] It is closely[how?] associated with such characteristically human activities as philosophy, science, language, mathematics, and art, and is normally considered to be a distinguishing ability possessed by humans.[3] Reason is sometimes referred to as rationality.[4]
Reasoning is associated with the acts of thinking and cognition, and involves the use of one's intellect. The field of logic studies the ways in which humans can use formal reasoning to produce logically valid arguments.[5] Reasoning may be subdivided into forms of logical reasoning, such as deductive reasoning, inductive reasoning, and abductive reasoning. Aristotle drew a distinction between logical discursive reasoning (reason proper), and intuitive reasoning,[6] in which the reasoning process through intuition—however valid—may tend toward the personal and the subjectively opaque. In some social and political settings logical and intuitive modes of reasoning may clash, while in other contexts intuition and formal reason are seen as complementary rather than adversarial. For example, in mathematics, intuition is often necessary for the creative processes involved with arriving at a formal proof, arguably the most difficult of formal reasoning tasks.
Reasoning, like habit or intuition, is one of the ways by which thinking moves from one idea to a related idea. For example, reasoning is the means by which rational individuals understand sensory information from their environments, or conceptualize abstract dichotomies such as cause and effect, truth and falsehood, or ideas regarding notions of good or evil. Reasoning, as a part of executive decision making, is also closely identified with the ability to self-consciously change, in terms of goals, beliefs, attitudes, traditions, and institutions, and therefore with the capacity for freedom and self-determination.[7]
In contrast to the use of "reason" as an abstract noun, a reason is a consideration given which either explains or justifies events, phenomena, or behavior.[8] Reasons justify decisions, reasons support explanations of natural phenomena; reasons can be given to explain the actions (conduct) of individuals.
Using reason, or reasoning, can also be described more plainly as providing good, or the best, reasons. For example, when evaluating a moral decision, "morality is, at the very least, the effort to guide one's conduct by reason—that is, doing what there are the best reasons for doing—while giving equal [and impartial] weight to the interests of all those affected by what one does."[9]
Psychologists and cognitive scientists have attempted to study and explain how people reason, e.g. which cognitive and neural processes are engaged, and how cultural factors affect the inferences that people draw. The field of automated reasoning studies how reasoning may or may not be modeled computationally. Animal psychology considers the question of whether animals other than humans can reason.
https://en.wikipedia.org/wiki/Reason
Reality is the sum or aggregate of all that is real or existent within a system, as opposed to that which is only imaginary, nonexistent or nonactual. The term is also used to refer to the ontological status of things, indicating their existence.[1] In physical terms, reality is the totality of a system, known and unknown.[2]
Philosophical questions about the nature of reality or existence or being are considered under the rubric of ontology, which is a major branch of metaphysics in the Western philosophical tradition. Ontological questions also feature in diverse branches of philosophy, including the philosophy of science, of religion, of mathematics, and philosophical logic. These include questions about whether only physical objects are real (i.e., physicalism), whether reality is fundamentally immaterial (e.g. idealism), whether hypothetical unobservable entities posited by scientific theories exist, whether a 'God' exists, whether numbers and other abstract objects exist, and whether possible worlds exist. Epistemology is concerned with what can be known or inferred as likely and how, whereby in the modern world emphasis is put on reason, empirical evidence and science as sources and methods to determine or investigate reality.
https://en.wikipedia.org/wiki/Reality
Rational reconstruction is a philosophical term with several distinct meanings. It is found in the work of Jürgen Habermas and Imre Lakatos.
https://en.wikipedia.org/wiki/Rational_reconstruction
Rational ignorance is refraining from acquiring knowledge when the supposed cost of educating oneself on an issue exceeds the expected potential benefit that the knowledge would provide.
Ignorance about an issue is said to be "rational" when the cost of educating oneself about the issue sufficiently to make an informed decision can outweigh any potential benefit one could reasonably expect to gain from that decision, and so it would be irrational to waste time doing so. This has consequences for the quality of decisions made by large numbers of people, such as in general elections, where the probability of any one vote changing the outcome is very small.
The term is most often found in economics, particularly public choice theory, but also used in other disciplines which study rationality and choice, including philosophy (epistemology) and game theory.
The term was coined by Anthony Downs in An Economic Theory of Democracy.[1]
https://en.wikipedia.org/wiki/Rational_ignorance
In philosophy and rhetoric, eristic (from Eris, the ancient Greek goddess of chaos, strife, and discord) refers to an argument that aims to successfully dispute another's argument, rather than searching for truth. According to T.H. Irwin, "It is characteristic of the eristic to think of some arguments as a way of defeating the other side, by showing that an opponent must assent to the negation of what he initially took himself to believe."[1] Eristic is arguing for the sake of conflict, as opposed to resolving conflict.[2]
https://en.wikipedia.org/wiki/Eristic
Eschatological verification describes a process whereby a proposition can be verified after death. A proposition such as "there is an afterlife" is verifiable if true but not falsifiable if false (if it's false, the individual will not know it's false, because they have no state of being). The term is most commonly used in relation to God and the afterlife, although there may be other propositions - such as moral propositions - which may also be verified after death.
John Hick has expressed the premise as an allegory of a quest to a Celestial City. In this parable, a theist and an atheist are both walking down the same road. The theist believes there is a destination, the atheist believes there is not. If they reach the destination, the theist will have been proven right; but if there is no destination on an endless road, this can never be verified. This is an attempt to explain how a theist expects some form of life or existence after death and an atheist does not. They both have separate belief systems and live life accordingly, but logically one is right and the other is not. If the theist is right, he will be proven so when he arrives in the afterlife. But if the atheist is right, they will simply both be dead and nothing will be verified.
This acts as a response to Verificationism. Under Hick's analogy claims about the afterlife are verifiable in principle because the truth becomes clear after death. To some extent it is therefore wrong to claim that religious language cannot be verified because it can (when you're dead).
https://en.wikipedia.org/wiki/Eschatological_verification
The evil demon, also known as Deus deceptor,[1] malicious demon,[2] and evil genius,[1][3] is an epistemological concept that features prominently in Cartesian philosophy.[1] In the first of his 1641 Meditations on First Philosophy, Descartes imagines that a malevolent God[1] or an evil demon, of "utmost power and cunning has employed all his energies in order to deceive me." This malevolent God or evil demon is imagined to present a complete illusion of an external world, so that Descartes can say, "I shall think that the sky, the air, the earth, colours, shapes, sounds and all external things are merely the delusions of dreams which he has devised to ensnare my judgement. I shall consider myself as not having hands or eyes, or flesh, or blood or senses, but as falsely believing that I have all these things."
https://en.wikipedia.org/wiki/Evil_demon
In the case of uncertainty, expectation is what is considered the most likely to happen. An expectation, which is a belief that is centered on the future, may or may not be realistic. A less advantageous result gives rise to the emotion of disappointment. If something happens that is not at all expected, it is a surprise. An expectation about the behavior or performance of another person, expressed to that person, may have the nature of a strong request, or an order; this kind of expectation is called a social norm. The degree to which something is expected to be true can be expressed using fuzzy logic. Anticipation is the emotion corresponding to expectation.
https://en.wikipedia.org/wiki/Expectation_(epistemic)
Experiential knowledge is knowledge gained through experience, as opposed to a priori (before experience) knowledge: it can also be contrasted both with propositional (textbook) knowledge, and with practical knowledge.[1]
Experiential knowledge is cognate to Michael Polanyi's personal knowledge, as well as to Bertrand Russell's contrast of Knowledge by Acquaintance and by Description.[2]
https://en.wikipedia.org/wiki/Experiential_knowledge
Suspended judgment is a cognitive process and a rational state of mind in which one withholds judgments, particularly on the drawing of moral or ethical conclusions. The opposite of suspension of judgment is premature judgement, usually shortened to prejudice. While prejudgment involves drawing a conclusion or making a judgment before having the information relevant to such a judgment, suspension of judgment involves waiting for all the facts before making a decision.
https://en.wikipedia.org/wiki/Suspension_of_judgment
In epistemology, the regress argument is the argument that any proposition requires a justification. However, any justification itself requires support. This means that any proposition whatsoever can be endlessly (infinitely) questioned, resulting in infinite regress. It is a problem in epistemology and in any general situation where a statement has to be justified.[1][2][3]
The argument is also known as diallelus[4] (Latin) or diallelon, from Greek di' allelon "through or by means of one another" and as the epistemic regress problem. It is an element of the Münchhausen trilemma.[5]
Structure
Assuming that knowledge is justified true belief, then:
- Suppose that P is some piece of knowledge. Then P is a justified true belief.
- The only thing that can justify P is another statement – let's call it P1; so P1 justifies P.
- But if P1 is to be a satisfactory justification for P, then we must know that P1 is true.
- But for P1 to be known, it must also be a justified true belief.
- That justification will be another statement - let's call it P2; so P2 justifies P1.
- But if P2 is to be a satisfactory justification for P1, then we must know that P2 is true
- But for P2 to count as knowledge, it must itself be a justified true belief.
- That justification will in turn be another statement - let's call it P3; so P3 justifies P2.
- and so on, ad infinitum.
Responses
Throughout history many responses to this problem have been generated. The major counter-arguments are
- some statements do not need justification,
- the chain of reasoning loops back on itself,
- the sequence never finishes,
- belief cannot be justified as beyond doubt.
Foundationalism
Perhaps the chain begins with a belief that is justified, but which is not justified by another belief. Such beliefs are called basic beliefs. In this solution, which is called foundationalism, all beliefs are justified by basic beliefs. Foundationalism seeks to escape the regress argument by claiming that there are some beliefs for which it is improper to ask for a justification. (See also a priori.)
Foundationalism is the belief that a chain of justification begins with a belief that is justified, but which is not justified by another belief. Thus, a belief is justified if and only if:
- it is a basic/foundational belief, or
- it is justified by a basic belief
- it is justified by a chain of beliefs that is ultimately justified by a basic belief or beliefs.
Foundationalism can be compared to a building. Ordinary individual beliefs occupy the upper stories of the building; basic, or foundational beliefs are down in the basement, in the foundation of the building, holding everything else up. In a similar way, individual beliefs, say about economics or ethics, rest on more basic beliefs, say about the nature of human beings; and those rest on still more basic beliefs, say about the mind; and in the end the entire system rests on a set of basic beliefs which are not justified by other beliefs.
Coherentism
Alternatively, the chain of reasoning may loop around on itself, forming a circle. In this case, the justification of any statement is used, perhaps after a long chain of reasoning, in justifying itself, and the argument is circular. This is a version of coherentism.
Coherentism is the belief that an idea is justified if and only if it is part of a coherent system of mutually supporting beliefs (i.e., beliefs that support each other). In effect Coherentism denies that justification can only take the form of a chain. Coherentism replaces the chain with a holistic web.
The most common objection to naïve Coherentism is that it relies on the idea that circular justification is acceptable. In this view, P ultimately supports P, begging the question. Coherentists reply that it is not just P that is supporting P, but P along with the totality of the other statements in the whole system of belief.
Coherentism accepts any belief that is part of a coherent system of beliefs. In contrast, P can cohere with P1 and P2 without P, P1 or P2 being true. Instead, Coherentists might say that it is very unlikely that the whole system would be both untrue and consistent, and that if some part of the system was untrue, it would almost certainly be inconsistent with some other part of the system.
A third objection is that some beliefs arise from experience and not from other beliefs. An example is that one is looking into a room which is totally dark. The lights turn on momentarily and one sees a white canopy bed in the room. The belief that there is a white canopy bed in this room is based entirely on experience and not on any other belief. Of course other possibilities exist, such as that the white canopy bed is entirely an illusion or that one is hallucinating, but the belief remains well-justified. Coherentists might respond that the belief which supports the belief that there is a white canopy bed in this room is that one saw the bed, however briefly. This appears to be an immediate qualifier which does not depend on other beliefs, and thus seems to prove that Coherentism is not true because beliefs can be justified by concepts other than beliefs. But others have argued that the experience of seeing the bed is indeed dependent on other beliefs, about what a bed, a canopy and so on, actually look like.
Another objection is that the rule demanding "coherence" in a system of ideas seems to be an unjustified belief.
Infinitism
Infinitism argues that the chain can go on forever. Critics argue that this means there is never adequate justification for any statement in the chain.
Skepticism
Skeptics reject the three above responses and argue that beliefs cannot be justified as beyond doubt. Note that many skeptics do not deny that things may appear in a certain way. However, such sense impressions cannot, in the skeptical view, be used to find beliefs that cannot be doubted. Also, skeptics do not deny that, for example, many laws of nature give the appearance of working or that doing certain things give the appearance of producing pleasure/pain or even that reason and logic seem to be useful tools. Skepticism is in this view valuable since it encourages continued investigation.[6]
Synthesized approaches
Common sense
The method of common sense espoused by such philosophers as Thomas Reid and G. E. Moore points out that whenever we investigate anything at all, whenever we start thinking about some subject, we have to make assumptions. When one tries to support one's assumptions with reasons, one must make yet more assumptions. Since it is inevitable that we will make some assumptions, why not assume those things that are most obvious: the matters of common sense that no one ever seriously doubts.
"Common sense" here does not mean old adages like "Chicken soup is good for colds" but statements about the background in which our experiences occur. Examples would be "Human beings typically have two eyes, two ears, two hands, two feet", or "The world has a ground and a sky" or "Plants and animals come in a wide variety of sizes and colors" or "I am conscious and alive right now". These are all the absolutely most obvious sorts of claims that one could possibly make; and, said Reid and Moore, these are the claims that make up common sense.
This view can be seen as either a version of foundationalism, with common sense statements taking the role of basic statements, or as a version of Coherentism. In this case, commonsense statements are statements that are so crucial to keeping the account coherent that they are all but impossible to deny.
If the method of common sense is correct, then philosophers may take the principles of common sense for granted. They do not need criteria in order to judge whether a proposition is true or not. They can also take some justifications for granted, according to common sense. They can get around Sextus' problem of the criterion because there is no infinite regress or circle of reasoning, because the principles of common sense ground the entire chain of reasoning.
Critical philosophy
Another escape from the diallelus is critical philosophy, which denies that beliefs should ever be justified at all. Rather, the job of philosophers is to subject all beliefs (including beliefs about truth criteria) to criticism, attempting to discredit them rather than justifying them. Then, these philosophers say, it is rational to act on those beliefs that have best withstood criticism, whether or not they meet any specific criterion of truth. Karl Popper expanded on this idea to include a quantitative measurement he called verisimilitude, or truth-likeness. He showed that even if one could never justify a particular claim, one can compare the verisimilitude of two competing claims by criticism to judge which is superior to the other.
Pragmatism
The pragmatist philosopher William James suggests that, ultimately, everyone settles at some level of explanation based on one's personal preferences that fit the particular individual's psychological needs. People select whatever level of explanation fits their needs, and things other than logic and reason determine those needs. In The Sentiment of Rationality, James compares the philosopher, who insists on a high degree of justification, and the boor, who accepts or rejects ideals without much thought:
The philosopher's logical tranquillity is thus in essence no other than the boor's. They differ only as to the point at which each refuses to let further considerations upset the absoluteness of the data he assumes.
See also
- Epistemology – Branch of philosophy concerning knowledge
- Cogito, ergo sum – Philosophical statement made by René Descartes
- Infinite regress – Philosophical problem
- Münchhausen trilemma – A thought experiment used to demonstrate the impossibility of proving any truth
- Plato's Theaetetus – Socratic dialogue by Plato concerning the nature of knowledge
- Problem of the criterion – Philosophical problem about the starting point of knowledge
- Turtles all the way down – Expression of the problem of infinite regress
- What the Tortoise Said to Achilles – Allegorical dialogue by Lewis Carroll
References
- skepticism on Stanford Encyclopedia of Philosophy
https://en.wikipedia.org/wiki/Regress_argument
The simulation hypothesis proposes that all of our existence is a simulated reality, such as a computer simulation.[1][2][3] This simulation could contain conscious minds that may or may not know that they live inside a simulation. This is quite different from the current, technologically achievable concept of virtual reality, which is easily distinguished from the experience of actuality. Simulated reality, by contrast, would be hard or impossible to separate from "true" reality. There has been much debate over this topic, ranging from philosophical discourse to practical applications in computing.
The simulation hypothesis, which was popularized in its current form by Nick Bostrom,[4] bears a close resemblance to various other skeptical scenarios from throughout the history of philosophy.
The suggestion that such a hypothesis is compatible with all human perceptual experiences is thought to have significant epistemological consequences in the form of philosophical skepticism. Versions of the hypothesis have also been featured in science fiction, appearing as a central plot device in many stories and films.[5]
The hypothesis popularized by Bostrom is very disputed, with, for example, theoretical physicist Sabine Hossenfelder, who called it pseudoscience[6] and cosmologist George F. R. Ellis, who stated that "[the hypothesis] is totally impracticable from a technical viewpoint" and that "protagonists seem to have confused science fiction with science. Late-night pub discussion is not a viable theory."[7]
Origins
There is a long philosophical and scientific history to the underlying thesis that reality is an illusion. This skeptical hypothesis can be traced back to antiquity; for example, to the "Butterfly Dream" of Zhuangzi,[8] or the Indian philosophy of Maya, or in Ancient Greek philosophy Anaxarchus and Monimus likened existing things to a scene-painting and supposed them to resemble the impressions experienced in sleep or madness.[9]
Aztec philosophical texts theorized that the world was a painting or book written by the Teotl.[10]
Nietzsche, in his 1886 book Beyond Good and Evil, chastised philosophers for seeking to find the true world behind the deceptive world of appearances.
It is nothing more than a moral prejudice that truth is worth more than semblance; it is, in fact, the worst proved supposition in the world.... Why might not the world which concerns us—be a fiction?[11]
https://en.wikipedia.org/wiki/Simulation_hypothesis
Hauntology (a portmanteau of haunting and ontology, also spectral studies, spectralities, or the spectral turn) is a range of ideas referring to the return or persistence of elements from the social or cultural past, as in the manner of a ghost. The term is a neologism first introduced by French philosopher Jacques Derrida in his 1993 book Specters of Marx. It has since been invoked in fields such as visual arts, philosophy, electronic music, anthropology, politics, fiction, and literary criticism.[1]
While Christine Brooke-Rose had previously punned "dehauntological" (on "deontological") in Amalgamemnon (1984),[2] Derrida initially used "hauntology" for his idea of the atemporal nature of Marxism and its tendency to "haunt Western society from beyond the grave".[3] It describes a situation of temporal and ontological disjunction in which presence, especially socially and culturally, is replaced by a deferred non-origin.[1] The concept is derived from deconstruction, in which any attempt to locate the origin of identity or history must inevitably find itself dependent on an always-already existing set of linguistic conditions.[4] Despite being the central focus of Spectres of Marx, the word hauntology appears only three times in the book, and there is little consistency in how other writers define the term.[5]
In the 2000s, the term was applied to musicians by theorists Simon Reynolds and Mark Fisher, who were said to explore ideas related to temporal disjunction, retrofuturism, cultural memory, and the persistence of the past. Hauntology has been used as a critical lens in various forms of media and theory, including music, aesthetics, political theory, architecture, Africanfuturism, Afrofuturism, Neo-futurism, Metamodernism, anthropology, and psychoanalysis.[1][failed verification][6][page needed] Due to the difficulty in understanding the concept, there is little consistency in how other writers define the term.[5]
https://en.wikipedia.org/wiki/Hauntology
"Fusion of horizons" (German: Horizontverschmelzung) is a dialectical concept which results from the rejection of two alternatives: objectivism, whereby the objectification of the other is premised on the forgetting of oneself; and absolute knowledge, according to which universal history can be articulated within a single horizon. Therefore, it argues that we exist neither in closed horizons, nor within a horizon that is unique.
People come from different backgrounds and it is not possible to totally remove oneself from one's background, history, culture, gender, language, education, etc. to an entirely different system of attitudes, beliefs and ways of thinking.[1] People may be looking for a way to be engaged in understanding a conversation or dialogue about different cultures and the speaker interprets texts or stories based on his or her past experience and prejudice. Therefore, "hermeneutic reflection and determination of one's own present life interpretation calls for the unfolding of one's 'effective-historical' consciousness."[2] During the discourse, a fusion of "horizons" takes place between the speaker and listeners.
https://en.wikipedia.org/wiki/Fusion_of_horizons
The fact–value distinction is a fundamental epistemological distinction described between:[1]
- Statements of fact (positive or descriptive statements), based upon reason and physical observation, and which are examined via the empirical method.
- Statements of value (normative or prescriptive statements), which encompass ethics and aesthetics, and are studied via axiology.
This barrier between fact and value, as construed in epistemology, implies it is impossible to derive ethical claims from factual arguments, or to defend the former using the latter.[2]
The fact–value distinction is closely related to, and derived from, the is–ought problem in moral philosophy, characterized by David Hume.[3] The terms are often used interchangeably, though philosophical discourse concerning the is–ought problem does not usually encompass aesthetics.
If values do not arise from facts, it opens questions about whether these have distinct spheres of origin. Evolutionary psychology implicitly challenges the fact–value distinction to the extent that cognitive psychology is positioned as foundational in human ethics, as a single sphere of origin. Among the new atheists, who often lean toward evolutionary psychology, Sam Harris in particular endorses and promotes a science of morality as outlined in his book The Moral Landscape (2010).
Religion and science
In his essay Science as a Vocation (1917) Max Weber draws a distinction between facts and values. He argues that facts can be determined through the methods of a value-free, objective social science, while values are derived through culture and religion, the truth of which cannot be known through science. He writes, "it is one thing to state facts, to determine mathematical or logical relations or the internal structure of cultural values, while it is another thing to answer questions of the value of culture and its individual contents and the question of how one should act in the cultural community and in political associations. These are quite heterogeneous problems."[4] In his 1919 essay Politics as a Vocation, he argues that facts, like actions, do not in themselves contain any intrinsic meaning or power: "any ethic in the world could establish substantially identical commandments applicable to all relationships."[5]
To MLK Jr., "Science deals mainly with facts; religion deals mainly with values. The two are not rivals. They are complementary."[6][7][8] He stated that science keeps religion from"crippling irrationalism and paralyzing obscurantism" whereas Religion prevents science from "falling into ... obsolete materialism and moral nihilism."[9]
Albert Einstein remarked that
the realms of religion and science in themselves are clearly marked off from each other, nevertheless there exist between the two strong reciprocal relationships and dependencies. Though religion may be that which determines the goal, it has, nevertheless, learned from science, in the broadest sense, what means will contribute to the attainment of the goals it has set up. But science can only be created by those who are thoroughly imbued with the aspiration toward truth and understanding. This source of feeling, however, springs from the sphere of religion. To this there also belongs the faith in the possibility that the regulations valid for the world of existence are rational, that is, comprehensible to reason. I cannot conceive of a genuine scientist without that profound faith. The situation may be expressed by an image: science without religion is lame, religion without science is blind.[10]
David Hume's skepticism
In A Treatise of Human Nature (1739), David Hume discusses the problems in grounding normative statements in positive statements, that is in deriving ought from is. It is generally regarded that Hume considered such derivations untenable, and his 'is–ought' problem is considered a principal question of moral philosophy.[11]
Hume shared a political viewpoint with early Enlightenment philosophers such as Thomas Hobbes (1588–1679) and John Locke (1632–1704). Specifically, Hume, at least to some extent, argued that religious and national hostilities that divided European society were based on unfounded beliefs. In effect, Hume contended that such hostilities are not found in nature, but are a human creation, depending on a particular time and place, and thus unworthy of mortal conflict.
Prior to Hume, Aristotelian philosophy maintained that all actions and causes were to be interpreted teleologically. This rendered all facts about human action examinable under a normative framework defined by cardinal virtues and capital vices. "Fact" in this sense was not value-free, and the fact-value distinction was an alien concept. The decline of Aristotelianism in the 16th century set the framework in which those theories of knowledge could be revised.[12]
Naturalistic fallacy
The fact–value distinction is closely related to the naturalistic fallacy, a topic debated in ethical and moral philosophy. G. E. Moore believed it essential to all ethical thinking.[13] However, contemporary philosophers like Philippa Foot have called into question the validity of such assumptions. Others, such as Ruth Anna Putnam, argue that even the most "scientific" of disciplines are affected by the "values" of those who research and practice the vocation.[14][15] Nevertheless, the difference between the naturalistic fallacy and the fact–value distinction is derived from the manner in which modern social science has used the fact–value distinction, and not the strict naturalistic fallacy to articulate new fields of study and create academic disciplines.
Moralistic fallacy
The fact–value distinction is also closely related to the moralistic fallacy, an invalid inference of factual conclusions from purely evaluative premises. For example, an invalid inference "Because everybody ought to be equal, there are no innate genetic differences between people" is an instance of the moralistic fallacy. As for the naturalistic fallacy one attempts to move from an "is" to an "ought" statement, with the moralistic fallacy one attempts to move from an "ought" to an "is" statement.
Nietzsche's table of values
Friedrich Nietzsche (1844–1900) in Thus Spoke Zarathustra said that a table of values hangs above every great people. Nietzsche argues that what is common among different peoples is the act of esteeming, of creating values, even if the values are different from one people to the next. Nietzsche asserts that what made people great was not the content of their beliefs, but the act of valuing. Thus the values a community strives to articulate are not as important as the collective will to act on those values.[16] The willing is more essential than the intrinsic worth of the goal itself, according to Nietzsche.[17] "A thousand goals have there been so far," says Zarathustra, "for there are a thousand peoples. Only the yoke for the thousand necks is still lacking: the one goal is lacking. Humanity still has no goal." Hence, the title of the aphorism, "On The Thousand And One Goals." The idea that one value-system is no more worthy than the next, although it may not be directly ascribed to Nietzsche, has become a common premise in modern social science. Max Weber and Martin Heidegger absorbed it and made it their own. It shaped their philosophical endeavor, as well as their political understanding.
Criticisms
Virtually all modern philosophers affirm some sort of fact–value distinction, insofar as they distinguish between science and "valued" disciplines such as ethics, aesthetics, or the fine arts. However, philosophers such as Hilary Putnam argue that the distinction between fact and value is not as absolute as Hume envisioned.[18] Philosophical pragmatists, for instance, believe that true propositions are those that are useful or effective in predicting future (empirical) states of affairs.[19] Far from being value-free, the pragmatists' conception of truth or facts directly relates to an end (namely, empirical predictability) that human beings regard as normatively desirable. Other thinkers, such as N. Hanson among others, talk of theory-ladenness, and reject an absolutist fact–value distinction by contending that our senses are imbued with prior conceptualizations, making it impossible to have any observation that is totally value-free, which is how Hume and the later positivists conceived of facts.
Functionalist counterexamples
Several counterexamples have been offered by philosophers claiming to show that there are cases when an evaluative statement does indeed logically follow from a factual statement. A. N. Prior argues, from the statement "He is a sea captain," that it logically follows, "He ought to do what a sea captain ought to do."[20] Alasdair MacIntyre argues, from the statement "This watch is grossly inaccurate and irregular in time-keeping and too heavy to carry about comfortably," that the evaluative conclusion validly follows, "This is a bad watch."[21] John Searle argues, from the statement "Jones promised to pay Smith five dollars," that it logically follows that "Jones ought to pay Smith five dollars", such that the act of promising by definition places the promiser under obligation.[22]
Moral realism
Philippa Foot adopts a moral realist position, criticizing the idea that when evaluation is superposed on fact there has been a "committal in a new dimension".[23] She introduces, by analogy, the practical implications of using the word "injury". Not just anything counts as an injury. There must be some impairment. When we suppose a man wants the things the injury prevents him from obtaining, haven’t we fallen into the old naturalist fallacy?
It may seem that the only way to make a necessary connection between 'injury' and the things that are to be avoided, is to say that it is only used in an 'action-guiding sense' when applied to something the speaker intends to avoid. But we should look carefully at the crucial move in that argument, and query the suggestion that someone might happen not to want anything for which he would need the use of hands or eyes. Hands and eyes, like ears and legs, play a part in so many operations that a man could only be said not to need them if he had no wants at all.[24]
Foot argues that the virtues, like hands and eyes in the analogy, play so large a part in so many operations that it is implausible to suppose that a committal in a non-naturalist dimension is necessary to demonstrate their goodness.
Philosophers who have supposed that actual action was required if 'good' were to be used in a sincere evaluation have got into difficulties over weakness of will, and they should surely agree that enough has been done if we can show that any man has reason to aim at virtue and avoid vice. But is this impossibly difficult if we consider the kinds of things that count as virtue and vice? Consider, for instance, the cardinal virtues, prudence, temperance, courage and justice. Obviously any man needs prudence, but does he not also need to resist the temptation of pleasure when there is harm involved? And how could it be argued that he would never need to face what was fearful for the sake of some good? It is not obvious what someone would mean if he said that temperance or courage were not good qualities, and this not because of the 'praising' sense of these words, but because of the things that courage and temperance are.[25]
Of Weber
Philosopher Leo Strauss criticizes Weber for attempting to isolate reason completely from opinion. Strauss acknowledges the philosophical trouble of deriving "ought" from "is", but argues that what Weber has done in his framing of this puzzle is in fact deny altogether that the "ought" is within reach of human reason.[26]: 66 Strauss worries that if Weber is right, we are left with a world in which the knowable truth is a truth that cannot be evaluated according to ethical standards. This conflict between ethics and politics would mean that there can be no grounding for any valuation of the good, and without reference to values, facts lose their meaning.[26]: 72
See also
References
- Strauss, Leo (2008). Natural right and history. University of Chicago Press. ISBN 978-0226776941. OCLC 551845170.
Bibliography
- Hume, David. A Treatise of Human Nature. 1739–1740
- Hume, David. An Enquiry Concerning Human Understanding. 1748
- Nietzsche, Friedrich. Thus Spoke Zarathustra. 1883–1891. Translated by R.J. Hollingdale. New York: Penguin, 1969
- Putnam, Hilary. The Collapse of the Fact/Value Dichotomy and Other Essays. Harvard University Press, 2002. ISBN 0-674-01380-8
- Silvestri P. (ed.), L. Einaudi, On Abstract and Historical Hypotheses and on Value judgments in Economic Sciences, Critical edition with an Introduction and Afterword by Paolo Silvestri, Routledge, London – New York, 2017. DOI: https://doi.org/10.4324/9781315457932
https://en.wikipedia.org/wiki/Fact%E2%80%93value_distinction
Time is the continued sequence of existence and events that occurs in an apparently irreversible succession from the past, through the present, into the future.[1][2][3] It is a component quantity of various measurements used to sequence events, to compare the duration of events or the intervals between them, and to quantify rates of change of quantities in material reality or in the conscious experience.[4][5][6][7] Time is often referred to as a fourth dimension, along with three spatial dimensions.[8]
Time has long been an important subject of study in religion, philosophy, and science, but defining it in a manner applicable to all fields without circularity has consistently eluded scholars.[7][9] Nevertheless, diverse fields such as business, industry, sports, the sciences, and the performing arts all incorporate some notion of time into their respective measuring systems.[10][11][12]
Time in physics is operationally defined as "what a clock reads".[6][13][14]
The physical nature of time is addressed by general relativity with respect to events in spacetime. Examples of events are the collision of two particles, the explosion of a supernova, or the arrival of a rocket ship. Every event can be assigned four numbers representing its time and position (the event's coordinates). However, the numerical values are different for different observers. In general relativity, the question of what time it is now only has meaning relative to a particular observer. Distance and time are intimately related, and the time required for light to travel a specific distance is the same for all observers, as first publicly demonstrated by Michelson and Morley. General relativity does not address the nature of time for extremely small intervals where quantum mechanics holds. As of 2023, there is no generally accepted theory of quantum general relativity.[15]
Time is one of the seven fundamental physical quantities in both the International System of Units (SI) and International System of Quantities. The SI base unit of time is the second, which is defined by measuring the electronic transition frequency of caesium atoms. Time is used to define other quantities, such as velocity, so defining time in terms of such quantities would result in circularity of definition.[16] An operational definition of time, wherein one says that observing a certain number of repetitions of one or another standard cyclical event (such as the passage of a free-swinging pendulum) constitutes one standard unit such as the second, is highly useful in the conduct of both advanced experiments and everyday affairs of life. To describe observations of an event, a location (position in space) and time are typically noted.
The operational definition of time does not address what the fundamental nature of time is. It does not address why events can happen forward and backward in space, whereas events only happen in the forward progress of time. Investigations into the relationship between space and time led physicists to define the spacetime continuum. General relativity is the primary framework for understanding how spacetime works.[17] Through advances in both theoretical and experimental investigations of spacetime, it has been shown that time can be distorted and dilated, particularly at the edges of black holes.
Temporal measurement has occupied scientists and technologists and was a prime motivation in navigation and astronomy. Periodic events and periodic motion have long served as standards for units of time. Examples include the apparent motion of the sun across the sky, the phases of the moon, and the swing of a pendulum. Time is also of significant social importance, having economic value ("time is money") as well as personal value, due to an awareness of the limited time in each day and in human life spans.
There are many systems for determining what time it is, including the Global Positioning System, other satellite systems, Coordinated Universal Time and mean solar time. In general, the numbers obtained from different time systems differ from one another.
Measurement
Generally speaking, methods of temporal measurement, or chronometry, take two distinct forms: the calendar, a mathematical tool for organising intervals of time,[18] and the clock, a physical mechanism that counts the passage of time. In day-to-day life, the clock is consulted for periods less than a day, whereas the calendar is consulted for periods longer than a day. Increasingly, personal electronic devices display both calendars and clocks simultaneously. The number (as on a clock dial or calendar) that marks the occurrence of a specified event as to hour or date is obtained by counting from a fiducial epoch – a central reference point.
History of the calendar
Artifacts from the Paleolithic suggest that the moon was used to reckon time as early as 6,000 years ago.[19] Lunar calendars were among the first to appear, with years of either 12 or 13 lunar months (either 354 or 384 days). Without intercalation to add days or months to some years, seasons quickly drift in a calendar based solely on twelve lunar months. Lunisolar calendars have a thirteenth month added to some years to make up for the difference between a full year (now known to be about 365.24 days) and a year of just twelve lunar months. The numbers twelve and thirteen came to feature prominently in many cultures, at least partly due to this relationship of months to years. Other early forms of calendars originated in Mesoamerica, particularly in ancient Mayan civilization. These calendars were religiously and astronomically based, with 18 months in a year and 20 days in a month, plus five epagomenal days at the end of the year.[20]
The reforms of Julius Caesar in 45 BC put the Roman world on a solar calendar. This Julian calendar was faulty in that its intercalation still allowed the astronomical solstices and equinoxes to advance against it by about 11 minutes per year. Pope Gregory XIII introduced a correction in 1582; the Gregorian calendar was only slowly adopted by different nations over a period of centuries, but it is now by far the most commonly used calendar around the world.
During the French Revolution, a new clock and calendar were invented in an attempt to de-Christianize time and create a more rational system in order to replace the Gregorian calendar. The French Republican Calendar's days consisted of ten hours of a hundred minutes of a hundred seconds, which marked a deviation from the base 12 (duodecimal) system used in many other devices by many cultures. The system was abolished in 1806.[21]
History of other devices
A large variety of devices have been invented to measure time. The study of these devices is called horology.[22]
An Egyptian device that dates to c. 1500 BC, similar in shape to a bent T-square, measured the passage of time from the shadow cast by its crossbar on a nonlinear rule. The T was oriented eastward in the mornings. At noon, the device was turned around so that it could cast its shadow in the evening direction.[23]
A sundial uses a gnomon to cast a shadow on a set of markings calibrated to the hour. The position of the shadow marks the hour in local time. The idea to separate the day into smaller parts is credited to Egyptians because of their sundials, which operated on a duodecimal system. The importance of the number 12 is due to the number of lunar cycles in a year and the number of stars used to count the passage of night.[24]
The most precise timekeeping device of the ancient world was the water clock, or clepsydra, one of which was found in the tomb of Egyptian pharaoh Amenhotep I. They could be used to measure the hours even at night but required manual upkeep to replenish the flow of water. The ancient Greeks and the people from Chaldea (southeastern Mesopotamia) regularly maintained timekeeping records as an essential part of their astronomical observations. Arab inventors and engineers, in particular, made improvements on the use of water clocks up to the Middle Ages.[25] In the 11th century, Chinese inventors and engineers invented the first mechanical clocks driven by an escapement mechanism.
The hourglass uses the flow of sand to measure the flow of time. They were used in navigation. Ferdinand Magellan used 18 glasses on each ship for his circumnavigation of the globe (1522).[26]
Incense sticks and candles were, and are, commonly used to measure time in temples and churches across the globe. Waterclocks, and later, mechanical clocks, were used to mark the events of the abbeys and monasteries of the Middle Ages. Richard of Wallingford (1292–1336), abbot of St. Alban's abbey, famously built a mechanical clock as an astronomical orrery about 1330.[27][28]
Great advances in accurate time-keeping were made by Galileo Galilei and especially Christiaan Huygens with the invention of pendulum-driven clocks along with the invention of the minute hand by Jost Burgi.[29]
The English word clock probably comes from the Middle Dutch word klocke which, in turn, derives from the medieval Latin word clocca, which ultimately derives from Celtic and is cognate with French, Latin, and German words that mean bell. The passage of the hours at sea was marked by bells and denoted the time (see ship's bell). The hours were marked by bells in abbeys as well as at sea.
Clocks can range from watches to more exotic varieties such as the Clock of the Long Now. They can be driven by a variety of means, including gravity, springs, and various forms of electrical power, and regulated by a variety of means such as a pendulum.
Alarm clocks first appeared in ancient Greece around 250 BC with a water clock that would set off a whistle. This idea was later mechanized by Levi Hutchins and Seth E. Thomas.[29]
A chronometer is a portable timekeeper that meets certain precision standards. Initially, the term was used to refer to the marine chronometer, a timepiece used to determine longitude by means of celestial navigation, a precision firstly achieved by John Harrison. More recently, the term has also been applied to the chronometer watch, a watch that meets precision standards set by the Swiss agency COSC.
The most accurate timekeeping devices are atomic clocks, which are accurate to seconds in many millions of years,[31] and are used to calibrate other clocks and timekeeping instruments.
Atomic clocks use the frequency of electronic transitions in certain atoms to measure the second. One of the atoms used is caesium, most modern atomic clocks probe caesium with microwaves to determine the frequency of these electron vibrations.[32] Since 1967, the International System of Measurements bases its unit of time, the second, on the properties of caesium atoms. SI defines the second as 9,192,631,770 cycles of the radiation that corresponds to the transition between two electron spin energy levels of the ground state of the 133Cs atom.
Today, the Global Positioning System in coordination with the Network Time Protocol can be used to synchronize timekeeping systems across the globe.
In medieval philosophical writings, the atom was a unit of time referred to as the smallest possible division of time. The earliest known occurrence in English is in Byrhtferth's Enchiridion (a science text) of 1010–1012,[33] where it was defined as 1/564 of a momentum (11⁄2 minutes),[34] and thus equal to 15/94 of a second. It was used in the computus, the process of calculating the date of Easter.
As of May 2010, the smallest time interval uncertainty in direct measurements is on the order of 12 attoseconds (1.2 × 10−17 seconds), about 3.7 × 1026 Planck times.[35]
Units
The second (s) is the SI base unit. A minute (min) is 60 seconds in length, and an hour is 60 minutes or 3600 seconds in length. A day is usually 24 hours or 86,400 seconds in length; however, the duration of a calendar day can vary due to Daylight saving time and Leap seconds.
Definitions and standards
A time standard is a specification for measuring time: assigning a number or calendar date to an instant (point in time), quantifying the duration of a time interval, and establishing a chronology (ordering of events). In modern times, several time specifications have been officially recognized as standards, where formerly they were matters of custom and practice. The invention in 1955 of the caesium atomic clock has led to the replacement of older and purely astronomical time standards such as sidereal time and ephemeris time, for most practical purposes, by newer time standards based wholly or partly on atomic time using the SI second.
International Atomic Time (TAI) is the primary international time standard from which other time standards are calculated. Universal Time (UT1) is mean solar time at 0° longitude, computed from astronomical observations. It varies from TAI because of the irregularities in Earth's rotation. Coordinated Universal Time (UTC) is an atomic time scale designed to approximate Universal Time. UTC differs from TAI by an integral number of seconds. UTC is kept within 0.9 second of UT1 by the introduction of one-second steps to UTC, the "leap second". The Global Positioning System broadcasts a very precise time signal based on UTC time.
The surface of the Earth is split up into a number of time zones. Standard time or civil time in a time zone deviates a fixed, round amount, usually a whole number of hours, from some form of Universal Time, usually UTC. Most time zones are exactly one hour apart, and by convention compute their local time as an offset from UTC. For example, time zones at sea are based on UTC. In many locations (but not at sea) these offsets vary twice yearly due to daylight saving time transitions.
Some other time standards are used mainly for scientific work. Terrestrial Time is a theoretical ideal scale realized by TAI. Geocentric Coordinate Time and Barycentric Coordinate Time are scales defined as coordinate times in the context of the general theory of relativity. Barycentric Dynamical Time is an older relativistic scale that is still in use.
https://en.wikipedia.org/wiki/Time
In philosophy, transcendence is the basic ground concept from the word's literal meaning (from Latin), of climbing or going beyond, albeit with varying connotations in its different historical and cultural stages. It includes philosophies, systems, and approaches that describe the fundamental structures of being, not as an ontology (theory of being), but as the framework of emergence and validation of knowledge of being. These definitions are generally grounded in reason and empirical observation, and seek to provide a framework for understanding the world that is not reliant on religious beliefs or supernatural forces. [1][2][3] "Transcendental" is a word derived from the scholastic, designating the extra-categorical attributes of beings.[4][5]
https://en.wikipedia.org/wiki/Transcendence_(philosophy)
In epistemology, transparency is a property of epistemic states defined as follows: An epistemic state E is "weakly transparent" to a subject S if and only if when S is in state E, S can know that S is in state E; an epistemic state E is "strongly transparent" to a subject S if and only if when S is in state E, S can know that S is in state E, AND when S is not in state E, S can know S is not in state E.
Pain is usually considered to be strongly transparent: when someone is in pain, they know immediately that they are in pain, and if they are not in pain, they will know they are not. Transparency is important in the study of self-knowledge and meta-knowledge.
https://en.wikipedia.org/wiki/Transparency_(philosophy)
Trope denotes figurative and metaphorical language and one which has been used in various technical senses. The term trope derives from the Greek τρόπος (tropos), "a turn, a change",[1] related to the root of the verb τρέπειν (trepein), "to turn, to direct, to alter, to change";[2] this means that the term is used metaphorically to denote, among other things, metaphorical language.
The term is also used in technical senses, which do not always correspond to its linguistic origin. Its meaning has to be judged from the context, some of which are given below.
https://en.wikipedia.org/wiki/Trope_(philosophy)
A consensus theory of truth is the process of taking statements to be true simply because people generally agree upon them.
Varieties of consensus
Consensus gentium
An ancient criterion of truth, the consensus gentium (Latin for agreement of the people), states "that which is universal among men carries the weight of truth" (Ferm, 64). A number of consensus theories of truth are based on variations of this principle. In some criteria the notion of universal consent is taken strictly, while others qualify the terms of consensus in various ways. There are versions of consensus theory in which the specific population weighing in on a given question, the proportion of the population required for consent, and the period of time needed to declare consensus vary from the classical norm.
Consensus as a regulative ideal
A descriptive theory is one that tells how things are, while a normative theory tells how things ought to be. Expressed in practical terms, a normative theory, more properly called a policy, tells agents how they ought to act. A policy can be an absolute imperative, telling agents how they ought to act in any case, or it can be a contingent directive, telling agents how they ought to act if they want to achieve a particular goal. A policy is frequently stated in the form of a piece of advice called a heuristic, a maxim, a norm, a rule, a slogan, and so on. Other names for a policy are a recommendation and a regulative principle.
A regulative ideal can be expressed in the form of a description, but what it describes is an ideal state of affairs, a condition of being that constitutes its aim, end, goal, intention, or objective. It is not the usual case for the actual case to be the ideal case, or else there would hardly be much call for a policy aimed at achieving an ideal.
Corresponding to the distinction between actual conditions and ideal conditions there is a distinction between actual consensus and ideal consensus. A theory of truth founded on a notion of actual consensus is a very different thing from a theory of truth founded on a notion of ideal consensus. Moreover, an ideal consensus may be ideal in several different ways. The state of consensus may be ideal in its own nature, conceived in the matrix of actual experience by way of intellectual operations like abstraction, extrapolation, and limit formation. Or the conditions under which the consensus is conceived to be possible may be formulated as idealizations of actual conditions. A very common type of ideal consensus theory refers to a community that is an idealization of actual communities in one or more respects.
Critiques
It is very difficult to find any philosopher of note who asserts a bare, naive, or pure consensus theory of truth, in other words, a treatment of truth that is based on actual consensus in an actual community without further qualification. One obvious critique is that not everyone agrees to consensus theory, implying that it may not be true by its own criteria. Another problem is defining how we know that consensus is achieved without falling prey to an infinite regress. Even if everyone agrees to a particular proposition, we may not know that it is true until everyone agrees that everyone agrees to it. Bare consensus theories are frequent topics of discussion, however, evidently because they serve the function of reference points for the discussion of alternative theories.
If consensus equals truth, then truth can be made by forcing or organizing a consensus, rather than being discovered through experiment or observation, or existing separately from consensus. The principles of mathematics also do not hold under consensus truth because mathematical propositions build on each other. If the consensus declared 2+2=5 it would render the practice of mathematics where 2+2=4 impossible.
Imre Lakatos characterizes it as a "watered down" form of provable truth propounded by some sociologists of knowledge, particularly Thomas Kuhn and Michael Polanyi.[1]
Philosopher Nigel Warburton
argues that the truth by consensus process is not a reliable way of
discovering truth, that there is general agreement upon something does
not make it actually true.
There are two main reasons for this:[2]
- One reason Warburton discusses is that people are prone to wishful thinking. People can believe an assertion and espouse it as truth in the face of overwhelming evidence and facts to the contrary, simply because they wish that things were so.
- The other one is that people are gullible, and easily misled.
See also
- Argumentum ad populum
- Coherentism
- Common knowledge
- Confirmation holism
- Consensus reality
- Conventional wisdom
- Jury trial
- Philosophy of history § History as propaganda: Is history always written by the victors?
- Philosophy of history
- Truth by consensus
- Truthiness
- Wikiality
Related topics
|
References
- Nigel Warburton (2000). "truth by consensus". Thinking from A to Z. Routledge. pp. 134–135. ISBN 0-415-22281-8.
Sources
- Ferm, Vergilius (1962), "Consensus Gentium", p. 64 in Runes (1962).
- Haack, Susan (1993), Evidence and Inquiry: Towards Reconstruction in Epistemology, Blackwell Publishers, Oxford, UK.
- Habermas, Jürgen (1976), "What Is Universal Pragmatics?", 1st published, "Was heißt Universalpragmatik?", Sprachpragmatik und Philosophie, Karl-Otto Apel (ed.), Suhrkamp Verlag, Frankfurt am Main. Reprinted, pp. 1–68 in Jürgen Habermas, Communication and the Evolution of Society, Thomas McCarthy (trans.), Beacon Press, Boston, Massachusetts, 1979.
- Habermas, Jürgen (1979), Communication and the Evolution of Society, Thomas McCarthy (trans.), Beacon Press, Boston, Massachusetts.
- Habermas, Jürgen (1990), Moral Consciousness and Communicative Action, Christian Lenhardt and Shierry Weber Nicholsen (trans.), Thomas McCarthy (intro.), MIT Press, Cambridge, Massachusetts.
- Habermas, Jürgen (2003), Truth and Justification, Barbara Fultner (trans.), MIT Press, Cambridge, Massachusetts.
- James, William (1907), Pragmatism, A New Name for Some Old Ways of Thinking, Popular Lectures on Philosophy, Longmans, Green, and Company, New York, New York.
- James, William (1909), The Meaning of Truth, A Sequel to 'Pragmatism', Longmans, Green, and Company, New York, New York.
- Kant, Immanuel (1800), Introduction to Logic. Reprinted, Thomas Kingsmill Abbott (trans.), Dennis Sweet (intro.), Barnes and Noble, New York, New York, 2005.
- Kirkham, Richard L. (1992), Theories of Truth: A Critical Introduction, MIT Press, Cambridge, Massachusetts.
- Rescher, Nicholas (1995), Pluralism: Against the Demand for Consensus, Oxford University Press, Oxford, UK.
- Runes, Dagobert D. (ed., 1962), Dictionary of Philosophy, Littlefield, Adams, and Company, Totowa, New Jersey.
https://en.wikipedia.org/wiki/Consensus_theory_of_truth
The principle of truth-value links is a concept in metaphysics discussed in debates between philosophical realism and anti-realism. Philosophers who appeal to truth-value links in order to explain how individuals can come to understand parts of the world that are apparently cognitively inaccessible (the past, the feelings of others, etc.) are called truth-value link realists.
Truth-value link realism
Proponents of truth-value link realism argue that our understanding of past-tense statements allows us to grasp the truth-conditions of the statements, even if they are evidence-transcendent. They explain this by noting that it is unproblematic for us to conceptualize a present-tense true statement being true in the future. In other words, if "It is raining today" is true today, then "It was raining yesterday" will be true tomorrow. Truth-value link realists argue that this same construction can be applied to past-tense statements. For example, "It was raining yesterday" is true today if and only if "It is raining today" was true yesterday.
The truth-value link allows us to understand the following. First, suppose that we can understand, in an unproblematic way, truth about a present-tense statement. Assume that it is true, now, when one claims "On 22 May 2006, Student X is writing a paper for her philosophy seminar," and call it statement A. Suppose that, a year later, someone claims, "On 22 May 2006, Student X was writing a paper for her philosophy seminar," and call it statement B. According to Michael Dummett’s explication of the truth-value link, "since the statement A is now true, the statement B, made in one year’s time, is likewise true."[1] It is in understanding the truth-value link that one is able to understand what it is for a statement in the past-tense to be true. The truth-value persists in the tense shift – hence the "link."
Criticisms
Some philosophers, including Michael Dummett and John McDowell, have criticized truth-value link realism. They argue that it is unintelligible to suppose that training in a language can give someone more than what is involved in the training, i.e. access to inaccessible realms like the past and the minds of others. More important, they suggest that the realist appeal to the principle of truth-value links does not actually explain how the inaccessible can be cognized. When the truth-value link realist claims that, if "It is raining today" was true yesterday, then "It was raining yesterday" is true today, he or she is still appealing to an inaccessible realm of the past. In brief, one is attempting to access that which one has already conceded as being inaccessible.
See also
References
- Dummett, Michael, "The Reality of the Past," in Truth and Other Enigmas (Harvard University Press, 1978), 363.
https://en.wikipedia.org/wiki/Truth-value_link
An unobservable (also called impalpable) is an entity whose existence, nature, properties, qualities or relations are not directly observable by humans. In philosophy of science, typical examples of "unobservables" are the force of gravity, causation and beliefs or desires.[1]: 7 [2] The distinction between observable and unobservable plays a central role in Immanuel Kant's distinction between noumena and phenomena as well as in John Locke's distinction between primary and secondary qualities. The theory that unobservables posited by scientific theories exist is referred to as scientific realism. It contrasts with instrumentalism, which asserts that we should withhold ontological commitments to unobservables even though it is useful for scientific theories to refer to them. There is considerable disagreement about which objects should be classified as unobservable, for example, whether bacteria studied using microscopes or positrons studied using cloud chambers count as unobservable. Different notions of unobservability have been formulated corresponding to different types of obstacles to their observation.
Kant on noumena
The distinction between "observable" and "unobservable" is similar to Immanuel Kant's distinction between noumena and phenomena. Noumena are the things-in-themselves, i.e., raw things in their necessarily unknowable state,[3] before they pass through the formalizing apparatus of the senses and the mind in order to become perceived objects, which he refers to as "phenomena". According to Kant, humans can never know noumena; all that humans know is the phenomena.
Locke on primary and secondary qualities
Kant's distinction is similar to John Locke's distinction between primary and secondary qualities. Secondary qualities are what humans perceive such as redness, chirping, heat, mustiness or sweetness. Primary qualities would be the actual qualities of the things themselves which give rise to the secondary qualities which humans perceive.
Philosophy of science
The ontological nature and epistemological issues concerning unobservables are central topics in philosophy of science. The theory that unobservables posited by scientific theories exist is referred to as scientific realism. It contrasts with instrumentalism, which asserts that we should withhold ontological commitments to unobservables even though it is useful for scientific theories to refer to them.
The notion of observability plays a central role in constructive empiricism. According to van Fraassen, the goal of scientific theories is not truth about all entities but only truth about all observable entities.[4] If a theory is true in this restricted sense, it is called an empirically adequate theory. Van Fraassen characterizes observability counterfactually: "X is observable if there are circumstances which are such that, if X is present to us under those circumstances, then we observe it".[5]
A problem with this and similar characterizations is to determine the exact extension of what is unobservable. There is little controversy that regular everyday-objects that we can perceive without any aids are observable. Such objects include e.g. trees, chairs or dogs. But controversy starts with cases where unaided perception fails. This includes cases like using telescopes to study distant galaxies,[6] using microscopes to study bacteria or using cloud chambers to study positrons.[5]
Some philosophers have been motivated by these and similar examples to question the value of the distinction between observable and unobservable in general.[7]
Kinds of unobservables
W. V. Metcalf distinguishes three kinds of unobservables.[8] One is the logically unobservable, which involves a contradiction. An example would be a length which is both longer and shorter than a given length. The second is the practically unobservable, that which we can conceive of as observable by the known sense-faculties of man but we are prevented from observing by practical difficulties. The third kind is the physically unobservable, that which can never be observed by any existing sense-faculties of man.
See also
- Empiricism
- Logical positivism
- Phenomenology
- Rationalism
- Hidden variable theory
- Object of the mind
- If a tree falls in a forest
- Unobservable chaos
- Proxy (statistics), for an unobservable variable
References
https://en.wikipedia.org/wiki/Unobservable
In the philosophy of science, underdetermination or the underdetermination of theory by data (sometimes abbreviated UTD) is the idea that evidence available to us at a given time may be insufficient to determine what beliefs we should hold in response to it.[1] The underdetermination thesis says that all evidence necessarily underdetermines any scientific theory.[2]
Underdetermination exists when available evidence is insufficient to identify which belief one should hold about that evidence. For example, if all that was known was that exactly $10 was spent on apples and oranges, and that apples cost $1 and oranges $2, then one would know enough to eliminate some possibilities (e.g., 6 oranges could not have been purchased), but one would not have enough evidence to know which specific combination of apples and oranges was purchased. In this example, one would say that belief in what combination was purchased is underdetermined by the available evidence.
In contrast, overdetermination in philosophy of science means that more evidence is available than is necessary to justify a conclusion.
Origin
Ancient Greek skeptics argued for equipollence, the view that reasons for and against claims are equally balanced. This captures at least one sense of saying that the claims themselves are underdetermined.
Underdetermination, again under different labels, arises in the modern period in the work of René Descartes. Among other skeptical arguments, Descartes presents two arguments involving underdetermination. His dream argument points out that experiences perceived while dreaming (for example, falling) do not necessarily contain sufficient information to deduce the true situation (being in bed). He concluded that since one cannot always distinguish dreams from reality, one cannot rule out the possibility that one is dreaming rather than having veridical experiences; thus the conclusion that one is having a veridical experience is underdetermined. His demon argument posits that all of one's experiences and thoughts might be manipulated by a very powerful and deceptive "evil demon". Once again, so long as the perceived reality appears internally consistent to the limits of one's limited ability to tell, the situation is indistinguishable from reality and one cannot logically determine that such a demon does not exist.
Underdetermination and evidence
To show that a conclusion is underdetermined, one must show that there is a rival conclusion that is equally well supported by the standards of evidence. A trivial example of underdetermination is the addition of the statement "whenever we look for evidence" (or more generally, any statement which cannot be falsified). For example, the conclusion "objects near earth fall toward it when dropped" might be opposed by "objects near earth fall toward it when dropped but only when one checks to see that they do." Since one may append this to any conclusion, all conclusions are at least trivially underdetermined. If one considers such statements to be illegitimate, e.g. by applying Occam's Razor, then such "tricks" are not considered demonstrations of underdetermination.
This concept also applies to scientific theories: for example, it is similarly trivial to find situations that a theory does not address. For example, classical mechanics did not distinguish between non-accelerating reference frames. As a result, any conclusion about such a reference frame was underdetermined; it was equally consistent with the theory to say that the solar system is at rest, as it is to say that it moves at any constant velocity in any particular direction. Newton himself stated that these possibilities were indistinguishable. More generally, evidence may not always be sufficient to distinguish between competing theories (or to determine a different theory that will unify both), as is the case with general relativity and quantum mechanics.
Another example is provided by Johann Wolfgang von Goethe's Theory of Colours: "Newton believed that with the help of his prism experiments, he could prove that sunlight was composed of variously coloured rays of light. Goethe showed that this step from observation to theory is more problematic than Newton wanted to admit. By insisting that the step to theory is not forced upon us by the phenomena, Goethe revealed our own free, creative contribution to theory construction. And Goethe's insight is surprisingly significant, because he correctly claimed that all of the results of Newton's prism experiments fit a theoretical alternative equally well. If this is correct, then by suggesting an alternative to a well-established physical theory, Goethe developed the problem of underdetermination a century before Duhem and Quine's famous argument." (Mueller, 2016)[3] Hermann von Helmholtz says of this, "And I for one do not know how anyone, regardless of what his views about colours are, can deny that the theory in itself is fully consequent, that its assumptions, once granted, explain the facts treated completely and indeed simply".[4]
Experimental violations of Bell inequality show, that there are some limitations to underdetermination – every theory exhibiting local realism and statistical independence was disproved by this tests. Analogus limitations follow from Kochen–Specker experiments.[5][6][7] These tests employ only correlations between results of measurements and therefore are able to bypass the issue of theory-ladenness of observation.[7][8]
Arguments involving underdetermination
Arguments involving underdetermination attempt to show that there is no reason to believe some conclusion because it is underdetermined by the evidence. Then, if the evidence available at a particular time can be equally well explained by at least one other hypothesis, there is no reason to believe it rather than the equally supported rival, which can be considered observationally equivalent (although many other hypotheses may still be eliminated).
Because arguments involving underdetermination involve both a claim about what the evidence is and that such evidence underdetermines a conclusion, it is often useful to separate these two claims within the underdetermination argument as follows:
- All the available evidence of a certain type underdetermines which of several rival conclusions is correct.
- Only evidence of that type is relevant to believing one of these conclusions.
- Therefore, there is no evidence for believing one among the rival conclusions.
The first premise makes the claim that a theory is underdetermined. The second says that rational decision (i.e. using available evidence) depends upon insufficient evidence.
Epistemological problem of the indeterminacy of data to theory
Any phenomenon can be explained by a multiplicity of hypotheses. How, then, can data ever be sufficient to prove a theory? This is the "epistemological problem of the indeterminacy of data to theory".
The poverty of the stimulus argument and W.V.O. Quine's 1960 'Gavagai' example are perhaps the most commented variants of the epistemological problem of the indeterminacy of data to theory.
General skeptical arguments
Some skeptical arguments appeal to the fact that no possible evidence could be incompatible with 'skeptical hypotheses' like the maintenance of a complex illusion by Descartes' evil demon or (in a modern version) the machines who run the Matrix. A skeptic may argue that this undermines any claims to knowledge, or even (by internalist definitions), justification.
Philosophers have found this argument very powerful. Hume felt it was unanswerable, but observed that it was in practice impossible to accept its conclusions. Influenced by this, Kant held that while the nature of the 'noumenal' world was indeed unknowable, we could aspire to knowledge of the 'phenomenal' world. A similar response has been advocated by modern anti-realists.
Underdetermined ideas are not implied to be incorrect (taking into account present evidence); rather, we cannot know if they are correct.
Philosophy of science
In the philosophy of science, underdetermination is often presented as a problem for scientific realism, which holds that we have reason to believe in entities that are not directly observable talked about by scientific theories. One such argument proceeds as follows (to be compared to the previous one):
- All the available observational evidence for such entities underdetermines the claims of a scientific theory about such entities.
- Only the observational evidence is relevant to believing a scientific theory.
- Therefore, there is no evidence for believing what scientific theories say about such entities.
Particular responses to this argument attack both the first and the second premise (1 and 2). It is argued against the first premise that the underdetermination must be strong and/or inductive. It is argued against the second premise that there is evidence for a theory's truth besides observations; for example, it is argued that simplicity, explanatory power or some other feature of a theory is evidence for it over its rivals.
A more general response from the scientific realist is to argue that underdetermination is no special problem for science, because, as indicated earlier in this article, all knowledge that is directly or indirectly supported by evidence suffers from it—for example, conjectures concerning unobserved observables. It is therefore too powerful an argument to have any significance in the philosophy of science, since it does not cast doubt uniquely on conjectured unobservables.
See also
- Poverty of the stimulus
- Reference class problem
- Scientific method
- Instrumentalism
- Confirmation holism
- Equifinality
- Metaphysics
- Occam's razor
- Observational equivalence
- Overdetermination
Notes and references
- Philosophical consequences of quantum theory : reflections on Bell's theorem. James T. Cushing, Ernan McMullin. Notre Dame, Ind.: University of Notre Dame Press. 1989. p. 65. ISBN 0-268-01578-3. OCLC 19352903.
External links
- Underdetermination and the Claims of Science by P. D. Magnus
- Stanford, Kyle (Winter 2000), Edward N. Zalta (ed.), Underdetermination of Scientific Theory, The Stanford Encyclopedia of Philosophy
https://en.wikipedia.org/wiki/Underdetermination
ntentionality is the power of minds to be about something: to represent or to stand for things, properties and states of affairs.[1] Intentionality is primarily ascribed to mental states, like perceptions, beliefs or desires, which is why it has been regarded as the characteristic mark of the mental by many philosophers. A central issue for theories of intentionality has been the problem of intentional inexistence: to determine the ontological status of the entities which are the objects of intentional states.
An early theory of intentionality is associated with Anselm of Canterbury's ontological argument for the existence of God, and with his tenets distinguishing between objects that exist in the understanding and objects that exist in reality.[2] The idea fell out of discussion with the end of the medieval scholastic period, but in recent times was resurrected by empirical psychologist Franz Brentano and later adopted by contemporary phenomenological philosopher Edmund Husserl. Today, intentionality is a live concern among philosophers of mind and language.[3] A common dispute is between naturalism, the view that intentional properties are reducible to natural properties as studied by the natural sciences, and the phenomenal intentionality theory, the view that intentionality is grounded in consciousness.
https://en.wikipedia.org/wiki/Intentionality
Intellectual responsibility (also known as epistemic responsibility) is a philosophical concept related to that of epistemic justification.[1] According to Frederick F. Schmitt, "the conception of justified belief as epistemically responsible belief has been endorsed by a number of philosophers, including Roderick Chisholm (1977), Hilary Kornblith (1983), and Lorraine Code (1983)."[2]
https://en.wikipedia.org/wiki/Intellectual_responsibility
Insight is the understanding of a specific cause and effect within a particular context. The term insight can have several related meanings:
- a piece of information
- the act or result of understanding the inner nature of things or of seeing intuitively (called noesis in Greek)
- an introspection
- the power of acute observation and deduction, discernment, and perception, called intellection or noesis
- An understanding of cause and effect based on the identification of relationships and behaviors within a model, context, or scenario (see artificial intelligence)
An insight that manifests itself suddenly, such as understanding how to solve a difficult problem, is sometimes called by the German word Aha-Erlebnis. The term was coined by the German psychologist and theoretical linguist Karl Bühler. It is also known as an epiphany, eureka moment or (for cross word solvers) the penny dropping moment (PDM).[1] Sudden sickening realisations often identify a problem rather than solving it, so Uh-oh rather than Aha moments are further seen in negative insight.[2] A further example of negative insight is chagrin which is annoyance at the obviousness of a solution missed up until the point of insight,[3] an example of this being the Homer Simpson's catchphrase exclamation, D'oh!.
Psychology
In psychology, insight occurs when a solution to a problem presents itself quickly and without warning.[5] It is the sudden discovery of the correct solution following incorrect attempts based on trial and error.[6][7] Solutions via insight have been proven to be more accurate than non-insight solutions.[6]
Insight was first studied by Gestalt psychology, in the early part of the 20th century, during the search for an alternative to associationism and the associationistic view of learning.[8] Some proposed potential mechanisms for insight include: suddenly seeing the problem in a new way, connecting the problem to another relevant problem/solution pair, releasing past experiences that are blocking the solution, or seeing problem in a larger, coherent context.[8]
Classic methods
Generally, methodological approaches to the study of insight in the laboratory involve presenting participants with problems and puzzles that cannot be solved in a conventional or logical manner.[8] Problems of insight commonly fall into three types.[8]
Breaking functional fixedness
The first type of problem forces participants to use objects in a way they are not accustomed to (thus, breaking their functional fixedness), like the "Duncker candle problem".[8] In the "Duncker candle problem", individuals are given matches and a box of tacks and asked to find a way to attach a candle to the wall to light the room.[4] The solution requires the participants to empty the box of tacks, set the candle inside the box, tack the box to the wall, and light the candle with the matches.
Spatial ability
The second type of insight problem requires spatial ability to solve, like the "Nine-dot problem".[8] The famous "Nine-dot problem" requires participants to draw four lines, through nine dots, without picking their pencil up.[9][page needed]
Using verbal ability
The third and final type of problem requires verbal ability to solve, like the Remote Associates Test (RAT).[8] In the RAT, individuals must think of a word that connects three, seemingly unrelated, words.[10] RAT are often used in experiments, because they can be solved both with and without insight.[11]
Specific results
Versus non-insight problems
Two clusters of problems, those solvable by insight and those not requiring insight to solve, have been observed.[12] An individual’s cognitive flexibility, fluency, and vocabulary ability are predictive of performance on insight problems, but not on non-insight problems.[12] In contrast, fluid intelligence is mildly predictive of performance on non-insight problems, but not on insight problems.[12] More recent research suggests that rather than insight versus search, that the subjective feeling of insight varies, with some solutions experienced with a stronger feeling of Aha than others.[13][14]
Emotion
People in a better mood are more likely to solve problems using insight.[15] Research demonstrated that self-reported positive affect of participants uniquely increased insight before and during the solving of a problem,[16] as indicated by differing brain activity patterns.[15] People experiencing anxiety showed the opposite effect, and solved fewer problems by insight.[15] Emotion can also be considered in terms of the insight experience and whether this is a positive Aha or negative Uh-oh moment.[2] Research demonstrate that in order to have insights it is important to have a good degree of access to one's own emotions and sensations, that can cause insights. To the degree that individuals have limited introspective access to these underlying causes, they have only limited control over these processes as well. [17]
Incubation
Using a geometric and spatial insight problem, it was found that providing participants with breaks improved their performance as compared to participants who did not receive a break.[18] However, the length of incubation between problems did not matter. Thus, participants' performance on insight problems improved just as much with a short break (4 minutes) as it did with a long break (12 minutes).[18]
Sleep
Research has shown sleep to help produce insight.[19] Individuals were initially trained on insight problems. Following training, one group was tested on the insight problems after sleeping for eight hours at night, one group was tested after staying awake all night, and one group was tested after staying awake all day. Those that slept performed twice as well on the insight problems than those who stayed awake.[19]
In the brain
Differences in brain activation in the left and right hemisphere seem to be indicative of insight versus non-insight solutions.[20] Using RAT’s that were either presented to the left or right visual field, it was shown that participants having solved the problem with insight were more likely to have been shown the RAT on the left visual field, indicating right hemisphere processing. This provides evidence that the right hemisphere plays a unique role in insight.[20]
fMRI and EEG scans of participants completing RAT's demonstrated unique brain activity corresponding to problems solved by insight.[11] For one, there is high EEG activity in the alpha- and gamma-band about 300 milliseconds before participants indicated a solution to insight problems, but not to non-insight problems.[11] Additionally, problems solved by insight corresponded to increased activity in the temporal lobes and mid-frontal cortex, while more activity in the posterior cortex corresponded to non-insight problems.[11] The data suggests there is something different occurring in the brain when solving insight versus non-insight problems that happens right before the solving of the problem. This conclusion has been supported also by eye-tracking data which shows an increased eye blink duration and frequency when people solve problems via Insight. This latter result, paired with an eye pattern oriented to look away from sources of visual inputs (such as looking at blank wall, or out the window at the sky) proves different attention involvement in Insight problem solving vs problem solving via analysis.[21]
Group insight
It was found that groups typically perform better on insight problems (in the form of rebus puzzles with either helpful or unhelpful clues) than individuals.[22]
Additionally, while incubation improves insight performance for individuals, it improves insight performance for groups even more.[22] Thus, after a 15-minute break, individual performance improved for the rebus puzzles with unhelpful clues, and group performance improved for rebus puzzles with both unhelpful and helpful clues.[22]
Individual differences
Personality and gender, as they relate to performance on insight problems, was studied using a variety of insight problems. It was found that participants who ranked lower on emotionality and higher on openness to experience performed better on insight problems. Men outperformed women on insight problems, and women outperformed men on non-insight problems.[23]
Higher intelligence (higher IQ) has also been found to be associated with better performance on insight problems. However, those of lower intelligence benefit more than those of higher intelligence from being provided with cues and hints for insight problems.[8]
A recent large-scale study in Australia suggests that insight may not be universally experienced, with almost 20% of respondents reporting that they had not experienced insight.[24]
Metacognition
Individuals are poorer at predicting their own metacognition for insight problems, than for non-insight problems.[25] Individuals were asked to indicate how "hot" or "cold" to a solution they felt. Generally, they were able to predict this fairly well for non-insight problems, but not for insight problems.[25] This provides evidence for the suddenness involved during insight.
Naturalistic settings
Recently, insight was studied in a non-laboratory setting.[26] Accounts of insight that have been reported in the media, such as in interviews, etc., were examined and coded. It was found that insights that occur in the field are typically reported to be associated with a sudden "change in understanding" and with "seeing connections and contradictions" in the problem.[26] It was also found that insight in nature differed from insight in the laboratory. For example, insight in nature was often rather gradual, not sudden, and incubation was not as important.[26] Other studies used online questionnaires to further explore insight outside of the laboratory,[27][2] verifying the notion that insight often happens in situations such as in the shower,[24] echoing the idea that creative ideas occur in situations where divergent thought is more likely, sometimes called the Three B's of Creativity, in Bed, on the Bus or in the Bath.
Non-Human Animals
Studies on primate cognition have provided evidence of what may be interpreted as insight in animals. In 1917, Wolfgang Köhler published his book The Mentality of Apes, having studied primates on the island of Tenerife for six years. In one of his experiments, apes were presented with an insight problem that required the use of objects in new and original ways, in order to win a prize (usually, some kind of food). He observed that the animals would continuously fail to get the food, and this process occurred for quite some time; however, rather suddenly, they would purposefully use the object in the way needed to get the food, as if the realization had occurred out of nowhere. He interpreted this behavior as something resembling insight in apes.[28] A more recent study suggested that elephants might also experience insight, showing that a young male elephant was able to identify and move a large cube under food that was out of reach so that he could stand on it to get the reward.[29]
Theories
There are a number of theories representing insight; at present, no one theory dominates interpretation.[8]
Dual-process theory
According to the dual-process theory, there are two systems used to solve problems.[23] The first involves logical and analytical thought processes based on reason, while the second involves intuitive and automatic processes based on experience.[23] Research has demonstrated that insight probably involves both processes; however, the second process is more influential.[23]
Three-process theory
According to the three-process theory, intelligence plays a large role in insight.[30] Specifically, insight involves three different processes (selective encoding, combination, and comparison), which require intelligence to apply to problems.[30] Selective encoding is the process of focusing attention on ideas relevant to a solution, while ignoring features that are irrelevant.[30] Selective combination is the process of combining the information previously deemed relevant.[30] Finally, selective comparison is the use of past experience with problems and solutions that are applicable to the current problem and solution.[30]
Four-stage model
According to the four-stage model of insight, there are four stages to problem solving.[31] First, the individual prepares to solve a problem.[31] Second, the individual incubates on the problem, which encompasses trial-and-error, etc.[31] Third, the insight occurs, and the solution is illuminated.[31] Finally, the verification of the solution to the problem is experienced.[31] Since this model was proposed, other similar models have been explored that contain two or three similar stages.[8]
Psychiatry
In psychology and psychiatry, insight can mean the ability to recognize one's own mental illness.[32][33] Psychiatric insight is typically measured with the Beck cognitive insight scale (BCIS), named after American psychiatrist Aaron Beck.[34] This form of insight has multiple dimensions, such as recognizing the need for treatment, and recognizing consequences of one's behavior as stemming from an illness.[35] A person with very poor recognition or acknowledgment is referred to as having "poor insight" or "lack of insight". The most extreme form is anosognosia, the total absence of insight into one's own mental illness. Many mental illnesses are associated with varying levels of insight. For example, people with obsessive compulsive disorder and various phobias tend to have relatively good insight that they have a problem and that their thoughts and/or actions are unreasonable, yet are compelled to carry out the thoughts and actions regardless.[36] Patients with schizophrenia, and various psychotic conditions tend to have very poor awareness that anything is wrong with them.[37] Psychiatric insight favourably predicts outcomes in cognitive behavioural therapy for psychosis.[38] Today some psychiatrists believe psychiatric medication may contribute to the patient's lack of insight.[39][40][41]
Spirituality
The Pali word for "insight" is vipassana, which has been adopted as the name of a kind of Buddhist mindfulness meditation. Recent research indicates that mindfulness meditation does facilitate solving of insight problems with dosage of 20 minutes.[42]
Marketing
Pat Conroy[citation needed] points out that an insight is a statement based on a deep understanding of your target consumers' attitudes and beliefs, which connect at an emotional level with your consumer, provoking a clear response (This brand understands me! That is exactly how I feel! — even if they've never thought about it quite like that) which, when leveraged, has the power to change consumer behavior. Insights must effect a change in consumer behavior that benefits your brand, leading to the achievement of the marketing objective.[citation needed]
Insights can be based on:
- Real or perceived weakness to be exploited in competitive product performance or value
- Attitudinal or perceived barrier in the minds of consumers, regarding your brand
- Untapped or compelling belief or practice
Insights are most effective when they are/do one of the following:
- Unexpected
- Create a disequilibrium
- Change momentum
- Exploited via a benefit or point of difference that your brand can deliver
In order to be actionable, as the expression of a consumer truth, an insight is to be stated as an articulated sentence, containing:[43]
- An observation or a wish, e.g. "I would like to ...."
- A motivation explaining the wish, e.g. " because ..."
- A barrier preventing the consumer from being satisfied with the fulfillment of his/her motivation, e.g. " but..."
The gap between the second and the third term offers a tension, which constitutes a potential for a brand. Like there are concept writers for copies, there are insight writers.
In technical terminology of insight in market research is the understanding of local market by referring different source of information (such as quantitative research and qualitative research) proving for the consumers' insight.
See also
References
- "Qu'est-ce qu'un insight consommateur ?". insightquest.fr (in French). Retrieved January 14, 2021.
Further reading
- Bradley, Nigel, (2007), Marketing Research: Tools and Techniques, Oxford: Oxford University Press. ISBN 978-0-19-928196-1
External links
- The dictionary definition of insight at Wiktionary
https://en.wikipedia.org/wiki/Insight
Illative sense is an epistemological concept coined by John Henry Newman (1801–1890) in his Grammar of Assent. For him it is the unconscious process of the mind, by which probabilities converge into certainty.
https://en.wikipedia.org/wiki/Illative_sense
Inferences are steps in reasoning, moving from premises to logical consequences; etymologically, the word infer means to "carry forward". Inference is theoretically traditionally divided into deduction and induction, a distinction that in Europe dates at least to Aristotle (300s BCE). Deduction is inference deriving logical conclusions from premises known or assumed to be true, with the laws of valid inference being studied in logic. Induction is inference from particular evidence to a universal conclusion. A third type of inference is sometimes distinguished, notably by Charles Sanders Peirce, contradistinguishing abduction from induction.
Various fields study how inference is done in practice. Human inference (i.e. how humans draw conclusions) is traditionally studied within the fields of logic, argumentation studies, and cognitive psychology; artificial intelligence researchers develop automated inference systems to emulate human inference. Statistical inference uses mathematics to draw conclusions in the presence of uncertainty. This generalizes deterministic reasoning, with the absence of uncertainty as a special case. Statistical inference uses quantitative or qualitative (categorical) data which may be subject to random variations.
Definition
The process by which a conclusion is inferred from multiple observations is called inductive reasoning. The conclusion may be correct or incorrect, or correct to within a certain degree of accuracy, or correct in certain situations. Conclusions inferred from multiple observations may be tested by additional observations.
This definition is disputable (due to its lack of clarity. Ref: Oxford English dictionary: "induction ... 3. Logic the inference of a general law from particular instances."[clarification needed]) The definition given thus applies only when the "conclusion" is general.
Two possible definitions of "inference" are:
- A conclusion reached on the basis of evidence and reasoning.
- The process of reaching such a conclusion.
Examples
Example for definition #1
Ancient Greek philosophers defined a number of syllogisms, correct three part inferences, that can be used as building blocks for more complex reasoning. We begin with a famous example:
- All humans are mortal.
- All Greeks are humans.
- All Greeks are mortal.
The reader can check that the premises and conclusion are true, but logic is concerned with inference: does the truth of the conclusion follow from that of the premises?
The validity of an inference depends on the form of the inference. That is, the word "valid" does not refer to the truth of the premises or the conclusion, but rather to the form of the inference. An inference can be valid even if the parts are false, and can be invalid even if some parts are true. But a valid form with true premises will always have a true conclusion.
For example, consider the form of the following symbological track:
- All meat comes from animals.
- All beef is meat.
- Therefore, all beef comes from animals.
If the premises are true, then the conclusion is necessarily true, too.
Now we turn to an invalid form.
- All A are B.
- All C are B.
- Therefore, all C are A.
To show that this form is invalid, we demonstrate how it can lead from true premises to a false conclusion.
- All apples are fruit. (True)
- All bananas are fruit. (True)
- Therefore, all bananas are apples. (False)
A valid argument with a false premise may lead to a false conclusion, (this and the following examples do not follow the Greek syllogism):
- All tall people are French. (False)
- John Lennon was tall. (True)
- Therefore, John Lennon was French. (False)
When a valid argument is used to derive a false conclusion from a false premise, the inference is valid because it follows the form of a correct inference.
A valid argument can also be used to derive a true conclusion from a false premise:
- All tall people are musicians. (Valid, False)
- John Lennon was tall. (Valid, True)
- Therefore, John Lennon was a musician. (Valid, True)
In this case we have one false premise and one true premise where a true conclusion has been inferred.
Example for definition #2
Evidence: It is the early 1950s and you are an American stationed in the Soviet Union. You read in the Moscow newspaper that a soccer team from a small city in Siberia starts winning game after game. The team even defeats the Moscow team. Inference: The small city in Siberia is not a small city anymore. The Soviets are working on their own nuclear or high-value secret weapons program.
Knowns: The Soviet Union is a command economy: people and material are told where to go and what to do. The small city was remote and historically had never distinguished itself; its soccer season was typically short because of the weather.
Explanation: In a command economy, people and material are moved where they are needed. Large cities might field good teams due to the greater availability of high quality players; and teams that can practice longer (weather, facilities) can reasonably be expected to be better. In addition, you put your best and brightest in places where they can do the most good—such as on high-value weapons programs. It is an anomaly for a small city to field such a good team. The anomaly (i.e. the soccer scores and great soccer team) indirectly described a condition by which the observer inferred a new meaningful pattern—that the small city was no longer small. Why would you put a large city of your best and brightest in the middle of nowhere? To hide them, of course.
Incorrect inference
An incorrect inference is known as a fallacy. Philosophers who study informal logic have compiled large lists of them, and cognitive psychologists have documented many biases in human reasoning that favor incorrect reasoning.
Applications
Inference engines
AI systems first provided automated logical inference and these were once extremely popular research topics, leading to industrial applications under the form of expert systems and later business rule engines. More recent work on automated theorem proving has had a stronger basis in formal logic.
An inference system's job is to extend a knowledge base automatically. The knowledge base (KB) is a set of propositions that represent what the system knows about the world. Several techniques can be used by that system to extend KB by means of valid inferences. An additional requirement is that the conclusions the system arrives at are relevant to its task.
Prolog engine
Prolog (for "Programming in Logic") is a programming language based on a subset of predicate calculus. Its main job is to check whether a certain proposition can be inferred from a KB (knowledge base) using an algorithm called backward chaining.
Let us return to our Socrates syllogism. We enter into our Knowledge Base the following piece of code:
mortal(X) :- man(X). man(socrates).
( Here :- can be read as "if". Generally, if P Q (if P then Q) then in Prolog we would code Q:-P (Q if P).)
This states that all men are mortal and that Socrates is a man. Now we can ask the Prolog system about Socrates:
?- mortal(socrates).
(where ?- signifies a query: Can mortal(socrates). be deduced from the KB using the rules) gives the answer "Yes".
On the other hand, asking the Prolog system the following:
?- mortal(plato).
gives the answer "No".
This is because Prolog does not know anything about Plato, and hence defaults to any property about Plato being false (the so-called closed world assumption). Finally
?- mortal(X) (Is anything mortal) would result in "Yes" (and in some implementations: "Yes": X=socrates)
Prolog can be used for vastly more complicated inference tasks. See the corresponding article for further examples.
Semantic web
Recently automatic reasoners found in semantic web a new field of application. Being based upon description logic, knowledge expressed using one variant of OWL can be logically processed, i.e., inferences can be made upon it.
Bayesian statistics and probability logic
Philosophers and scientists who follow the Bayesian framework for inference use the mathematical rules of probability to find this best explanation. The Bayesian view has a number of desirable features—one of them is that it embeds deductive (certain) logic as a subset (this prompts some writers to call Bayesian probability "probability logic", following E. T. Jaynes).
Bayesians identify probabilities with degrees of beliefs, with certainly true propositions having probability 1, and certainly false propositions having probability 0. To say that "it's going to rain tomorrow" has a 0.9 probability is to say that you consider the possibility of rain tomorrow as extremely likely.
Through the rules of probability, the probability of a conclusion and of alternatives can be calculated. The best explanation is most often identified with the most probable (see Bayesian decision theory). A central rule of Bayesian inference is Bayes' theorem.
Fuzzy logic
This section needs expansion. You can help by adding to it. (October 2016) |
Non-monotonic logic
A relation of inference is monotonic if the addition of premises does not undermine previously reached conclusions; otherwise the relation is non-monotonic. Deductive inference is monotonic: if a conclusion is reached on the basis of a certain set of premises, then that conclusion still holds if more premises are added.
By contrast, everyday reasoning is mostly non-monotonic because it involves risk: we jump to conclusions from deductively insufficient premises. We know when it is worth or even necessary (e.g. in medical diagnosis) to take the risk. Yet we are also aware that such inference is defeasible—that new information may undermine old conclusions. Various kinds of defeasible but remarkably successful inference have traditionally captured the attention of philosophers (theories of induction, Peirce's theory of abduction, inference to the best explanation, etc.). More recently logicians have begun to approach the phenomenon from a formal point of view. The result is a large body of theories at the interface of philosophy, logic and artificial intelligence.
See also
- A priori and a posteriori
- Abductive reasoning
- Deductive reasoning
- Inductive reasoning
- Entailment
- Epilogism
- Analogy
- Axiom system
- Immediate inference
- Inferential programming
- Inquiry
- Logic
- Logic of information
- Logical assertion
- Logical graph
- Rule of inference
- List of rules of inference
- Theorem
- Transduction (machine learning)
References
- Fuhrmann, André. Nonmonotonic Logic (PDF). Archived from the original (PDF) on 9 December 2003.
Further reading
- Hacking, Ian (2001). An Introduction to Probability and Inductive Logic. Cambridge University Press. ISBN 978-0-521-77501-4.
- Jaynes, Edwin Thompson (2003). Probability Theory: The Logic of Science. Cambridge University Press. ISBN 978-0-521-59271-0. Archived from the original on 2004-10-11. Retrieved 2004-11-29.
- McKay, David J.C. (2003). Information Theory, Inference, and Learning Algorithms. Cambridge University Press. ISBN 978-0-521-64298-9.
- Russell, Stuart J.; Norvig, Peter (2003), Artificial Intelligence: A Modern Approach (2nd ed.), Upper Saddle River, New Jersey: Prentice Hall, ISBN 0-13-790395-2
- Tijms, Henk (2004). Understanding Probability. Cambridge University Press. ISBN 978-0-521-70172-3.
Inductive inference:
- Carnap, Rudolf; Jeffrey, Richard C., eds. (1971). Studies in Inductive Logic and Probability. Vol. 1. The University of California Press.
- Jeffrey, Richard C., ed. (1980). Studies in Inductive Logic and Probability. Vol. 2. The University of California Press. ISBN 9780520038264.
- Angluin, Dana (1976). An Application of the Theory of Computational Complexity to the Study of Inductive Inference (Ph.D.). University of California at Berkeley.
- Angluin, Dana (1980). "Inductive Inference of Formal Languages from Positive Data" (PDF). Information and Control. 45 (2): 117–135. doi:10.1016/s0019-9958(80)90285-5.
- Angluin, Dana; Smith, Carl H. (Sep 1983). "Inductive Inference: Theory and Methods" (PDF). Computing Surveys. 15 (3): 237–269. doi:10.1145/356914.356918. S2CID 3209224.
- Gabbay, Dov M.; Hartmann, Stephan; Woods, John, eds. (2009). Inductive Logic. Handbook of the History of Logic. Vol. 10. Elsevier.
- Goodman, Nelson (1983). Fact, Fiction, and Forecast. Harvard University Press. ISBN 9780674290716.
Abductive inference:
- O'Rourke, P.; Josephson, J., eds. (1997). Automated abduction: Inference to the best explanation. AAAI Press.
- Psillos, Stathis (2009). Gabbay, Dov M.; Hartmann, Stephan; Woods, John (eds.). An Explorer upon Untrodden Ground: Peirce on Abduction (PDF). Handbook of the History of Logic. Vol. 10. Elsevier. pp. 117–152.
- Ray, Oliver (Dec 2005). Hybrid Abductive Inductive Learning (Ph.D.). University of London, Imperial College. CiteSeerX 10.1.1.66.1877.
Psychological investigations about human reasoning:
- deductive:
- Johnson-Laird, Philip Nicholas; Byrne, Ruth M. J. (1992). Deduction. Erlbaum.
- Byrne, Ruth M. J.; Johnson-Laird, P. N. (2009). ""If" and the Problems of Conditional Reasoning" (PDF). Trends in Cognitive Sciences. 13 (7): 282–287. doi:10.1016/j.tics.2009.04.003. PMID 19540792. S2CID 657803. Archived from the original (PDF) on 2014-04-07. Retrieved 2013-08-09.
- Knauff, Markus; Fangmeier, Thomas; Ruff, Christian C.; Johnson-Laird, P. N. (2003). "Reasoning, Models, and Images: Behavioral Measures and Cortical Activity" (PDF). Journal of Cognitive Neuroscience. 15 (4): 559–573. CiteSeerX 10.1.1.318.6615. doi:10.1162/089892903321662949. hdl:11858/00-001M-0000-0013-DC8B-C. PMID 12803967. S2CID 782228. Archived from the original (PDF) on 2015-05-18. Retrieved 2013-08-09.
- Johnson-Laird, Philip N. (1995). Gazzaniga, M. S. (ed.). Mental Models, Deductive Reasoning, and the Brain (PDF). MIT Press. pp. 999–1008.
- Khemlani, Sangeet; Johnson-Laird, P. N. (2008). "Illusory Inferences about Embedded Disjunctions" (PDF). Proceedings of the 30th Annual Conference of the Cognitive Science Society. Washington/DC. pp. 2128–2133.
- statistical:
- McCloy, Rachel; Byrne, Ruth M. J.; Johnson-Laird, Philip N. (2009). "Understanding Cumulative Risk" (PDF). The Quarterly Journal of Experimental Psychology. 63 (3): 499–515. doi:10.1080/17470210903024784. PMID 19591080. S2CID 7741180. Archived from the original (PDF) on 2015-05-18. Retrieved 2013-08-09.
- Johnson-Laird, Philip N. (1994). "Mental Models and Probabilistic Thinking" (PDF). Cognition. 50 (1–3): 189–209. doi:10.1016/0010-0277(94)90028-0. PMID 8039361. S2CID 9439284.,
- analogical:
- Burns, B. D. (1996). "Meta-Analogical Transfer: Transfer Between Episodes of Analogical Reasoning". Journal of Experimental Psychology: Learning, Memory, and Cognition. 22 (4): 1032–1048. doi:10.1037/0278-7393.22.4.1032.
- spatial:
- Jahn, Georg; Knauff, Markus; Johnson-Laird, P. N. (2007). "Preferred mental models in reasoning about spatial relations" (PDF). Memory & Cognition. 35 (8): 2075–2087. doi:10.3758/bf03192939. PMID 18265622. S2CID 25356700.
- Knauff, Markus; Johnson-Laird, P. N. (2002). "Visual imagery can impede reasoning" (PDF). Memory & Cognition. 30 (3): 363–371. doi:10.3758/bf03194937. PMID 12061757. S2CID 7330724.
- Waltz, James A.; Knowlton, Barbara J.; Holyoak, Keith J.; Boone, Kyle B.; Mishkin, Fred S.; de Menezes Santos, Marcia; Thomas, Carmen R.; Miller, Bruce L. (Mar 1999). "A System for Relational Reasoning in Human Prefrontal Cortex". Psychological Science. 10 (2): 119–125. doi:10.1111/1467-9280.00118. S2CID 44019775.
- moral:
- Bucciarelli, Monica; Khemlani, Sangeet; Johnson-Laird, P. N. (Feb 2008). "The Psychology of Moral Reasoning" (PDF). Judgment and Decision Making. 3 (2): 121–139. doi:10.1017/S1930297500001479. S2CID 327124.
External links
- Inference at PhilPapers
- Inference example and definition
- Inference at the Indiana Philosophy Ontology Project
- Inference
- Concepts in epistemology
- Concepts in logic
- Concepts in metaphilosophy
- Concepts in metaphysics
- Concepts in the philosophy of mind
- History of logic
- Intellectual history
- Logic
- Logic and statistics
- Logical consequence
- Metaphysics of mind
- Reasoning
- Sources of knowledge
- Thought
https://en.wikipedia.org/wiki/Inference
No comments:
Post a Comment