Excellent post. Are you hip to Gilovitch's book: How We Know What isn't So, The fallibility of human reason in everyday life? He studies human cognitive error at Cornell.
http://www.amazon.com/gp/product/0029117062/sr=8-1/qid=1149893839/ref=pd_bbs_1/102-4458199-6191348?%5Fencoding=UTF8 --- In FairfieldLife@yahoogroups.com, new_morning_blank_slate <[EMAIL PROTECTED]> wrote: > > We all make them. To the extent that we are aware of their existence > and structure, we can avoid them in our own internal reasoning, and in > communications. > > Whoever has more than 20 in any post, gets a gallon of woowoo juice. > > > http://en.wikipedia.org/wiki/List_of_cognitive_biases > > Cognitive bias is distortion in the way we perceive reality (see also > cognitive distortion). > > Some of these have been verified empirically in the field of > psychology, others are considered general categories of bias. > > This is an incomplete list, which may never be able to satisfy > certain standards for completeness. > > > > Decision making and behavioral biases > > Many of these biases are studied for how they affect belief formation > and business decisions and scientific research > > * Bandwagon effect - the tendency to do (or believe) things > because many other people do (or believe) the same. > * Bias blind spot - the tendency not to compensate for one's own > cognitive biases. > * Choice-supportive bias - the tendency to remember one's choices > as better than they actually were. > * Confirmation bias - the tendency to search for or interpret > information in a way that confirms one's preconceptions. > * Congruence bias - the tendency to test hypotheses exclusively > through direct testing > * Contrast effect - the enhancement or diminishment of a weight or > other measurement when compared with recently observed contrasting object. > * Disconfirmation bias - the tendency for people to extend > critical scrutiny to information which contradicts their prior beliefs > and accept uncritically information that is congruent with their prior > beliefs. > * Endowment effect - the tendency for people to value something > more as soon as they own it. > * Focusing effect - prediction bias occurring when people place > too much importance on one aspect of an event; causes error in > accurately predicting the utility of a future outcome. > * Hyperbolic discounting - the tendency for people to have a > stronger preference for more immediate payoffs relative to later > payoffs, the closer to the present both payoffs are. > * Illusion of control - the tendency for human beings to believe > they can control or at least influence outcomes which they clearly cannot. > * Impact bias - the tendency for people to overestimate the length > or the intensity of the impact of future feeling states. > * Information bias - the tendency to seek information even when it > cannot affect action > * Loss aversion - the tendency for people to strongly prefer > avoiding losses over acquiring gains (see also sunk cost effects) > * Neglect of Probability - the tendency to completely disregard > probability when making a decision under uncertainty. > * Mere exposure effect - the tendency for people to express undue > liking for things merely because they are familiar with them. > * Color psychology - the tendency for cultural symbolism of > certain colors to affect affective reasoning. > * Omission Bias - The tendency to judge harmful actions as worse, > or less moral than equally harmful omissions (inactions.) > * Outcome Bias - the tendency to judge a decision by its eventual > outcome instead of based on the quality of the decision at the time it > was made. > * Planning fallacy - the tendency to underestimate task-completion > times. > * Post-purchase rationalization - the tendency to persuade oneself > through rational argument that a purchase was good value. > * Pseudocertainty effect - the tendency to make risk-averse > choices if the expected outcome is positive, but risk-seeking choices > to avoid negative outcomes. > * Rosy retrospection - the tendency to rate past events more > positively than they had actually rated them when the event occurred. > * Selective perception - the tendency for expectations to affect > perception. > * Status quo bias - the tendency for people to like things to stay > relatively the same. > * Von Restorff effect - the tendency for an item that "stands out > like a sore thumb" to be more likely to be remembered than other items. > * Zeigarnik effect - the tendency for people to remember > uncompleted or interrupted tasks better than completed ones. > * Zero-risk bias - preference for reducing a small risk to zero > over a greater reduction in a larger risk. > > > Biases in probability and belief > > Many of these biases are often studied for how they affect business > and economic decisions and how they affect experimental research. > > * Affective forecasting > Affective forecasting is the forecasting of one's affect (emotional > state) in the future. This kind of prediction is affected by various > kinds of cognitive biases, i.e. systematic errors of thought. Daniel > Gilbert of the department of social psychology at Harvard University > and other researchers in the field, such as Timothy Wilson of the > University of Virginia and George Loewenstein of Carnegie Mellon > University, have studied those cognitive biases and given them names > like "empathy gap" and "impact bias" and the like. > > Affective forecasting is an important concept in psychology, because > psychologists try to study what situations in life are important to > humans, and how they change their views with time. > > > * Ambiguity effect - the avoidance of options for which missing > information makes the probability seem "unknown" > > The ambiguity effect is a cognitive bias where decision-making is > affected due to a lack of information, or an "ambiguity." > > For example, picture an urn with 90 balls inside of it. The balls are > colored red, black and yellow. 30 of the balls are red, and the other > 60 are some combination of black and yellow balls, with all > combinations being equally likely. In option X, drawing a red ball > would earn you the $100, and in option Y, drawing a black ball would > earn you the $100. The difference between the two options is that the > number of red balls is certain for option X, but the number of black > balls for option Y is uncertain. > > Which option gives you the best chance at picking out a winning ball? > The truth is that the probability of picking a winning ball is > identical for both options X and Y. In option X, where the number of > red balls is certain, the probability of selecting a winning ball is > 1/3 (30 red balls out of 90 total balls). In option Y, despite the > fact that the number of black balls is not certain, the probability of > selecting a winning ball is also 1/3. This is because the range of > possibilities as to the number of black balls is some amount between 0 > and 60. This means that the probability of there being more than 30 > black balls is the same as there being less than 30 black balls. > Because of this, according to what is known as the expected-utility > theory, one should be indifferent between the two options. As a > result, the chances of winning the $100 are the same for both urns. > > People are much more likely to want to select a ball under option X, > where the probability of selecting a winning ball is, in their minds, > more certain. The question as to the number of black balls under > scenario Y turns people off to that option. Despite the fact that > there could possibly be double the black balls to red balls, people > tend to not want to take the opposing risk that there may be less than > 30 black balls. The "ambiguity" behind option Y makes people want to > select option X, even when they are theoretically equivalent. > > This bias was discovered by Daniel Ellsberg in 1961. Ellsberg deemed > these situations where the "probability is unknown" as "ambiguous," > hence the "ambiguity effect." > > One explanation of the effect is that people follow a heuristic, a > rule of thumb, of avoiding options about what information is missing > (Frisch & Baron, 1988; Ritov & Baron, 1990). This is usually a good > rule because it leads us to look for the information. In many cases, > though, the information cannot be obtained. Information is almost > always missing, and the effect is often the result of calling some > particular missing piece to our attention. > > * Anchoring - the tendency to rely too heavily, or "anchor," on > one trait or piece of information when making decisions > > * Anthropic bias - the tendency for one's evidence to be biased by > observation selection effects > * Attentional bias - neglect of relevant data when making > judgments of a correlation or association > * Availability error - the distortion of one's perceptions of > reality, due to the tendency to remember one alternative outcome of a > situation much more easily than another > * Belief bias - the tendency to base assessments on personal > beliefs (see also belief perseverance and Experimenter's regress) > * Belief Overkill - the tendency to bring beliefs and values > together so that they all point to the same conclusion > * Clustering illusion - the tendency to see patterns where > actually none exist > * Conjunction fallacy - the tendency to assume that specific > conditions are more probable than general ones > * Gambler's fallacy - the tendency to assume that individual > random events are influenced by previous random events "the coin has > a memory" > * Hindsight bias - sometimes called the "I-knew-it-all-along" > effect, the inclination to see past events as being predictable > * Illusory correlation - beliefs that inaccurately suppose a > relationship between a certain type of action and an effect > * Myside bias - the tendency for people to fail to look for or to > ignore evidence against what they already favor > * Neglect of prior base rates effect - the tendency to fail to > incorporate prior known probabilities which are pertinent to the > decision at hand > * Observer-expectancy effect - when a researcher expects a given > result and therefore unconsciously manipulates an experiment or > misinterprets data in order to find it. (see also subject-expectancy > effect) > * Overconfidence effect - the tendency to overestimate one's own > abilities > * Polarization effect - increase in strength of belief on both > sides of an issue after presentation of neutral or mixed evidence, > resulting from biased assimilation of the evidence. > * Positive outcome bias (prediction) - a tendency in prediction to > overestimate the probability of good things happening to them. (see > also wishful thinking and valence effect) > * Recency effect - the tendency to weigh recent events more than > earlier events (see also peak-end rule) > * Primacy effect - the tendency to weigh initial events more than > subsequent events > * Subadditivity effect - the tendency to judge probability of the > whole to be less than the probabilities of the parts. > > > > Social biases > > Most of these biases are labeled as attributional biases. > > * Barnum effect (or Forer Effect) - the tendency to give high > accuracy ratings to descriptions of their personality that supposedly > are tailored specifically for them, but are in fact vague and general > enough to apply to a wide range of people. > * Egocentric bias - occurs when people claim more responsibility > for themselves for the results of a joint action than an outside > observer would. > * False consensus effect - the tendency for people to overestimate > the degree to which others agree with them. > * Fundamental attribution error - the tendency for people to > over-emphasize personality-based explanations for behaviors observed > in others while under-emphasizing the role and power of situational > influences on the same behavior. (see also group attribution error, > positivity effect, and negativity effect) > * Halo effect - the tendency for a person's positive or negative > traits to "spill over" from one area of their personality to another > in others' perceptions of them. (see also physical attractiveness > stereotype) > * Illusion of asymmetic insight - people perceive their knowledge > of their peers to surpass their peers' knowledge of them. > * Ingroup bias - preferential treatment people give to whom they > perceive to be members of their own groups. > * Just-world phenomenon - the tendency for people to believe the > world is "just" and so therefore people "get what they deserve." > * Lake Wobegon effect - the human tendency to report flattering > beliefs about oneself and believe that one is above average (see also > worse-than-average effect, and overconfidence effect). > * Notational bias - a form of cultural bias in which a notation > induces the appearance of a nonexistent natural law. > * Outgroup homogeneity bias - individuals see members of their own > group as being relatively more varied than members of other groups. > * Projection bias - the tendency to unconsciously assume that > others share the same or similar thoughts, beliefs, values, or positions. > * Self-serving bias - the tendency to claim more responsibility > for successes than failures. It may also manifest itself as a tendency > for people to evaluate ambiguous information in a way beneficial to > their interests. (see also group-serving bias) > * Trait ascription bias - the tendency for people to view > themselves as relatively variable in terms of personality, behavior > and mood while viewing others as much more predictable. > * Self-fulfilling prophecy - the tendency to engage in behaviors > that elicit results which will (consciously or subconsciously) confirm > our beliefs. > > ========== > > Other Cognitive Biases > > http://en.wikipedia.org/wiki/Category:Cognitive_biases > (Some duplicates with above) > > * Adaptive Bias > Adaptive Bias is the idea that the human brain has evolved to reason > adaptively, rather than truthfully or even rationally, and that > Cognitive bias may have evolved as a mechanism to reduce the overall > cost of cognitive errors as opposed to merely reducing the number of > cognitive errors, when faced with making a decision under conditions > of uncertainty. > > When making decisions under conditions of uncertainty, two kinds of > errors need to be taken into account - "false positives", i.e. > deciding that a risk or benefit exists when it does not, and "false > negatives", i.e. failing to notice a risk or benefit that exists. > False positives are also commonly called "Type 1 errors", and false > negatives are called "Type 2 errors". > > Where the cost or impact of a type 1 error is much greater than the > cost of a type 2 error (e.g. the water is safe to drink), it can be > worthwhile to bias the decision making system towards making fewer > type 1 errors, i.e. making it less likely to conclude that a > particular situation exists. This by definition would also increase > the number of type 2 errors. Conversely, where a false positive is > much less costly than a false negative (blood tests, smoke detectors), > it makes sense to bias the system towards maximising the probablility > that a particular (very costly) situation will be recognised, even if > this often leads to the (relatively un-costly) event of noticing > something that is not actually there. > > Martie G. Haselton and David M. Buss (2003) state that Cognitive Bias > can be expected to have developed in humans for cognitive tasks where: > > * Decision making is complicated by a significant signal-detection > problem (i.e. when there is uncertainty) > * The solution to the particular kind of decision making problem > has had a recurrent effect on survival and fitness throughout > evolutionary history > * The costs of a "false positive" or "false negative" error > dramatically outweighs the cost of the alternative type of error > > > * Affective forecasting > Affective forecasting is the forecasting of one's affect (emotional > state) in the future. This kind of prediction is affected by various > kinds of cognitive biases, i.e. systematic errors of thought. Daniel > Gilbert of the department of social psychology at Harvard University > and other researchers in the field, such as Timothy Wilson of the > University of Virginia and George Loewenstein of Carnegie Mellon > University, have studied those cognitive biases and given them names > like "empathy gap" and "impact bias" and the like. > > Affective forecasting is an important concept in psychology, because > psychologists try to study what situations in life are important to > humans, and how they change their views with time. > > * Anchor (NLP) > * Anthropic bias > * Apophenia > * Appeal to pity > * Attributional bias > * Availability error > * Availability heuristic > > B > > * Base rate fallacy > * Belief Overkill > * Bias blind spot > > C > > * Choice blindness > * Choice-supportive bias > * Clustering illusion > * Confirmation bias > * Conjunction fallacy > * Contrast effect > * Cultural bias > > D > > * Data dredging > * Disconfirmation bias > > E > > * Egocentric bias > * Empathy gap > * Endowment effect > * Errors in Syllogisms > > > E cont. > > * Exposure effect > > F > > * False consensus effect > * Forer effect > * Fundamental attribution error > > G > > * Gambler's fallacy > * Group attribution error > * Group-serving bias > * Groupthink > > H > > * Halo effect > * Hindsight bias > * Hostile media effect > * Hyperbolic discounting > > I > > * Illusion of control > * Impact bias > * Ingroup bias > > J > > * Just-world phenomenon > > K > > * Kuleshov Effect > > L > > * Lake Wobegon effect > * Loss aversion > > M > > * Memory bias > * Mindset > * Misinformation effect > > N > > * Negativity effect > * Neglect of Probability > * Notational bias > > O > > * Observer-expectancy effect > * Omission Bias > * Outgroup homogeneity bias > * Overconfidence effect > > P > > * Pareidolia > > > P cont. > > * Peak-end rule > * Physical attractiveness stereotype > * Picture superiority effect > * Planning fallacy > * Pollyanna principle > * Positivity effect > * Primacy effect > * Publication bias > > R > > * Recall bias > * Recency effect > * Regression fallacy > * Response bias > * Rosy retrospection > > S > > * Selective perception > * Self-deception > * Self-serving bias > * Serial position effect > * Spacing effect > * Status quo bias > * Subject-expectancy effect > * Sunk cost > * Superstition > * Suspension of judgment > > T > > * Trait ascription bias > > V > > * Valence effect > * Von Restorff effect > > W > > * Wishful thinking > * Worse-than-average effect > > Z > > * Zeigarnik effect > * Zero-risk bias > > > Memory biases may either enhance or impair the recall of memory, or > they may alter the content of what we report remembering. > > List of memory biases > > * Choice-supportive bias - states that chosen options are > remembered as better than rejected options (Mather, Shafir & Johnson, > 2000). > * Classroom effect - states that some portion of student > performance is explained by the classroom environment and teacher as > opposed to purely individual factors. > * Context effect - states that cognition and memory are dependent > on context, such that out-of-context memories are more difficult to > retrieve than in-context memories (i.e, recall time and accuracy for a > work-related memory will be lower at home, and vice versa). > * Hindsight bias - sometimes called the "I-knew-it-all-along" > effect, is the inclination to see past events as being predictable. > * Humor effect - states that humorous items are more easily > remembered than non-humorous ones, which might be explained by the > distinctiveness of humor, the increased cognitive processing time to > understand the humor, or the emotional arousal caused by the humor. > * Infantile amnesia - states that few memories are retained from > before age 2. > * Generation effect - states that self-generated information is > remembered best. > * Lag effect > * Levels-of-processing effect - states that different methods of > encoding information into memory have different levels of > effectiveness (Craik & Lockhart, 1972). > * List-length effect > * Mere exposure effect - states that familiarity increases liking. > * Misinformation effect - states that misinformation affects > people's reports of their own memory. > * Modality effect - states that memory recall is higher for the > last items of a list when the list items were received auditorily > versus visually. > * Mood congruent memory bias - states that information congruent > with one's current mood is remembered best. > * Next-in-line effect > * Part-list cueing effect - states that being shown some items > from a list makes it harder to retrieve the other items. > * Picture superiority effect - states that concepts are much more > likely to be remembered experimentally if they are presented as > pictures rather than as words. > * Positivity effect - states that older adults favor positive over > negative information in their memories. > * Processing difficulty effect - see Levels-of-processing effect. > * Primacy effect - states that the first items on a list show an > advantage in memory. > * Recency effect - states that the last items on a list show an > advantage in memory. > * Rosy retrospection - states that the past is remembered as > better than it really was. > * Serial position effect - states that items at the beginning of a > list are the easiest to recall, followed by the items near the end of > a list; items in the middle are the least likely to be remembered. > * Self-generation effect - states that people are better able to > recall memories of statements that they have generated than similar > statements generated by others. > * Self-relevance effect - states that memories considered > self-relevent are better recalled that other, similar information > * Spacing effect - states that while you are more likely to > remember material if exposed to it many times, you will be much more > likely to remember it if the exposures are repeated over a longer span > of time. > * Suffix effect - states that there is considerable impairment of > the Recency effect, if a redundant suffix item is added to a list, > which the subject is not required to recall (Morton, Crowder & > Prussin, 1972). > * Testing effect - states that frequent testing of material that > has been committed to memory improves memory recall more than simply > study of the material without testing. > * Time-of-day effect > * Verbatim effect - states that the "gist" of what someone has > said is better remembered than the verbatim wording (Poppenk, Walia, > Joanisse, Danckert, & Köhler, 2006) > * Von Restorff effect - states that an item that "stands out like > a sore thumb" is more likely to be remembered than other items (von > Restorff, 1933). > * Zeigarnik effect - states that people remember uncompleted or > interrupted tasks better than completed ones. > > > Recall bias > From Wikipedia, the free encyclopedia > Jump to: navigation, search > > Taken generally, recall bias is a type of statistical bias which > occurs when the way a survey respondent answers a question is affected > not just by the correct answer, but also by the respondent's memory. > [1] [2] This can affect the results of the survey. As a hypothetical > example, suppose that a survey in 2005 asked respondents whether they > believed that O. J. Simpson had killed his wife. Respondents who > believed him innocent might be more likely to have forgotten about the > case, and therefore to state no opinion, than respondents who thought > him guilty. If this is the case, then the survey would find a > higher-than-accurate proportion of people who believed that O.J. did > kill his wife. > > Relatedly but distinctly, the term might also be used to describe an > instance where a survey respondent intentionally responds incorrectly > to a question about their personal history which results in response > bias. As a hypothetical example, suppose that a researcher conducts a > survey among women of group A, asking whether they have had an > abortion, and the same survey among women of group B. > > If the results are different between the two groups, it might be that > women of one group are less likely to have had an abortion, or it > might simply be that women of one group who have had abortions are > less likely to admit to it. If the latter is the case, then this would > skew the survey results; this is a kind of response bias. (It is also > possible that both are the case: women of one group are less likely to > have had abortions, and women of one group who have had abortions are > less likely to admit to it. This would still affect the survey > statistics.) > > ==== > > > Logical Fallacies > > Aristotelian fallacies > [edit] > > Material fallacies > > The classification of material fallacies widely adopted by modern > logicians and based on that of Aristotle, Organon (Sophistici > elenchi), is as follows: > > * Fallacy of Accident (also called destroying the exception or a > dicto simpliciter ad dictum secundum quid) meaning to argue > erroneously from a general rule to a particular case, without proper > regard to particular conditions that vitiate the application of the > general rule; e.g. if manhood suffrage be the law, arguing that a > criminal or a lunatic must, therefore, have a vote. > > * Converse Fallacy of Accident (also called reverse accident, > destroying the exception, or a dicto secundum quid ad dictum > simpliciter) meaning to argue from a special case to a general rule. > > * Irrelevant Conclusion (also called Ignoratio Elenchi), wherein, > instead of proving the fact in dispute, the arguer seeks to gain his > point by diverting attention to some extraneous fact (as in the legal > story of "No case. Abuse the plaintiff's attorney"). The fallacies are > common in platform oratory, in which the speaker obscures the real > issue by appealing to his audience on the grounds of > o purely personal considerations (argumentum ad hominem) > o popular sentiment (argumentum ad populum, appeal to the > majority) > o fear (argumentum ad baculum) > o conventional propriety (argumentum ad verecundiam) > > This fallacy has been illustrated by ethical or theological > arguments wherein the fear of punishment is subtly substituted for > abstract right as the sanction of moral obligation. > > * Begging the question (also called Petitio Principii or Circulus > in Probando--arguing in a circle) consists in demonstrating a > conclusion by means of premises that pre-suppose that conclusion. > Jeremy Bentham points out that this fallacy may lurk in a single word, > especially in an epithet, e.g. if a measure were condemned simply on > the ground that it is alleged to be "un-English". > > * Fallacy of the Consequent, really a species of Irrelevant > Conclusion, wherein a conclusion is drawn from premises that do not > really support it. > > * Fallacy of False Cause, or Non Sequitur (L., it does not > follow), wherein one thing is incorrectly assumed as the cause of > another, as when the ancients attributed a public calamity to a > meteorological phenomenon (a special case of this fallacy also goes by > the Latin term post hoc ergo propter hoc; the fallacy of believing > that temporal succession implies a causal relation). > > * Fallacy of Many Questions (Plurium Interrogationum), wherein > several questions are improperly grouped in the form of one, and a > direct categorical answer is demanded, e.g. if a prosecuting counsel > asked the prisoner " What time was it when you met this man? " with > the intention of eliciting the tacit admission that such a meeting had > taken place. Another example is the classic line, "Is it true that you > no longer beat your wife?" > > [edit] > > Verbal fallacies > > Verbal fallacies are those in which a false conclusion is obtained by > improper or ambiguous use of words. They are generally classified as > follows. > > * Equivocation consists in employing the same word in two or more > senses, e.g. in a syllogism, the middle term being used in one sense > in the major and another in the minor premise, so that in fact there > are four not three terms ("All fair things are honourable; This woman > is fair; therefore this woman is honourable," the second "fair" being > in reference to complexion). > * Amphibology is the result of ambiguity of grammatical structure, > e.g. of the position of the adverb "only" in careless writers ("He > only said that," in which sentence, as experience shows, the adverb > has been intended to qualify any one of the other three words). > * Fallacy of Composition is a species of Amphibology that results > from the confused use of collective terms. e.g. "The angles of a > triangle are less than two right angles" might refer to the angles > separately or added together. > * Division, the converse of the preceding, which consists in > employing the middle term distributively in the minor and collectively > in the major premise. > * Accent, which occurs only in speaking and consists of > emphasizing the wrong word in a sentence. e.g., "He is a fairly good > pianist," according to the emphasis on the words, may imply praise of > a beginner's progress, or an expert's depreciation of a popular hero, > or it may imply that the person in question is a deplorable violinist. > * Figure of Speech, the confusion between the metaphorical and > ordinary uses of a word or phrase. > > Logical Fallacies > > http://en.wikipedia.org/wiki/Fallacy > > The standard Aristotelian logical fallacies are: > > * Fallacy of Four Terms (Quaternio terminorum) > * Fallacy of Undistributed Middle > * Fallacy of Illicit process of the major or the Illicit minor term; > * Fallacy of Negative Premises. > > [edit] > > Other systems of classification > > Of other classifications of fallacies in general the most famous are > those of Francis Bacon and J. S. Mill. Bacon (Novum Organum, Aph. 33, > 38 sqq.) divided fallacies into four Idola (Idols, i.e. False > Appearances), which summarize the various kinds of mistakes to which > the human intellect is prone. With these should be compared the > Offendicula of Roger Bacon, contained in the Opus maius, pt. i. J. S. > Mill discussed the subject in book v. of his Logic, and Jeremy > Bentham's Book of Fallacies (1824) contains valuable remarks. See Rd. > Whateley's Logic, bk. v.; A. de Morgan, Formal Logic (1847) ; A. > Sidgwick, Fallacies (1883) and other textbooks. > [edit] > > Fallacies in the media and politics > > Fallacies are used frequently by pundits in the media and politics. > When one politician says to another, "You don't have the moral > authority to say X", this could be an example of the argumentum ad > hominem or personal attack fallacy; that is, attempting to disprove X, > not by addressing validity of X but by attacking the person who > asserted X. Arguably, the politician is not even attempting to make an > argument against X, but is instead offering a moral rebuke against the > interlocutor. For instance, if X is the assertion: > > The military uniform is a symbol of national strength and honor. > > Then ostensibly, the politician is not trying to prove the contrary > assertion. If this is the case, then there is no logically fallacious > argument, but merely a personal opinion about moral worth. Thus > identifying logical fallacies may be difficult and dependent upon context. > > In the opposite direction is the fallacy of argument from authority. A > classic example is the ipse dixit"He himself said it" argumentused > throughout the Middle Ages in reference to Aristotle. A modern > instance is "celebrity spokespersons" in advertisements: a product is > good and you should buy/use/support it because your favorite celebrity > endorses it. > > An appeal to authority is always a logical fallacy, though it can be > an appropriate form of rational argument if, for example, it is an > appeal to expert testimony. In this case, the expert witness must be > recognized as such and all parties must agree that the testimony is > appropriate to the circumstances. This form of argument is common in > legal situations. > > By definition, arguments with logical fallacies are invalid, but they > can often be (re)written in such a way that they fit a valid argument > form. The challenge to the interlocutor is, of course, to discover the > false premise, i.e. the premise that makes the argument unsound. > [edit] > > General list of fallacies > > The entries in the following list are neither exhaustive nor mutually > exclusive; that is, several distinct entries may refer to the same > pattern. As noted in the introduction, these fallacies describe > erroneous or at least suspect patterns of argument in general, not > necessarily argument based on formal logic. Many of the fallacies > listed are traditionally recognized and discussed in works on critical > thinking; others are more specialized. > > * Ad hominem (also called argumentum ad hominem or personal > attack) including: > o ad hominem abusive (also called argumentum ad personam) > o ad hominem circumstantial (also called ad hominem > circumstantiae) > o ad hominem tu quoque (also called you-too argument) > * Amphibology (also called amphiboly) > * Appeal to authority (also called argumentum ad verecundiam or > argument by authority) > * Appeal to emotion including: > o Appeal to consequences (also called argumentum ad > consequentiam) > o Appeal to fear (also called argumentum ad metum or > argumentum in terrorem) > o Appeal to flattery > o Appeal to pity (also called argumentum ad misericordiam) > o Appeal to ridicule > o Appeal to spite (also called argumentum ad odium) > o Two wrongs make a right > o Wishful thinking > * Appeal to the majority (also called Appeal to belief, Argumentum > ad numerum, Appeal to popularity, Appeal to the people, Bandwagon > fallacy, Argumentum ad populum, Authority of the many, Consensus > gentium, Argument by consensus) > * Appeal to motive > * Appeal to novelty (also called argumentum ad novitatem) > * Appeal to probability > * Appeal to tradition (also called argumentum ad antiquitatem or > appeal to common practice) > * Argument from fallacy (also called argumentum ad logicam) > * Argument from ignorance (also called argumentum ad ignorantiam > or argument by lack of imagination) > * Argument from silence (also called argumentum ex silentio) > * Appeal to force (also called argumentum ad baculum) > * Appeal to wealth (also called argumentum ad crumenam) > * Appeal to poverty (also called argumentum ad lazarum) > * Argument from repetition (also called argumentum ad nauseam) > * Base rate fallacy > * Begging the question (also called petitio principii, circular > argument or circular reasoning) > * Conjunction fallacy > * Continuum fallacy (also called fallacy of the beard) > * Correlative based fallacies including: > o Fallacy of many questions (also called complex question, > fallacy of presupposition, loaded question or plurium interrogationum) > o False dilemma (also called false dichotomy or bifurcation) > o Denying the correlative > o Suppressed correlative > * Definist fallacy > * Dicto simpliciter, including: > o Accident (also called a dicto simpliciter ad dictum > secundum quid) > o Converse accident (also called a dicto secundum quid ad > dictum simpliciter) > * Equivocation > * Engineering Fallacy > * Fallacies of distribution: > o Composition > o Division > o Ecological fallacy > * Fallacies of Presumption > * False analogy > * False premise > * False compromise > * Faulty generalization including: > o Biased sample > o Hasty generalization (also called fallacy of insufficient > statistics, fallacy of insufficient sample, fallacy of the lonely > fact, leaping to a conclusion, hasty induction, secundum quid) > o Overwhelming exception > o Statistical special pleading > * Gambler's fallacy/Inverse gambler's fallacy > * Genetic fallacy > * Guilt by association > * Historian's fallacy > * Homunculus fallacy > * If-by-whiskey (argues both sides) > * Ignoratio elenchi (also called irrelevant conclusion) > * Inappropriate interpretations or applications of statistics > including: > o Biased sample > o Correlation implies causation > o Gambler's fallacy > o Prosecutor's fallacy > o Screening test fallacy > * Incomplete comparison > * Inconsistent comparison > * Invalid proof > * Judgemental language > * Juxtaposition > * Lump of labour fallacy (also called the fallacy of labour scarcity) > * Meaningless statement > * Middle ground (also called argumentum ad temperantiam) > * Misleading vividness > * Naturalistic fallacy > * Negative proof > * Non sequitur including: > o Affirming the consequent > o Denying the antecedent > * No true Scotsman > * Package deal fallacy > * Perfect solution fallacy > * Poisoning the well > * Progressive fallacy ("New is improved") > * Proof by assertion > * Questionable cause (also called non causa pro causa) including: > o Correlation implies causation (also called cum hoc ergo > propter hoc) > o Fallacy of the single cause > o Joint effect > o Post hoc (also called post hoc ergo propter hoc) > o Regression fallacy > o Texas sharpshooter fallacy > o Wrong direction > * Red herring (also called irrelevant conclusion) > * Reification (also called hypostatization) > * Relativist fallacy (also called subjectivist fallacy) > * Retrospective determinism (it happened so it was bound to) > * Shifting the burden of proof > * Slippery slope > * Special pleading > * Straw man > * Style over substance fallacy > * Sunk cost fallacy > * Syllogistic fallacies, including: > o Affirming a disjunct > o Affirmative conclusion from a negative premise > o Existential fallacy > o Fallacy of exclusive premises > o Fallacy of four terms (also called quaternio terminorum) > o Fallacy of the undistributed middle > o Illicit major > o Illicit minor > > [edit] > > General examples > > Fallacious arguments involve not only formal logic but also causality. > Others involve psychological ploys such as use of power relationships > between proposer and interlocutor, appeals to patriotism and morality, > appeals to ego etc., to establish necessary intermediate (explicit or > implicit) premises for an argument. Indeed, fallacies very often lay > in unstated assumptions or implied premises in arguments that are not > always obvious at first glance. One way to obscure a premise is > through enthymeme. > > We now give a few examples illustrating common errors in reasoning. > Note that providing a critique of an argument has no relation to the > truth of the conclusion. The conclusion could very well be true, while > the argument itself is not valid. See argument from fallacy. > > In the following, we view an argument as a dialogue between a proposer > and an interlocutor. > [edit] > > Example 1: Material Fallacy > > James argues: > > 1. Cheese is food. > 2. Food is delicious. > 3. Therefore, cheese is delicious. > > This argument claims to prove that cheese is delicious. This > particular argument has the form of a categorical syllogism. Any > argument must have premises as well as a conclusion. In this case we > need to ask what the premises are, that is the set of assumptions the > proposer of the argument can expect the interlocutor to grant. The > first assumption is almost true by definition: cheese is a foodstuff > edible by humans. The second assumption is less clear as to its > meaning. Since the assertion has no quantifiers of any kind, it could > mean any one of the following: > > * All food is delicious. > * Most food is delicious. > * All food is delicious, except for spoiled or moldy food. > * Some food is disgusting. > > In any of the last three interpretations, the above syllogism would > then fail to have validated its second premise. James may try to > assume that his interlocutor believes that all food is delicious; if > the interlocutor grants this then the argument is valid. In this case, > the interlocutor is essentially conceding the point to James. However, > the interlocutor is more likely to believe that some food is > disgusting, such as a sheep's liver white chocolate torte; and in this > case James is not much better off than he was before he formulated the > argument, since he now has to prove the assertion that cheese is a > unique type of universally delicious food, which is a disguised form > of the original thesis. From the point of view of the interlocutor, > James commits the logical fallacy of begging the question. > [edit] > > Example 2: Verbal Fallacy > > Barbara argues: > > 1. Andre is a good tennis player. > 2. Therefore, Andre is 'good', that is to say a morally good person. > > Here the problem is that the word good has different meanings, which > is to say that it is an ambiguous word. In the premise, Barbara says > that Andre is good at some particular activity, in this case tennis. > In the conclusion, she says that Andre is a morally good person. These > are clearly two different senses of the word "good". The premise might > be true but the conclusion can still be false: Andre might be the best > tennis player in the world but a rotten person morally. However, it is > not legitimate to infer he is a bad person on the ground there has > been a fallacious argument on the part of Barbara. Nothing concerning > Andre's moral qualities is to be inferred from the premise. > Appropriately, since it plays on an ambiguity, this sort of fallacy is > called the fallacy of equivocation, that is, equating two incompatible > terms or claims. > [edit] > > Example 3: Verbal Fallacy > > Ramesh argues: > > 1. Nothing is better than eternal happiness. > 2. Eating a hamburger is better than nothing. > 3. Therefore, eating a hamburger is better than eternal happiness. > > This argument has the appearance of an inference that applies > transitivity of the two-placed relation is better than, which in this > critique we grant is a valid property. The argument is an example of > syntactic ambiguity. In fact, the first premise semantically does not > predicate an attribute of the subject, as would for instance the assertion > > A potato is better than eternal happiness. > > In fact it is semantically equivalent to the following universal > quantification: > > Everything fails to be better than eternal happiness. > > So instantiating this fact with eating a hamburger, it logically > follows that > > Eating a hamburger fails to be better than eternal happiness. > > Note that the premise A hamburger is better than nothing does not > provide anything to this argument. This fact really means something > such as > > Eating a hamburger is better than eating nothing at all. > > Thus this is a fallacy of composition. > [edit] > > Example 4: Logical Fallacy > > In the strictest sense, a logical fallacy is the incorrect application > of a valid logical principle or an application of a nonexistent principle: > > 1. Some drivers are men. > 2. Some drivers are women. > 3. Therefore, some drivers are both men and women. > > This is fallacious. Indeed, there is no logical principle that states > > 1. For some x, P(x). > 2. For some x, Q(x). > 3. Therefore for some x, P(x) and Q(x). > > An easy way to show the above inference is invalid is by using Venn > diagrams. In logical parlance, the inference is invalid, since under > at least one interpretation of the predicates it is not validity > preserving. > ------------------------ Yahoo! Groups Sponsor --------------------~--> Everything you need is one click away. Make Yahoo! your home page now. http://us.click.yahoo.com/AHchtC/4FxNAA/yQLSAA/UlWolB/TM --------------------------------------------------------------------~-> To subscribe, send a message to: [EMAIL PROTECTED] Or go to: http://groups.yahoo.com/group/FairfieldLife/ and click 'Join This Group!' Yahoo! Groups Links <*> To visit your group on the web, go to: http://groups.yahoo.com/group/FairfieldLife/ <*> To unsubscribe from this group, send an email to: [EMAIL PROTECTED] <*> Your use of Yahoo! Groups is subject to: http://docs.yahoo.com/info/terms/