Irritating shit on facebook

Off Topic (Everything besides dubstep)
Forum rules
Please read and follow this sub-forum's specific rules listed HERE, as well as our sitewide rules listed HERE.

Link to the Secret Ninja Sessions community ustream channel - info in this thread
Locked
User avatar
ezza
THE POOREST
Posts: 9934
Joined: Sat Apr 28, 2012 8:19 pm
Location: brizzy

Re: Irritating shit on facebook

Post by ezza » Thu Jan 17, 2013 2:14 pm

Image
Have a history teacher explain this if they can.

Abraham Lincoln was elected to Congress in 1846.

John F. Kennedy was elected to Congress in 1946.

Abraham Lincoln was elected President in 1860.

John F. Kennedy was elected President in 1960.

Both were particularly concerned with civil rights.

Both wives lost a child while living in the White House.

Both Presidents were shot on a Friday.

Both Presidents were shot in the head.

Now it gets really weird.

Lincoln's secretary was named Kennedy.

Kennedy's Secretary was named Lincoln.

Both were assassinated by Southerners.

Both were succeeded by Southerners named Johnson.

Andrew Johnson, who succeeded Lincoln, was born in 1808.

Lyndon Johnson, who succeeded Kennedy, was born in 1908.

John Wilkes Booth, who assassinated Lincoln, was born in 1839.

Lee Harvey Oswald, who assassinated Kennedy, was born in 1939.

Both assassins were known by their three names.

Both names are composed of fifteen letters.

Now hang on to your seat.

Lincoln was shot at the theater named "Ford."

Kennedy was shot in a car called "Lincoln" made by "Ford."

Booth and Oswald were assassinated before their trials.

And here's the "kicker":

A week before Lincoln was shot, he was in Monroe, Maryland.

A week before Kennedy was shot, he was with Marilyn Monroe.

AND...................:

Lincoln was shot in a theater and the assassin ran to a warehouse...

Kennedy was shot from a warehouse and the assassin ran to a theater..
:o lolol
DiegoSapiens wrote:thats so industrial
soronery wrote:New low

wub
Posts: 34156
Joined: Tue Feb 26, 2008 3:11 pm
Location: Madrid
Contact:

Re: Irritating shit on facebook

Post by wub » Thu Jan 17, 2013 2:24 pm

Agent 47 wrote:Image
Have a history teacher explain this if they can.

Abraham Lincoln was elected to Congress in 1846.

John F. Kennedy was elected to Congress in 1946.

Abraham Lincoln was elected President in 1860.

John F. Kennedy was elected President in 1960.

Both were particularly concerned with civil rights.

Both wives lost a child while living in the White House.

Both Presidents were shot on a Friday.

Both Presidents were shot in the head.

Now it gets really weird.

Lincoln's secretary was named Kennedy.

Kennedy's Secretary was named Lincoln.

Both were assassinated by Southerners.

Both were succeeded by Southerners named Johnson.

Andrew Johnson, who succeeded Lincoln, was born in 1808.

Lyndon Johnson, who succeeded Kennedy, was born in 1908.

John Wilkes Booth, who assassinated Lincoln, was born in 1839.

Lee Harvey Oswald, who assassinated Kennedy, was born in 1939.

Both assassins were known by their three names.

Both names are composed of fifteen letters.

Now hang on to your seat.

Lincoln was shot at the theater named "Ford."

Kennedy was shot in a car called "Lincoln" made by "Ford."

Booth and Oswald were assassinated before their trials.

And here's the "kicker":

A week before Lincoln was shot, he was in Monroe, Maryland.

A week before Kennedy was shot, he was with Marilyn Monroe.

AND...................:

Lincoln was shot in a theater and the assassin ran to a warehouse...

Kennedy was shot from a warehouse and the assassin ran to a theater..
:o lolol
http://www.snopes.com/history/american/ ... ennedy.asp


Let's examine them one at a time:

Abraham Lincoln was elected to Congress in 1846.
John F. Kennedy was elected to Congress in 1946.


This statement is literally true: both Lincoln and Kennedy were first elected to Congress one hundred years apart. Aside from that minor coincidence, however, their political careers bore little resemblance to each other.

Lincoln was an Illinois state legislator who, outside of his election to a single term in the House of Representatives, failed in his every attempt to gain national political office until he was elected President in 1860, including an unsuccessful bid for the Senate in 1854, a unsuccessful bid to become the Republican vice-presidential candidate in 1856, and another unsuccessful bid for a Senate seat in 1858.

Kennedy, on the other hand, enjoyed an unbroken string of political successes at the national level when he entered the political arena after World War II. He was elected to the House of Representatives in 1946, re-elected in 1948, re-elected again in 1950, won a Senate seat in 1952, was re-elected to the Senate in 1958, and was elected President in 1960.

Abraham Lincoln was elected President in 1860.
John F. Kennedy was elected President in 1960.


It's hardly surprising that two men who (as noted above) both achieved their first political successes at the national level a hundred years apart would also ascend to the Presidency a hundred years apart. This "coincidence" is even less surprising when we consider that presidential elections are held only once every four years. Lincoln couldn't possibly have been elected President in 1857 or 1858 or 1859 or 1861 or 1862 or 1863, because no presidential elections were held in those years. Likewise, Kennedy couldn't possibly have been elected
President in the non-election years of 1957, 1958, 1959, 1961, 1962, or 1963. So, even though both men were politically active at the national level during eight-year spans when they might have been elected President, circumstances dictated that the only years during those spans when they both could have been elected were exactly one hundred years apart.

We're supposed to be amazed at minor happenstances such as the two men's being elected exactly one hundred years apart, but we're supposed to think nothing of the numerous non-coincidences: Lincoln was born in 1809; Kennedy was born in 1917. Lincoln died in 1865; Kennedy died in 1963. Lincoln was 56 years old at the time of his death; Kennedy was 46 years old at the time of his death. No striking coincidences or convenient hundred-year differences in any of those facts. Even when we consider that, absent all other factors, the two men had a one in twelve chance of dying in the same month, we find no coincidence there: Lincoln was killed in April; Kennedy was killed in November. Also unmentioned here is the fact that Lincoln was re-elected to a second term as President, but Kennedy was killed before the completion of his first term.

The names Lincoln and Kennedy each contain seven letters.

Surely this is the most trivial of coincidences, especially when once considers that the average length of presidential surnames is 6.64 letters. No mention is made of the fact that the two men's first names contain different numbers of letters, and that Kennedy had a middle name (Fitzgerald) while Lincoln had none.


Both were particularly concerned with civil rights.

Saying that Lincoln and Kennedy were both "particularly concerned with civil rights" is like saying that Woodrow Wilson and Franklin Roosevelt were both "particularly concerned with war," or that Herbert Hoover and Ronald Reagan were both "particularly concerned with economics." Those weren't subjects these men evinced a particular interest in; those were issues they were forced to deal with due to events currently taking place in the U.S. which were beyond their control.

Both wives lost their children while living in the White House.

Another statement that, while literally true, encompasses events that were completely different in circumstance and nature.

All of Lincoln's children were born before he entered the White House, and the Lincolns actually lost two children, not just one (although only one died during Lincoln's tenure as President). Edward Lincoln died of tuberculosis in 1850, just before his fourth birthday, and the Lincolns' eleven-year-old son Willie succumbed to typhoid at the end of their first year in the White House.

The Kennedys, on the other hand, were the rare Presidential couple still young enough to be bearing children after entering the White House, and a premature child born to Mrs. Kennedy in 1963 died two days later.

Other substantial differences not mentioned: The Lincolns had four children, all boys, only one of whom lived past his teens. The Kennedys had three children, two boys and a girl, two of whom survived well into adulthood.

Both Presidents were shot on a Friday.

Another non-surprise. Absent all other factors, the odds were already one in seven that both killings would have occurred on the same day of the week. (Don't even think about writing to tell us that we're wrong and the odds are really one in forty-nine. If you think we're wrong, you don't understand the question.) Add to that the obvious notions that the best chance the average person has to shoot a President is at a public function and that most public functions are held on weekends, and it becomes even more likely that a President would be killed on a Friday, Saturday, or Sunday. (Indeed, an earlier plot by Booth to kidnap Lincoln while the latter was attending a play at the Campbell Hospital was slated for March 17, also a Friday.)

Both were shot in the head.

This "coincidence" is another one which is exceedingly trivial in nature. The only two types of shots which reasonably assure a dead victim are chest shots and head shots, so two assassinations committed by head shots aren't the least bit coincidental — especially when one considers that since both Lincoln and Kennedy were shot from behind and while seated, so their assassins had no other practical choice of target. And the "coincidence" here is even less surprising when we note the substantial differences: Lincoln was killed indoors with a small handgun at point blank range; Kennedy was shot outdoors with a rifle from several hundred feet away.

Lincoln's secretary, Kennedy, warned him not to go to Ford's Theatre.
Kennedy's secretary, Lincoln, warned him not to go to Dallas.


This is one of those coincidences that isn't a coincidence at all; it's simply wrong. John Kennedy did have a secretary named Evelyn Lincoln (who may or may not have warned him about going to Dallas), but one searches in vain to find a Lincoln secretary named Kennedy. (Lincoln's White House secretaries were John G. Nicolay and John Hay.)

The more important point is that since Presidents are frequent recipients of assassination threats, they rarely make any public appearances without somebody's warning them of potential danger. Only on the extemely rare occasions when a tragedy actually occurs do we later take note of the warnings; in all other cases the failed "prophecies" are quickly forgotten. (Lincoln received "an unusual number of letters about plots to kidnap or assassinate him," said to have numbered at least eighty, yet none of those plots were enacted.) Nor does anyone think to mention other attempts at kidnap or assassination that were not preceded by any recorded warnings to the victims. (Lincoln was shot at on at least one other occasion.)

Yes, Lincoln was warned not to go to Ford's Theatre by persons concerned for his safety, just as he had been warned not to visit Richmond a week earlier, and just as he had been warned not to attend his own inauguration in 1861. Obviously, only one of the myriad of warnings he received throughout his four years in office was on the mark. Likewise, Kennedy was warned not to visit San Antonio the day before his trip to Dallas (and undoubtedly before a host of other appearances as well), but only the last warning he allegedly received is considered significant, because it coincidentally happened to come true. As Jeane Dixon and other "psychics" have demonstrated, if you make enough predictions, one of them is eventually bound to come true; the public remembers only that and forgets about all the others failed predictions.

Both were assassinated by Southerners.

A dubious use of the term "Southerner." John Wilkes Booth was undeniably a Southern sympathizer, but he was born in Maryland, which (along with Delaware) was the northernmost of the border slave states and remained part of the Union throughout the Civil War. Additionally, Booth spent a good deal of his life in the North and "thought of himself as a Northerner who understood the South."

Oswald was nominally a Southerner by virtue of his having been born in New Orleans; he spent his youth being shuttled between Lousiana, Texas, and New York before finally joining the Marines. But Oswald's "Southerness" is of no real import, because, unlike Booth, Oswald was not motivated by a regional affiliation.

Both were succeeded by Southerners.

Both Lincoln and Kennedy were "succeeded by Southerners" because both had Southerners as vice-president, another fact hardly surprising considering the historical circumstances of their times. Lincoln was a Northern Republican running for re-election while the country was in the midst of a civil war and needed a Southerner and a Democrat to balance the ticket, hence his choice of Tennessean Andrew Johnson. Kennedy, represented New England and therefore needed a vice-presidential candidate who could appeal to the populous Southern and Western regions, hence his choice of a Southwesterner, Texan Lyndon Johnson.

The identification of Andrew Johnson as a "Southerner" is also a bit problematic here. Although Johnson was born in North Carolina and spent his adult life in Tennessee (both slave states), Johnson was also the only Southern senator who refused to follow his state when it seceded, and he remained loyal to the Union.

Both successors were named Johnson.

Given the high frequency of "Johnson" (literally "son of John") as a surname in both Lincoln's and Kennedy's time, this "coincidence" should be no real surprise to anyone.

Andrew Johnson, who succeeded Lincoln, was born in 1808.
Lyndon Johnson, who succeeded Kennedy, was born in 1908.


Another hundred-year coincidence that is hardly surprising, since nearly all American politicians have attained high office (President or Vice-President) while in the 50-70 age range (and Andrew Johnson and Lyndon Johnson were, obviously, contemporaries of Lincoln and Kennedy, respectively). Perhaps it's time to point out that there's nothing "coincidental" about events merely because they somehow involve the number 100. If we sifted through all the Lincoln/Kennedy data, we could produce multiple instances of events involving the number 17 or 49 or 116, but nobody would consider those "coincidences" because they don't yield nice round numbers that have any significance to us, even though they're all just as "coincidental" as the number 100.

And once again, let's consider all the differences between the two Johnsons, such as that one hailed from North Carolina while the other was from Texas, or that one supported slavery while the other championed civil rights, or that one was never elected President in his own right while the other won the biggest presidential landslide in history, or that one was impeached while the other wasn't, or that one became President at the end of a war while the other became President at the beginning of a war.

John Wilkes Booth was born in 1839.
Lee Harvey Oswald was born in 1939.


Another coincidence that is no coincidence because it's plain wrong: Booth was born in 1838, not 1839.

Both assassins were known by their three names.

Another "coincidence" of dubious veracity. John Wilkes Booth was often billed as "J. Wilkes Booth" or simply "John Wilkes" (primarily to distinguish himself from his father and brother — both named Junius — and his brother Edwin, all three of whom were also actors), and as a prominent actor, his name was already familiar to the general public at the time of Lincoln's assassination. Lee Oswald was generally referred to as "Lee" (not "Lee Harvey") before Kennedy's assassination and was unknown to the general public until his arrest; the common usage of his full name only came about after the assassination because his habitual employment of false names (including several variations on his real name) and his possession of forged identification cards made it difficult for the Dallas police to initially identify him.

Both names are comprised of fifteen letters

Coincidence? Neither their first nor last names have the same number of letters. And why should it be significant that both assassins had the same number of letters in their full names when the same wasn't true of Abraham Lincoln and John Fitzgerald Kennedy, or of Andrew Johnson and Lyndon Baines Johnson?

Once again, perhaps we should focus on the substantive differences between the two men: Booth was born into a prominent family and, like his father, was a well-known, popular, gregarious actor. Oswald was born (and lived most of his life) in near poverty-level circumstances, never knew his father (who died two months before Oswald was born) and was an obscure, moody malcontent who never had any close friends or a steady job. Oswald was married with two children; Booth had neither wife nor offspring. Oswald enlisted in the Marines, but Booth kept a promise to his mother not to join the Confederate army.

Booth ran from the theater and was caught in a warehouse.
Oswald ran from a warehouse and was caught in a theater.


Another "coincidence" that is both inaccurate and superficial.

Booth shot Lincoln in a theatre of the type where live stage shows are held, then fled across state lines before being trapped and killed in a tobacco shed several days later.

Oswald shot Kennedy from (not in) a textbook warehouse, then remained in Dallas and was caught and taken alive in a movie theater a little over an hour later.

Booth and Oswald were assassinated before their trials.

Another superficial similarity with much more significant underlying differences, and a potentially dubious use of the word "assassinated."

After Booth shot Lincoln, he fled the scene and eventually (with co-conspirator, David Herold) crossed the Potomac from Maryland into Virginia, eluding capture for a total of eleven days before federal troops finally discovered him to be hiding on a farm belonging to Richard Garrett and surrounded the barn in which he and Herold were sleeping. The two men were ordered to surrender: Herold complied, but when Booth failed to drop his weapon and come out, the barn was set ablaze. A trooper named Boston Corbett, who was watching Booth through a gap in the barn's siding, shot the assassin. Whether Corbett can be said to have "assassinated" Booth is problematic — the deeply religious Corbett sometimes claimed that he had shot Booth because "Providence directed" him to do it or because he "did not want Booth to be roasted alive," but he also testified that he shot Booth because he "saw [Booth] in the act of stooping or springing and concluded he was going to use his weapons."

Oswald left the warehouse from which he shot Kennedy and was arrested in a movie theater a little over an hour later by police officers who had no idea who he was. (Oswald was initially arrested only for the murder of Dallas police officer J.D. Tippit, whom he shot while in flight; his connection to the Kennedy assassination was not established until later.) Oswald was captured alive and remained in custody for two days before being gunned down by Jack Ruby, a private citizen.

Other differences: Booth was shot in the back in the neck and lived for another three hours; Oswald was shot in the abdomen and died within minutes of his arrival at Parkland Hospital.

A month before Lincoln was assassinated he was in Monroe, Maryland.
A month before Kennedy was assassinated he was with Marilyn Monroe.


Marilyn Monroe died well over a year before Kennedy's assassination.

cryptical
Posts: 1144
Joined: Mon Oct 01, 2012 9:21 pm
Location: Birmingham

Re: Irritating shit on facebook

Post by cryptical » Thu Jan 17, 2013 2:35 pm

wub wrote:
Agent 47 wrote:Image
Have a history teacher explain this if they can.

Abraham Lincoln was elected to Congress in 1846.

John F. Kennedy was elected to Congress in 1946.

Abraham Lincoln was elected President in 1860.

John F. Kennedy was elected President in 1960.

Both were particularly concerned with civil rights.

Both wives lost a child while living in the White House.

Both Presidents were shot on a Friday.

Both Presidents were shot in the head.

Now it gets really weird.

Lincoln's secretary was named Kennedy.

Kennedy's Secretary was named Lincoln.

Both were assassinated by Southerners.

Both were succeeded by Southerners named Johnson.

Andrew Johnson, who succeeded Lincoln, was born in 1808.

Lyndon Johnson, who succeeded Kennedy, was born in 1908.

John Wilkes Booth, who assassinated Lincoln, was born in 1839.

Lee Harvey Oswald, who assassinated Kennedy, was born in 1939.

Both assassins were known by their three names.

Both names are composed of fifteen letters.

Now hang on to your seat.

Lincoln was shot at the theater named "Ford."

Kennedy was shot in a car called "Lincoln" made by "Ford."

Booth and Oswald were assassinated before their trials.

And here's the "kicker":

A week before Lincoln was shot, he was in Monroe, Maryland.

A week before Kennedy was shot, he was with Marilyn Monroe.

AND...................:

Lincoln was shot in a theater and the assassin ran to a warehouse...

Kennedy was shot from a warehouse and the assassin ran to a theater..
:o lolol
http://www.snopes.com/history/american/ ... ennedy.asp


Let's examine them one at a time:

Abraham Lincoln was elected to Congress in 1846.
John F. Kennedy was elected to Congress in 1946.


This statement is literally true: both Lincoln and Kennedy were first elected to Congress one hundred years apart. Aside from that minor coincidence, however, their political careers bore little resemblance to each other.

Lincoln was an Illinois state legislator who, outside of his election to a single term in the House of Representatives, failed in his every attempt to gain national political office until he was elected President in 1860, including an unsuccessful bid for the Senate in 1854, a unsuccessful bid to become the Republican vice-presidential candidate in 1856, and another unsuccessful bid for a Senate seat in 1858.

Kennedy, on the other hand, enjoyed an unbroken string of political successes at the national level when he entered the political arena after World War II. He was elected to the House of Representatives in 1946, re-elected in 1948, re-elected again in 1950, won a Senate seat in 1952, was re-elected to the Senate in 1958, and was elected President in 1960.

Abraham Lincoln was elected President in 1860.
John F. Kennedy was elected President in 1960.


It's hardly surprising that two men who (as noted above) both achieved their first political successes at the national level a hundred years apart would also ascend to the Presidency a hundred years apart. This "coincidence" is even less surprising when we consider that presidential elections are held only once every four years. Lincoln couldn't possibly have been elected President in 1857 or 1858 or 1859 or 1861 or 1862 or 1863, because no presidential elections were held in those years. Likewise, Kennedy couldn't possibly have been elected
President in the non-election years of 1957, 1958, 1959, 1961, 1962, or 1963. So, even though both men were politically active at the national level during eight-year spans when they might have been elected President, circumstances dictated that the only years during those spans when they both could have been elected were exactly one hundred years apart.

We're supposed to be amazed at minor happenstances such as the two men's being elected exactly one hundred years apart, but we're supposed to think nothing of the numerous non-coincidences: Lincoln was born in 1809; Kennedy was born in 1917. Lincoln died in 1865; Kennedy died in 1963. Lincoln was 56 years old at the time of his death; Kennedy was 46 years old at the time of his death. No striking coincidences or convenient hundred-year differences in any of those facts. Even when we consider that, absent all other factors, the two men had a one in twelve chance of dying in the same month, we find no coincidence there: Lincoln was killed in April; Kennedy was killed in November. Also unmentioned here is the fact that Lincoln was re-elected to a second term as President, but Kennedy was killed before the completion of his first term.

The names Lincoln and Kennedy each contain seven letters.

Surely this is the most trivial of coincidences, especially when once considers that the average length of presidential surnames is 6.64 letters. No mention is made of the fact that the two men's first names contain different numbers of letters, and that Kennedy had a middle name (Fitzgerald) while Lincoln had none.


Both were particularly concerned with civil rights.

Saying that Lincoln and Kennedy were both "particularly concerned with civil rights" is like saying that Woodrow Wilson and Franklin Roosevelt were both "particularly concerned with war," or that Herbert Hoover and Ronald Reagan were both "particularly concerned with economics." Those weren't subjects these men evinced a particular interest in; those were issues they were forced to deal with due to events currently taking place in the U.S. which were beyond their control.

Both wives lost their children while living in the White House.

Another statement that, while literally true, encompasses events that were completely different in circumstance and nature.

All of Lincoln's children were born before he entered the White House, and the Lincolns actually lost two children, not just one (although only one died during Lincoln's tenure as President). Edward Lincoln died of tuberculosis in 1850, just before his fourth birthday, and the Lincolns' eleven-year-old son Willie succumbed to typhoid at the end of their first year in the White House.

The Kennedys, on the other hand, were the rare Presidential couple still young enough to be bearing children after entering the White House, and a premature child born to Mrs. Kennedy in 1963 died two days later.

Other substantial differences not mentioned: The Lincolns had four children, all boys, only one of whom lived past his teens. The Kennedys had three children, two boys and a girl, two of whom survived well into adulthood.

Both Presidents were shot on a Friday.

Another non-surprise. Absent all other factors, the odds were already one in seven that both killings would have occurred on the same day of the week. (Don't even think about writing to tell us that we're wrong and the odds are really one in forty-nine. If you think we're wrong, you don't understand the question.) Add to that the obvious notions that the best chance the average person has to shoot a President is at a public function and that most public functions are held on weekends, and it becomes even more likely that a President would be killed on a Friday, Saturday, or Sunday. (Indeed, an earlier plot by Booth to kidnap Lincoln while the latter was attending a play at the Campbell Hospital was slated for March 17, also a Friday.)

Both were shot in the head.

This "coincidence" is another one which is exceedingly trivial in nature. The only two types of shots which reasonably assure a dead victim are chest shots and head shots, so two assassinations committed by head shots aren't the least bit coincidental — especially when one considers that since both Lincoln and Kennedy were shot from behind and while seated, so their assassins had no other practical choice of target. And the "coincidence" here is even less surprising when we note the substantial differences: Lincoln was killed indoors with a small handgun at point blank range; Kennedy was shot outdoors with a rifle from several hundred feet away.

Lincoln's secretary, Kennedy, warned him not to go to Ford's Theatre.
Kennedy's secretary, Lincoln, warned him not to go to Dallas.


This is one of those coincidences that isn't a coincidence at all; it's simply wrong. John Kennedy did have a secretary named Evelyn Lincoln (who may or may not have warned him about going to Dallas), but one searches in vain to find a Lincoln secretary named Kennedy. (Lincoln's White House secretaries were John G. Nicolay and John Hay.)

The more important point is that since Presidents are frequent recipients of assassination threats, they rarely make any public appearances without somebody's warning them of potential danger. Only on the extemely rare occasions when a tragedy actually occurs do we later take note of the warnings; in all other cases the failed "prophecies" are quickly forgotten. (Lincoln received "an unusual number of letters about plots to kidnap or assassinate him," said to have numbered at least eighty, yet none of those plots were enacted.) Nor does anyone think to mention other attempts at kidnap or assassination that were not preceded by any recorded warnings to the victims. (Lincoln was shot at on at least one other occasion.)

Yes, Lincoln was warned not to go to Ford's Theatre by persons concerned for his safety, just as he had been warned not to visit Richmond a week earlier, and just as he had been warned not to attend his own inauguration in 1861. Obviously, only one of the myriad of warnings he received throughout his four years in office was on the mark. Likewise, Kennedy was warned not to visit San Antonio the day before his trip to Dallas (and undoubtedly before a host of other appearances as well), but only the last warning he allegedly received is considered significant, because it coincidentally happened to come true. As Jeane Dixon and other "psychics" have demonstrated, if you make enough predictions, one of them is eventually bound to come true; the public remembers only that and forgets about all the others failed predictions.

Both were assassinated by Southerners.

A dubious use of the term "Southerner." John Wilkes Booth was undeniably a Southern sympathizer, but he was born in Maryland, which (along with Delaware) was the northernmost of the border slave states and remained part of the Union throughout the Civil War. Additionally, Booth spent a good deal of his life in the North and "thought of himself as a Northerner who understood the South."

Oswald was nominally a Southerner by virtue of his having been born in New Orleans; he spent his youth being shuttled between Lousiana, Texas, and New York before finally joining the Marines. But Oswald's "Southerness" is of no real import, because, unlike Booth, Oswald was not motivated by a regional affiliation.

Both were succeeded by Southerners.

Both Lincoln and Kennedy were "succeeded by Southerners" because both had Southerners as vice-president, another fact hardly surprising considering the historical circumstances of their times. Lincoln was a Northern Republican running for re-election while the country was in the midst of a civil war and needed a Southerner and a Democrat to balance the ticket, hence his choice of Tennessean Andrew Johnson. Kennedy, represented New England and therefore needed a vice-presidential candidate who could appeal to the populous Southern and Western regions, hence his choice of a Southwesterner, Texan Lyndon Johnson.

The identification of Andrew Johnson as a "Southerner" is also a bit problematic here. Although Johnson was born in North Carolina and spent his adult life in Tennessee (both slave states), Johnson was also the only Southern senator who refused to follow his state when it seceded, and he remained loyal to the Union.

Both successors were named Johnson.

Given the high frequency of "Johnson" (literally "son of John") as a surname in both Lincoln's and Kennedy's time, this "coincidence" should be no real surprise to anyone.

Andrew Johnson, who succeeded Lincoln, was born in 1808.
Lyndon Johnson, who succeeded Kennedy, was born in 1908.


Another hundred-year coincidence that is hardly surprising, since nearly all American politicians have attained high office (President or Vice-President) while in the 50-70 age range (and Andrew Johnson and Lyndon Johnson were, obviously, contemporaries of Lincoln and Kennedy, respectively). Perhaps it's time to point out that there's nothing "coincidental" about events merely because they somehow involve the number 100. If we sifted through all the Lincoln/Kennedy data, we could produce multiple instances of events involving the number 17 or 49 or 116, but nobody would consider those "coincidences" because they don't yield nice round numbers that have any significance to us, even though they're all just as "coincidental" as the number 100.

And once again, let's consider all the differences between the two Johnsons, such as that one hailed from North Carolina while the other was from Texas, or that one supported slavery while the other championed civil rights, or that one was never elected President in his own right while the other won the biggest presidential landslide in history, or that one was impeached while the other wasn't, or that one became President at the end of a war while the other became President at the beginning of a war.

John Wilkes Booth was born in 1839.
Lee Harvey Oswald was born in 1939.


Another coincidence that is no coincidence because it's plain wrong: Booth was born in 1838, not 1839.

Both assassins were known by their three names.

Another "coincidence" of dubious veracity. John Wilkes Booth was often billed as "J. Wilkes Booth" or simply "John Wilkes" (primarily to distinguish himself from his father and brother — both named Junius — and his brother Edwin, all three of whom were also actors), and as a prominent actor, his name was already familiar to the general public at the time of Lincoln's assassination. Lee Oswald was generally referred to as "Lee" (not "Lee Harvey") before Kennedy's assassination and was unknown to the general public until his arrest; the common usage of his full name only came about after the assassination because his habitual employment of false names (including several variations on his real name) and his possession of forged identification cards made it difficult for the Dallas police to initially identify him.

Both names are comprised of fifteen letters

Coincidence? Neither their first nor last names have the same number of letters. And why should it be significant that both assassins had the same number of letters in their full names when the same wasn't true of Abraham Lincoln and John Fitzgerald Kennedy, or of Andrew Johnson and Lyndon Baines Johnson?

Once again, perhaps we should focus on the substantive differences between the two men: Booth was born into a prominent family and, like his father, was a well-known, popular, gregarious actor. Oswald was born (and lived most of his life) in near poverty-level circumstances, never knew his father (who died two months before Oswald was born) and was an obscure, moody malcontent who never had any close friends or a steady job. Oswald was married with two children; Booth had neither wife nor offspring. Oswald enlisted in the Marines, but Booth kept a promise to his mother not to join the Confederate army.

Booth ran from the theater and was caught in a warehouse.
Oswald ran from a warehouse and was caught in a theater.


Another "coincidence" that is both inaccurate and superficial.

Booth shot Lincoln in a theatre of the type where live stage shows are held, then fled across state lines before being trapped and killed in a tobacco shed several days later.

Oswald shot Kennedy from (not in) a textbook warehouse, then remained in Dallas and was caught and taken alive in a movie theater a little over an hour later.

Booth and Oswald were assassinated before their trials.

Another superficial similarity with much more significant underlying differences, and a potentially dubious use of the word "assassinated."

After Booth shot Lincoln, he fled the scene and eventually (with co-conspirator, David Herold) crossed the Potomac from Maryland into Virginia, eluding capture for a total of eleven days before federal troops finally discovered him to be hiding on a farm belonging to Richard Garrett and surrounded the barn in which he and Herold were sleeping. The two men were ordered to surrender: Herold complied, but when Booth failed to drop his weapon and come out, the barn was set ablaze. A trooper named Boston Corbett, who was watching Booth through a gap in the barn's siding, shot the assassin. Whether Corbett can be said to have "assassinated" Booth is problematic — the deeply religious Corbett sometimes claimed that he had shot Booth because "Providence directed" him to do it or because he "did not want Booth to be roasted alive," but he also testified that he shot Booth because he "saw [Booth] in the act of stooping or springing and concluded he was going to use his weapons."

Oswald left the warehouse from which he shot Kennedy and was arrested in a movie theater a little over an hour later by police officers who had no idea who he was. (Oswald was initially arrested only for the murder of Dallas police officer J.D. Tippit, whom he shot while in flight; his connection to the Kennedy assassination was not established until later.) Oswald was captured alive and remained in custody for two days before being gunned down by Jack Ruby, a private citizen.

Other differences: Booth was shot in the back in the neck and lived for another three hours; Oswald was shot in the abdomen and died within minutes of his arrival at Parkland Hospital.

A month before Lincoln was assassinated he was in Monroe, Maryland.
A month before Kennedy was assassinated he was with Marilyn Monroe.


Marilyn Monroe died well over a year before Kennedy's assassination.
qft

faultier
Posts: 1230
Joined: Tue Jan 12, 2010 10:11 am

Re: Irritating shit on facebook

Post by faultier » Thu Jan 17, 2013 2:37 pm

just posting the wikipedia article for "confirmation bias" would have been enough imo

cryptical
Posts: 1144
Joined: Mon Oct 01, 2012 9:21 pm
Location: Birmingham

Re: Irritating shit on facebook

Post by cryptical » Thu Jan 17, 2013 2:39 pm

dfaultuzr wrote:just posting the wikipedia article for "confirmation bias" would have been enough imo

Confirmation bias
From Wikipedia, the free encyclopedia
Jump to: navigation, search
It has been suggested that Backfire effect be merged into this article or section. (Discuss) Proposed since September 2012.

Confirmation bias (also called confirmatory bias or myside bias) is a tendency of people to favor information that confirms their beliefs or hypotheses.[Note 1][1] People display this bias when they gather or remember information selectively, or when they interpret it in a biased way. The effect is stronger for emotionally charged issues and for deeply entrenched beliefs. For example, in reading about current political issues, people usually prefer sources that affirm their existing attitudes. They also tend to interpret ambiguous evidence as supporting their existing position. Biased search, interpretation and memory have been invoked to explain attitude polarization (when a disagreement becomes more extreme even though the different parties are exposed to the same evidence), belief perseverance (when beliefs persist after the evidence for them is shown to be false), the irrational primacy effect (a greater reliance on information encountered early in a series) and illusory correlation (when people falsely perceive an association between two events or situations).

A series of experiments in the 1960s suggested that people are biased toward confirming their existing beliefs. Later work re-interpreted these results as a tendency to test ideas in a one-sided way, focusing on one possibility and ignoring alternatives. In certain situations, this tendency can bias people's conclusions. Explanations for the observed biases include wishful thinking and the limited human capacity to process information. Another explanation is that people show confirmation bias because they are weighing up the costs of being wrong, rather than investigating in a neutral, scientific way.

Confirmation biases contribute to overconfidence in personal beliefs and can maintain or strengthen beliefs in the face of contrary evidence. Poor decisions due to these biases have been found in military, political, and organizational contexts.
Contents

1 Types
1.1 Biased search for information
1.2 Biased interpretation
1.3 Biased memory
2 Related effects
2.1 Backfire effect
2.2 Polarization of opinion
2.3 Persistence of discredited beliefs
2.4 Preference for early information
2.5 Illusory association between events
3 History
3.1 Informal observation
3.2 Wason's research on hypothesis-testing
3.3 Klayman and Ha's critique
4 Explanations
5 Consequences
5.1 In finance
5.2 In physical and mental health
5.3 In politics and law
5.4 In the paranormal
5.5 In scientific procedure
5.6 In self-image
6 See also
7 Notes
8 Footnotes
9 References
10 Further reading
11 External links

Types

Confirmation biases are effects in information processing, distinct from the behavioral confirmation effect, also called "self-fulfilling prophecy", in which people's expectations affect their behaviour to make the expectations come true.[2] Some psychologists use "confirmation bias" to refer to any way in which people avoid rejecting a belief, whether in searching for evidence, interpreting it, or recalling it from memory. Others restrict the term to selective collection of evidence.[3][Note 2]
Biased search for information
A drawing of a man sitting on a stool at a writing desk
Confirmation bias has been described as an internal "yes man", echoing back a person's beliefs like Charles Dickens' character Uriah Heep.[4]

Experiments have repeatedly found that people tend to test hypotheses in a one-sided way, by searching for evidence consistent with the hypothesis they hold at a given time.[5][6] Rather than searching through all the relevant evidence, they ask questions that are phrased so that an affirmative answer supports their hypothesis.[7] They look for the consequences that they would expect if their hypothesis were true, rather than what would happen if it were false.[7] For example, someone who is trying to identify a number using yes/no questions and suspects that the number is 3 might ask, "Is it an odd number?" People prefer this sort of question, called a "positive test", even when a negative test such as "Is it an even number?" would yield exactly the same information.[8] However, this does not mean that people seek tests that are guaranteed to give a positive answer. In studies where subjects could select either such pseudo-tests or genuinely diagnostic ones, they favored the genuinely diagnostic.[9][10]

The preference for positive tests is not itself a bias, since positive tests can be highly informative.[11] However, in conjunction with other effects, this strategy can confirm existing beliefs or assumptions, independently of whether they are true.[12] In real-world situations, evidence is often complex and mixed. For example, various contradictory ideas about someone could each be supported by concentrating on one aspect of his or her behavior.[6] Thus any search for evidence in favor of a hypothesis is likely to succeed.[12] One illustration of this is the way the phrasing of a question can significantly change the answer.[6] For example, people who are asked, "Are you happy with your social life?" report greater satisfaction than those asked, "Are you unhappy with your social life?"[13]

Even a small change in the wording of a question can affect how people search through available information, and hence the conclusions they reach. This was shown using a fictional child custody case.[14] Subjects read that Parent A was moderately suitable to be the guardian in multiple ways. Parent B had a mix of salient positive and negative qualities: a close relationship with the child but a job that would take him or her away for long periods. When asked, "Which parent should have custody of the child?" the subjects looked for positive attributes and a majority chose Parent B. However, when the question was, "Which parent should be denied custody of the child?" they looked for negative attributes, but again a majority answered Parent B, implying that Parent A should have custody.[14]

Similar studies have demonstrated how people engage in biased search for information, but also that this phenomenon may be limited by a preference for genuine diagnostic tests, where they are available. In an initial experiment, subjects had to rate another person on the introversion-extroversion personality dimension on the basis of an interview. They chose the interview questions from a given list. When the interviewee was introduced as an introvert, the subjects chose questions that presumed introversion, such as, "What do you find unpleasant about noisy parties?" When the interviewee was described as extroverted, almost all the questions presumed extroversion, such as, "What would you do to liven up a dull party?" These loaded questions gave the interviewees little or no opportunity to falsify the hypothesis about them.[15] However, a later version of the experiment gave the subjects less presumptive questions to choose from, such as, "Do you shy away from social interactions?"[16] Subjects preferred to ask these more diagnostic questions, showing only a weak bias towards positive tests. This pattern, of a main preference for diagnostic tests and a weaker preference for positive tests, has been replicated in other studies.[16]

Another experiment gave subjects a particularly complex rule-discovery task involving moving objects simulated by a computer.[17] Objects on the computer screen followed specific laws, which the subjects had to figure out. They could "fire" objects across the screen to test their hypotheses. Despite making many attempts over a ten hour session, none of the subjects worked out the rules of the system. They typically sought to confirm rather than falsify their hypotheses, and were reluctant to consider alternatives. Even after seeing evidence that objectively refuted their working hypotheses, they frequently continued doing the same tests. Some of the subjects were instructed in proper hypothesis-testing, but these instructions had almost no effect.[17]
Biased interpretation

"Smart people believe weird things because they are skilled at defending beliefs they arrived at for non-smart reasons."
—Michael Shermer[18]

Confirmation biases are not limited to the collection of evidence. Even if two individuals have the same information, the way they interpret it can be biased.

A team at Stanford University ran an experiment with subjects who felt strongly about capital punishment, with half in favor and half against.[19][20] Each of these subjects read descriptions of two studies; a comparison of U.S. states with and without the death penalty, and a comparison of murder rates in a state before and after the introduction of the death penalty. After reading a quick description of each study, the subjects were asked whether their opinions had changed. They then read a much more detailed account of each study's procedure and had to rate how well-conducted and convincing that research was.[19] In fact, the studies were fictional. Half the subjects were told that one kind of study supported the deterrent effect and the other undermined it, while for other subjects the conclusions were swapped.[19][20]

The subjects, whether proponents or opponents, reported shifting their attitudes slightly in the direction of the first study they read. Once they read the more detailed descriptions of the two studies, they almost all returned to their original belief regardless of the evidence provided, pointing to details that supported their viewpoint and disregarding anything contrary. Subjects described studies supporting their pre-existing view as superior to those that contradicted it, in detailed and specific ways.[19][21] Writing about a study that seemed to undermine the deterrence effect, a death penalty proponent wrote, "The research didn't cover a long enough period of time", while an opponent's comment on the same study said, "No strong evidence to contradict the researchers has been presented".[19] The results illustrated that people set higher standards of evidence for hypotheses that go against their current expectations. This effect, known as "disconfirmation bias", has been supported by other experiments.[22]
A large round machine with a hole in the middle, with a platter for a person to lie on so that their head can fit into the hole
An MRI scanner allowed researchers to examine how the human brain deals with unwelcome information.

A study of biased interpretation took place during the 2004 US presidential election and involved subjects who described themselves as having strong feelings about the candidates. They were shown apparently contradictory pairs of statements, either from Republican candidate George W. Bush, Democratic candidate John Kerry or a politically neutral public figure. They were also given further statements that made the apparent contradiction seem reasonable. From these three pieces of information, they had to decide whether or not each individual's statements were inconsistent. There were strong differences in these evaluations, with subjects much more likely to interpret statements by the candidate they opposed as contradictory.[23]

In this experiment, the subjects made their judgments while in a magnetic resonance imaging (MRI) scanner which monitored their brain activity. As subjects evaluated contradictory statements by their favored candidate, emotional centers of their brains were aroused. This did not happen with the statements by the other figures. The experimenters inferred that the different responses to the statements were not due to passive reasoning errors. Instead, the subjects were actively reducing the cognitive dissonance induced by reading about their favored candidate's irrational or hypocritical behavior.[23]

Biased interpretation is not restricted to emotionally significant topics. In another experiment, subjects were told a story about a theft. They had to rate the evidential importance of statements arguing either for or against a particular character being responsible. When they hypothesized that character's guilt, they rated statements supporting that hypothesis as more important than conflicting statements.[24]
Biased memory

Even if someone has sought and interpreted evidence in a neutral manner, they may still remember it selectively to reinforce their expectations. This effect is called "selective recall", "confirmatory memory" or "access-biased memory".[25] Psychological theories differ in their predictions about selective recall. Schema theory predicts that information matching prior expectations will be more easily stored and recalled.[26] Some alternative approaches say that surprising information stands out more and so is more memorable.[26] Predictions from both these theories have been confirmed in different experimental contexts, with no theory winning outright.[27]

In one study, subjects read a profile of a woman which described a mix of introverted and extroverted behaviors.[28] They later had to recall examples of her introversion and extroversion. One group was told this was to assess the woman for a job as a librarian, while a second group were told it was for a job in real estate sales. There was a significant difference between what these two groups recalled, with the "librarian" group recalling more examples of introversion and the "sales" groups recalling more extroverted behavior.[28] A selective memory effect has also been shown in experiments that manipulate the desirability of personality types.[26][29] In one of these, a group of subjects were shown evidence that extroverted people are more successful than introverts. Another group were told the opposite. In a subsequent, apparently unrelated, study, they were asked to recall events from their lives in which they had been either introverted or extroverted. Each group of subjects provided more memories connecting themselves with the more desirable personality type, and recalled those memories more quickly.[30]

One study showed how selective memory can maintain belief in extrasensory perception (ESP).[31] Believers and disbelievers were each shown descriptions of ESP experiments. Half of each group were told that the experimental results supported the existence of ESP, while the others were told they did not. In a subsequent test, subjects recalled the material accurately, apart from believers who had read the non-supportive evidence. This group remembered significantly less information and some of them incorrectly remembered the results as supporting ESP.[31]
Related effects
Backfire effect

A similar cognitive bias found in individuals is the Backfire effect. Here, individuals challenged with evidence contradictory to their beliefs tend to reject the evidence and instead become an even firmer supporter of the initial belief.[32][33] The phrase was first coined by Brendan Nyhan and Jason Reifler in a paper entitled "When Corrections Fail: The persistence of political misperceptions".[34]
Polarization of opinion
Main article: Attitude polarization

When people with opposing views interpret new information in a biased way, their views can move even further apart. This is called "attitude polarization".[35] The effect was demonstrated by an experiment that involved drawing a series of red and black balls from one of two concealed "bingo baskets". Subjects knew that one basket contained 60% black and 40% red balls; the other, 40% black and 60% red. The experimenters looked at what happened when balls of alternating color were drawn in turn, a sequence that does not favor either basket. After each ball was drawn, subjects in one group were asked to state out loud their judgments of the probability that the balls were being drawn from one or the other basket. These subjects tended to grow more confident with each successive draw—whether they initially thought the basket with 60% black balls or the one with 60% red balls was the more likely source, their estimate of the probability increased. Another group of subjects were asked to state probability estimates only at the end of a sequence of drawn balls, rather than after each ball. They did not show the polarization effect, suggesting that it does not necessarily occur when people simply hold opposing positions, but rather when they openly commit to them.[36]
A collection of eight different handguns resting on the ground
Strong opinions on an issue such as gun ownership can bias how someone interprets new evidence.

A less abstract study was the Stanford biased interpretation experiment in which subjects with strong opinions about the death penalty read about mixed experimental evidence. Twenty-three percent of the subjects reported that their views had become more extreme, and this self-reported shift correlated strongly with their initial attitudes.[19] In later experiments, subjects also reported their opinions becoming more extreme in response to ambiguous information. However, comparisons of their attitudes before and after the new evidence showed no significant change, suggesting that the self-reported changes might not be real.[22][35][37] Based on these experiments, Deanna Kuhn and Joseph Lao concluded that polarization is a real phenomenon but far from inevitable, only happening in a small minority of cases. They found that it was prompted not only by considering mixed evidence, but by merely thinking about the topic.[35]

Charles Taber and Milton Lodge argued that the Stanford team's result had been hard to replicate because the arguments used in later experiments were too abstract or confusing to evoke an emotional response. The Taber and Lodge study used the emotionally charged topics of gun control and affirmative action.[22] They measured the attitudes of their subjects towards these issues before and after reading arguments on each side of the debate. Two groups of subjects showed attitude polarization; those with strong prior opinions and those who were politically knowledgeable. In part of this study, subjects chose which information sources to read, from a list prepared by the experimenters. For example they could read the National Rifle Association's and the Brady Anti-Handgun Coalition's arguments on gun control. Even when instructed to be even-handed, subjects were more likely to read arguments that supported their existing attitudes. This biased search for information correlated well with the polarization effect.[22]
Persistence of discredited beliefs

"eliefs can survive potent logical or empirical challenges. They can survive and even be bolstered by evidence that most uncommitted observers would agree logically demands some weakening of such beliefs. They can even survive the total destruction of their original evidential bases."
—Lee Ross and Craig Anderson[38]

Confirmation biases can be used to explain why some beliefs remain when the initial evidence for them is removed.[39] This belief perseverance effect has been shown by a series of experiments using what is called the "debriefing paradigm": subjects read fake evidence for a hypothesis, their attitude change is measured, then the fakery is exposed in detail. Their attitudes are then measured once more to see if their belief returns to its previous level.[38]

A typical finding is that at least some of the initial belief remains even after a full debrief.[40] In one experiment, subjects had to distinguish between real and fake suicide notes. The feedback was random: some were told they had done well while others were told they had performed badly. Even after being fully debriefed, subjects were still influenced by the feedback. They still thought they were better or worse than average at that kind of task, depending on what they had initially been told.[41]

In another study, subjects read job performance ratings of two firefighters, along with their responses to a risk aversion test.[38] These fictional data were arranged to show either a negative or positive association: some subjects were told that a risk-taking firefighter did better, while others were told they did less well than a risk-averse colleague.[42] Even if these two case studies had been true, they would have been scientifically poor evidence for a conclusion about firefighters in general. However, the subjects found them subjectively persuasive.[42] When the case studies were shown to be fictional, subjects' belief in a link diminished, but around half of the original effect remained.[38] Follow-up interviews established that the subjects had understood the debriefing and taken it seriously. Subjects seemed to trust the debriefing, but regarded the discredited information as irrelevant to their personal belief.[42]
Preference for early information

Experiments have shown that information is weighted more strongly when it appears early in a series, even when the order is unimportant. For example, people form a more positive impression of someone described as "intelligent, industrious, impulsive, critical, stubborn, envious" than when they are given the same words in reverse order.[43] This irrational primacy effect is independent of the primacy effect in memory in which the earlier items in a series leave a stronger memory trace.[43] Biased interpretation offers an explanation for this effect: seeing the initial evidence, people form a working hypothesis that affects how they interpret the rest of the information.[39]

One demonstration of irrational primacy involved colored chips supposedly drawn from two urns. Subjects were told the color distributions of the urns, and had to estimate the probability of a chip being drawn from one of them.[43] In fact, the colors appeared in a pre-arranged order. The first thirty draws favored one urn and the next thirty favored the other.[39] The series as a whole was neutral, so rationally, the two urns were equally likely. However, after sixty draws, subjects favored the urn suggested by the initial thirty.[43]

Another experiment involved a slide show of a single object, seen as just a blur at first and in slightly better focus with each succeeding slide.[43] After each slide, subjects had to state their best guess of what the object was. Subjects whose early guesses were wrong persisted with those guesses, even when the picture was sufficiently in focus that other people could readily identify the object.[39]
Illusory association between events
Main article: Illusory correlation

Illusory correlation is the tendency to see non-existent correlations in a set of data.[44] This tendency was first demonstrated in a series of experiments in the late 1960s.[45] In one experiment, subjects read a set of psychiatric case studies, including responses to the Rorschach inkblot test. They reported that the homosexual men in the set were more likely to report seeing buttocks, anuses or sexually ambiguous figures in the inkblots. In fact the case studies were fictional and, in one version of the experiment, had been constructed so that the homosexual men were less likely to report this imagery.[44] In a survey, a group of experienced psychoanalysts reported the same set of illusory associations with homosexuality.[44][45]

Another study recorded the symptoms experienced by arthritic patients, along with weather conditions over a 15-month period. Nearly all the patients reported that their pains were correlated with weather conditions, although the real correlation was zero.[46]

This effect is a kind of biased interpretation, in that objectively neutral or unfavorable evidence is interpreted to support existing beliefs. It is also related to biases in hypothesis-testing behavior.[47] In judging whether two events, such as illness and bad weather, are correlated, people rely heavily on the number of positive-positive cases: in this example, instances of both pain and bad weather. They pay relatively little attention to the other kinds of observation (of no pain and/or good weather).[48] This parallels the reliance on positive tests in hypothesis testing.[47] It may also reflect selective recall, in that people may have a sense that two events are correlated because it is easier to recall times when they happened together.[47]
Example Days Rain No rain
Arthritis 14 6
No arthritis 7 2

In the above fictional example, arthritic symptoms are more likely on days with no rain. However, people are likely to focus on the relatively large number of days which have both rain and symptoms. By concentrating on one cell of the table rather than all four, people can misperceive the relationship, in this case associating rain with arthritic symptoms.[49]
History
Engraved head-and-shoulders portrait of Francis Bacon wearing a hat and ruff.
Francis Bacon
Informal observation

Before psychological research on confirmation bias, the phenomenon had been observed anecdotally by writers, including the Greek historian Thucydides (c. 460 BC – c. 395 BC), Italian poet Dante Alighieri (1265–1321), English philosopher and scientist Francis Bacon (1561–1626),[50] and Russian author Leo Tolstoy (1828–1910). Thucydides, in the History of the Peloponnesian War wrote, "it is a habit of mankind ... to use sovereign reason to thrust aside what they do not fancy."[51] In the Divine Comedy, St. Thomas Aquinas cautions Dante when they meet in Paradise, "opinion—hasty—often can incline to the wrong side, and then affection for one's own opinion binds, confines the mind."[52] Bacon, in the Novum Organum, wrote,

The human understanding when it has once adopted an opinion ... draws all things else to support and agree with it. And though there be a greater number and weight of instances to be found on the other side, yet these it either neglects or despises, or else by some distinction sets aside or rejects[.][53]

Bacon said that biased assessment of evidence drove "all superstitions, whether in astrology, dreams, omens, divine judgments or the like".[53] In his essay "What Is Art?", Tolstoy wrote,

I know that most men—not only those considered clever, but even those who are very clever, and capable of understanding most difficult scientific, mathematical, or philosophic problems—can very seldom discern even the simplest and most obvious truth if it be such as to oblige them to admit the falsity of conclusions they have formed, perhaps with much difficulty—conclusions of which they are proud, which they have taught to others, and on which they have built their lives.[54]

Wason's research on hypothesis-testing

The term "confirmation bias" was coined by English psychologist Peter Wason.[55] For an experiment published in 1960, he challenged subjects to identify a rule applying to triples of numbers. At the outset, they were told that (2,4,6) fits the rule. Subjects could generate their own triples and the experimenter told them whether or not each triple conformed to the rule.[56][57]

While the actual rule was simply "any ascending sequence", the subjects had a great deal of difficulty in finding it, often announcing rules that were far more specific, such as "the middle number is the average of the first and last".[56] The subjects seemed to test only positive examples—triples that obeyed their hypothesized rule. For example, if they thought the rule was, "Each number is two greater than its predecessor", they would offer a triple that fit this rule, such as (11,13,15) rather than a triple that violates it, such as (11,12,19).[58]

Wason accepted falsificationism, according to which a scientific test of a hypothesis is a serious attempt to falsify it. He interpreted his results as showing a preference for confirmation over falsification, hence the term "confirmation bias".[Note 3][59] Wason also used confirmation bias to explain the results of his selection task experiment.[60] In this task, subjects are given partial information about a set of objects, and have to specify what further information they would need to tell whether or not a conditional rule ("If A, then B") applies. It has been found repeatedly that people perform badly on various forms of this test, in most cases ignoring information that could potentially refute the rule.[61][62]
Klayman and Ha's critique

A 1987 paper by Joshua Klayman and Young-Won Ha argued that the Wason experiments had not actually demonstrated a bias towards confirmation. Instead, Klayman and Ha interpreted the results in terms of a tendency to make tests that are consistent with the working hypothesis.[63] They called this the "positive test strategy".[6] This strategy is an example of a heuristic: a reasoning shortcut that is imperfect but easy to compute.[1] Klayman and Ha used Bayesian probability and information theory as their standard of hypothesis-testing, rather than the falsificationism used by Wason. According to these ideas, each answer to a question yields a different amount of information, which depends on the person's prior beliefs. Thus a scientific test of a hypothesis is one that is expected to produce the most information. Since the information content depends on initial probabilities, a positive test can either be highly informative or uninformative. Klayman and Ha argued that when people think about realistic problems, they are looking for a specific answer with a small initial probability. In this case, positive tests are usually more informative than negative tests.[11] However, in Wason's rule discovery task the answer—three numbers in ascending order—is very broad, so positive tests are unlikely to yield informative answers. Klayman and Ha supported their analysis by citing an experiment that used the labels "DAX" and "MED" in place of "fits the rule" and "doesn't fit the rule". This avoided implying that the aim was to find a low-probability rule. Subjects had much more success with this version of the experiment.[64][65]
Within the universe of all possible triples, those that fit the true rule are shown schematically as a circle. The hypothesized rule is a smaller circle enclosed within it.
If the true rule (T) encompasses the current hypothesis (H), then positive tests (examining an H to see if it is T) will not show that the hypothesis is false.

Two overlapping circles represent the true rule and the hypothesized rule. Any observation falling in the non-overlapping parts of the circles shows that the two rules are not exactly the same. In other words, those observations falsify the hypothesis.
If the true rule (T) overlaps the current hypothesis (H), then either a negative test or a positive test can potentially falsify H.

The triples fitting the hypothesis are represented as a circle within the universe of all triples. The true rule is a smaller circle within this.
When the working hypothesis (H) includes the true rule (T) then positive tests are the only way to falsify H.

In light of this and other critiques, the focus of research moved away from confirmation versus falsification to examine whether people test hypotheses in an informative way, or an uninformative but positive way. The search for "true" confirmation bias led psychologists to look at a wider range of effects in how people process information.[66]
Explanations

Confirmation bias is often described as a result of automatic, unintentional strategies rather than deliberate deception.[12][67] According to Robert Maccoun, most biased evidence processing occurs through a combination of both "cold" (cognitive) and "hot" (motivated) mechanisms.[68]

Cognitive explanations for confirmation bias are based on limitations in people's ability to handle complex tasks, and the shortcuts, called heuristics, that they use.[69] For example, people may judge the reliability of evidence by using the availability heuristic, i.e. how readily a particular idea comes to mind.[70] It is also possible that people can only focus on one thought at a time, so find it difficult to test alternative hypotheses in parallel.[71] Another heuristic is the positive test strategy identified by Klayman and Ha, in which people test a hypothesis by examining cases where they expect a property or event to occur. This heuristic avoids the difficult or impossible task of working out how diagnostic each possible question will be. However, it is not universally reliable, so people can overlook challenges to their existing beliefs.[11][72]

Motivational explanations involve an effect of desire on belief, sometimes called "wishful thinking".[73][74] It is known that people prefer pleasant thoughts over unpleasant ones in a number of ways: this is called the "Pollyanna principle".[75] Applied to arguments or sources of evidence, this could explain why desired conclusions are more likely to be believed true.[73] According to experiments that manipulate the desirability of the conclusion, people demand a high standard of evidence for unpalatable ideas and a low standard for preferred ideas. In other words, they ask, "Can I believe this?" for some suggestions and, "Must I believe this?" for others.[76][77] Although consistency is a desirable feature of attitudes, an excessive drive for consistency is another potential source of bias because it may prevent people from neutrally evaluating new, surprising information.[73] Social psychologist Ziva Kunda combines the cognitive and motivational theories, arguing that motivation creates the bias, but cognitive factors determine the size of the effect.[78]

Explanations in terms of cost-benefit analysis assume that people do not just test hypotheses in a disinterested way, but assess the costs of different errors.[79] Using ideas from evolutionary psychology, James Friedrich suggests that people do not primarily aim at truth in testing hypotheses, but try to avoid the most costly errors. For example, employers might ask one-sided questions in job interviews because they are focused on weeding out unsuitable candidates.[80] Yaacov Trope and Akiva Liberman's refinement of this theory assumes that people compare the two different kinds of error: accepting a false hypothesis or rejecting a true hypothesis. For instance, someone who underestimates a friend's honesty might treat him or her suspiciously and so undermine the friendship. Overestimating the friend's honesty may also be costly, but less so. In this case, it would be rational to seek, evaluate or remember evidence of their honesty in a biased way.[81] When someone gives an initial impression of being introverted or extroverted, questions that match that impression come across as more empathic.[82] This suggests that when talking to someone who seems to be an introvert, it is a sign of better social skills to ask, "Do you feel awkward in social situations?" rather than, "Do you like noisy parties?" The connection between confirmation bias and social skills was corroborated by a study of how college students get to know other people. Highly self-monitoring students, who are more sensitive to their environment and to social norms, asked more matching questions when interviewing a high-status staff member than when getting to know fellow students.[82]

Psychologists Jennifer Lerner and Philip Tetlock distinguish two different kinds of thinking process. Exploratory thought neutrally considers multiple points of view and tries to anticipate all possible objections to a particular position, while confirmatory thought seeks to justify a specific point of view. Lerner and Tetlock say that when people expect to need to justify their position to other people, whose views they already know, they will tend to adopt a similar position to those people, and then use confirmatory thought to bolster their own credibility. However, if the external parties are overly aggressive or critical, people will disengage from thought altogether, and simply assert their personal opinions without justification.[83] Lerner and Tetlock say that people only push themselves to think critically and logically when they know in advance they will need to explain themselves to others who are well-informed, genuinely interested in the truth, and whose views they don't already know.[84] Because those conditions rarely exist, they argue, most people are using confirmatory thought most of the time.[85]
Consequences
In finance

Confirmation bias can lead investors to be overconfident, ignoring evidence that their strategies will lose money.[4][86] In studies of political stock markets, investors made more profit when they resisted bias. For example, participants who interpreted a candidate's debate performance in a neutral rather than partisan way were more likely to profit.[87] To combat the effect of confirmation bias, investors can try to adopt a contrary viewpoint "for the sake of argument".[88] In one technique, they imagine that their investments have collapsed and ask themselves why this might happen.[4]
In physical and mental health

Raymond Nickerson, a psychologist, blames confirmation bias for the ineffective medical procedures that were used for centuries before the arrival of scientific medicine.[89] If a patient recovered, medical authorities counted the treatment as successful, rather than looking for alternative explanations such as that the disease had run its natural course.[89] Biased assimilation is a factor in the modern appeal of alternative medicine, whose proponents are swayed by positive anecdotal evidence but treat scientific evidence hyper-critically.[90][91][92]

Cognitive therapy was developed by Aaron T. Beck in the early 1960s and has become a popular approach.[93] According to Beck, biased information processing is a factor in depression.[94] His approach teaches people to treat evidence impartially, rather than selectively reinforcing negative outlooks.[50] Phobias and hypochondria have also been shown to involve confirmation bias for threatening information.[95]
In politics and law
A woman and a man reading a document in a courtroom
Mock trials allow researchers to examine confirmation biases in a realistic setting.

Nickerson argues that reasoning in judicial and political contexts is sometimes subconsciously biased, favoring conclusions that judges, juries or governments have already committed to.[96] Since the evidence in a jury trial can be complex, and jurors often reach decisions about the verdict early on, it is reasonable to expect an attitude polarization effect. The prediction that jurors will become more extreme in their views as they see more evidence has been borne out in experiments with mock trials.[97][98] Both inquisitorial and adversarial criminal justice systems are affected by confirmation bias.[99]

Confirmation bias can be a factor in creating or extending conflicts, from emotionally charged debates to wars: by interpreting the evidence in their favor, each opposing party can become overconfident that it is in the stronger position.[100] On the other hand, confirmation bias can result in people ignoring or misinterpreting the signs of an imminent or incipient conflict. For example, psychologists Stuart Sutherland and Thomas Kida have each argued that US Admiral Husband E. Kimmel showed confirmation bias when playing down the first signs of the Japanese attack on Pearl Harbor.[61][101]

A two-decade study of political pundits by Philip E. Tetlock found that, on the whole, their predictions were not much better than chance. Tetlock divided experts into "foxes" who maintained multiple hypotheses, and "hedgehogs" who were more dogmatic. In general, the hedgehogs were much less accurate. Tetlock blamed their failure on confirmation bias—specifically, their inability to make use of new information that contradicted their existing theories.[102]
In the paranormal

One factor in the appeal of psychic "readings" is that listeners apply a confirmation bias which fits the psychic's statements to their own lives.[103] By making a large number of ambiguous statements in each sitting, the psychic gives the client more opportunities to find a match. This is one of the techniques of cold reading, with which a psychic can deliver a subjectively impressive reading without any prior information about the client.[103] Investigator James Randi compared the transcript of a reading to the client's report of what the psychic had said, and found that the client showed a strong selective recall of the "hits".[104]

As a striking illustration of confirmation bias in the real world, Nickerson mentions numerological pyramidology: the practice of finding meaning in the proportions of the Egyptian pyramids.[105] There are many different length measurements that can be made of, for example, the Great Pyramid of Giza and many ways to combine or manipulate them. Hence it is almost inevitable that people who look at these numbers selectively will find superficially impressive correspondences, for example with the dimensions of the Earth.[105]
In scientific procedure

A distinguishing feature of scientific thinking is the search for falsifying as well as confirming evidence.[106] However, many times in the history of science, scientists have resisted new discoveries by selectively interpreting or ignoring unfavorable data.[106] Previous research has shown that the assessment of the quality of scientific studies seems to be particularly vulnerable to confirmation bias. It has been found several times that scientists rate studies that report findings consistent with their prior beliefs more favorably than studies reporting findings inconsistent with their previous beliefs.[67][107][108] However, assuming that the research question is relevant, the experimental design adequate and the data are clearly and comprehensively described, the found results should be of importance to the scientific community and should not be viewed prejudicially, regardless of whether they conform to current theoretical predictions.[108]

Confirmation bias may thus be especially harmful to objective evaluations regarding nonconforming results since biased individuals may regard opposing evidence to be weak in principle and give little serious thought to revising their beliefs.[107] Scientific innovators often meet with resistance from the scientific community, and research presenting controversial results frequently receives harsh peer review.[109]

In the context of scientific research, confirmation biases can sustain theories or research programs in the face of inadequate or even contradictory evidence;[61][110] the field of parapsychology has been particularly affected.[111]

An experimenter's confirmation bias can potentially affect which data are reported. Data that conflict with the experimenter's expectations may be more readily discarded as unreliable, producing the so-called file drawer effect. To combat this tendency, scientific training teaches ways to prevent bias.[112] For example, experimental design of randomized controlled trials (coupled with their systematic review) aims to minimize sources of bias.[112][113] The social process of peer review is thought to mitigate the effect of individual scientists' biases,[114] even though the peer review process itself may be susceptible to such biases.[108][115]
In self-image

Social psychologists have identified two tendencies in the way people seek or interpret information about themselves. Self-verification is the drive to reinforce the existing self-image and self-enhancement is the drive to seek positive feedback. Both are served by confirmation biases.[116] In experiments where people are given feedback that conflicts with their self-image, they are less likely to attend to it or remember it than when given self-verifying feedback.[117][118][119] They reduce the impact of such information by interpreting it as unreliable.[117][120][121] Similar experiments have found a preference for positive feedback, and the people who give it, over negative feedback.[116]
See also
Portal icon Psychology portal
Portal icon Thinking portal

Backfire effect
Cherry picking (fallacy)
Cognitive bias mitigation
Cognitive inertia
Denial
Denialism
Experimenter's bias
Filter bubble
Hostile media effect
List of biases in judgment and decision making
List of memory biases
Observer-expectancy effect
Reporting bias
Selective exposure theory
Semmelweis reflex
Woozle effect

wub
Posts: 34156
Joined: Tue Feb 26, 2008 3:11 pm
Location: Madrid
Contact:

Re: Irritating shit on facebook

Post by wub » Thu Jan 17, 2013 2:41 pm

cryptical wrote:
dfaultuzr wrote:just posting the wikipedia article for "confirmation bias" would have been enough imo

Confirmation bias
From Wikipedia, the free encyclopedia
Jump to: navigation, search
It has been suggested that Backfire effect be merged into this article or section. (Discuss) Proposed since September 2012.

Confirmation bias (also called confirmatory bias or myside bias) is a tendency of people to favor information that confirms their beliefs or hypotheses.[Note 1][1] People display this bias when they gather or remember information selectively, or when they interpret it in a biased way. The effect is stronger for emotionally charged issues and for deeply entrenched beliefs. For example, in reading about current political issues, people usually prefer sources that affirm their existing attitudes. They also tend to interpret ambiguous evidence as supporting their existing position. Biased search, interpretation and memory have been invoked to explain attitude polarization (when a disagreement becomes more extreme even though the different parties are exposed to the same evidence), belief perseverance (when beliefs persist after the evidence for them is shown to be false), the irrational primacy effect (a greater reliance on information encountered early in a series) and illusory correlation (when people falsely perceive an association between two events or situations).

A series of experiments in the 1960s suggested that people are biased toward confirming their existing beliefs. Later work re-interpreted these results as a tendency to test ideas in a one-sided way, focusing on one possibility and ignoring alternatives. In certain situations, this tendency can bias people's conclusions. Explanations for the observed biases include wishful thinking and the limited human capacity to process information. Another explanation is that people show confirmation bias because they are weighing up the costs of being wrong, rather than investigating in a neutral, scientific way.

Confirmation biases contribute to overconfidence in personal beliefs and can maintain or strengthen beliefs in the face of contrary evidence. Poor decisions due to these biases have been found in military, political, and organizational contexts.
Contents

1 Types
1.1 Biased search for information
1.2 Biased interpretation
1.3 Biased memory
2 Related effects
2.1 Backfire effect
2.2 Polarization of opinion
2.3 Persistence of discredited beliefs
2.4 Preference for early information
2.5 Illusory association between events
3 History
3.1 Informal observation
3.2 Wason's research on hypothesis-testing
3.3 Klayman and Ha's critique
4 Explanations
5 Consequences
5.1 In finance
5.2 In physical and mental health
5.3 In politics and law
5.4 In the paranormal
5.5 In scientific procedure
5.6 In self-image
6 See also
7 Notes
8 Footnotes
9 References
10 Further reading
11 External links

Types

Confirmation biases are effects in information processing, distinct from the behavioral confirmation effect, also called "self-fulfilling prophecy", in which people's expectations affect their behaviour to make the expectations come true.[2] Some psychologists use "confirmation bias" to refer to any way in which people avoid rejecting a belief, whether in searching for evidence, interpreting it, or recalling it from memory. Others restrict the term to selective collection of evidence.[3][Note 2]
Biased search for information
A drawing of a man sitting on a stool at a writing desk
Confirmation bias has been described as an internal "yes man", echoing back a person's beliefs like Charles Dickens' character Uriah Heep.[4]

Experiments have repeatedly found that people tend to test hypotheses in a one-sided way, by searching for evidence consistent with the hypothesis they hold at a given time.[5][6] Rather than searching through all the relevant evidence, they ask questions that are phrased so that an affirmative answer supports their hypothesis.[7] They look for the consequences that they would expect if their hypothesis were true, rather than what would happen if it were false.[7] For example, someone who is trying to identify a number using yes/no questions and suspects that the number is 3 might ask, "Is it an odd number?" People prefer this sort of question, called a "positive test", even when a negative test such as "Is it an even number?" would yield exactly the same information.[8] However, this does not mean that people seek tests that are guaranteed to give a positive answer. In studies where subjects could select either such pseudo-tests or genuinely diagnostic ones, they favored the genuinely diagnostic.[9][10]

The preference for positive tests is not itself a bias, since positive tests can be highly informative.[11] However, in conjunction with other effects, this strategy can confirm existing beliefs or assumptions, independently of whether they are true.[12] In real-world situations, evidence is often complex and mixed. For example, various contradictory ideas about someone could each be supported by concentrating on one aspect of his or her behavior.[6] Thus any search for evidence in favor of a hypothesis is likely to succeed.[12] One illustration of this is the way the phrasing of a question can significantly change the answer.[6] For example, people who are asked, "Are you happy with your social life?" report greater satisfaction than those asked, "Are you unhappy with your social life?"[13]

Even a small change in the wording of a question can affect how people search through available information, and hence the conclusions they reach. This was shown using a fictional child custody case.[14] Subjects read that Parent A was moderately suitable to be the guardian in multiple ways. Parent B had a mix of salient positive and negative qualities: a close relationship with the child but a job that would take him or her away for long periods. When asked, "Which parent should have custody of the child?" the subjects looked for positive attributes and a majority chose Parent B. However, when the question was, "Which parent should be denied custody of the child?" they looked for negative attributes, but again a majority answered Parent B, implying that Parent A should have custody.[14]

Similar studies have demonstrated how people engage in biased search for information, but also that this phenomenon may be limited by a preference for genuine diagnostic tests, where they are available. In an initial experiment, subjects had to rate another person on the introversion-extroversion personality dimension on the basis of an interview. They chose the interview questions from a given list. When the interviewee was introduced as an introvert, the subjects chose questions that presumed introversion, such as, "What do you find unpleasant about noisy parties?" When the interviewee was described as extroverted, almost all the questions presumed extroversion, such as, "What would you do to liven up a dull party?" These loaded questions gave the interviewees little or no opportunity to falsify the hypothesis about them.[15] However, a later version of the experiment gave the subjects less presumptive questions to choose from, such as, "Do you shy away from social interactions?"[16] Subjects preferred to ask these more diagnostic questions, showing only a weak bias towards positive tests. This pattern, of a main preference for diagnostic tests and a weaker preference for positive tests, has been replicated in other studies.[16]

Another experiment gave subjects a particularly complex rule-discovery task involving moving objects simulated by a computer.[17] Objects on the computer screen followed specific laws, which the subjects had to figure out. They could "fire" objects across the screen to test their hypotheses. Despite making many attempts over a ten hour session, none of the subjects worked out the rules of the system. They typically sought to confirm rather than falsify their hypotheses, and were reluctant to consider alternatives. Even after seeing evidence that objectively refuted their working hypotheses, they frequently continued doing the same tests. Some of the subjects were instructed in proper hypothesis-testing, but these instructions had almost no effect.[17]
Biased interpretation

"Smart people believe weird things because they are skilled at defending beliefs they arrived at for non-smart reasons."
—Michael Shermer[18]

Confirmation biases are not limited to the collection of evidence. Even if two individuals have the same information, the way they interpret it can be biased.

A team at Stanford University ran an experiment with subjects who felt strongly about capital punishment, with half in favor and half against.[19][20] Each of these subjects read descriptions of two studies; a comparison of U.S. states with and without the death penalty, and a comparison of murder rates in a state before and after the introduction of the death penalty. After reading a quick description of each study, the subjects were asked whether their opinions had changed. They then read a much more detailed account of each study's procedure and had to rate how well-conducted and convincing that research was.[19] In fact, the studies were fictional. Half the subjects were told that one kind of study supported the deterrent effect and the other undermined it, while for other subjects the conclusions were swapped.[19][20]

The subjects, whether proponents or opponents, reported shifting their attitudes slightly in the direction of the first study they read. Once they read the more detailed descriptions of the two studies, they almost all returned to their original belief regardless of the evidence provided, pointing to details that supported their viewpoint and disregarding anything contrary. Subjects described studies supporting their pre-existing view as superior to those that contradicted it, in detailed and specific ways.[19][21] Writing about a study that seemed to undermine the deterrence effect, a death penalty proponent wrote, "The research didn't cover a long enough period of time", while an opponent's comment on the same study said, "No strong evidence to contradict the researchers has been presented".[19] The results illustrated that people set higher standards of evidence for hypotheses that go against their current expectations. This effect, known as "disconfirmation bias", has been supported by other experiments.[22]
A large round machine with a hole in the middle, with a platter for a person to lie on so that their head can fit into the hole
An MRI scanner allowed researchers to examine how the human brain deals with unwelcome information.

A study of biased interpretation took place during the 2004 US presidential election and involved subjects who described themselves as having strong feelings about the candidates. They were shown apparently contradictory pairs of statements, either from Republican candidate George W. Bush, Democratic candidate John Kerry or a politically neutral public figure. They were also given further statements that made the apparent contradiction seem reasonable. From these three pieces of information, they had to decide whether or not each individual's statements were inconsistent. There were strong differences in these evaluations, with subjects much more likely to interpret statements by the candidate they opposed as contradictory.[23]

In this experiment, the subjects made their judgments while in a magnetic resonance imaging (MRI) scanner which monitored their brain activity. As subjects evaluated contradictory statements by their favored candidate, emotional centers of their brains were aroused. This did not happen with the statements by the other figures. The experimenters inferred that the different responses to the statements were not due to passive reasoning errors. Instead, the subjects were actively reducing the cognitive dissonance induced by reading about their favored candidate's irrational or hypocritical behavior.[23]

Biased interpretation is not restricted to emotionally significant topics. In another experiment, subjects were told a story about a theft. They had to rate the evidential importance of statements arguing either for or against a particular character being responsible. When they hypothesized that character's guilt, they rated statements supporting that hypothesis as more important than conflicting statements.[24]
Biased memory

Even if someone has sought and interpreted evidence in a neutral manner, they may still remember it selectively to reinforce their expectations. This effect is called "selective recall", "confirmatory memory" or "access-biased memory".[25] Psychological theories differ in their predictions about selective recall. Schema theory predicts that information matching prior expectations will be more easily stored and recalled.[26] Some alternative approaches say that surprising information stands out more and so is more memorable.[26] Predictions from both these theories have been confirmed in different experimental contexts, with no theory winning outright.[27]

In one study, subjects read a profile of a woman which described a mix of introverted and extroverted behaviors.[28] They later had to recall examples of her introversion and extroversion. One group was told this was to assess the woman for a job as a librarian, while a second group were told it was for a job in real estate sales. There was a significant difference between what these two groups recalled, with the "librarian" group recalling more examples of introversion and the "sales" groups recalling more extroverted behavior.[28] A selective memory effect has also been shown in experiments that manipulate the desirability of personality types.[26][29] In one of these, a group of subjects were shown evidence that extroverted people are more successful than introverts. Another group were told the opposite. In a subsequent, apparently unrelated, study, they were asked to recall events from their lives in which they had been either introverted or extroverted. Each group of subjects provided more memories connecting themselves with the more desirable personality type, and recalled those memories more quickly.[30]

One study showed how selective memory can maintain belief in extrasensory perception (ESP).[31] Believers and disbelievers were each shown descriptions of ESP experiments. Half of each group were told that the experimental results supported the existence of ESP, while the others were told they did not. In a subsequent test, subjects recalled the material accurately, apart from believers who had read the non-supportive evidence. This group remembered significantly less information and some of them incorrectly remembered the results as supporting ESP.[31]
Related effects
Backfire effect

A similar cognitive bias found in individuals is the Backfire effect. Here, individuals challenged with evidence contradictory to their beliefs tend to reject the evidence and instead become an even firmer supporter of the initial belief.[32][33] The phrase was first coined by Brendan Nyhan and Jason Reifler in a paper entitled "When Corrections Fail: The persistence of political misperceptions".[34]
Polarization of opinion
Main article: Attitude polarization

When people with opposing views interpret new information in a biased way, their views can move even further apart. This is called "attitude polarization".[35] The effect was demonstrated by an experiment that involved drawing a series of red and black balls from one of two concealed "bingo baskets". Subjects knew that one basket contained 60% black and 40% red balls; the other, 40% black and 60% red. The experimenters looked at what happened when balls of alternating color were drawn in turn, a sequence that does not favor either basket. After each ball was drawn, subjects in one group were asked to state out loud their judgments of the probability that the balls were being drawn from one or the other basket. These subjects tended to grow more confident with each successive draw—whether they initially thought the basket with 60% black balls or the one with 60% red balls was the more likely source, their estimate of the probability increased. Another group of subjects were asked to state probability estimates only at the end of a sequence of drawn balls, rather than after each ball. They did not show the polarization effect, suggesting that it does not necessarily occur when people simply hold opposing positions, but rather when they openly commit to them.[36]
A collection of eight different handguns resting on the ground
Strong opinions on an issue such as gun ownership can bias how someone interprets new evidence.

A less abstract study was the Stanford biased interpretation experiment in which subjects with strong opinions about the death penalty read about mixed experimental evidence. Twenty-three percent of the subjects reported that their views had become more extreme, and this self-reported shift correlated strongly with their initial attitudes.[19] In later experiments, subjects also reported their opinions becoming more extreme in response to ambiguous information. However, comparisons of their attitudes before and after the new evidence showed no significant change, suggesting that the self-reported changes might not be real.[22][35][37] Based on these experiments, Deanna Kuhn and Joseph Lao concluded that polarization is a real phenomenon but far from inevitable, only happening in a small minority of cases. They found that it was prompted not only by considering mixed evidence, but by merely thinking about the topic.[35]

Charles Taber and Milton Lodge argued that the Stanford team's result had been hard to replicate because the arguments used in later experiments were too abstract or confusing to evoke an emotional response. The Taber and Lodge study used the emotionally charged topics of gun control and affirmative action.[22] They measured the attitudes of their subjects towards these issues before and after reading arguments on each side of the debate. Two groups of subjects showed attitude polarization; those with strong prior opinions and those who were politically knowledgeable. In part of this study, subjects chose which information sources to read, from a list prepared by the experimenters. For example they could read the National Rifle Association's and the Brady Anti-Handgun Coalition's arguments on gun control. Even when instructed to be even-handed, subjects were more likely to read arguments that supported their existing attitudes. This biased search for information correlated well with the polarization effect.[22]
Persistence of discredited beliefs

"eliefs can survive potent logical or empirical challenges. They can survive and even be bolstered by evidence that most uncommitted observers would agree logically demands some weakening of such beliefs. They can even survive the total destruction of their original evidential bases."
—Lee Ross and Craig Anderson[38]

Confirmation biases can be used to explain why some beliefs remain when the initial evidence for them is removed.[39] This belief perseverance effect has been shown by a series of experiments using what is called the "debriefing paradigm": subjects read fake evidence for a hypothesis, their attitude change is measured, then the fakery is exposed in detail. Their attitudes are then measured once more to see if their belief returns to its previous level.[38]

A typical finding is that at least some of the initial belief remains even after a full debrief.[40] In one experiment, subjects had to distinguish between real and fake suicide notes. The feedback was random: some were told they had done well while others were told they had performed badly. Even after being fully debriefed, subjects were still influenced by the feedback. They still thought they were better or worse than average at that kind of task, depending on what they had initially been told.[41]

In another study, subjects read job performance ratings of two firefighters, along with their responses to a risk aversion test.[38] These fictional data were arranged to show either a negative or positive association: some subjects were told that a risk-taking firefighter did better, while others were told they did less well than a risk-averse colleague.[42] Even if these two case studies had been true, they would have been scientifically poor evidence for a conclusion about firefighters in general. However, the subjects found them subjectively persuasive.[42] When the case studies were shown to be fictional, subjects' belief in a link diminished, but around half of the original effect remained.[38] Follow-up interviews established that the subjects had understood the debriefing and taken it seriously. Subjects seemed to trust the debriefing, but regarded the discredited information as irrelevant to their personal belief.[42]
Preference for early information

Experiments have shown that information is weighted more strongly when it appears early in a series, even when the order is unimportant. For example, people form a more positive impression of someone described as "intelligent, industrious, impulsive, critical, stubborn, envious" than when they are given the same words in reverse order.[43] This irrational primacy effect is independent of the primacy effect in memory in which the earlier items in a series leave a stronger memory trace.[43] Biased interpretation offers an explanation for this effect: seeing the initial evidence, people form a working hypothesis that affects how they interpret the rest of the information.[39]

One demonstration of irrational primacy involved colored chips supposedly drawn from two urns. Subjects were told the color distributions of the urns, and had to estimate the probability of a chip being drawn from one of them.[43] In fact, the colors appeared in a pre-arranged order. The first thirty draws favored one urn and the next thirty favored the other.[39] The series as a whole was neutral, so rationally, the two urns were equally likely. However, after sixty draws, subjects favored the urn suggested by the initial thirty.[43]

Another experiment involved a slide show of a single object, seen as just a blur at first and in slightly better focus with each succeeding slide.[43] After each slide, subjects had to state their best guess of what the object was. Subjects whose early guesses were wrong persisted with those guesses, even when the picture was sufficiently in focus that other people could readily identify the object.[39]
Illusory association between events
Main article: Illusory correlation

Illusory correlation is the tendency to see non-existent correlations in a set of data.[44] This tendency was first demonstrated in a series of experiments in the late 1960s.[45] In one experiment, subjects read a set of psychiatric case studies, including responses to the Rorschach inkblot test. They reported that the homosexual men in the set were more likely to report seeing buttocks, anuses or sexually ambiguous figures in the inkblots. In fact the case studies were fictional and, in one version of the experiment, had been constructed so that the homosexual men were less likely to report this imagery.[44] In a survey, a group of experienced psychoanalysts reported the same set of illusory associations with homosexuality.[44][45]

Another study recorded the symptoms experienced by arthritic patients, along with weather conditions over a 15-month period. Nearly all the patients reported that their pains were correlated with weather conditions, although the real correlation was zero.[46]

This effect is a kind of biased interpretation, in that objectively neutral or unfavorable evidence is interpreted to support existing beliefs. It is also related to biases in hypothesis-testing behavior.[47] In judging whether two events, such as illness and bad weather, are correlated, people rely heavily on the number of positive-positive cases: in this example, instances of both pain and bad weather. They pay relatively little attention to the other kinds of observation (of no pain and/or good weather).[48] This parallels the reliance on positive tests in hypothesis testing.[47] It may also reflect selective recall, in that people may have a sense that two events are correlated because it is easier to recall times when they happened together.[47]
Example Days Rain No rain
Arthritis 14 6
No arthritis 7 2

In the above fictional example, arthritic symptoms are more likely on days with no rain. However, people are likely to focus on the relatively large number of days which have both rain and symptoms. By concentrating on one cell of the table rather than all four, people can misperceive the relationship, in this case associating rain with arthritic symptoms.[49]
History
Engraved head-and-shoulders portrait of Francis Bacon wearing a hat and ruff.
Francis Bacon
Informal observation

Before psychological research on confirmation bias, the phenomenon had been observed anecdotally by writers, including the Greek historian Thucydides (c. 460 BC – c. 395 BC), Italian poet Dante Alighieri (1265–1321), English philosopher and scientist Francis Bacon (1561–1626),[50] and Russian author Leo Tolstoy (1828–1910). Thucydides, in the History of the Peloponnesian War wrote, "it is a habit of mankind ... to use sovereign reason to thrust aside what they do not fancy."[51] In the Divine Comedy, St. Thomas Aquinas cautions Dante when they meet in Paradise, "opinion—hasty—often can incline to the wrong side, and then affection for one's own opinion binds, confines the mind."[52] Bacon, in the Novum Organum, wrote,

The human understanding when it has once adopted an opinion ... draws all things else to support and agree with it. And though there be a greater number and weight of instances to be found on the other side, yet these it either neglects or despises, or else by some distinction sets aside or rejects[.][53]

Bacon said that biased assessment of evidence drove "all superstitions, whether in astrology, dreams, omens, divine judgments or the like".[53] In his essay "What Is Art?", Tolstoy wrote,

I know that most men—not only those considered clever, but even those who are very clever, and capable of understanding most difficult scientific, mathematical, or philosophic problems—can very seldom discern even the simplest and most obvious truth if it be such as to oblige them to admit the falsity of conclusions they have formed, perhaps with much difficulty—conclusions of which they are proud, which they have taught to others, and on which they have built their lives.[54]

Wason's research on hypothesis-testing

The term "confirmation bias" was coined by English psychologist Peter Wason.[55] For an experiment published in 1960, he challenged subjects to identify a rule applying to triples of numbers. At the outset, they were told that (2,4,6) fits the rule. Subjects could generate their own triples and the experimenter told them whether or not each triple conformed to the rule.[56][57]

While the actual rule was simply "any ascending sequence", the subjects had a great deal of difficulty in finding it, often announcing rules that were far more specific, such as "the middle number is the average of the first and last".[56] The subjects seemed to test only positive examples—triples that obeyed their hypothesized rule. For example, if they thought the rule was, "Each number is two greater than its predecessor", they would offer a triple that fit this rule, such as (11,13,15) rather than a triple that violates it, such as (11,12,19).[58]

Wason accepted falsificationism, according to which a scientific test of a hypothesis is a serious attempt to falsify it. He interpreted his results as showing a preference for confirmation over falsification, hence the term "confirmation bias".[Note 3][59] Wason also used confirmation bias to explain the results of his selection task experiment.[60] In this task, subjects are given partial information about a set of objects, and have to specify what further information they would need to tell whether or not a conditional rule ("If A, then B") applies. It has been found repeatedly that people perform badly on various forms of this test, in most cases ignoring information that could potentially refute the rule.[61][62]
Klayman and Ha's critique

A 1987 paper by Joshua Klayman and Young-Won Ha argued that the Wason experiments had not actually demonstrated a bias towards confirmation. Instead, Klayman and Ha interpreted the results in terms of a tendency to make tests that are consistent with the working hypothesis.[63] They called this the "positive test strategy".[6] This strategy is an example of a heuristic: a reasoning shortcut that is imperfect but easy to compute.[1] Klayman and Ha used Bayesian probability and information theory as their standard of hypothesis-testing, rather than the falsificationism used by Wason. According to these ideas, each answer to a question yields a different amount of information, which depends on the person's prior beliefs. Thus a scientific test of a hypothesis is one that is expected to produce the most information. Since the information content depends on initial probabilities, a positive test can either be highly informative or uninformative. Klayman and Ha argued that when people think about realistic problems, they are looking for a specific answer with a small initial probability. In this case, positive tests are usually more informative than negative tests.[11] However, in Wason's rule discovery task the answer—three numbers in ascending order—is very broad, so positive tests are unlikely to yield informative answers. Klayman and Ha supported their analysis by citing an experiment that used the labels "DAX" and "MED" in place of "fits the rule" and "doesn't fit the rule". This avoided implying that the aim was to find a low-probability rule. Subjects had much more success with this version of the experiment.[64][65]
Within the universe of all possible triples, those that fit the true rule are shown schematically as a circle. The hypothesized rule is a smaller circle enclosed within it.
If the true rule (T) encompasses the current hypothesis (H), then positive tests (examining an H to see if it is T) will not show that the hypothesis is false.

Two overlapping circles represent the true rule and the hypothesized rule. Any observation falling in the non-overlapping parts of the circles shows that the two rules are not exactly the same. In other words, those observations falsify the hypothesis.
If the true rule (T) overlaps the current hypothesis (H), then either a negative test or a positive test can potentially falsify H.

The triples fitting the hypothesis are represented as a circle within the universe of all triples. The true rule is a smaller circle within this.
When the working hypothesis (H) includes the true rule (T) then positive tests are the only way to falsify H.

In light of this and other critiques, the focus of research moved away from confirmation versus falsification to examine whether people test hypotheses in an informative way, or an uninformative but positive way. The search for "true" confirmation bias led psychologists to look at a wider range of effects in how people process information.[66]
Explanations

Confirmation bias is often described as a result of automatic, unintentional strategies rather than deliberate deception.[12][67] According to Robert Maccoun, most biased evidence processing occurs through a combination of both "cold" (cognitive) and "hot" (motivated) mechanisms.[68]

Cognitive explanations for confirmation bias are based on limitations in people's ability to handle complex tasks, and the shortcuts, called heuristics, that they use.[69] For example, people may judge the reliability of evidence by using the availability heuristic, i.e. how readily a particular idea comes to mind.[70] It is also possible that people can only focus on one thought at a time, so find it difficult to test alternative hypotheses in parallel.[71] Another heuristic is the positive test strategy identified by Klayman and Ha, in which people test a hypothesis by examining cases where they expect a property or event to occur. This heuristic avoids the difficult or impossible task of working out how diagnostic each possible question will be. However, it is not universally reliable, so people can overlook challenges to their existing beliefs.[11][72]

Motivational explanations involve an effect of desire on belief, sometimes called "wishful thinking".[73][74] It is known that people prefer pleasant thoughts over unpleasant ones in a number of ways: this is called the "Pollyanna principle".[75] Applied to arguments or sources of evidence, this could explain why desired conclusions are more likely to be believed true.[73] According to experiments that manipulate the desirability of the conclusion, people demand a high standard of evidence for unpalatable ideas and a low standard for preferred ideas. In other words, they ask, "Can I believe this?" for some suggestions and, "Must I believe this?" for others.[76][77] Although consistency is a desirable feature of attitudes, an excessive drive for consistency is another potential source of bias because it may prevent people from neutrally evaluating new, surprising information.[73] Social psychologist Ziva Kunda combines the cognitive and motivational theories, arguing that motivation creates the bias, but cognitive factors determine the size of the effect.[78]

Explanations in terms of cost-benefit analysis assume that people do not just test hypotheses in a disinterested way, but assess the costs of different errors.[79] Using ideas from evolutionary psychology, James Friedrich suggests that people do not primarily aim at truth in testing hypotheses, but try to avoid the most costly errors. For example, employers might ask one-sided questions in job interviews because they are focused on weeding out unsuitable candidates.[80] Yaacov Trope and Akiva Liberman's refinement of this theory assumes that people compare the two different kinds of error: accepting a false hypothesis or rejecting a true hypothesis. For instance, someone who underestimates a friend's honesty might treat him or her suspiciously and so undermine the friendship. Overestimating the friend's honesty may also be costly, but less so. In this case, it would be rational to seek, evaluate or remember evidence of their honesty in a biased way.[81] When someone gives an initial impression of being introverted or extroverted, questions that match that impression come across as more empathic.[82] This suggests that when talking to someone who seems to be an introvert, it is a sign of better social skills to ask, "Do you feel awkward in social situations?" rather than, "Do you like noisy parties?" The connection between confirmation bias and social skills was corroborated by a study of how college students get to know other people. Highly self-monitoring students, who are more sensitive to their environment and to social norms, asked more matching questions when interviewing a high-status staff member than when getting to know fellow students.[82]

Psychologists Jennifer Lerner and Philip Tetlock distinguish two different kinds of thinking process. Exploratory thought neutrally considers multiple points of view and tries to anticipate all possible objections to a particular position, while confirmatory thought seeks to justify a specific point of view. Lerner and Tetlock say that when people expect to need to justify their position to other people, whose views they already know, they will tend to adopt a similar position to those people, and then use confirmatory thought to bolster their own credibility. However, if the external parties are overly aggressive or critical, people will disengage from thought altogether, and simply assert their personal opinions without justification.[83] Lerner and Tetlock say that people only push themselves to think critically and logically when they know in advance they will need to explain themselves to others who are well-informed, genuinely interested in the truth, and whose views they don't already know.[84] Because those conditions rarely exist, they argue, most people are using confirmatory thought most of the time.[85]
Consequences
In finance

Confirmation bias can lead investors to be overconfident, ignoring evidence that their strategies will lose money.[4][86] In studies of political stock markets, investors made more profit when they resisted bias. For example, participants who interpreted a candidate's debate performance in a neutral rather than partisan way were more likely to profit.[87] To combat the effect of confirmation bias, investors can try to adopt a contrary viewpoint "for the sake of argument".[88] In one technique, they imagine that their investments have collapsed and ask themselves why this might happen.[4]
In physical and mental health

Raymond Nickerson, a psychologist, blames confirmation bias for the ineffective medical procedures that were used for centuries before the arrival of scientific medicine.[89] If a patient recovered, medical authorities counted the treatment as successful, rather than looking for alternative explanations such as that the disease had run its natural course.[89] Biased assimilation is a factor in the modern appeal of alternative medicine, whose proponents are swayed by positive anecdotal evidence but treat scientific evidence hyper-critically.[90][91][92]

Cognitive therapy was developed by Aaron T. Beck in the early 1960s and has become a popular approach.[93] According to Beck, biased information processing is a factor in depression.[94] His approach teaches people to treat evidence impartially, rather than selectively reinforcing negative outlooks.[50] Phobias and hypochondria have also been shown to involve confirmation bias for threatening information.[95]
In politics and law
A woman and a man reading a document in a courtroom
Mock trials allow researchers to examine confirmation biases in a realistic setting.

Nickerson argues that reasoning in judicial and political contexts is sometimes subconsciously biased, favoring conclusions that judges, juries or governments have already committed to.[96] Since the evidence in a jury trial can be complex, and jurors often reach decisions about the verdict early on, it is reasonable to expect an attitude polarization effect. The prediction that jurors will become more extreme in their views as they see more evidence has been borne out in experiments with mock trials.[97][98] Both inquisitorial and adversarial criminal justice systems are affected by confirmation bias.[99]

Confirmation bias can be a factor in creating or extending conflicts, from emotionally charged debates to wars: by interpreting the evidence in their favor, each opposing party can become overconfident that it is in the stronger position.[100] On the other hand, confirmation bias can result in people ignoring or misinterpreting the signs of an imminent or incipient conflict. For example, psychologists Stuart Sutherland and Thomas Kida have each argued that US Admiral Husband E. Kimmel showed confirmation bias when playing down the first signs of the Japanese attack on Pearl Harbor.[61][101]

A two-decade study of political pundits by Philip E. Tetlock found that, on the whole, their predictions were not much better than chance. Tetlock divided experts into "foxes" who maintained multiple hypotheses, and "hedgehogs" who were more dogmatic. In general, the hedgehogs were much less accurate. Tetlock blamed their failure on confirmation bias—specifically, their inability to make use of new information that contradicted their existing theories.[102]
In the paranormal

One factor in the appeal of psychic "readings" is that listeners apply a confirmation bias which fits the psychic's statements to their own lives.[103] By making a large number of ambiguous statements in each sitting, the psychic gives the client more opportunities to find a match. This is one of the techniques of cold reading, with which a psychic can deliver a subjectively impressive reading without any prior information about the client.[103] Investigator James Randi compared the transcript of a reading to the client's report of what the psychic had said, and found that the client showed a strong selective recall of the "hits".[104]

As a striking illustration of confirmation bias in the real world, Nickerson mentions numerological pyramidology: the practice of finding meaning in the proportions of the Egyptian pyramids.[105] There are many different length measurements that can be made of, for example, the Great Pyramid of Giza and many ways to combine or manipulate them. Hence it is almost inevitable that people who look at these numbers selectively will find superficially impressive correspondences, for example with the dimensions of the Earth.[105]
In scientific procedure

A distinguishing feature of scientific thinking is the search for falsifying as well as confirming evidence.[106] However, many times in the history of science, scientists have resisted new discoveries by selectively interpreting or ignoring unfavorable data.[106] Previous research has shown that the assessment of the quality of scientific studies seems to be particularly vulnerable to confirmation bias. It has been found several times that scientists rate studies that report findings consistent with their prior beliefs more favorably than studies reporting findings inconsistent with their previous beliefs.[67][107][108] However, assuming that the research question is relevant, the experimental design adequate and the data are clearly and comprehensively described, the found results should be of importance to the scientific community and should not be viewed prejudicially, regardless of whether they conform to current theoretical predictions.[108]

Confirmation bias may thus be especially harmful to objective evaluations regarding nonconforming results since biased individuals may regard opposing evidence to be weak in principle and give little serious thought to revising their beliefs.[107] Scientific innovators often meet with resistance from the scientific community, and research presenting controversial results frequently receives harsh peer review.[109]

In the context of scientific research, confirmation biases can sustain theories or research programs in the face of inadequate or even contradictory evidence;[61][110] the field of parapsychology has been particularly affected.[111]

An experimenter's confirmation bias can potentially affect which data are reported. Data that conflict with the experimenter's expectations may be more readily discarded as unreliable, producing the so-called file drawer effect. To combat this tendency, scientific training teaches ways to prevent bias.[112] For example, experimental design of randomized controlled trials (coupled with their systematic review) aims to minimize sources of bias.[112][113] The social process of peer review is thought to mitigate the effect of individual scientists' biases,[114] even though the peer review process itself may be susceptible to such biases.[108][115]
In self-image

Social psychologists have identified two tendencies in the way people seek or interpret information about themselves. Self-verification is the drive to reinforce the existing self-image and self-enhancement is the drive to seek positive feedback. Both are served by confirmation biases.[116] In experiments where people are given feedback that conflicts with their self-image, they are less likely to attend to it or remember it than when given self-verifying feedback.[117][118][119] They reduce the impact of such information by interpreting it as unreliable.[117][120][121] Similar experiments have found a preference for positive feedback, and the people who give it, over negative feedback.[116]
See also
Portal icon Psychology portal
Portal icon Thinking portal

Backfire effect
Cherry picking (fallacy)
Cognitive bias mitigation
Cognitive inertia
Denial
Denialism
Experimenter's bias
Filter bubble
Hostile media effect
List of biases in judgment and decision making
List of memory biases
Observer-expectancy effect
Reporting bias
Selective exposure theory
Semmelweis reflex
Woozle effect



qft

faultier
Posts: 1230
Joined: Tue Jan 12, 2010 10:11 am

Re: Irritating shit on facebook

Post by faultier » Thu Jan 17, 2013 2:41 pm

:Q:

User avatar
Johnlenham
Posts: 6067
Joined: Fri Nov 18, 2011 11:08 pm
Location: London

Re: Irritating shit on facebook

Post by Johnlenham » Thu Jan 17, 2013 2:55 pm

The quoting
Image

wub
Posts: 34156
Joined: Tue Feb 26, 2008 3:11 pm
Location: Madrid
Contact:

Re: Irritating shit on facebook

Post by wub » Thu Jan 17, 2013 2:56 pm

Johnlenham wrote:The quoting
Image
Image

User avatar
garethom
Posts: 13426
Joined: Wed Dec 22, 2010 12:55 pm
Location: Birmz
Contact:

Re: Irritating shit on facebook

Post by garethom » Thu Jan 17, 2013 5:04 pm

cryptical, can you go back to being idontreallygiveashit, you weren't really annoying when you didn'tgiveashit

User avatar
dickman69
Posts: 14517
Joined: Wed Mar 24, 2010 3:58 am
Location: Pittsburgh

Re: Irritating shit on facebook

Post by dickman69 » Thu Jan 17, 2013 5:07 pm

irritating shit on snh
every Tuesday 11pm EST on http://cosmicsound.club

buy my tunes pls
Soundcloud

User avatar
wilson
Posts: 3471
Joined: Sat May 23, 2009 6:51 pm
Location: East London/Essex

Re: Irritating shit on facebook

Post by wilson » Thu Jan 17, 2013 5:08 pm

Good to have you back, wub!

cryptical
Posts: 1144
Joined: Mon Oct 01, 2012 9:21 pm
Location: Birmingham

Re: Irritating shit on facebook

Post by cryptical » Thu Jan 17, 2013 5:15 pm

garethom wrote:cryptical, can you go back to being idontreallygiveashit, you weren't really annoying when you didn'tgiveashit
:oops:
and i cant its been banned :(
i cut down on posting until like two days ago tbf can i have a little credit for that?
i apologise for hampering your forum experience though.

edit: not being sarcastic by the way

User avatar
EliteLennon117
Posts: 2870
Joined: Wed Apr 04, 2012 12:35 am

Re: Irritating shit on facebook

Post by EliteLennon117 » Thu Jan 17, 2013 10:11 pm

cryptical wrote:
garethom wrote:cryptical, can you go back to being idontreallygiveashit, you weren't really annoying when you didn'tgiveashit
:oops:
and i cant its been banned :(
i cut down on posting until like two days ago tbf can i have a little credit for that?
i apologise for hampering your forum experience though.

edit: not being sarcastic by the way
fuck the h8s do you mane
Don't Snitch

User avatar
ezza
THE POOREST
Posts: 9934
Joined: Sat Apr 28, 2012 8:19 pm
Location: brizzy

Re: Irritating shit on facebook

Post by ezza » Thu Jan 17, 2013 10:17 pm

wub wrote:
Agent 47 wrote:Image
Have a history teacher explain this if they can.

Abraham Lincoln was elected to Congress in 1846.

John F. Kennedy was elected to Congress in 1946.

Abraham Lincoln was elected President in 1860.

John F. Kennedy was elected President in 1960.

Both were particularly concerned with civil rights.

Both wives lost a child while living in the White House.

Both Presidents were shot on a Friday.

Both Presidents were shot in the head.

Now it gets really weird.

Lincoln's secretary was named Kennedy.

Kennedy's Secretary was named Lincoln.

Both were assassinated by Southerners.

Both were succeeded by Southerners named Johnson.

Andrew Johnson, who succeeded Lincoln, was born in 1808.

Lyndon Johnson, who succeeded Kennedy, was born in 1908.

John Wilkes Booth, who assassinated Lincoln, was born in 1839.

Lee Harvey Oswald, who assassinated Kennedy, was born in 1939.

Both assassins were known by their three names.

Both names are composed of fifteen letters.

Now hang on to your seat.

Lincoln was shot at the theater named "Ford."

Kennedy was shot in a car called "Lincoln" made by "Ford."

Booth and Oswald were assassinated before their trials.

And here's the "kicker":

A week before Lincoln was shot, he was in Monroe, Maryland.

A week before Kennedy was shot, he was with Marilyn Monroe.

AND...................:

Lincoln was shot in a theater and the assassin ran to a warehouse...

Kennedy was shot from a warehouse and the assassin ran to a theater..
:o lolol
http://www.snopes.com/history/american/ ... ennedy.asp


Let's examine them one at a time:

Abraham Lincoln was elected to Congress in 1846.
John F. Kennedy was elected to Congress in 1946.


This statement is literally true: both Lincoln and Kennedy were first elected to Congress one hundred years apart. Aside from that minor coincidence, however, their political careers bore little resemblance to each other.

Lincoln was an Illinois state legislator who, outside of his election to a single term in the House of Representatives, failed in his every attempt to gain national political office until he was elected President in 1860, including an unsuccessful bid for the Senate in 1854, a unsuccessful bid to become the Republican vice-presidential candidate in 1856, and another unsuccessful bid for a Senate seat in 1858.

Kennedy, on the other hand, enjoyed an unbroken string of political successes at the national level when he entered the political arena after World War II. He was elected to the House of Representatives in 1946, re-elected in 1948, re-elected again in 1950, won a Senate seat in 1952, was re-elected to the Senate in 1958, and was elected President in 1960.

Abraham Lincoln was elected President in 1860.
John F. Kennedy was elected President in 1960.


It's hardly surprising that two men who (as noted above) both achieved their first political successes at the national level a hundred years apart would also ascend to the Presidency a hundred years apart. This "coincidence" is even less surprising when we consider that presidential elections are held only once every four years. Lincoln couldn't possibly have been elected President in 1857 or 1858 or 1859 or 1861 or 1862 or 1863, because no presidential elections were held in those years. Likewise, Kennedy couldn't possibly have been elected
President in the non-election years of 1957, 1958, 1959, 1961, 1962, or 1963. So, even though both men were politically active at the national level during eight-year spans when they might have been elected President, circumstances dictated that the only years during those spans when they both could have been elected were exactly one hundred years apart.

We're supposed to be amazed at minor happenstances such as the two men's being elected exactly one hundred years apart, but we're supposed to think nothing of the numerous non-coincidences: Lincoln was born in 1809; Kennedy was born in 1917. Lincoln died in 1865; Kennedy died in 1963. Lincoln was 56 years old at the time of his death; Kennedy was 46 years old at the time of his death. No striking coincidences or convenient hundred-year differences in any of those facts. Even when we consider that, absent all other factors, the two men had a one in twelve chance of dying in the same month, we find no coincidence there: Lincoln was killed in April; Kennedy was killed in November. Also unmentioned here is the fact that Lincoln was re-elected to a second term as President, but Kennedy was killed before the completion of his first term.

The names Lincoln and Kennedy each contain seven letters.

Surely this is the most trivial of coincidences, especially when once considers that the average length of presidential surnames is 6.64 letters. No mention is made of the fact that the two men's first names contain different numbers of letters, and that Kennedy had a middle name (Fitzgerald) while Lincoln had none.


Both were particularly concerned with civil rights.

Saying that Lincoln and Kennedy were both "particularly concerned with civil rights" is like saying that Woodrow Wilson and Franklin Roosevelt were both "particularly concerned with war," or that Herbert Hoover and Ronald Reagan were both "particularly concerned with economics." Those weren't subjects these men evinced a particular interest in; those were issues they were forced to deal with due to events currently taking place in the U.S. which were beyond their control.

Both wives lost their children while living in the White House.

Another statement that, while literally true, encompasses events that were completely different in circumstance and nature.

All of Lincoln's children were born before he entered the White House, and the Lincolns actually lost two children, not just one (although only one died during Lincoln's tenure as President). Edward Lincoln died of tuberculosis in 1850, just before his fourth birthday, and the Lincolns' eleven-year-old son Willie succumbed to typhoid at the end of their first year in the White House.

The Kennedys, on the other hand, were the rare Presidential couple still young enough to be bearing children after entering the White House, and a premature child born to Mrs. Kennedy in 1963 died two days later.

Other substantial differences not mentioned: The Lincolns had four children, all boys, only one of whom lived past his teens. The Kennedys had three children, two boys and a girl, two of whom survived well into adulthood.

Both Presidents were shot on a Friday.

Another non-surprise. Absent all other factors, the odds were already one in seven that both killings would have occurred on the same day of the week. (Don't even think about writing to tell us that we're wrong and the odds are really one in forty-nine. If you think we're wrong, you don't understand the question.) Add to that the obvious notions that the best chance the average person has to shoot a President is at a public function and that most public functions are held on weekends, and it becomes even more likely that a President would be killed on a Friday, Saturday, or Sunday. (Indeed, an earlier plot by Booth to kidnap Lincoln while the latter was attending a play at the Campbell Hospital was slated for March 17, also a Friday.)

Both were shot in the head.

This "coincidence" is another one which is exceedingly trivial in nature. The only two types of shots which reasonably assure a dead victim are chest shots and head shots, so two assassinations committed by head shots aren't the least bit coincidental — especially when one considers that since both Lincoln and Kennedy were shot from behind and while seated, so their assassins had no other practical choice of target. And the "coincidence" here is even less surprising when we note the substantial differences: Lincoln was killed indoors with a small handgun at point blank range; Kennedy was shot outdoors with a rifle from several hundred feet away.

Lincoln's secretary, Kennedy, warned him not to go to Ford's Theatre.
Kennedy's secretary, Lincoln, warned him not to go to Dallas.


This is one of those coincidences that isn't a coincidence at all; it's simply wrong. John Kennedy did have a secretary named Evelyn Lincoln (who may or may not have warned him about going to Dallas), but one searches in vain to find a Lincoln secretary named Kennedy. (Lincoln's White House secretaries were John G. Nicolay and John Hay.)

The more important point is that since Presidents are frequent recipients of assassination threats, they rarely make any public appearances without somebody's warning them of potential danger. Only on the extemely rare occasions when a tragedy actually occurs do we later take note of the warnings; in all other cases the failed "prophecies" are quickly forgotten. (Lincoln received "an unusual number of letters about plots to kidnap or assassinate him," said to have numbered at least eighty, yet none of those plots were enacted.) Nor does anyone think to mention other attempts at kidnap or assassination that were not preceded by any recorded warnings to the victims. (Lincoln was shot at on at least one other occasion.)

Yes, Lincoln was warned not to go to Ford's Theatre by persons concerned for his safety, just as he had been warned not to visit Richmond a week earlier, and just as he had been warned not to attend his own inauguration in 1861. Obviously, only one of the myriad of warnings he received throughout his four years in office was on the mark. Likewise, Kennedy was warned not to visit San Antonio the day before his trip to Dallas (and undoubtedly before a host of other appearances as well), but only the last warning he allegedly received is considered significant, because it coincidentally happened to come true. As Jeane Dixon and other "psychics" have demonstrated, if you make enough predictions, one of them is eventually bound to come true; the public remembers only that and forgets about all the others failed predictions.

Both were assassinated by Southerners.

A dubious use of the term "Southerner." John Wilkes Booth was undeniably a Southern sympathizer, but he was born in Maryland, which (along with Delaware) was the northernmost of the border slave states and remained part of the Union throughout the Civil War. Additionally, Booth spent a good deal of his life in the North and "thought of himself as a Northerner who understood the South."

Oswald was nominally a Southerner by virtue of his having been born in New Orleans; he spent his youth being shuttled between Lousiana, Texas, and New York before finally joining the Marines. But Oswald's "Southerness" is of no real import, because, unlike Booth, Oswald was not motivated by a regional affiliation.

Both were succeeded by Southerners.

Both Lincoln and Kennedy were "succeeded by Southerners" because both had Southerners as vice-president, another fact hardly surprising considering the historical circumstances of their times. Lincoln was a Northern Republican running for re-election while the country was in the midst of a civil war and needed a Southerner and a Democrat to balance the ticket, hence his choice of Tennessean Andrew Johnson. Kennedy, represented New England and therefore needed a vice-presidential candidate who could appeal to the populous Southern and Western regions, hence his choice of a Southwesterner, Texan Lyndon Johnson.

The identification of Andrew Johnson as a "Southerner" is also a bit problematic here. Although Johnson was born in North Carolina and spent his adult life in Tennessee (both slave states), Johnson was also the only Southern senator who refused to follow his state when it seceded, and he remained loyal to the Union.

Both successors were named Johnson.

Given the high frequency of "Johnson" (literally "son of John") as a surname in both Lincoln's and Kennedy's time, this "coincidence" should be no real surprise to anyone.

Andrew Johnson, who succeeded Lincoln, was born in 1808.
Lyndon Johnson, who succeeded Kennedy, was born in 1908.


Another hundred-year coincidence that is hardly surprising, since nearly all American politicians have attained high office (President or Vice-President) while in the 50-70 age range (and Andrew Johnson and Lyndon Johnson were, obviously, contemporaries of Lincoln and Kennedy, respectively). Perhaps it's time to point out that there's nothing "coincidental" about events merely because they somehow involve the number 100. If we sifted through all the Lincoln/Kennedy data, we could produce multiple instances of events involving the number 17 or 49 or 116, but nobody would consider those "coincidences" because they don't yield nice round numbers that have any significance to us, even though they're all just as "coincidental" as the number 100.

And once again, let's consider all the differences between the two Johnsons, such as that one hailed from North Carolina while the other was from Texas, or that one supported slavery while the other championed civil rights, or that one was never elected President in his own right while the other won the biggest presidential landslide in history, or that one was impeached while the other wasn't, or that one became President at the end of a war while the other became President at the beginning of a war.

John Wilkes Booth was born in 1839.
Lee Harvey Oswald was born in 1939.


Another coincidence that is no coincidence because it's plain wrong: Booth was born in 1838, not 1839.

Both assassins were known by their three names.

Another "coincidence" of dubious veracity. John Wilkes Booth was often billed as "J. Wilkes Booth" or simply "John Wilkes" (primarily to distinguish himself from his father and brother — both named Junius — and his brother Edwin, all three of whom were also actors), and as a prominent actor, his name was already familiar to the general public at the time of Lincoln's assassination. Lee Oswald was generally referred to as "Lee" (not "Lee Harvey") before Kennedy's assassination and was unknown to the general public until his arrest; the common usage of his full name only came about after the assassination because his habitual employment of false names (including several variations on his real name) and his possession of forged identification cards made it difficult for the Dallas police to initially identify him.

Both names are comprised of fifteen letters

Coincidence? Neither their first nor last names have the same number of letters. And why should it be significant that both assassins had the same number of letters in their full names when the same wasn't true of Abraham Lincoln and John Fitzgerald Kennedy, or of Andrew Johnson and Lyndon Baines Johnson?

Once again, perhaps we should focus on the substantive differences between the two men: Booth was born into a prominent family and, like his father, was a well-known, popular, gregarious actor. Oswald was born (and lived most of his life) in near poverty-level circumstances, never knew his father (who died two months before Oswald was born) and was an obscure, moody malcontent who never had any close friends or a steady job. Oswald was married with two children; Booth had neither wife nor offspring. Oswald enlisted in the Marines, but Booth kept a promise to his mother not to join the Confederate army.

Booth ran from the theater and was caught in a warehouse.
Oswald ran from a warehouse and was caught in a theater.


Another "coincidence" that is both inaccurate and superficial.

Booth shot Lincoln in a theatre of the type where live stage shows are held, then fled across state lines before being trapped and killed in a tobacco shed several days later.

Oswald shot Kennedy from (not in) a textbook warehouse, then remained in Dallas and was caught and taken alive in a movie theater a little over an hour later.

Booth and Oswald were assassinated before their trials.

Another superficial similarity with much more significant underlying differences, and a potentially dubious use of the word "assassinated."

After Booth shot Lincoln, he fled the scene and eventually (with co-conspirator, David Herold) crossed the Potomac from Maryland into Virginia, eluding capture for a total of eleven days before federal troops finally discovered him to be hiding on a farm belonging to Richard Garrett and surrounded the barn in which he and Herold were sleeping. The two men were ordered to surrender: Herold complied, but when Booth failed to drop his weapon and come out, the barn was set ablaze. A trooper named Boston Corbett, who was watching Booth through a gap in the barn's siding, shot the assassin. Whether Corbett can be said to have "assassinated" Booth is problematic — the deeply religious Corbett sometimes claimed that he had shot Booth because "Providence directed" him to do it or because he "did not want Booth to be roasted alive," but he also testified that he shot Booth because he "saw [Booth] in the act of stooping or springing and concluded he was going to use his weapons."

Oswald left the warehouse from which he shot Kennedy and was arrested in a movie theater a little over an hour later by police officers who had no idea who he was. (Oswald was initially arrested only for the murder of Dallas police officer J.D. Tippit, whom he shot while in flight; his connection to the Kennedy assassination was not established until later.) Oswald was captured alive and remained in custody for two days before being gunned down by Jack Ruby, a private citizen.

Other differences: Booth was shot in the back in the neck and lived for another three hours; Oswald was shot in the abdomen and died within minutes of his arrival at Parkland Hospital.

A month before Lincoln was assassinated he was in Monroe, Maryland.
A month before Kennedy was assassinated he was with Marilyn Monroe.


Marilyn Monroe died well over a year before Kennedy's assassination.
fuck that musta took long :Q:
DiegoSapiens wrote:thats so industrial
soronery wrote:New low

User avatar
Muncey
Posts: 6580
Joined: Tue Aug 09, 2011 6:30 pm
Location: Northants/Manchester

Re: Irritating shit on facebook

Post by Muncey » Thu Jan 17, 2013 10:35 pm

Agent 47 wrote:Now hang on to your seat.

Lincoln was shot at the theater named "Ford."

Kennedy was shot in a car called "Lincoln" made by "Ford."
Holy shit.

BTW Wub are you a history teacher? If not, you was not asked to explain it. Your sensible, logical, thought out answers aren't welcome round these parts.

User avatar
fuz
Posts: 89
Joined: Tue Aug 14, 2012 3:15 pm
Location: Paris 17th, France
Contact:

Re: Irritating shit on facebook

Post by fuz » Fri Jan 18, 2013 2:05 am

tLDR, but I'm still impressed. B)

@Muncey : this was reversed trolling imho. And you're feeding it.

User avatar
finji
Posts: 2922
Joined: Wed Sep 09, 2009 8:38 am

Re: Irritating shit on facebook

Post by finji » Fri Jan 18, 2013 2:46 am

Image


:cornlol: YOB
#cyber

User avatar
zerbaman
Posts: 2783
Joined: Tue Oct 20, 2009 9:05 am

Re: Irritating shit on facebook

Post by zerbaman » Fri Jan 18, 2013 3:19 am

garethom wrote:why do you even have fucking absolute stnuc on facebook?
Really?
How many times have you made that god damned post in this thread?

You seem like the kind of dude who would piss me off on facebook.
Do you say zerbaman? Or do you say zebraman?
Soundcloud

Dystinkt
Posts: 1744
Joined: Fri Oct 22, 2010 8:36 pm
Location: Bradford, UK

Re: Irritating shit on facebook

Post by Dystinkt » Fri Jan 18, 2013 3:59 am

a mate of mine put this gem of a status on facebook a while ago

'If the guy from 127 hours had gone to gym more he'd have got out'

Locked

Who is online

Users browsing this forum: No registered users and 0 guests