Saturday, February 11, 2017

Convincing People You Are Right (And They Are Wrong) - Part Four

We have gone over a few methods to determine if your views are true, and then discussed how to break down and analyze your views. In this post, we are going to briefly cover a variety of points to keep in mind when considering if your viewpoints are true. The points I cover are going to be somewhat all over the board, touching on a variety of random things. We can then start trying to convince people they are wrong.

When is the last time you were wrong? Not with something small like mixing up what time you were meeting your friends for dinner, but with something that you had put a fair amount of thought into? Something you had defended and strongly believed in? When did you last change your mind on something you had a strong opinion on, and thought people who disagreed with you didn’t know what they were talking about?

For most people, I suspect coming up with an example isn’t quite so easy. This could be for several reasons. One, you’ve never made a mistake and have never had to change your mind (unlikely). Two, you are too stubborn to admit when you have made a mistake (possible). Three, you made a mistake, changed your mind, and then forgot about it (probable).

To help battle this (probable) selective memory, I find it helpful to make a list of all the large things I have been wrong about, as a reminder that just because I’m absolutely certain about something, in no way means I am right. Unfortunately, certainty and reality are not as related as we would like to think. Keeping a list helps with the selective memory, and hopefully will serve as a reminder to take a breath and ease up—because you might be wrong.1

What is something that is extremely important to you? Perhaps you identify as part of a certain religion, or political party, or some other group. If you were to fill in the blank, “I am a _______”, what would you say? Now ask yourself “what is more important: advancing the cause of this group, or advancing truth?” If you are a good member of your group, you will probably think “my group does have the truth… I wouldn’t be part of it if that wasn’t the case.” Fair. But pause for a second. You’ve just decided that your group determines the truth. What about when it’s wrong? If you associate truth with what your group says, how would you know if it ever was in error? No group identity, beliefs, values, etc ever stays the same—the conservatives of today are not the same as conservatives of a few decades ago. Heck, liberals of the past used to advocate for eugenics. Aligning yourself with a group and deciding that they have it right is very dangerous, as it puts you in a position where being part of the group is more important than what is true.

To battle this, you have to make a commitment to what is true—and not to a group you currently identify with. Truth has to be more important to you than your religious beliefs, your political viewpoints, etc. You might be thinking, “Zak, I’m intelligent enough to be able to see when something the group identifies with isn’t true.” Maybe… but research suggests otherwise. The psychologist Geoffrey Cohen wanted to investigate the link between an individual’s opinion and their group identity, so he rounded up two groups of volunteers who described themselves as either strong conservatives or strong liberals. Cohen asked the members of each group to review and give their opinion on a welfare reform policy. The members of the conservative group were given a policy proposed by a republican leader, and the members of the liberal group were given a policy that had been proposed by a democratic leader. No surprise, the conservatives all agreed with the republican policy, and then liberals all agreed with the democratic policy.

Cohen then swapped the policies and asked each member of both groups to review them. Again, to no one’s surprise, the members of both groups found that the policy proposed by the other party was terrible. The policies were described as immoral, impractical, unrealistic, etc. Finally, Cohen asked the members of both groups if they had reached their conclusions about the policies because of the details of the policy itself, or because of the group the policy was associated with? Everyone scoffed at the suggestion that their opinion of the policies would have anything to do with the group associated with it—their conclusions were based solely on the details of the policy. Both groups agreed, however, that the OTHER group would definitely fall for such group-think silliness.

Well, surprise surprise! The policy that the conservatives had first reviewed and agreed with was NOT from a republican—but from a democrat. Likewise, the liberals had actually reviewed and approved of a policy from a republican. Cohen had tricked everyone into doing what they both claimed they would never do—support a policy simply because of which group they viewed the policy to be associated with.

Your social identity is a HUGE source of bias. And the more you associate with groups that confirm your beliefs, the more difficult it will be for you to change your mind if you come across disconfirming evidence. The stronger you associate with the group, the more likely you are to dismiss evidence, rationalize it away, or just ignore it. 

There are two ways to try and prevent being misled in such a way. First, don’t identify with a group—just be you. Soon as you decide you are part of a group, you will want to defend that group, and defending the group, rather than defending what is true, is a huge misstep.2 Rather than being “Zak the liberal atheist”, it’s better to just be Zak, and agree that atheism reflects certain views I have, and that my political views tend to fall on the liberal side of things. Like with the Sam Harris quote, don’t join a tribe. That brings up to the second step: if you do find yourself in part of a tribe (group), such as being a registered democrat, remind yourself of aspects of that group that you don’t identify with (or even actively disagree with). This will help you keep a skeptical sense about you.

Next up is the concept of Illusory Superiority, which is the tendency for people to overestimate everything about themselves. Compared to others, people tend to view themselves as being healthier, better looking, better drivers, more popular, having happier relationships, etc. One study found that 90% of professors polled rated themselves as above average teachers. Of that 90%, 65% rated themselves in the top 25% for teaching ability. Something is clearly askew!3

To make matters even worse, people also tend to view themselves as being more rational and less prone to blind spots than others! In a study with over 600 participants, only one person stated that she felt she was more prone to biases than average. Some people thought they were just as biased as everyone else, but a whopping 85% claimed that they were less prone to bias than everyone else. You might think “I am calm, I don’t freak out, I listen to points being made and don’t fall for rhetoric. I am absolutely more critical and less biased than your average person.” Maybe you are. But it’s more likely that you’re falling for the Bias Blind-Spot and just don’t know it. Simply put, “I am not biased” is exactly what someone blind to their biases would say.

To add one more layer to this, when people differ on opinions, everyone starts accusing the other party as being biased, but view themselves as the rational deliberator. “You believe in Global warming!? That’s just because you listen to libtards and shill scientists. Go watch Fox News, check out these Breitbart articles and get off the fake news train.” On the other hand, “You think vaccines are GOOD!? Of course the scientists who are in Big Pharma’s pocket would say that. You should listen to what Natural News and the Food Babe say—they do their own research, and aren’t corrupted by the system.”

Everyone is biased—EVERYONE. It’s just a part of being human. But with a little practice, and knowledge of how blind spots work, we can take a step in the right direction of being a little less biased than before.

In conclusion of this section, remember: keep a list of things you’ve changed your mind about to remind yourself that you’re not as omniscient as you’d like to think. Reject the desire to identify as part of a social group—especially groups with strong ideological foundations (political, social, religious, etc)—just be YOU. Lastly, always keep in mind of how biased you are about everything!

1. The two things I have been most certain about—bet my life on certain—turned out to be wrong. Both instances involved women I thought I would eventually marry. It’s no coincidence that I was so certain, seeing that I was heavily emotionally invested in both relationships. The more emotion involved, the easier it is to be convinced of something, even if you shouldn’t be.
Of course, one might argue that relationships are a whole different ball game, seeing as they are based primarily on emotion. That’s fair. We talk about love in absolute terms, and I can attest that having those feelings completely makes you believe that you will be together forever, etc. Several years ago, the musician Katie Melua changed the lyrics to one of her songs to reflect more scientific accuracy. The results were quite amusing, simply because we never hear people speak about love in the language of science.

Don’t take this to mean that you should be emotionless when it comes to decision making. Contrary to what was once thought, some emotion can be very helpful in making decisions. People who are sociopaths and don’t let emotion factor into their reasoning are unable to make ethical decision. Likewise, people with brain damage to areas of the brain that regulate emotion struggle to make even the simplest of decisions, such as “when should I schedule my dentist appointment: Wednesday or Thursday?” They tend to get bogged down in the minor logical details (such as “what will traffic be like on both days? Will it be different?”) and can’t just say “Thursday will be fine.”

2. The ease of getting people to align with groups, even arbitrarily defined, is extremely easy. Researchers have found that dividing people by shirt color, or even a flip of a coin, will produce hostility towards the other group (this is called Minimal Group Paradigm). Research has found that groups will even do things that help create a strong definition between them and the other groups, even when doing so hurts their own group!

3. Most of the research on Illusory Superiority have been done in the United States. There is some evidence that such an effect may be caused by culture to some extent. Reason being, there is evidence that Asians tend to view themselves as lower in ability than the rest of the population.

No comments:

Post a Comment