The art of persuasion
 
I'm not sure about the conclusions in the following article. Unsaid is the  
factor
of indoctrination. That is, some people hold political views the way  that
Christians or Jews hold religious views  -as articles of faith. Along  with
belief in particulars goes a whole catechism of  "reasons why." 
 
True believers memorize all the "necessary" talking points. They sound  
informed 
and in a sense they are informed. Its just that their views are all  
one-sided.
Regardless they hold these opinions because (1) they reinforce one's
sense of identity and (2) it is far better the do a little memorization  and
sound informed than appear to be ignorant in front of others and,  anyway,
who has the time to actually be informed  -in a scholarly of  scientific 
sense-
and shortcuts are a decent enough solution.
 
This said, the article does apply to a lot of people, possibly 80% of  the
vast majority of everyone, and as such it has a good deal of  usefulness.
 
The point about teaching is, of course, all-too-true. Want to learn  
something well?
The very best approach is to teach it. Right away you start to learn as  
never before.
 
But persuading others by asking them to teach has one  limitation:  You 
also need
some kind of "prime value" that tells you it is good to change your mind if 
 the
facts warrant doing so. You get this not only  in RC but also  classical 
liberalism
and traditional conservatism. You do not get this in any ideological  
political system,
viz., Libertarianism, "Greenism," Marxism, Fascism, etc. Nor do you get  
this in
over-the-top partisanship among Democrats or Republicans, think hard  core
Yellow Dogs or GOP "rock ribs." Such people get all their information  and
views from other yellow dogs or rock ribs and nowhere else. To say the  
least
this is a perfect formula for closed-mindedness.
 
 
This ought to be a good thing to remember for RC purposes.
 
Finally, I wonder if it always is the case that pointing out facts  
necessarily
cannot be persuasive. In some settings it seems to work well enough,
such as university seminars and science laboratories. It also may
be sufficient in some military, engineering, and business  settings.  
When this is true it would seem to be a good idea to 
take advantage of the effect.
 
In persuasion it has always seemed to me smart  to observe a principle  
that was 
first expressed by Sun Tzu:  Leave your opponent a  way out, a line of 
retreat.
In any person-to-person discussion this means leave him or her some way to  
save
face even if a change of opinion is the clear goal. Doing so allows  the
other person to protect his or her ego. If the price of winning an  argument
is damaging someone's self image then that price is too high and 
instead of winning you lose  -forfeiting future cooperation or  friendship.
That is, it is rare when any argument should be considered in  isolation.
Usually arguments are related to each other.
 
 
Billy
 
 
========================
 
 
BBC Future
_Neurohacks_ (http://www.bbc.com/future/columns/neurohacks)  | 21 May  2014 
 
 
IN ASSOCIATION WITH 

The best way to win  an argument

Tom Stafford
 
 
 
How do you change someone’s mind  if you think you are right and they are 
wrong? Psychology reveals the last thing  to do is the tactic we usually 
resort to. 

 
You are, I'm afraid to say, mistaken. The position you are taking makes no  
logical sense. Just listen up and I'll be more than happy to elaborate on 
the  many, many reasons why I'm right and you are wrong. Are you feeling 
ready to be  convinced? 
Whether the subject is climate change, the Middle East or forthcoming 
holiday  plans, this is the approach many of us adopt when we try to convince 
others to  change their minds. It's also an approach that, more often than not, 
leads to  the person on the receiving end hardening their existing 
position. Fortunately  research suggests there is a better way – one that 
involves 
more listening, and  less trying to bludgeon your opponent into submission. 
A little over a decade ago Leonid Rozenblit and Frank Keil from Yale  
University suggested that in many instances people believe they understand how  
something works when in fact their understanding is superficial at best. They 
 called this phenomenon "_the  illusion of explanatory depth_ 
(http://www.yale.edu/cogdevlab/aarticles/IOED%20proofs.pdf%201.pdf) ". They 
began by 
asking their study  participants to rate how well they understood how things 
like flushing toilets,  car speedometers and sewing machines worked, before 
asking them to explain what  they understood and then answer questions on it. 
The effect they revealed was  that, on average, people in the experiment 
rated their understanding as much  worse after it had been put to the test. 
What happens, argued the researchers, is that we mistake our familiarity 
with  these things for the belief that we have a detailed understanding of how 
they  work. Usually, nobody tests us and if we have any questions about 
them we can  just take a look. Psychologists call this idea that humans have a 
tendency to  take mental short cuts when making decisions or assessments the 
_"cognitive miser"  theory_ (https://en.wikipedia.org/wiki/Cognitive_miser) 
. 
Why would we bother expending the effort to really understand things when 
we  can get by without doing so? The interesting thing is that we manage to 
hide  from ourselves exactly how shallow our understanding is. 
It's a phenomenon that will be familiar to anyone who has ever had to teach 
 something. Usually, it only takes the first moments when you start to 
rehearse  what you'll say to explain a topic, or worse, the first student 
question, for  you to realise that you don't truly understand it. All over the 
world, teachers  say to each other "I didn't really understand this until I had 
to teach it". Or  as researcher and inventor _Mark Changizi  quipped_ 
(http://www.changizi.com/) : "I find that no matter how badly I teach I still 
learn  something". 
Explain yourself 
Research _published  last year_ 
(http://pss.sagepub.com/content/24/6/939.short)  on this illusion of 
understanding shows how the effect might be  used 
to convince others they are wrong. The research team, led by Philip  
Fernbach, of the University of Colorado, reasoned that the phenomenon might 
hold  
as much for political understanding as for things like how toilets work.  
Perhaps, they figured, people who have strong political opinions would be more  
open to other viewpoints, if asked to explain exactly how they thought the  
policy they were advocating would bring about the effects they claimed it  
would. 
Recruiting a sample of Americans via the internet, they polled participants 
 on a set of contentious US policy issues, such as imposing sanctions on 
Iran,  healthcare and approaches to carbon emissions. One group was asked to 
give their  opinion and then provide reasons for why they held that view. 
This group got the  opportunity to put their side of the issue, in the same way 
anyone in an  argument or debate has a chance to argue their case. 
Those in the second group did something subtly different. Rather that 
provide  reasons, they were asked to explain how the policy they were 
advocating 
would  work. They were asked to trace, step by step, from start to finish, 
the causal  path from the policy to the effects it was supposed to have. 
The results were clear. People who provided reasons remained as convinced 
of  their positions as they had been before the experiment. Those who were 
asked to  provide explanations softened their views, and reported a 
correspondingly larger  drop in how they rated their understanding of the 
issues. 
People who had  previously been strongly for or against carbon emissions 
trading, for example,  tended to became more moderate – ranking themselves as 
less 
certain in their  support or opposition to the policy. 
So this is something worth bearing in mind next time you're _trying  to 
convince a friend_ 
(https://www.contributoria.com/issue/2014-05/5319c4add63a707e780000cd)  that we 
should build more nuclear power stations, that  the 
collapse of capitalism is inevitable, or that dinosaurs co-existed with  humans 
10,000 years ago. Just remember, however, there's a chance you might need  
to be able to explain precisely why you think you are correct. Otherwise you  
might end up being the one who changes their  mind.

-- 
-- 
Centroids: The Center of the Radical Centrist Community 
<[email protected]>
Google Group: http://groups.google.com/group/RadicalCentrism
Radical Centrism website and blog: http://RadicalCentrism.org

--- 
You received this message because you are subscribed to the Google Groups 
"Centroids: The Center of the Radical Centrist Community" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
For more options, visit https://groups.google.com/d/optout.

Reply via email to