<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>social learning &#8211; Spencer Greenberg</title>
	<atom:link href="https://www.spencergreenberg.com/tag/social-learning/feed/" rel="self" type="application/rss+xml" />
	<link>https://www.spencergreenberg.com</link>
	<description></description>
	<lastBuildDate>Mon, 20 Jan 2025 15:27:24 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.9.4</generator>

<image>
	<url>https://i0.wp.com/www.spencergreenberg.com/wp-content/uploads/2024/05/cropped-icon.png?fit=32%2C32&#038;ssl=1</url>
	<title>social learning &#8211; Spencer Greenberg</title>
	<link>https://www.spencergreenberg.com</link>
	<width>32</width>
	<height>32</height>
</image> 
<site xmlns="com-wordpress:feed-additions:1">23753251</site>	<item>
		<title>Why do people often disagree about what&#8217;s immoral?</title>
		<link>https://www.spencergreenberg.com/2024/11/why-do-people-often-disagree-about-whats-immoral/</link>
					<comments>https://www.spencergreenberg.com/2024/11/why-do-people-often-disagree-about-whats-immoral/#comments</comments>
		
		<dc:creator><![CDATA[Admin]]></dc:creator>
		<pubDate>Sat, 16 Nov 2024 15:24:00 +0000</pubDate>
				<category><![CDATA[Essays]]></category>
		<category><![CDATA[character]]></category>
		<category><![CDATA[character judgments]]></category>
		<category><![CDATA[consequentialism]]></category>
		<category><![CDATA[disgust]]></category>
		<category><![CDATA[ethics]]></category>
		<category><![CDATA[fundamentally unethical]]></category>
		<category><![CDATA[generalizations]]></category>
		<category><![CDATA[heuristics]]></category>
		<category><![CDATA[immorality]]></category>
		<category><![CDATA[intrinsically unethical]]></category>
		<category><![CDATA[intuition]]></category>
		<category><![CDATA[morality]]></category>
		<category><![CDATA[shortcuts]]></category>
		<category><![CDATA[social learning]]></category>
		<category><![CDATA[system 1]]></category>
		<category><![CDATA[system 2]]></category>
		<category><![CDATA[virtue ethics]]></category>
		<guid isPermaLink="false">https://www.spencergreenberg.com/?p=4252</guid>

					<description><![CDATA[One reason people often disagree about what&#8217;s immoral is that they have different values. But there&#8217;s another important reason that I think few are aware of: there are at least four different kinds of moral evaluations of behavior, and it&#8217;s easy to conflate them. I argue that only one of these categories is actually sufficient [&#8230;]]]></description>
										<content:encoded><![CDATA[
<p>One reason people often disagree about what&#8217;s immoral is that they have different values. But there&#8217;s another important reason that I think few are aware of: there are at least four different kinds of moral evaluations of behavior, and it&#8217;s easy to conflate them. I argue that only one of these categories is actually sufficient grounds for judging an *action* as immoral, despite many people using the other categories to evaluate the morality of actions. I think they are making a subtle (and common) mistake when they do so.</p>



<p>These four categories of moral reactions to behavior are:</p>



<p><strong>1) Disgust:</strong> A visceral, emotional reaction to a behavior (which appears connected to the moral realm) that&#8217;s perceived as gross or disgusting.</p>



<p>Examples:</p>



<p>• The disgust that most people would have at the idea of someone eating human flesh (even in a survival scenario where there is no other food available and a person has died of natural causes).</p>



<p>• The disgusted reaction that some (but not most) people have to the idea of gay sex. For instance, in one study I ran, about 40% of Americans said that imagining themselves having sex with someone of their own gender caused them &#8220;an emotion of disgust.&#8221;&nbsp;&nbsp;</p>



<p>I think that people sometimes confuse the visceral emotion of disgust in response to a behavior with the behavior itself actually being immoral. Jonathan Haidt&#8217;s &#8220;moral dumbfounding&#8221; experiments support this point of view, where people insist a situation is immoral but can&#8217;t explain what is immoral about it (because the situations were carefully crafted not to violate moral principles and not to involve harm).</p>



<p></p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><strong>2) Character judgments:</strong> when a behavior is seen as indicative of an unethical character, even in cases where the behavior itself has no actual effects.</p>



<p>Examples:</p>



<p>• A person going to watch the daily operations of a slaughterhouse because they are intrigued by the idea of watching an animal die</p>



<p>• A person who enjoys daydreaming about stealing items from people they know, even though they have never stolen before</p>



<p>Cases like these provide evidence (some would argue, though it&#8217;s perhaps debatable) that a person has bad moral character, even if the behaviors themselves are not immoral. But people can jump to thinking an action is immoral because the sort of person that does it is more likely to be immoral, which I&#8217;d argue is, a mistake. While immoral actions are evidence of bad moral character, some actions that are evidence of bad moral character are not themselves immoral actions!</p>



<p></p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><strong>3) Heuristics: </strong>when a behavior is (collectively) judged as being &#8220;bad&#8221; because it <em>often </em>either causes harm, involves unethical behavior, or involves defecting on a social contract of some sort. These negative judgments can apply even if the behavior in question isn&#8217;t causing any harm in the specific situation where it&#8217;s being witnessed.</p>



<p>Examples:</p>



<p>• Driving fifty miles an hour over the speed limit is generally categorized as bad behavior, but if someone did it because they were driving a dying person to the hospital, we&#8217;d say that was actually okay.</p>



<p>• The head of an organization dating one of their employees is generally viewed as bad behavior because it often leads to harm, but there are instances where many people would say that in that particular case, it was ethically okay, such as cases where the employee insistently initiated the relationship leading to the pair ending up happily married</p>



<p>It&#8217;s easy to think of violations of heuristics like these as being bad, but really what&#8217;s going on is that we&#8217;re socially agreeing they are bad because it&#8217;s a good and helpful rule of thumb to treat them that way. But special circumstances can make them fine. Because of this, it&#8217;s not unreasonable to judge actions as probably bad when they fall in these categories (when we lack other information), but we should be sensitive to the specific details of the case since they are not necessarily bad.</p>



<p></p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><strong>4) Fundamentally unethical:</strong> a behavior that is, in and of itself, actually unethical according to at least some people&#8217;s deep ethical values. These actions don&#8217;t necessarily cause a feeling of disgust, and don&#8217;t merely appear bad because they are the sort of thing bad people do, and aren&#8217;t merely matching a heuristic about what&#8217;s bad &#8211; they are actually bad because of the precise action for some ethical values that people hold.</p>



<p>Examples:</p>



<p>• Poisoning your toddler because you don&#8217;t want a child</p>



<p>• Pretending you love someone when you don&#8217;t because you lack the courage to be honest</p>



<p>• Secretly spying on someone so you can see them naked</p>



<p>• Violating a promise you swore to uphold merely because you&#8217;re feeling lazy</p>



<p></p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p>When we treat an action as fundamentally unethical merely because it produces a feeling of disgust, or because it&#8217;s the sort of thing that provides evidence of bad character, or only because we have a societal heuristic against it because actions in that category tend to be harmful, I think we&#8217;re making a mistake. These categories are easy to conflate with an action being immoral, but they aren&#8217;t the same thing.</p>



<p></p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><em>This piece was first written on November 16, 2024, and first appeared on my website on January 20, 2025.</em></p>



<p></p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.spencergreenberg.com/2024/11/why-do-people-often-disagree-about-whats-immoral/feed/</wfw:commentRss>
			<slash:comments>1</slash:comments>
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">4252</post-id>	</item>
		<item>
		<title>A major (overlooked) reason why smart people fall for stupid things</title>
		<link>https://www.spencergreenberg.com/2024/09/a-major-overlooked-reason-why-smart-people-fall-for-stupid-things/</link>
					<comments>https://www.spencergreenberg.com/2024/09/a-major-overlooked-reason-why-smart-people-fall-for-stupid-things/#comments</comments>
		
		<dc:creator><![CDATA[Admin]]></dc:creator>
		<pubDate>Fri, 13 Sep 2024 12:47:50 +0000</pubDate>
				<category><![CDATA[Essays]]></category>
		<category><![CDATA[belief formation]]></category>
		<category><![CDATA[epistemics]]></category>
		<category><![CDATA[mentors]]></category>
		<category><![CDATA[mentorship]]></category>
		<category><![CDATA[naivete]]></category>
		<category><![CDATA[recommendations]]></category>
		<category><![CDATA[scams]]></category>
		<category><![CDATA[social learning]]></category>
		<category><![CDATA[social proof]]></category>
		<category><![CDATA[trrusting]]></category>
		<category><![CDATA[trust]]></category>
		<guid isPermaLink="false">https://www.spencergreenberg.com/?p=4117</guid>

					<description><![CDATA[Why do smart people fall for stupid things? Here is what I think is an important part of the answer that almost never gets discussed. It&#8217;s easy to look around at the stupid seeming things that other people believe (e.g., people who join harmful cults, get scammed by a con artist, become vocal evangelists for [&#8230;]]]></description>
										<content:encoded><![CDATA[
<p>Why do smart people fall for stupid things? Here is what I think is an important part of the answer that almost never gets discussed.</p>



<p>It&#8217;s easy to look around at the stupid seeming things that other people believe (e.g., people who join harmful cults, get scammed by a con artist, become vocal evangelists for a placebo treatment, or jump on the hype train of some outrageous new bubble) and wonder: &#8220;How on earth can they be so dumb?&#8221;</p>



<p>The answer, a lot of times, is simply the trust they have in someone else.<br>In other words, if a person were to evaluate the bad idea itself &#8211; call it X &#8211; they may well see it as dumb, dangerous, or full of hot air.</p>



<p>Instead, someone that person sees as impressive and totally trustworthy (or someone they just really like and respect) tells them that X is the next big thing. Or that X will change their life. Or that X will make them rich. Or that X will solve a problem for them that they desperately want solved.</p>



<p>This puts their brain in a predicament. They can either believe:<br>(1) That this impressive person who they deeply trust is deceiving them<br>Or<br>(2) That his impressive person who they trust is right &#8211; and their life will be way better off because of it!</p>



<p>If their trust in the person is great enough, or, at least, greater than their level of skepticism, (2) may win them over simply for that reason.</p>



<p>But (2) may also win them over for one or more of these reasons:</p>



<ul class="wp-block-list">
<li>they so desperately want this to be real &#8211; they so want to be special, or rich, or to have their biggest problems finally solved</li>



<li>it&#8217;s difficult and painful to believe this person they trust so much is deceiving them or so wrong about something important</li>



<li>they sense it will damage the relationship if they refuse to believe, and they care deeply about the relationship</li>



<li>they have a hard time saying &#8216;no&#8217; &#8211; perhaps it makes them very anxious to do so</li>
</ul>



<p>In other words, there are a great many dumb things that even smart people end up believing in simply because people believe people. To be clear, this is not the only mechanism by which smart people fall for dumb things. Being smart is not the same as acting rationally all the time. But this trust-based force is, I think, an important mechanism.</p>



<p>While a belief in others is wonderful and admirable in many instances, it can also be a chink in our skepticism and rationality. It can lead us to believe in crazy and dangerous things that we wouldn&#8217;t be likely to believe without that trust. We see this when people get scammed by their favorite influencer or when they become true believers in quack cures because they have a friend who says it changed their life.</p>



<p>While this effect often happens when one person we trust causes us to believe in X, the effect is magnified when more people around us believe. Being recruited into a harmful cult by a trusted friend can be difficult, but leaving a cult &#8211; at which point all of our close friends are believers in X &#8211; is far more difficult. And growing up in an authoritarian regime &#8211; where EVERYONE we&#8217;ve met seems to believe in X, makes X that much more impossible to resist.</p>



<p>When rationality is discussed, it&#8217;s often talked about at the level of the individual. But quite a bit of our thinking we necessarily outsource to others &#8211; we can&#8217;t make sense of everything ourselves. When we allow someone into our circle of trust who doesn&#8217;t deserve to be there, that can jeopardize our rationality. Hence, an important meta-skill of rationality is knowing who to trust &#8211; and not being suckered into trusting those who don&#8217;t deserve it.</p>



<p>Almost everyone is susceptible to this phenomenon of being duped because of our trust in people, but that doesn&#8217;t mean there&#8217;s nothing we can do to avoid it.</p>



<p>One thing that I think helps is to treat trust as being multi-factor. I can trust a person in one way or in one domain but not another. Or, put another way, earning trust is multi-dimensional. I can see someone as trustworthy because:</p>



<ul class="wp-block-list">
<li>I know they wouldn&#8217;t betray me and that they care a lot about me</li>



<li>I know that they vet evidence carefully, come to their beliefs in a rigorous way, and approach new information skeptically</li>



<li>I know that they are extremely knowledgeable about a specific topic area</li>
</ul>



<p>Being strong in one of these domains doesn&#8217;t automatically make someone strong in another. So, viewing someone as trustworthy in one of these domains shouldn&#8217;t cause you to view them as trustworthy in the other ones. And yet, that&#8217;s what many people do.</p>



<p>If you track trust in a one-dimensional way, it puts you at a lot of risk because someone you trust may have a very bad idea that they really want you to believe in. It may be hard to reject that idea because you trust them so much &#8211; and that may mean joining a harmful cult, buying into the peak of the next bubble, putting stock in an ineffective treatment, or being scammed.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><em>This piece was first written on September 13, 2024, and first appeared on my website on September 22, 2024.</em></p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.spencergreenberg.com/2024/09/a-major-overlooked-reason-why-smart-people-fall-for-stupid-things/feed/</wfw:commentRss>
			<slash:comments>2</slash:comments>
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">4117</post-id>	</item>
		<item>
		<title>Human behavior makes more sense when you understand &#8220;Anchor Beliefs&#8221;</title>
		<link>https://www.spencergreenberg.com/2021/11/human-behavior-makes-more-sense-when-you-understand-anchor-beliefs/</link>
					<comments>https://www.spencergreenberg.com/2021/11/human-behavior-makes-more-sense-when-you-understand-anchor-beliefs/#comments</comments>
		
		<dc:creator><![CDATA[admin]]></dc:creator>
		<pubDate>Sun, 21 Nov 2021 15:31:00 +0000</pubDate>
				<category><![CDATA[Essays]]></category>
		<category><![CDATA[anchor beliefs]]></category>
		<category><![CDATA[Bayesianism]]></category>
		<category><![CDATA[confirmation bias]]></category>
		<category><![CDATA[core beleifs]]></category>
		<category><![CDATA[delusions]]></category>
		<category><![CDATA[fixed false beliefs]]></category>
		<category><![CDATA[learning]]></category>
		<category><![CDATA[lines of retreat]]></category>
		<category><![CDATA[sacred values]]></category>
		<category><![CDATA[scout mindset]]></category>
		<category><![CDATA[social learning]]></category>
		<category><![CDATA[soldier mindset]]></category>
		<category><![CDATA[trapped priors]]></category>
		<category><![CDATA[updating beliefs]]></category>
		<category><![CDATA[worldviews]]></category>
		<guid isPermaLink="false">https://www.spencergreenberg.com/?p=2556</guid>

					<description><![CDATA[There&#8217;s an important type of belief most of us have, which we call &#8220;Anchor Beliefs.&#8221; These beliefs are, by definition, those beliefs we hold that are almost impossible to change. To the believer, an Anchor Belief doesn&#8217;t feel like a mere belief &#8211; it feels like an&#160;undeniable truth. These beliefs are often too deeply rooted [&#8230;]]]></description>
										<content:encoded><![CDATA[
<p>There&#8217;s an important type of belief most of us have, which we call &#8220;Anchor Beliefs.&#8221; These beliefs are, by definition, those beliefs we hold that are almost impossible to change. To the believer, an Anchor Belief doesn&#8217;t feel like a mere belief &#8211; it feels like an&nbsp;<em>undeniable truth</em>. These beliefs are often too deeply rooted to change, and the cost of giving them up may be extremely high (e.g., questioning the belief might cause you to lose your family, friends, livelihood, or your understanding of what reality looks like). </p>



<p>Whereas with most of our beliefs, when we get strong counter-evidence, we become at least a bit less confident in those beliefs, with anchor beliefs, strong counter-evidence doesn&#8217;t budge us at all &#8211; it just bounces off. That means we either ignore the evidence (e.g., just don&#8217;t let ourselves think about it), explain it away (e.g., &#8220;I must have misunderstood what happened&#8221;), or change our mind about other facts to incorporate the counter-evidence without our anchor belief budging (e.g., &#8220;they must have been lying to me.&#8221;)</p>



<p>Understanding the role that Anchor Beliefs play in human psychology &#8211; and identifying your own personal Anchor Beliefs &#8211; can help you make better sense of the world around you. Additionally, such an understanding can help you search for&nbsp;false&nbsp;Anchor Beliefs &#8211; those apparently unquestionable truths that make up the foundations of some people&#8217;s worldviews despite being wrong! Challenging your own false anchors is very difficult, but the consequences may be life-changing.</p>



<p>This article provides an introduction to Anchor Beliefs, including an explanation of how they differ from other beliefs, what can make them so hard to change, and a list of common categories of Anchor Beliefs that can help you identify your own. We also give some proposals for how false Anchor Beliefs form, how you can identify these false beliefs in yourself, and what you can do to question them in a safe and productive way. If you care about understanding your own mind and the minds of other people, we think you may find this write-up valuable!&nbsp;</p>



<hr class="wp-block-separator has-css-opacity"/>



<h2 class="wp-block-heading">How do Anchor Beliefs work? Instead of shifting with evidence, they shape how we&nbsp;<em>see</em>&nbsp;evidence.</h2>



<p>With many low-stakes, shallow-rooted beliefs (such as which turn to take to get to a restaurant), our commitment to the belief tends to adjust when we get counter-evidence (for example, &#8220;this doesn&#8217;t look familiar &#8211; I wonder if I took the wrong turn back there&#8221;). This behavior looks like an approximate form of&nbsp;<a target="_blank" href="https://programs.clearerthinking.org/question_of_evidence.html" rel="noreferrer noopener"><u>Bayesian updating</u></a>. Anchor Beliefs don&#8217;t work like this.&nbsp;</p>



<p>Anchor Beliefs almost never change, yet we still have to make sense of new information that we come across (some of which may strongly contradict our Anchor Beliefs). As mentioned, one solution is to warp the evidence that we receive such that we can fit it into our worldview AND keep our Anchor Belief intact at the same time. This is how Anchor Beliefs get their name: they are like huge, steel anchors securing boats to the ocean floor &#8211; only an enormously powerful current will be able to make them budge; any lesser current will simply swirl around the anchor. In this way, only incredibly powerful evidence can pose a threat to our Anchor Beliefs. And even then, our brains are highly adept at interpreting evidence so that our original Anchor Belief remains steadfast.&nbsp;</p>



<p>Here&#8217;s a silly example to show how this might look in reality. Imagine that you HAVE to believe the walls of your house are blue. If you don&#8217;t, then everyone you love will reject you (or something equally catastrophic). So it&#8217;s really important that you believe the walls of your house are blue. This means you must bend the evidence that you receive so that your perception is compatible with this belief. But the walls of your house don&#8217;t&nbsp;<em>look&nbsp;</em>blue. How can you make sense of this? Maybe there is some strange-colored light in the house that makes the walls appear white. Or maybe there&#8217;s something wrong with your vision. Or maybe the walls are just an incredibly pale shade of blue that is very close to white. It&#8217;s not clear which explanation is correct, but it&#8217;s not worth wasting your time worrying about why blue walls would appear white. </p>



<p>While you may think that you wouldn&#8217;t fall for a false Anchor Belief like this, being particularly smart or logical doesn&#8217;t necessarily help you challenge these kinds of beliefs. You&#8217;re more likely to come up with smarter and more logical reasons why your Anchor Belief&nbsp;<em>must&nbsp;</em>be correct (regardless of whether or not it is). If you search hard enough, there is almost always a way to reinterpret the evidence so that your Anchor Belief can remain steadfast.&nbsp;</p>



<hr class="wp-block-separator has-css-opacity"/>



<h2 class="wp-block-heading">Examples of common Anchor Beliefs</h2>



<p>Some Anchor Beliefs are profound (say, about the origins of life), but many are prosaic (say, about the earth being spherical rather than flat). We all have large numbers of boring and trivial-sounding Anchor Beliefs such as:</p>



<ul class="wp-block-list">
<li>1+1 = 2</li>



<li>humans have teeth</li>



<li>you live in [whatever country you believe you live in]</li>
</ul>



<p>Consider for a moment how hard it would be for someone to convince you that you were wrong about any of these beliefs! And consider for a moment how INSANE things would be for your worldview if you did correctly come to believe that these beliefs were false. The ramifications would be so shocking that it is hard to comprehend the implications of being wrong. </p>



<p>What makes the above Anchor Beliefs &#8220;trivial&#8221; is they have a really high probability of being true, and virtually everyone concurs about them. Much more interesting and important to consider are Anchor Beliefs that may be false. False Anchor Beliefs are often acquired through social and cultural influences, though they can also come about in other ways (e.g., as defense mechanisms or by generalizing from a small number of traumatic experiences). Here are some common categories of Anchor Beliefs that could be false:</p>



<ul class="wp-block-list">
<li>Things that almost everyone you know is taught</li>



<li>Certain religious beliefs learned in childhood</li>



<li>Perceptions of ourselves (e.g., as good/bad)</li>



<li>Views about one&#8217;s community</li>



<li>Views about &#8220;enemy&#8221; groups</li>



<li>Inferences from viscerally shocking first-hand experiences (e.g., &#8220;the world&#8217;s unsafe&#8221;)</li>



<li>Beliefs your social group REQUIRES</li>



<li>Claims that the reputation of your most trusted authority figures are staked on</li>



<li>Beliefs that, if you stopped believing them, would leave you very confused about what to believe or what to do</li>
</ul>



<p>The idea of an Anchor Belief is connected to (though not the same as) a number of other ideas, including:</p>



<ul class="wp-block-list">
<li>sacred values (social psychology)</li>



<li>soldier vs. scout mindset (see&nbsp;<a href="https://www.amazon.com/gp/product/0735217556/ref=as_li_tl?ie=UTF8&amp;camp=1789&amp;creative=9325&amp;creativeASIN=0735217556&amp;linkCode=as2&amp;tag=gimbeltechno-20&amp;linkId=0aee95a399d979bfe22a6f70def37ce1" target="_blank" rel="noreferrer noopener"><u>Julia Galef&#8217;s book</u></a>&nbsp;on the topic)</li>



<li>shibboleths (the Bible)</li>



<li>conflict vs. mistake theory (see&nbsp;<a href="https://slatestarcodex.com/2018/01/24/conflict-vs-mistake/" target="_blank" rel="noreferrer noopener"><u>Slate Star Codex</u></a>)</li>



<li>trapped priors (see&nbsp;<a href="https://astralcodexten.substack.com/p/trapped-priors-as-a-basic-problem" target="_blank" rel="noreferrer noopener"><u>Astral Codex Ten</u></a>)</li>



<li>belief updating (Connection Theory)</li>



<li>core beliefs (CBT)</li>



<li>leaving lines of retreat (<a href="https://www.lesswrong.com/posts/3XgYbghWruBMrPTAL/leave-a-line-of-retreat" target="_blank" rel="noreferrer noopener"><u>see LessWrong</u></a>)</li>
</ul>



<hr class="wp-block-separator has-css-opacity"/>



<h2 class="wp-block-heading">What does changing an Anchor Belief involve?</h2>



<p>While Anchor Beliefs almost never change, on rare occasions, we cut an anchor loose, and our boat suddenly lurches forward into the unknown. Jettisoning an anchor doesn&#8217;t necessarily take you where you want to go. It is scary, and it isn&#8217;t always safe. It may even lead you to abandon other, even more steadfast anchors. But abandoning an Anchor Belief is sometimes the only way to move forward. There&#8217;s no guarantee that the new beliefs you adopt will be correct. Consider your belief &#8220;the earth is round&#8221; (i.e., approximately an oblate spheroid, rather than flat like a disk).&nbsp;</p>



<p>Imagine, for a moment, what would happen if you came to believe in a flat earth, and you traced out the consequences of that belief. What is NASA, then? And SpaceX? What is the field of astronomy, or geology, or cartography? Surely our government must know &#8211; so why are they keeping it from us? How long have they known this? How do they prevent the truth from getting out? Is the whole world involved in this conspiracy? Am I in danger if I publicly say it&#8217;s a conspiracy? Why don&#8217;t more people speak out about this? Have all of my friends and family also been misled? Is gravity real (and if so, how does it work on a disk)? What are the stars in the sky?&nbsp;</p>



<p>It would be incredibly disorienting to work through all the implications that would follow from changing this once previously-unquestioned belief. Additionally, many of us have never checked whether the earth<em>&nbsp;is</em>&nbsp;actually round! Have you looked carefully at the arguments for this claim? And at the counterarguments to those arguments? Have you ever doubted (for even one minute of your entire adult life) that the earth is round? If we trust the evidence we receive from our eyes, many of us might assume that the earth is flat. So, how do we know that it is not? We learn this fact through our social world. (Unless, perhaps, you&#8217;ve lived by the ocean and had a habit of watching tall sailboats on the horizon.)&nbsp;</p>



<p>Fortunately, the earth isn&#8217;t flat. At least,&nbsp;<em>our</em>&nbsp;Anchor Beliefs say so &#8211; we&#8217;ve never personally run experiments to check, nor have we closely scrutinized the arguments for and against this claim. (If you wanted to, there are simple experiments that you can run &#8211; for example, you could attach a camera to a helium-filled high altitude weather balloon and&nbsp;<a target="_blank" href="https://www.bbc.com/news/av/uk-england-dorset-40706868" rel="noreferrer noopener"><u>see the round horizon from the vantage point of the camera</u></a>.) Notice that most scientific facts you believe are not as fundamental as &#8220;the earth is round,&#8221; which is a fact that has many important implications for our belief systems. Other scientific facts have fewer implications. For example, people were interested to learn that dinosaurs probably had feathers, but few (if any) had reasons to resist this update in their beliefs &#8211; because it didn&#8217;t challenge a core part of their worldview. In contrast, questioning a belief like &#8220;the earth is round&#8221; would throw most people&#8217;s belief systems into disarray.&nbsp;</p>



<p>Interestingly enough, if you already believed there was a cabal secretly ruling the world and that scientists were controlled by this cabal, then the earth being round may merely be an ordinary belief rather than an Anchor Belief. In that case, switching your view from believing in a round earth to believing in a flat earth is not likely to be perspective-shattering &#8211; it&#8217;s easy to incorporate it into your worldview as just another thing the cabal has manipulated people to believe.&nbsp;</p>



<hr class="wp-block-separator has-css-opacity"/>



<h2 class="wp-block-heading">What happens when our anchor beliefs are wrong?</h2>



<p>Anchor Beliefs aren&#8217;t necessarily false. A lot of them are true. They just aren&#8217;t&nbsp;<em>necessarily</em>&nbsp;true. So, it&#8217;s important to distinguish between two types of Anchor Beliefs: (1) &#8220;False Anchors&#8221; and (2) Anchor Beliefs that happen to be true. False Anchors are obviously much more worrisome. However, they often aren&#8217;t easy to spot. </p>



<p>Remember, Anchor Beliefs feel to us (the believer) not like mere beliefs but like&nbsp;<em>indisputable truths</em>. However, it&#8217;s almost certain that we learned them from the people around us or derived them from shoddy generalizations, and they may not be obvious truths at all (that doesn&#8217;t mean they are false, just that they may not be indisputable like they seem). Our own Anchor Beliefs are like the dark matter of the self. They flow through us without detection, and they influence our actions. </p>



<p>It&#8217;s not that doubting these beliefs is impossible, but we tend to automatically dismiss skepticism towards our Anchor Beliefs (or entirely ignore evidence that contradicts them) so that we don&#8217;t have to face abandoning our worldview. False Anchor Beliefs can have negative consequences the way any false belief can: by causing your predictions to be out of sync with reality. What makes them worse than your average false belief is that they are hard to change, largely because so many other beliefs tend to rely on them. But remember that not all Anchor Beliefs are false (so it wouldn&#8217;t make sense to give up a belief&nbsp;<em>merely</em>&nbsp;because it&#8217;s an Anchor).&nbsp;</p>



<hr class="wp-block-separator has-css-opacity"/>



<h2 class="wp-block-heading">Why is it that some of our beliefs become Anchor Beliefs?</h2>



<p>In the simplest examples (such as 1+1=2), a belief can become an Anchor Belief (i.e., almost impossible to change) because we have so much valid evidence for it being true that our prior probability of it being true is almost 100%. These are Anchor Beliefs that we&#8217;re almost certainly right about &#8211; we&#8217;ll call them&nbsp;<em>Steel&nbsp;</em>Anchors because they provide a sturdy foundation for an accurate worldview. But what about those Anchor Beliefs we may well be wrong about? Beliefs that are foundational to us but are not the result of witnessing tons of valid evidence? </p>



<p>Let&#8217;s call these Anchor Beliefs&nbsp;<em>Tin</em>&nbsp;Anchors. We are almost completely unable to change our mind about Tin Anchors, despite the fact that they don&#8217;t provide a valid foundation for an accurate worldview. Five reasons that we may form Tin Anchor Beliefs:</p>



<ol class="wp-block-list">
<li>It could be that Tin Anchors are a consequence of too many beliefs being piled on top of one fundamental belief, such that the fundamental belief can&#8217;t be questioned without challenging the whole pile of beliefs.</li>



<li>Tin Anchors might also be explained by the fact that it is better to have some model of the world than to have no model at all; without any kind of model, you fail to make predictions about the world, which is essential for understanding it. When an Anchor Belief falls, we&#8217;re often temporarily thrown into a state of confusion about what to believe, which makes predictions difficult. Tin Anchors, like real anchors, keep us stabilized.</li>



<li>Another explanation behind Tin Anchors is that people might grow up in social communities with poor epistemic standards; if everyone you trust tells you that something is true despite not having good evidence (especially if this starts in childhood and if this continues for a long time), these beliefs may become unmovable aspects of your worldview.</li>



<li>You might also find yourself forming Tin Anchors when you are in a situation where your brain predicts highly negative consequences from abandoning a belief. This makes changing our minds about that thing very painful and difficult. We will find all kinds of ways to cling to a belief if the alternative involves losing something that seems essential to our survival. Doubting one of these beliefs is like sticking your hand into a cage full of poisonous snakes: your mind wants to get away from the doubt as fast as possible (to avoid the severe predicted danger).</li>



<li>Tin Anchors sometimes occur when we generalize from shocking or traumatizing experiences. For instance, someone who is abused as a child might have an unshakable belief that they can&#8217;t trust other people not to hurt them (of course, some people truly cannot be trusted, but their brain may have overgeneralized this).</li>
</ol>



<hr class="wp-block-separator has-css-opacity"/>



<h2 class="wp-block-heading">Finding&nbsp;<em>your</em>&nbsp;Anchor Beliefs</h2>



<p>It may be valuable to ask yourself: &#8220;What are my own Tin Anchors?&#8221; If you want to consider what Tin Anchor Beliefs you may have, here are some questions that it might be helpful to ask yourself:&nbsp;</p>



<ul class="wp-block-list">
<li>&#8220;What beliefs did I pick up from those around me that I can&#8217;t imagine not believing (yet many people in other social groups somehow manage not to believe)?&#8221;</li>



<li>&#8220;What viscerally shocking experience might I have overgeneralized from that explains my worldview now?&#8221;</li>



<li>&#8220;What might other people from another community claim my Anchor Beliefs are?&#8221;</li>
</ul>



<p>These are pretty safe queries, as you&#8217;re very unlikely to stop believing your Tin Anchor Beliefs. And identifying one of your beliefs as a Tin Anchor doesn&#8217;t make it change, though it might be useful to know where your Anchors lie. Of course, it might be valuable (though costly) to try to change such an Anchor Belief if you want to. This might be something worth considering.&nbsp;</p>



<hr class="wp-block-separator has-css-opacity"/>



<h2 class="wp-block-heading">So, how do you challenge your Anchor Beliefs?</h2>



<p>Suppose you think that you&#8217;ve found one of your own Tin Anchors that you think has important implications for your life,&nbsp;<em>and</em>&nbsp;you actually want to examine whether it&#8217;s true. One strategy that may help is to try and clearly imagine the world where this Tin Anchor Belief turns out to be false. What is that world like? Can you deal with and accept that world? How would believing that you live in that world change your behavior and relationships? Can you accept those changes? </p>



<p>If you DO live in that world (where your Anchor Belief is false), would you want to believe you live in it, or would you rather pretend that your Anchor Belief isn&#8217;t false? If the answer is truly &#8220;yes&#8221; &#8211; you really would want to know if the belief is false, and you&#8217;re prepared to face the ramifications and consequences of losing that belief &#8211; then now you can truly start to put the belief to the test. Consider the strongest arguments on each side by, for instance, using&nbsp;<a rel="noreferrer noopener" target="_blank" href="https://programs.clearerthinking.org/challenge_your_deepest_beliefs.html"><u>our Clearer Thinking Belief Challenger program</u></a>. Seek out evidence that might disconfirm the belief. Ask people that disagree with this belief why they disagree with it. Surround yourself with some people who don&#8217;t have that Anchor Belief for a little while. See if the belief survives these tests.&nbsp;</p>



<hr class="wp-block-separator has-css-opacity"/>



<h2 class="wp-block-heading">How can you work around someone else&#8217;s false Anchor Beliefs?</h2>



<p>Suppose you know someone who you think has a harmful, false belief. If your goal is to help them understand the world accurately, it is simplest to first identify the relevant Anchor Belief that this false belief relies on and assume that it cannot be changed. (Changing someone else&#8217;s foundational Anchor Belief will require an IMMENSE current or, in other words, powerful evidence that is impossible to ignore or misinterpret.) Consider what you&#8217;d say to this person to nudge them towards truth GIVEN their Anchor Beliefs. Too often, we try to change other people&#8217;s minds by attempting to shift 20-ton anchors. Instead, it is more effective to identify those Anchor Beliefs and then act as best you can under the assumption that you will almost certainly not be able to change them. </p>



<p>Is it bad to have Tin Anchors? Some of these Anchor Beliefs are bad, but others are fine. In any event, it seems like we may not have a choice &#8211; Anchor Beliefs might be part of the construction of human minds. Once we acknowledge that people have these almost entirely unquestionable beliefs, the world &#8211; and how people act in it &#8211; begins to make much more sense. There&#8217;s a reason so many boats are just stuck where they are (no matter how much paddling appears to be happening on the surface).</p>



<hr class="wp-block-separator has-css-opacity"/>



<p>For more about Anchor Beliefs, see <a href="https://www.spencergreenberg.com/2024/08/what-happens-when-your-beliefs-cant-change/">part 2 of this series</a>.</p>



<p></p>



<p><em>This essay was first written on November 21, 2021, was published on Clearer Thinking on December 22, 2021, and was cross-posted here on December 24, 2021.</em></p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.spencergreenberg.com/2021/11/human-behavior-makes-more-sense-when-you-understand-anchor-beliefs/feed/</wfw:commentRss>
			<slash:comments>1</slash:comments>
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">2556</post-id>	</item>
		<item>
		<title>On Philosophical Disorders</title>
		<link>https://www.spencergreenberg.com/2020/07/on-philosophical-disorders/</link>
					<comments>https://www.spencergreenberg.com/2020/07/on-philosophical-disorders/#respond</comments>
		
		<dc:creator><![CDATA[Admin]]></dc:creator>
		<pubDate>Sat, 18 Jul 2020 23:54:00 +0000</pubDate>
				<category><![CDATA[Essays]]></category>
		<category><![CDATA[cults]]></category>
		<category><![CDATA[delusions]]></category>
		<category><![CDATA[drug use]]></category>
		<category><![CDATA[false beliefs]]></category>
		<category><![CDATA[group delusions]]></category>
		<category><![CDATA[hallucinogens]]></category>
		<category><![CDATA[harmful beliefs]]></category>
		<category><![CDATA[narcissism]]></category>
		<category><![CDATA[philosophical disorders]]></category>
		<category><![CDATA[philosophy]]></category>
		<category><![CDATA[psychological disorders]]></category>
		<category><![CDATA[psychopathy]]></category>
		<category><![CDATA[reality]]></category>
		<category><![CDATA[religion]]></category>
		<category><![CDATA[social learning]]></category>
		<category><![CDATA[sociopathy]]></category>
		<guid isPermaLink="false">https://www.spencergreenberg.com/?p=2510</guid>

					<description><![CDATA[I&#8217;d like to propose a new term: &#8220;philosophical disorder.&#8221; It&#8217;s when someone has a persistent belief that&#8217;s both highly inaccurate and substantially harmful. Here are some examples: A false belief that you are unlovable Being convinced that God punishes pre-marital with death Believing that &#8220;no usually means yes&#8221; in sexual encounters Whereas a psychological disorder [&#8230;]]]></description>
										<content:encoded><![CDATA[
<p>I&#8217;d like to propose a new term: &#8220;philosophical disorder.&#8221;</p>



<p>It&#8217;s when someone has a persistent belief that&#8217;s both highly inaccurate and substantially harmful. Here are some examples:</p>



<ul class="wp-block-list"><li>A false belief that you are unlovable</li><li>Being convinced that God punishes pre-marital with death</li><li>Believing that &#8220;no usually means yes&#8221; in sexual encounters</li></ul>



<p>Whereas a psychological disorder consists of emotions, thoughts, and personality traits creating distress or impairment, many WITHOUT psychological challenges have &#8220;Philosophical Disorders.&#8221; All it takes is being infected by false ideas that either harm the believer or lead them to harm others.</p>



<p>Since people will naturally disagree regarding which ideas are false or harmful, I think the term &#8220;Philosophical Disorder&#8221; is best reserved just for the extreme cases, where it&#8217;s easy for an outside observer to see both the falsity and harm of a belief.</p>



<p>Furthermore, much like it would rarely be a good idea to tell someone (outside of a therapeutic relationship or very trusting friendship) that you think they have a &#8220;psychological disorder,&#8221; telling someone you think they have a &#8220;philosophical disorder&#8221; is not advisable.</p>



<p>A true belief that causes harm at least has the virtue of being accurate, and trying to change it would imply some form of deception. Some people would rather believe a difficult truth than a comforting falsehood. On the other hand, a false belief that causes no problems can at least be said to be harmless, and one could argue that it&#8217;s not worth taking the time to correct it. Philosophical disorders, on the other hand, must (by definition) be both highly inaccurate and substantially harmful &#8211; they are the category of beliefs we can unequivocally say are worth correcting.</p>



<p>I require a belief to be &#8220;persistent&#8221; to meet the definition of philosophical disorder because if it is going to go away on its own anyway (e.g., a temporary harmful, false belief while someone is having a drug trip), it feels like it&#8217;s in a fundamentally different category (and action to change the belief is usually not as important since it is time-limited).</p>



<hr class="wp-block-separator"/>



<p><strong>— A Better Understanding of &#8220;Evil&#8221; —</strong></p>



<p>Pretty often, when large groups do what seems to be extreme evil, they are led by either a low-empathy narcissist or a sociopath. But chances are that most of the rank-and-file members of that group have philosophical disorders, not psychological disorders. Pretty often, even the leader has a philosophical disorder.</p>



<p>As an example, consider the case of religious zealots who are truly convinced that blowing up civilians is a holy act. Or the many groups whose members have become convinced they are inherently superior to other groups, whom they then kill or subjugate.</p>



<p>It&#8217;s understandable that people call those who commit atrocious acts &#8220;evil&#8221; regardless of their motivations, but there is a big difference between doing something highly harmful that you&#8217;re truly convinced is a good deed and doing a highly harmful act selfishly or with indifference towards the suffering of others.</p>



<p>While plenty of harm is caused by people due to their having some psychological traits, like sociopathy or low empathy narcissism, it may well be that as much or even more is caused by people who are pretty typical psychologically but who have philosophical disorders.</p>



<hr class="wp-block-separator"/>



<p><strong>— Philosophical Disorders vs. Psychological Disorders —</strong></p>



<p>Philosophical disorders can cause very bizarre behavior that is easily confused with a psychological disorder. As an example, some school shooters are well-characterized as having psychological disorders (e.g., showing signs of psychosis or sociopathy &#8211; like being convinced monsters lived in their basement at home or torturing animals for fun), whereas other school shooters may have been infected with harmful false belief systems (e.g., that others deserve death), hence, they might be better understood to have philosophical disorders. But without a careful inspection, the behavior (a &#8220;school shooting&#8221;) looks the same. The same outcome is coming from a different cause. I think it can be a clarifying question to ask: is this particular case best explained by a psychological disorder or philosophical disorder (or both simultaneously).</p>



<p>The link between psychological and philosophical disorders is complex. Philosophical disorders can both cause and be caused by psychological ones. For instance, falsely believing that nobody likes you could make you depressed. And schizophrenia can cause false, harmful beliefs (such as paranoid beliefs that others are out to get you).</p>



<p>In Cognitive Behavioral Therapy, &#8220;negative core beliefs&#8221; are sometimes uncovered as being a major factor causing a person&#8217;s depression or anxiety. That being said, typically, other factors are at play as well, such as patterns of behavior, reactivity to negative stimuli, disruptive thoughts, and so on.</p>



<p>But occasionally, a psychological disorder can be said to very directly result from a philosophical disorder (e.g., constant anxiety because of being convinced that lustful urges imply eternity in hell). In other cases, philosophical disorders are really not the right level of explanation for a psychological challenge (since emotional reactions, behavioral patterns, disruptive thoughts, etc., may better characterize what&#8217;s happening). Additionally, it&#8217;s possible to have a severe philosophical disorder without having a psychological disorder.</p>



<hr class="wp-block-separator"/>



<p><strong>— Categorizing Philosophical Disorders —</strong></p>



<p>As with psychological disorders, we might attempt to organize philosophical disorders. To do so, we might consider different aspects along which they can vary. Here&#8217;s a first attempt:</p>



<p><strong>(1) Who is substantially harmed?</strong></p>



<p>A. Self harmed (e.g., &#8220;I will fail at everything I try&#8221;)</p>



<p>B. Others harmed (e.g., &#8220;people of group X are inferior&#8221;)</p>



<p>C. Both harmed, meaning the belief harms both self and others (e.g., a man believing that &#8220;no men are trustworthy&#8221;)</p>



<hr class="wp-block-separator"/>



<p><strong>(2) What is the false belief about?</strong></p>



<p>A. Self (e.g., &#8220;My angry thoughts prove I am evil, even if I don&#8217;t act on them&#8221;)</p>



<p>B. Humanity (e.g., &#8220;Pretty much everyone is a bad person&#8221;)</p>



<p>C. People of importance in your life (e.g., a false belief that &#8220;my wife is cheating on me&#8221;)</p>



<p>D. Future events (e.g., &#8220;It is inevitable that the world ends in a nuclear holocaust sometime in the near future&#8221;)</p>



<p>E. Past events (e.g., &#8220;Mistakes my parents made when raising me have forever doomed me to being miserable&#8221;)</p>



<p>F. God (e.g., &#8220;God hates group X&#8221;)</p>



<p>G. Reality (e.g., &#8220;What if the whole world is just my imagination, and nobody else is real? I can&#8217;t prove that anyone else other than me exists.&#8221;)</p>



<hr class="wp-block-separator"/>



<p><strong>(3) How was this belief acquired?</strong></p>



<p>A. Caregivers who raised you (e.g., &#8220;you&#8217;re a bad child, nobody is ever going to love you&#8221;)</p>



<p>B. Culture of your region (e.g., &#8220;female circumcision is an important right of passage&#8221;)</p>



<p>C. Religion (e.g., &#8220;you will be rewarded in heaven if you die as a martyr killing the enemy&#8221;)</p>



<p>D. Group membership (e.g., falling into a social circle that has a shared philosophical disorder, for instance, a cult or movement)</p>



<p>E. Argumentation (e.g., reading an essay that convinces you that humanity should be wiped out)</p>



<p>F. Personality (e.g., narcissists tending to believe they are inherently superior to others)</p>



<p>G. Drugs (e.g., when someone forms an upsetting false belief while on drugs that persists afterward, such as that nothing is real)</p>



<p>H. Psychosis (e.g., the paranoid beliefs that schizophrenia sometimes causes)</p>



<p>I. Previously accurate but now resolved (e.g., you have a belief that people are not trustworthy because when you were young, the people around you really weren&#8217;t trustworthy, but you haven&#8217;t updated your belief even now that you are around mainly trustworthy people)</p>



<p>Of course, there are a lot of overlaps between the categories above.</p>



<hr class="wp-block-separator"/>



<p>When you hear of someone engaging in really a harmful behavior, it may be worth asking whether it better fits the hypothesis of a psychological disorder or of a philosophical one.</p>



<p><em>This essay was first written on July 20, 2020, and first appeared on this site on November 19, 2021.</em></p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.spencergreenberg.com/2020/07/on-philosophical-disorders/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">2510</post-id>	</item>
	</channel>
</rss>
