<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>delusions &#8211; Spencer Greenberg</title>
	<atom:link href="https://www.spencergreenberg.com/tag/delusions/feed/" rel="self" type="application/rss+xml" />
	<link>https://www.spencergreenberg.com</link>
	<description></description>
	<lastBuildDate>Sun, 22 Sep 2024 19:50:53 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.9.4</generator>

<image>
	<url>https://i0.wp.com/www.spencergreenberg.com/wp-content/uploads/2024/05/cropped-icon.png?fit=32%2C32&#038;ssl=1</url>
	<title>delusions &#8211; Spencer Greenberg</title>
	<link>https://www.spencergreenberg.com</link>
	<width>32</width>
	<height>32</height>
</image> 
<site xmlns="com-wordpress:feed-additions:1">23753251</site>	<item>
		<title>What happens when your beliefs can&#8217;t change?</title>
		<link>https://www.spencergreenberg.com/2024/08/what-happens-when-your-beliefs-cant-change/</link>
					<comments>https://www.spencergreenberg.com/2024/08/what-happens-when-your-beliefs-cant-change/#respond</comments>
		
		<dc:creator><![CDATA[admin]]></dc:creator>
		<pubDate>Tue, 13 Aug 2024 14:10:00 +0000</pubDate>
				<category><![CDATA[Essays]]></category>
		<category><![CDATA[anchor beliefs]]></category>
		<category><![CDATA[beliefs]]></category>
		<category><![CDATA[biases]]></category>
		<category><![CDATA[cognitive distortions]]></category>
		<category><![CDATA[criticism]]></category>
		<category><![CDATA[dedication]]></category>
		<category><![CDATA[deluded]]></category>
		<category><![CDATA[delusions]]></category>
		<category><![CDATA[denial]]></category>
		<category><![CDATA[faulty thinking]]></category>
		<category><![CDATA[imposter syndrome]]></category>
		<category><![CDATA[ingroup bias]]></category>
		<category><![CDATA[ingroup loyalty]]></category>
		<category><![CDATA[insight]]></category>
		<category><![CDATA[self-denial]]></category>
		<category><![CDATA[sunk cost fallacy]]></category>
		<category><![CDATA[updating]]></category>
		<guid isPermaLink="false">https://www.spencergreenberg.com/?p=4082</guid>

					<description><![CDATA[This is part 2 in my series about &#8220;anchor beliefs&#8221; &#8211; but you don&#8217;t need to read part 1 in order to understand it. I think that almost everyone has beliefs that are essentially unchangeable. These don&#8217;t feel to us like beliefs but like incontrovertible truths. Counter-evidence can&#8217;t touch them. They are beliefs we can&#8217;t [&#8230;]]]></description>
										<content:encoded><![CDATA[
<p>This is part 2 in my series about &#8220;anchor beliefs&#8221; &#8211; but you don&#8217;t need <a href="https://www.spencergreenberg.com/2021/11/human-behavior-makes-more-sense-when-you-understand-anchor-beliefs/">to read part 1</a> in order to understand it.</p>



<p>I think that almost everyone has beliefs that are essentially unchangeable. These don&#8217;t feel to us like beliefs but like incontrovertible truths. Counter-evidence can&#8217;t touch them. They are beliefs we can&#8217;t change our mind about. I call these &#8220;Anchor Beliefs.&#8221;</p>



<p>When Anchor Beliefs are false, we distort reality to fit them. So, what distortions do some reasonably common Anchor Beliefs cause?</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><strong>Anchor Belief 1: &#8220;I&#8217;m entirely good&#8221; or &#8220;I don&#8217;t do unethical things&#8221;</strong></p>



<p>What happens when someone with these Anchor Beliefs acts highly unethically? Well, since the Anchor Belief can&#8217;t change, that means the action must have been ethically okay to do, or else it was someone else&#8217;s fault or impossible to avoid. Victim blaming, denial, or shirking of responsibility ensues.</p>



<p>&#8220;My whole foundation, life, what I believed in, devotion to the company, was based on believing [Ramesh Balwani] was this person&#8230;He told me he didn&#8217;t know what I was doing in business, that my convictions were wrong&#8230;There was no way I could save our company if he was there…We were trying to do the right thing. We were trying to report results that we believed in and not report results if we thought there was any issue&#8221; -Elizabeth Holmes, who was found guilty on four counts of defrauding the investors in her company, Theranos</p>



<p>&#8220;All I ever wanted was to love women and, in turn, to be loved by them back. Their behavior towards me has only earned my hatred, and rightfully so! I am the true victim in all of this. I am the good guy.&#8221; -Elliot Rodger, in his manifesto about why he planned to commit murder before murdering six people.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><strong>Anchor Belief 2: &#8220;I&#8217;m not good enough&#8221;</strong></p>



<p>What happens when someone with this anchor belief gets a great job, performs really well, or achieves success? Well, it must have been a fluke or mistake; eventually, others will figure it out. Imposter syndrome ensues.</p>



<p>&#8220;No matter what we&#8217;ve done, there comes a point where you think, &#8216;How did I get here? When are they going to discover that I am, in fact, a fraud and take everything away from me?&#8221; &#8211; Tom Hanks, winner of two consecutive Academy Awards for Best Actor</p>



<p>&#8220;I have written 11 books, but each time I think, &#8216;Uh oh, they&#8217;re going to find out now. I&#8217;ve run a game on everybody, and they&#8217;re going to find me out.&#8221; &#8211; Maya Angelou, legendary poet and winner of the Presidential Medal of Freedom.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><strong>Anchor Belief 3: &#8220;This thing I&#8217;ve devoted a great deal of time/energy/identity into works and is good&#8221; [that doesn&#8217;t work or is harmful]</strong></p>



<p>What happens when it&#8217;s criticized? The criticism must be bad faith. Any imperfection in counter-evidence fully invalidates that evidence. Confirmation bias, cherry-picking, and motivated reasoning ensues.</p>



<p>&#8220;Those who have attacked my work on Vitamin C are scoundrels.&#8221; &#8211; Linus Pauling, two-time Nobel prize winner, defending his theory that vitamin C cures cancer and heart disease.</p>



<p>&#8220;We do not find critics of Scientology who do not have criminal pasts…Politician A stands up on his hind legs in a Parliament and brays for a condemnation of Scientology. When we look him over we find crimes &#8211; embezzled funds, moral lapses, a thirst for young boys &#8211; sordid stuff. Wife B howls at her husband for attending a Scientology group. We look her up and find she had a baby he didn&#8217;t know about.&#8221; &#8211; L. Ron Hubbard, founder of Scientology</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><strong>Anchor Belief 4: &#8220;My group is good.&#8221;</strong></p>



<p>What happens when their group does something really bad? The victims must be lying or have deserved it. Or acting badly must be justified in this case because it&#8217;s done for some more important greater good. Denial of and justification of immoral actions ensues.</p>



<p>&#8220;When we show a statement by Donald Trump that&#8217;s not truthful, Republicans will say it&#8217;s okay if it&#8217;s not true because it sends the right message, whereas Democrats will say that a statement needs to be factual&#8230;With a statement from Joe Biden, Democrats will say it&#8217;s okay if it&#8217;s not based on evidence, that it supports a generally true message, while Republicans will then have a higher bar and say every statement needs to be based on facts.&#8221; &#8211; Ethan Poskanzer, based on his studies on moral flexibility</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p>So, what are the takeaways here? I think that the following three things are important and true:</p>



<p>(1) Almost everyone has at least one Anchor Belief &#8211; a belief that is so sticky that it&#8217;s nearly impossible for it to change in the face of even extremely strong counter-evidence. Some people have more of these, and perhaps a small number of people have none, but I think Anchor beliefs are a near-universal among us humans.</p>



<p>(2) When our Anchor Beliefs are false (or partially false), because the beliefs won&#8217;t change, we distort reality when we get evidence against them in order to keep them intact while also somehow &#8220;making sense&#8221; of that counter-evidence.</p>



<p>(3) By looking at fairly common Anchor Beliefs people have, we can start to understand some recurring distortions in people&#8217;s thinking. Since people&#8217;s Anchor Beliefs are fixed but reality sometimes provides strong counter-evidence against these beliefs, that leads to predictable patterns of distortions that people&#8217;s minds deploy to keep the beliefs intact around those Anchor Beliefs.</p>



<p>In particular, I think that we find:</p>



<p>• Anchor Beliefs related to being good may lead to victim blaming and denial of responsibility.</p>



<p>• Anchor Beliefs about not being good enough may lead to imposter syndrome.</p>



<p>• Anchor Beliefs about something we&#8217;ve invested a lot of time/energy/identity into working on and being good may lead to confirmation bias, cherry-picking, and motivated reasoning.</p>



<p>• Anchor Beliefs about our group being good may lead us to deny or justify immoral actions by our group.</p>



<p>There are no strong studies that I&#8217;m aware of that identify or map out anchor beliefs and their frequency in the population &#8211; I believe the points above are true based on my experiences and observations.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><em>This piece was first written on August 13, 2024, and first appeared on my website on September 2, 2024.</em></p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.spencergreenberg.com/2024/08/what-happens-when-your-beliefs-cant-change/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">4082</post-id>	</item>
		<item>
		<title>Three motivations for believing </title>
		<link>https://www.spencergreenberg.com/2024/04/three-motivations-for-believing/</link>
					<comments>https://www.spencergreenberg.com/2024/04/three-motivations-for-believing/#respond</comments>
		
		<dc:creator><![CDATA[admin]]></dc:creator>
		<pubDate>Sat, 20 Apr 2024 14:04:00 +0000</pubDate>
				<category><![CDATA[Essays]]></category>
		<category><![CDATA[addiction]]></category>
		<category><![CDATA[belief]]></category>
		<category><![CDATA[delusional]]></category>
		<category><![CDATA[delusions]]></category>
		<category><![CDATA[denial]]></category>
		<category><![CDATA[dopamine]]></category>
		<category><![CDATA[epistemics]]></category>
		<category><![CDATA[hedonism]]></category>
		<category><![CDATA[hope]]></category>
		<category><![CDATA[motivated reasoning]]></category>
		<category><![CDATA[optimism]]></category>
		<category><![CDATA[pragmatism]]></category>
		<category><![CDATA[present bias]]></category>
		<category><![CDATA[rationalism]]></category>
		<category><![CDATA[religion]]></category>
		<category><![CDATA[self-sabotage]]></category>
		<category><![CDATA[utility]]></category>
		<category><![CDATA[values]]></category>
		<category><![CDATA[wishful thinking]]></category>
		<guid isPermaLink="false">https://www.spencergreenberg.com/?p=3929</guid>

					<description><![CDATA[There are three different motivations for belief, and it&#8217;s important to distinguish between them.&#160; 1) Belief because you think something&#8217;s true. For instance, you may think that the evidence supports the idea that you will eventually find love, or you may feel convinced by logical arguments you&#8217;ve heard in favor of god&#8217;s existence. 2) Belief [&#8230;]]]></description>
										<content:encoded><![CDATA[
<p>There are three different motivations for belief, and it&#8217;s important to distinguish between them.&nbsp;</p>



<p><strong>1) Belief because you think something&#8217;s true.</strong></p>



<p>For instance, you may think that the evidence supports the idea that you will eventually find love, or you may feel convinced by logical arguments you&#8217;ve heard in favor of god&#8217;s existence.</p>



<p><strong>2) Belief because you think it&#8217;s useful to believe.&nbsp;</strong></p>



<p>Regardless of whether you predict something&#8217;s true, you can predict that believing it will be more helpful than harmful to you in the long term, and so be motivated to believe for that pragmatic benefit.</p>



<p>For instance, you may intuit that you&#8217;ll be better off long-term believing that you will eventually find love (because that will make love more likely) or perceive that you&#8217;ll be happier believing in god (even if it turns out there is no god).</p>



<p><strong>3) Belief because it feels good in the moment.&nbsp;</strong></p>



<p>Regardless of whether it&#8217;s true or helpful to you in the long term, you may be motivated to believe something because it feels good right now (or prevents you from feeling bad).&nbsp;</p>



<p>For instance, you may feel comforted right now by thinking you&#8217;ll eventually find love or feel good in the moment, believing a god is watching over you.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><strong>Rationalists&nbsp;</strong>typically recommend striving to have your beliefs be of type 1: believing based on what&#8217;s most likely to be true.</p>



<p><strong>Pragmatists</strong>&nbsp;often recommend aiming for type 2 beliefs: believing based on what&#8217;s ultimately most useful to you.</p>



<p>I favor striving to have type 1 beliefs rather than type 2 beliefs, in part because I intrinsically value truth, but also because I think that for beliefs in category 2 that are *not* actually true, there are typically some beliefs in category 1 that will help you just as much, but which&nbsp;have the advantage of&nbsp;also&nbsp;being true.&nbsp;So often (but not always), there is a low cost to replacing beliefs from 2 with beliefs from 1 that have the added benefit of being true.</p>



<p>I also think that if you allow yourself&nbsp;to indiscriminately hold type 2 beliefs, it makes it hard to suddenly switch to rigorous truth-oriented thinking when it&#8217;s important to figure out the truth (e.g.,&nbsp;when you have to make a very important decision based on evidence).</p>



<p>On the other hand, many people have lots of type 3 beliefs, and all of us, myself included, have some type 3 beliefs. Whether you think that type 1 or type 2 beliefs are ultimately preferable, I think a valuable aspiration is to replace some of our type 3 beliefs with either 1s or 2s.</p>



<p>It&#8217;s very, very easy for us humans to delude ourselves based on what it feels good to believe at the moment because the reward cycle is so fast. Type 3 beliefs are immediately rewarding, incentivizing more such beliefs. But they are like the social media addiction version of believing, where you pursue what gives the greatest instantaneous reward rather than what&#8217;s actually good for you.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><em>This piece was first written on April 20, 2024, and first appeared on my website on May 7, 2024.</em></p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.spencergreenberg.com/2024/04/three-motivations-for-believing/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">3929</post-id>	</item>
		<item>
		<title>What is Lightgassing? A way we harm people by validating their false beliefs</title>
		<link>https://www.spencergreenberg.com/2023/10/what-is-lightgassing-a-way-we-harm-people-by-validating-their-beliefs/</link>
					<comments>https://www.spencergreenberg.com/2023/10/what-is-lightgassing-a-way-we-harm-people-by-validating-their-beliefs/#comments</comments>
		
		<dc:creator><![CDATA[admin]]></dc:creator>
		<pubDate>Mon, 02 Oct 2023 03:34:00 +0000</pubDate>
				<category><![CDATA[Essays]]></category>
		<category><![CDATA[delusions]]></category>
		<category><![CDATA[enablers]]></category>
		<category><![CDATA[enabling]]></category>
		<category><![CDATA[friendship]]></category>
		<category><![CDATA[gaslighting]]></category>
		<category><![CDATA[good intentions]]></category>
		<category><![CDATA[hallucinations]]></category>
		<category><![CDATA[light gassing]]></category>
		<category><![CDATA[lightgassing]]></category>
		<category><![CDATA[manipulation]]></category>
		<category><![CDATA[narcissism]]></category>
		<category><![CDATA[psychosis]]></category>
		<category><![CDATA[reality]]></category>
		<category><![CDATA[reinforcing]]></category>
		<category><![CDATA[relationships]]></category>
		<category><![CDATA[support]]></category>
		<category><![CDATA[victims]]></category>
		<guid isPermaLink="false">https://www.spencergreenberg.com/?p=3949</guid>

					<description><![CDATA[Gaslighting, where someone causes another person to doubt their sanity or senses, can cause psychological damage. There&#8217;s an opposite thing, though, that can also be damaging. As far as I know, it has no name. I call it &#8220;lightgassing&#8221; (or &#8220;light gassing&#8221;). Here, I explain how lightgassing works. Lightgassing is when one person agrees with [&#8230;]]]></description>
										<content:encoded><![CDATA[
<p></p>



<p><a href="https://en.wikipedia.org/wiki/Gaslighting">Gaslighting</a>, where someone causes another person to doubt their sanity or senses, can cause psychological damage.</p>



<p>There&#8217;s an opposite thing, though, that can also be damaging. As far as I know, it has no name. I call it &#8220;lightgassing&#8221; (or &#8220;light gassing&#8221;). Here, I explain how lightgassing works.</p>



<p>Lightgassing is when one person agrees with or validates another person&#8217;s false beliefs or misconceptions in order to be supportive.&nbsp;</p>



<p>Unlike gaslighting, a tactic of jerks and abusers, lightgassing is an (unintentionally harmful) tactic of friends and supporters.</p>



<p>Here are common examples I&#8217;ve seen that are sometimes, but obviously not always, lightgassing:</p>



<p>• &#8220;Since they did X, they don&#8217;t deserve to be with you.&#8221;</p>



<p>• &#8220;It was reasonable for you to do Y because they made you feel bad.&#8221;</p>



<p>• &#8220;You did nothing wrong. It was 100% their fault.&#8221;</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p>Ideally, when you&#8217;re upset, friends should validate your feelings and help you feel heard and understood, but they should do so without agreeing with statements they&nbsp;themselves&nbsp;know to be false.</p>



<p>We do a disservice to people when we encourage their false beliefs. Most people have a value of truth-telling (and knowing the truth), and by avoiding lightgassing, we stay truer to these values.</p>



<p>But how does one listen with openness and empathy to an upset friend and still validate <strong>*feelings*,</strong> without validating<strong> *false beliefs*</strong>?&nbsp;This&nbsp;can be a tricky maneuver, which I think is one reason people feel tempted to lightgas.</p>



<p>If you want to avoid lightgassing, the key is to validate those elements of a person&#8217;s <strong>*beliefs* </strong>that you know to be true while empathizing with them and validating that their *emotions* are understandable and okay to feel. But the key is to do this without reinforcing beliefs in false things.</p>



<p>I think it&#8217;s usually not helpful to challenge what another person believes is true right in the heat of emotion when that person is sad or upset. So avoiding lightgassing will often initially involve simply not validating/agreeing with what you believe is false. Later, when the person is feeling better, if they ask for your opinion on the facts (or you feel it&#8217;s important for them to hear your opinion), you can tell them what you believe to be true at that point.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p>A caveat that I think is worth mentioning is that sometimes, the only thing we know about a situation is what our upset friend or loved one has told us. In such cases, I think we should start with the assumption that what they have described is an accurate representation of what they experienced (unless any reason to doubt it emerges).</p>



<p>Lightgassing (or light gassing) typically happens in ordinary situations where someone feels hurt or upset. But it can also happen in more extreme situations, such as when you&#8217;re trying to help someone who is feeling upset due to severe delusions caused by psychosis.&nbsp;</p>



<p>In the case of someone with psychosis, the path of least resistance is to lightgas them by agreeing with their delusions, but this is not in their&nbsp;own&nbsp;interest. On the other hand, if you invalidate their emotions, you will likely make them more upset and may lose their trust. The tightrope to walk is to help them feel cared about, listened to, empathized with, and understood, without saying that their delusions are reality. In other words, to avoid lightgassing them while also not causing them to feel gaslit.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p>Does lightgassing really deserve its&nbsp;own&nbsp;term? Why not just call it &#8220;enabling&#8221;? Well, lightgassing can be a type of enabling, similar to how gaslighting is typically a form of manipulation. But lightgassing is much more specific than enabling, and enabling can include lots of things that are not lightgassing (e.g., buying an alcoholic some alcohol is a form of enabling but not lightgassing). Having a more specific term (lightgassing) makes it easier to spot and communicate about this specific pattern of behavior.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><em>This piece was first written on October 1, 2023, and first appeared on my website on May 13, 2024.</em></p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.spencergreenberg.com/2023/10/what-is-lightgassing-a-way-we-harm-people-by-validating-their-beliefs/feed/</wfw:commentRss>
			<slash:comments>1</slash:comments>
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">3949</post-id>	</item>
		<item>
		<title>Human behavior makes more sense when you understand &#8220;Anchor Beliefs&#8221;</title>
		<link>https://www.spencergreenberg.com/2021/11/human-behavior-makes-more-sense-when-you-understand-anchor-beliefs/</link>
					<comments>https://www.spencergreenberg.com/2021/11/human-behavior-makes-more-sense-when-you-understand-anchor-beliefs/#comments</comments>
		
		<dc:creator><![CDATA[admin]]></dc:creator>
		<pubDate>Sun, 21 Nov 2021 15:31:00 +0000</pubDate>
				<category><![CDATA[Essays]]></category>
		<category><![CDATA[anchor beliefs]]></category>
		<category><![CDATA[Bayesianism]]></category>
		<category><![CDATA[confirmation bias]]></category>
		<category><![CDATA[core beleifs]]></category>
		<category><![CDATA[delusions]]></category>
		<category><![CDATA[fixed false beliefs]]></category>
		<category><![CDATA[learning]]></category>
		<category><![CDATA[lines of retreat]]></category>
		<category><![CDATA[sacred values]]></category>
		<category><![CDATA[scout mindset]]></category>
		<category><![CDATA[social learning]]></category>
		<category><![CDATA[soldier mindset]]></category>
		<category><![CDATA[trapped priors]]></category>
		<category><![CDATA[updating beliefs]]></category>
		<category><![CDATA[worldviews]]></category>
		<guid isPermaLink="false">https://www.spencergreenberg.com/?p=2556</guid>

					<description><![CDATA[There&#8217;s an important type of belief most of us have, which we call &#8220;Anchor Beliefs.&#8221; These beliefs are, by definition, those beliefs we hold that are almost impossible to change. To the believer, an Anchor Belief doesn&#8217;t feel like a mere belief &#8211; it feels like an&#160;undeniable truth. These beliefs are often too deeply rooted [&#8230;]]]></description>
										<content:encoded><![CDATA[
<p>There&#8217;s an important type of belief most of us have, which we call &#8220;Anchor Beliefs.&#8221; These beliefs are, by definition, those beliefs we hold that are almost impossible to change. To the believer, an Anchor Belief doesn&#8217;t feel like a mere belief &#8211; it feels like an&nbsp;<em>undeniable truth</em>. These beliefs are often too deeply rooted to change, and the cost of giving them up may be extremely high (e.g., questioning the belief might cause you to lose your family, friends, livelihood, or your understanding of what reality looks like). </p>



<p>Whereas with most of our beliefs, when we get strong counter-evidence, we become at least a bit less confident in those beliefs, with anchor beliefs, strong counter-evidence doesn&#8217;t budge us at all &#8211; it just bounces off. That means we either ignore the evidence (e.g., just don&#8217;t let ourselves think about it), explain it away (e.g., &#8220;I must have misunderstood what happened&#8221;), or change our mind about other facts to incorporate the counter-evidence without our anchor belief budging (e.g., &#8220;they must have been lying to me.&#8221;)</p>



<p>Understanding the role that Anchor Beliefs play in human psychology &#8211; and identifying your own personal Anchor Beliefs &#8211; can help you make better sense of the world around you. Additionally, such an understanding can help you search for&nbsp;false&nbsp;Anchor Beliefs &#8211; those apparently unquestionable truths that make up the foundations of some people&#8217;s worldviews despite being wrong! Challenging your own false anchors is very difficult, but the consequences may be life-changing.</p>



<p>This article provides an introduction to Anchor Beliefs, including an explanation of how they differ from other beliefs, what can make them so hard to change, and a list of common categories of Anchor Beliefs that can help you identify your own. We also give some proposals for how false Anchor Beliefs form, how you can identify these false beliefs in yourself, and what you can do to question them in a safe and productive way. If you care about understanding your own mind and the minds of other people, we think you may find this write-up valuable!&nbsp;</p>



<hr class="wp-block-separator has-css-opacity"/>



<h2 class="wp-block-heading">How do Anchor Beliefs work? Instead of shifting with evidence, they shape how we&nbsp;<em>see</em>&nbsp;evidence.</h2>



<p>With many low-stakes, shallow-rooted beliefs (such as which turn to take to get to a restaurant), our commitment to the belief tends to adjust when we get counter-evidence (for example, &#8220;this doesn&#8217;t look familiar &#8211; I wonder if I took the wrong turn back there&#8221;). This behavior looks like an approximate form of&nbsp;<a target="_blank" href="https://programs.clearerthinking.org/question_of_evidence.html" rel="noreferrer noopener"><u>Bayesian updating</u></a>. Anchor Beliefs don&#8217;t work like this.&nbsp;</p>



<p>Anchor Beliefs almost never change, yet we still have to make sense of new information that we come across (some of which may strongly contradict our Anchor Beliefs). As mentioned, one solution is to warp the evidence that we receive such that we can fit it into our worldview AND keep our Anchor Belief intact at the same time. This is how Anchor Beliefs get their name: they are like huge, steel anchors securing boats to the ocean floor &#8211; only an enormously powerful current will be able to make them budge; any lesser current will simply swirl around the anchor. In this way, only incredibly powerful evidence can pose a threat to our Anchor Beliefs. And even then, our brains are highly adept at interpreting evidence so that our original Anchor Belief remains steadfast.&nbsp;</p>



<p>Here&#8217;s a silly example to show how this might look in reality. Imagine that you HAVE to believe the walls of your house are blue. If you don&#8217;t, then everyone you love will reject you (or something equally catastrophic). So it&#8217;s really important that you believe the walls of your house are blue. This means you must bend the evidence that you receive so that your perception is compatible with this belief. But the walls of your house don&#8217;t&nbsp;<em>look&nbsp;</em>blue. How can you make sense of this? Maybe there is some strange-colored light in the house that makes the walls appear white. Or maybe there&#8217;s something wrong with your vision. Or maybe the walls are just an incredibly pale shade of blue that is very close to white. It&#8217;s not clear which explanation is correct, but it&#8217;s not worth wasting your time worrying about why blue walls would appear white. </p>



<p>While you may think that you wouldn&#8217;t fall for a false Anchor Belief like this, being particularly smart or logical doesn&#8217;t necessarily help you challenge these kinds of beliefs. You&#8217;re more likely to come up with smarter and more logical reasons why your Anchor Belief&nbsp;<em>must&nbsp;</em>be correct (regardless of whether or not it is). If you search hard enough, there is almost always a way to reinterpret the evidence so that your Anchor Belief can remain steadfast.&nbsp;</p>



<hr class="wp-block-separator has-css-opacity"/>



<h2 class="wp-block-heading">Examples of common Anchor Beliefs</h2>



<p>Some Anchor Beliefs are profound (say, about the origins of life), but many are prosaic (say, about the earth being spherical rather than flat). We all have large numbers of boring and trivial-sounding Anchor Beliefs such as:</p>



<ul class="wp-block-list">
<li>1+1 = 2</li>



<li>humans have teeth</li>



<li>you live in [whatever country you believe you live in]</li>
</ul>



<p>Consider for a moment how hard it would be for someone to convince you that you were wrong about any of these beliefs! And consider for a moment how INSANE things would be for your worldview if you did correctly come to believe that these beliefs were false. The ramifications would be so shocking that it is hard to comprehend the implications of being wrong. </p>



<p>What makes the above Anchor Beliefs &#8220;trivial&#8221; is they have a really high probability of being true, and virtually everyone concurs about them. Much more interesting and important to consider are Anchor Beliefs that may be false. False Anchor Beliefs are often acquired through social and cultural influences, though they can also come about in other ways (e.g., as defense mechanisms or by generalizing from a small number of traumatic experiences). Here are some common categories of Anchor Beliefs that could be false:</p>



<ul class="wp-block-list">
<li>Things that almost everyone you know is taught</li>



<li>Certain religious beliefs learned in childhood</li>



<li>Perceptions of ourselves (e.g., as good/bad)</li>



<li>Views about one&#8217;s community</li>



<li>Views about &#8220;enemy&#8221; groups</li>



<li>Inferences from viscerally shocking first-hand experiences (e.g., &#8220;the world&#8217;s unsafe&#8221;)</li>



<li>Beliefs your social group REQUIRES</li>



<li>Claims that the reputation of your most trusted authority figures are staked on</li>



<li>Beliefs that, if you stopped believing them, would leave you very confused about what to believe or what to do</li>
</ul>



<p>The idea of an Anchor Belief is connected to (though not the same as) a number of other ideas, including:</p>



<ul class="wp-block-list">
<li>sacred values (social psychology)</li>



<li>soldier vs. scout mindset (see&nbsp;<a href="https://www.amazon.com/gp/product/0735217556/ref=as_li_tl?ie=UTF8&amp;camp=1789&amp;creative=9325&amp;creativeASIN=0735217556&amp;linkCode=as2&amp;tag=gimbeltechno-20&amp;linkId=0aee95a399d979bfe22a6f70def37ce1" target="_blank" rel="noreferrer noopener"><u>Julia Galef&#8217;s book</u></a>&nbsp;on the topic)</li>



<li>shibboleths (the Bible)</li>



<li>conflict vs. mistake theory (see&nbsp;<a href="https://slatestarcodex.com/2018/01/24/conflict-vs-mistake/" target="_blank" rel="noreferrer noopener"><u>Slate Star Codex</u></a>)</li>



<li>trapped priors (see&nbsp;<a href="https://astralcodexten.substack.com/p/trapped-priors-as-a-basic-problem" target="_blank" rel="noreferrer noopener"><u>Astral Codex Ten</u></a>)</li>



<li>belief updating (Connection Theory)</li>



<li>core beliefs (CBT)</li>



<li>leaving lines of retreat (<a href="https://www.lesswrong.com/posts/3XgYbghWruBMrPTAL/leave-a-line-of-retreat" target="_blank" rel="noreferrer noopener"><u>see LessWrong</u></a>)</li>
</ul>



<hr class="wp-block-separator has-css-opacity"/>



<h2 class="wp-block-heading">What does changing an Anchor Belief involve?</h2>



<p>While Anchor Beliefs almost never change, on rare occasions, we cut an anchor loose, and our boat suddenly lurches forward into the unknown. Jettisoning an anchor doesn&#8217;t necessarily take you where you want to go. It is scary, and it isn&#8217;t always safe. It may even lead you to abandon other, even more steadfast anchors. But abandoning an Anchor Belief is sometimes the only way to move forward. There&#8217;s no guarantee that the new beliefs you adopt will be correct. Consider your belief &#8220;the earth is round&#8221; (i.e., approximately an oblate spheroid, rather than flat like a disk).&nbsp;</p>



<p>Imagine, for a moment, what would happen if you came to believe in a flat earth, and you traced out the consequences of that belief. What is NASA, then? And SpaceX? What is the field of astronomy, or geology, or cartography? Surely our government must know &#8211; so why are they keeping it from us? How long have they known this? How do they prevent the truth from getting out? Is the whole world involved in this conspiracy? Am I in danger if I publicly say it&#8217;s a conspiracy? Why don&#8217;t more people speak out about this? Have all of my friends and family also been misled? Is gravity real (and if so, how does it work on a disk)? What are the stars in the sky?&nbsp;</p>



<p>It would be incredibly disorienting to work through all the implications that would follow from changing this once previously-unquestioned belief. Additionally, many of us have never checked whether the earth<em>&nbsp;is</em>&nbsp;actually round! Have you looked carefully at the arguments for this claim? And at the counterarguments to those arguments? Have you ever doubted (for even one minute of your entire adult life) that the earth is round? If we trust the evidence we receive from our eyes, many of us might assume that the earth is flat. So, how do we know that it is not? We learn this fact through our social world. (Unless, perhaps, you&#8217;ve lived by the ocean and had a habit of watching tall sailboats on the horizon.)&nbsp;</p>



<p>Fortunately, the earth isn&#8217;t flat. At least,&nbsp;<em>our</em>&nbsp;Anchor Beliefs say so &#8211; we&#8217;ve never personally run experiments to check, nor have we closely scrutinized the arguments for and against this claim. (If you wanted to, there are simple experiments that you can run &#8211; for example, you could attach a camera to a helium-filled high altitude weather balloon and&nbsp;<a target="_blank" href="https://www.bbc.com/news/av/uk-england-dorset-40706868" rel="noreferrer noopener"><u>see the round horizon from the vantage point of the camera</u></a>.) Notice that most scientific facts you believe are not as fundamental as &#8220;the earth is round,&#8221; which is a fact that has many important implications for our belief systems. Other scientific facts have fewer implications. For example, people were interested to learn that dinosaurs probably had feathers, but few (if any) had reasons to resist this update in their beliefs &#8211; because it didn&#8217;t challenge a core part of their worldview. In contrast, questioning a belief like &#8220;the earth is round&#8221; would throw most people&#8217;s belief systems into disarray.&nbsp;</p>



<p>Interestingly enough, if you already believed there was a cabal secretly ruling the world and that scientists were controlled by this cabal, then the earth being round may merely be an ordinary belief rather than an Anchor Belief. In that case, switching your view from believing in a round earth to believing in a flat earth is not likely to be perspective-shattering &#8211; it&#8217;s easy to incorporate it into your worldview as just another thing the cabal has manipulated people to believe.&nbsp;</p>



<hr class="wp-block-separator has-css-opacity"/>



<h2 class="wp-block-heading">What happens when our anchor beliefs are wrong?</h2>



<p>Anchor Beliefs aren&#8217;t necessarily false. A lot of them are true. They just aren&#8217;t&nbsp;<em>necessarily</em>&nbsp;true. So, it&#8217;s important to distinguish between two types of Anchor Beliefs: (1) &#8220;False Anchors&#8221; and (2) Anchor Beliefs that happen to be true. False Anchors are obviously much more worrisome. However, they often aren&#8217;t easy to spot. </p>



<p>Remember, Anchor Beliefs feel to us (the believer) not like mere beliefs but like&nbsp;<em>indisputable truths</em>. However, it&#8217;s almost certain that we learned them from the people around us or derived them from shoddy generalizations, and they may not be obvious truths at all (that doesn&#8217;t mean they are false, just that they may not be indisputable like they seem). Our own Anchor Beliefs are like the dark matter of the self. They flow through us without detection, and they influence our actions. </p>



<p>It&#8217;s not that doubting these beliefs is impossible, but we tend to automatically dismiss skepticism towards our Anchor Beliefs (or entirely ignore evidence that contradicts them) so that we don&#8217;t have to face abandoning our worldview. False Anchor Beliefs can have negative consequences the way any false belief can: by causing your predictions to be out of sync with reality. What makes them worse than your average false belief is that they are hard to change, largely because so many other beliefs tend to rely on them. But remember that not all Anchor Beliefs are false (so it wouldn&#8217;t make sense to give up a belief&nbsp;<em>merely</em>&nbsp;because it&#8217;s an Anchor).&nbsp;</p>



<hr class="wp-block-separator has-css-opacity"/>



<h2 class="wp-block-heading">Why is it that some of our beliefs become Anchor Beliefs?</h2>



<p>In the simplest examples (such as 1+1=2), a belief can become an Anchor Belief (i.e., almost impossible to change) because we have so much valid evidence for it being true that our prior probability of it being true is almost 100%. These are Anchor Beliefs that we&#8217;re almost certainly right about &#8211; we&#8217;ll call them&nbsp;<em>Steel&nbsp;</em>Anchors because they provide a sturdy foundation for an accurate worldview. But what about those Anchor Beliefs we may well be wrong about? Beliefs that are foundational to us but are not the result of witnessing tons of valid evidence? </p>



<p>Let&#8217;s call these Anchor Beliefs&nbsp;<em>Tin</em>&nbsp;Anchors. We are almost completely unable to change our mind about Tin Anchors, despite the fact that they don&#8217;t provide a valid foundation for an accurate worldview. Five reasons that we may form Tin Anchor Beliefs:</p>



<ol class="wp-block-list">
<li>It could be that Tin Anchors are a consequence of too many beliefs being piled on top of one fundamental belief, such that the fundamental belief can&#8217;t be questioned without challenging the whole pile of beliefs.</li>



<li>Tin Anchors might also be explained by the fact that it is better to have some model of the world than to have no model at all; without any kind of model, you fail to make predictions about the world, which is essential for understanding it. When an Anchor Belief falls, we&#8217;re often temporarily thrown into a state of confusion about what to believe, which makes predictions difficult. Tin Anchors, like real anchors, keep us stabilized.</li>



<li>Another explanation behind Tin Anchors is that people might grow up in social communities with poor epistemic standards; if everyone you trust tells you that something is true despite not having good evidence (especially if this starts in childhood and if this continues for a long time), these beliefs may become unmovable aspects of your worldview.</li>



<li>You might also find yourself forming Tin Anchors when you are in a situation where your brain predicts highly negative consequences from abandoning a belief. This makes changing our minds about that thing very painful and difficult. We will find all kinds of ways to cling to a belief if the alternative involves losing something that seems essential to our survival. Doubting one of these beliefs is like sticking your hand into a cage full of poisonous snakes: your mind wants to get away from the doubt as fast as possible (to avoid the severe predicted danger).</li>



<li>Tin Anchors sometimes occur when we generalize from shocking or traumatizing experiences. For instance, someone who is abused as a child might have an unshakable belief that they can&#8217;t trust other people not to hurt them (of course, some people truly cannot be trusted, but their brain may have overgeneralized this).</li>
</ol>



<hr class="wp-block-separator has-css-opacity"/>



<h2 class="wp-block-heading">Finding&nbsp;<em>your</em>&nbsp;Anchor Beliefs</h2>



<p>It may be valuable to ask yourself: &#8220;What are my own Tin Anchors?&#8221; If you want to consider what Tin Anchor Beliefs you may have, here are some questions that it might be helpful to ask yourself:&nbsp;</p>



<ul class="wp-block-list">
<li>&#8220;What beliefs did I pick up from those around me that I can&#8217;t imagine not believing (yet many people in other social groups somehow manage not to believe)?&#8221;</li>



<li>&#8220;What viscerally shocking experience might I have overgeneralized from that explains my worldview now?&#8221;</li>



<li>&#8220;What might other people from another community claim my Anchor Beliefs are?&#8221;</li>
</ul>



<p>These are pretty safe queries, as you&#8217;re very unlikely to stop believing your Tin Anchor Beliefs. And identifying one of your beliefs as a Tin Anchor doesn&#8217;t make it change, though it might be useful to know where your Anchors lie. Of course, it might be valuable (though costly) to try to change such an Anchor Belief if you want to. This might be something worth considering.&nbsp;</p>



<hr class="wp-block-separator has-css-opacity"/>



<h2 class="wp-block-heading">So, how do you challenge your Anchor Beliefs?</h2>



<p>Suppose you think that you&#8217;ve found one of your own Tin Anchors that you think has important implications for your life,&nbsp;<em>and</em>&nbsp;you actually want to examine whether it&#8217;s true. One strategy that may help is to try and clearly imagine the world where this Tin Anchor Belief turns out to be false. What is that world like? Can you deal with and accept that world? How would believing that you live in that world change your behavior and relationships? Can you accept those changes? </p>



<p>If you DO live in that world (where your Anchor Belief is false), would you want to believe you live in it, or would you rather pretend that your Anchor Belief isn&#8217;t false? If the answer is truly &#8220;yes&#8221; &#8211; you really would want to know if the belief is false, and you&#8217;re prepared to face the ramifications and consequences of losing that belief &#8211; then now you can truly start to put the belief to the test. Consider the strongest arguments on each side by, for instance, using&nbsp;<a rel="noreferrer noopener" target="_blank" href="https://programs.clearerthinking.org/challenge_your_deepest_beliefs.html"><u>our Clearer Thinking Belief Challenger program</u></a>. Seek out evidence that might disconfirm the belief. Ask people that disagree with this belief why they disagree with it. Surround yourself with some people who don&#8217;t have that Anchor Belief for a little while. See if the belief survives these tests.&nbsp;</p>



<hr class="wp-block-separator has-css-opacity"/>



<h2 class="wp-block-heading">How can you work around someone else&#8217;s false Anchor Beliefs?</h2>



<p>Suppose you know someone who you think has a harmful, false belief. If your goal is to help them understand the world accurately, it is simplest to first identify the relevant Anchor Belief that this false belief relies on and assume that it cannot be changed. (Changing someone else&#8217;s foundational Anchor Belief will require an IMMENSE current or, in other words, powerful evidence that is impossible to ignore or misinterpret.) Consider what you&#8217;d say to this person to nudge them towards truth GIVEN their Anchor Beliefs. Too often, we try to change other people&#8217;s minds by attempting to shift 20-ton anchors. Instead, it is more effective to identify those Anchor Beliefs and then act as best you can under the assumption that you will almost certainly not be able to change them. </p>



<p>Is it bad to have Tin Anchors? Some of these Anchor Beliefs are bad, but others are fine. In any event, it seems like we may not have a choice &#8211; Anchor Beliefs might be part of the construction of human minds. Once we acknowledge that people have these almost entirely unquestionable beliefs, the world &#8211; and how people act in it &#8211; begins to make much more sense. There&#8217;s a reason so many boats are just stuck where they are (no matter how much paddling appears to be happening on the surface).</p>



<hr class="wp-block-separator has-css-opacity"/>



<p>For more about Anchor Beliefs, see <a href="https://www.spencergreenberg.com/2024/08/what-happens-when-your-beliefs-cant-change/">part 2 of this series</a>.</p>



<p></p>



<p><em>This essay was first written on November 21, 2021, was published on Clearer Thinking on December 22, 2021, and was cross-posted here on December 24, 2021.</em></p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.spencergreenberg.com/2021/11/human-behavior-makes-more-sense-when-you-understand-anchor-beliefs/feed/</wfw:commentRss>
			<slash:comments>1</slash:comments>
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">2556</post-id>	</item>
		<item>
		<title>On Philosophical Disorders</title>
		<link>https://www.spencergreenberg.com/2020/07/on-philosophical-disorders/</link>
					<comments>https://www.spencergreenberg.com/2020/07/on-philosophical-disorders/#respond</comments>
		
		<dc:creator><![CDATA[Admin]]></dc:creator>
		<pubDate>Sat, 18 Jul 2020 23:54:00 +0000</pubDate>
				<category><![CDATA[Essays]]></category>
		<category><![CDATA[cults]]></category>
		<category><![CDATA[delusions]]></category>
		<category><![CDATA[drug use]]></category>
		<category><![CDATA[false beliefs]]></category>
		<category><![CDATA[group delusions]]></category>
		<category><![CDATA[hallucinogens]]></category>
		<category><![CDATA[harmful beliefs]]></category>
		<category><![CDATA[narcissism]]></category>
		<category><![CDATA[philosophical disorders]]></category>
		<category><![CDATA[philosophy]]></category>
		<category><![CDATA[psychological disorders]]></category>
		<category><![CDATA[psychopathy]]></category>
		<category><![CDATA[reality]]></category>
		<category><![CDATA[religion]]></category>
		<category><![CDATA[social learning]]></category>
		<category><![CDATA[sociopathy]]></category>
		<guid isPermaLink="false">https://www.spencergreenberg.com/?p=2510</guid>

					<description><![CDATA[I&#8217;d like to propose a new term: &#8220;philosophical disorder.&#8221; It&#8217;s when someone has a persistent belief that&#8217;s both highly inaccurate and substantially harmful. Here are some examples: A false belief that you are unlovable Being convinced that God punishes pre-marital with death Believing that &#8220;no usually means yes&#8221; in sexual encounters Whereas a psychological disorder [&#8230;]]]></description>
										<content:encoded><![CDATA[
<p>I&#8217;d like to propose a new term: &#8220;philosophical disorder.&#8221;</p>



<p>It&#8217;s when someone has a persistent belief that&#8217;s both highly inaccurate and substantially harmful. Here are some examples:</p>



<ul class="wp-block-list"><li>A false belief that you are unlovable</li><li>Being convinced that God punishes pre-marital with death</li><li>Believing that &#8220;no usually means yes&#8221; in sexual encounters</li></ul>



<p>Whereas a psychological disorder consists of emotions, thoughts, and personality traits creating distress or impairment, many WITHOUT psychological challenges have &#8220;Philosophical Disorders.&#8221; All it takes is being infected by false ideas that either harm the believer or lead them to harm others.</p>



<p>Since people will naturally disagree regarding which ideas are false or harmful, I think the term &#8220;Philosophical Disorder&#8221; is best reserved just for the extreme cases, where it&#8217;s easy for an outside observer to see both the falsity and harm of a belief.</p>



<p>Furthermore, much like it would rarely be a good idea to tell someone (outside of a therapeutic relationship or very trusting friendship) that you think they have a &#8220;psychological disorder,&#8221; telling someone you think they have a &#8220;philosophical disorder&#8221; is not advisable.</p>



<p>A true belief that causes harm at least has the virtue of being accurate, and trying to change it would imply some form of deception. Some people would rather believe a difficult truth than a comforting falsehood. On the other hand, a false belief that causes no problems can at least be said to be harmless, and one could argue that it&#8217;s not worth taking the time to correct it. Philosophical disorders, on the other hand, must (by definition) be both highly inaccurate and substantially harmful &#8211; they are the category of beliefs we can unequivocally say are worth correcting.</p>



<p>I require a belief to be &#8220;persistent&#8221; to meet the definition of philosophical disorder because if it is going to go away on its own anyway (e.g., a temporary harmful, false belief while someone is having a drug trip), it feels like it&#8217;s in a fundamentally different category (and action to change the belief is usually not as important since it is time-limited).</p>



<hr class="wp-block-separator"/>



<p><strong>— A Better Understanding of &#8220;Evil&#8221; —</strong></p>



<p>Pretty often, when large groups do what seems to be extreme evil, they are led by either a low-empathy narcissist or a sociopath. But chances are that most of the rank-and-file members of that group have philosophical disorders, not psychological disorders. Pretty often, even the leader has a philosophical disorder.</p>



<p>As an example, consider the case of religious zealots who are truly convinced that blowing up civilians is a holy act. Or the many groups whose members have become convinced they are inherently superior to other groups, whom they then kill or subjugate.</p>



<p>It&#8217;s understandable that people call those who commit atrocious acts &#8220;evil&#8221; regardless of their motivations, but there is a big difference between doing something highly harmful that you&#8217;re truly convinced is a good deed and doing a highly harmful act selfishly or with indifference towards the suffering of others.</p>



<p>While plenty of harm is caused by people due to their having some psychological traits, like sociopathy or low empathy narcissism, it may well be that as much or even more is caused by people who are pretty typical psychologically but who have philosophical disorders.</p>



<hr class="wp-block-separator"/>



<p><strong>— Philosophical Disorders vs. Psychological Disorders —</strong></p>



<p>Philosophical disorders can cause very bizarre behavior that is easily confused with a psychological disorder. As an example, some school shooters are well-characterized as having psychological disorders (e.g., showing signs of psychosis or sociopathy &#8211; like being convinced monsters lived in their basement at home or torturing animals for fun), whereas other school shooters may have been infected with harmful false belief systems (e.g., that others deserve death), hence, they might be better understood to have philosophical disorders. But without a careful inspection, the behavior (a &#8220;school shooting&#8221;) looks the same. The same outcome is coming from a different cause. I think it can be a clarifying question to ask: is this particular case best explained by a psychological disorder or philosophical disorder (or both simultaneously).</p>



<p>The link between psychological and philosophical disorders is complex. Philosophical disorders can both cause and be caused by psychological ones. For instance, falsely believing that nobody likes you could make you depressed. And schizophrenia can cause false, harmful beliefs (such as paranoid beliefs that others are out to get you).</p>



<p>In Cognitive Behavioral Therapy, &#8220;negative core beliefs&#8221; are sometimes uncovered as being a major factor causing a person&#8217;s depression or anxiety. That being said, typically, other factors are at play as well, such as patterns of behavior, reactivity to negative stimuli, disruptive thoughts, and so on.</p>



<p>But occasionally, a psychological disorder can be said to very directly result from a philosophical disorder (e.g., constant anxiety because of being convinced that lustful urges imply eternity in hell). In other cases, philosophical disorders are really not the right level of explanation for a psychological challenge (since emotional reactions, behavioral patterns, disruptive thoughts, etc., may better characterize what&#8217;s happening). Additionally, it&#8217;s possible to have a severe philosophical disorder without having a psychological disorder.</p>



<hr class="wp-block-separator"/>



<p><strong>— Categorizing Philosophical Disorders —</strong></p>



<p>As with psychological disorders, we might attempt to organize philosophical disorders. To do so, we might consider different aspects along which they can vary. Here&#8217;s a first attempt:</p>



<p><strong>(1) Who is substantially harmed?</strong></p>



<p>A. Self harmed (e.g., &#8220;I will fail at everything I try&#8221;)</p>



<p>B. Others harmed (e.g., &#8220;people of group X are inferior&#8221;)</p>



<p>C. Both harmed, meaning the belief harms both self and others (e.g., a man believing that &#8220;no men are trustworthy&#8221;)</p>



<hr class="wp-block-separator"/>



<p><strong>(2) What is the false belief about?</strong></p>



<p>A. Self (e.g., &#8220;My angry thoughts prove I am evil, even if I don&#8217;t act on them&#8221;)</p>



<p>B. Humanity (e.g., &#8220;Pretty much everyone is a bad person&#8221;)</p>



<p>C. People of importance in your life (e.g., a false belief that &#8220;my wife is cheating on me&#8221;)</p>



<p>D. Future events (e.g., &#8220;It is inevitable that the world ends in a nuclear holocaust sometime in the near future&#8221;)</p>



<p>E. Past events (e.g., &#8220;Mistakes my parents made when raising me have forever doomed me to being miserable&#8221;)</p>



<p>F. God (e.g., &#8220;God hates group X&#8221;)</p>



<p>G. Reality (e.g., &#8220;What if the whole world is just my imagination, and nobody else is real? I can&#8217;t prove that anyone else other than me exists.&#8221;)</p>



<hr class="wp-block-separator"/>



<p><strong>(3) How was this belief acquired?</strong></p>



<p>A. Caregivers who raised you (e.g., &#8220;you&#8217;re a bad child, nobody is ever going to love you&#8221;)</p>



<p>B. Culture of your region (e.g., &#8220;female circumcision is an important right of passage&#8221;)</p>



<p>C. Religion (e.g., &#8220;you will be rewarded in heaven if you die as a martyr killing the enemy&#8221;)</p>



<p>D. Group membership (e.g., falling into a social circle that has a shared philosophical disorder, for instance, a cult or movement)</p>



<p>E. Argumentation (e.g., reading an essay that convinces you that humanity should be wiped out)</p>



<p>F. Personality (e.g., narcissists tending to believe they are inherently superior to others)</p>



<p>G. Drugs (e.g., when someone forms an upsetting false belief while on drugs that persists afterward, such as that nothing is real)</p>



<p>H. Psychosis (e.g., the paranoid beliefs that schizophrenia sometimes causes)</p>



<p>I. Previously accurate but now resolved (e.g., you have a belief that people are not trustworthy because when you were young, the people around you really weren&#8217;t trustworthy, but you haven&#8217;t updated your belief even now that you are around mainly trustworthy people)</p>



<p>Of course, there are a lot of overlaps between the categories above.</p>



<hr class="wp-block-separator"/>



<p>When you hear of someone engaging in really a harmful behavior, it may be worth asking whether it better fits the hypothesis of a psychological disorder or of a philosophical one.</p>



<p><em>This essay was first written on July 20, 2020, and first appeared on this site on November 19, 2021.</em></p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.spencergreenberg.com/2020/07/on-philosophical-disorders/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">2510</post-id>	</item>
	</channel>
</rss>
