<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>false positives &#8211; Spencer Greenberg</title>
	<atom:link href="https://www.spencergreenberg.com/tag/false-positives/feed/" rel="self" type="application/rss+xml" />
	<link>https://www.spencergreenberg.com</link>
	<description></description>
	<lastBuildDate>Mon, 03 Apr 2023 01:47:24 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.9.4</generator>

<image>
	<url>https://i0.wp.com/www.spencergreenberg.com/wp-content/uploads/2024/05/cropped-icon.png?fit=32%2C32&#038;ssl=1</url>
	<title>false positives &#8211; Spencer Greenberg</title>
	<link>https://www.spencergreenberg.com</link>
	<width>32</width>
	<height>32</height>
</image> 
<site xmlns="com-wordpress:feed-additions:1">23753251</site>	<item>
		<title>Demystifying p-values</title>
		<link>https://www.spencergreenberg.com/2022/12/demystifying-p-values/</link>
					<comments>https://www.spencergreenberg.com/2022/12/demystifying-p-values/#comments</comments>
		
		<dc:creator><![CDATA[admin]]></dc:creator>
		<pubDate>Sat, 31 Dec 2022 20:40:00 +0000</pubDate>
				<category><![CDATA[Essays]]></category>
		<category><![CDATA[alpha]]></category>
		<category><![CDATA[alternative hypothesis]]></category>
		<category><![CDATA[Bayesianism]]></category>
		<category><![CDATA[false positives]]></category>
		<category><![CDATA[frequentism]]></category>
		<category><![CDATA[garden of forking paths]]></category>
		<category><![CDATA[multiple hypothesis testing]]></category>
		<category><![CDATA[null hypothesis]]></category>
		<category><![CDATA[null hypothesis significance testing]]></category>
		<category><![CDATA[p-hacking]]></category>
		<category><![CDATA[p-values]]></category>
		<category><![CDATA[probability]]></category>
		<category><![CDATA[publication bias]]></category>
		<category><![CDATA[random chance]]></category>
		<category><![CDATA[replication crisis]]></category>
		<category><![CDATA[statistical significance]]></category>
		<category><![CDATA[statistics]]></category>
		<category><![CDATA[underpowered]]></category>
		<guid isPermaLink="false">https://www.spencergreenberg.com/?p=3382</guid>

					<description><![CDATA[There is a tremendous amount of confusion around what a p-value actually is, despite their widespread use in science. Here is my attempt to explain the concept of p-values concisely and clearly (including why they are useful and what often goes wrong with them). — What&#8217;s a p-value? — If you run a study, then [&#8230;]]]></description>
										<content:encoded><![CDATA[
<p>There is a tremendous amount of confusion around what a p-value actually is, despite their widespread use in science. Here is my attempt to explain the concept of p-values concisely and clearly (including why they are useful and what often goes wrong with them).</p>



<p><strong>— What&#8217;s a p-value? —</strong></p>



<p>If you run a study, then (all else equal, aside from rare edge cases) the lower the p-value, the lower the chance that your results are due to random chance or luck.</p>



<p>More precisely: a p-value is the probability you&#8217;d get a result at least as extreme as what you got IF there were actually no effect (or if some other pre-specified &#8220;null hypothesis&#8221; is true).</p>



<p>So it&#8217;s a probability calculated based on assuming that there is no effect (or assuming that a pre-specified &#8220;null hypothesis&#8221; is true). Here the phrase &#8220;no effect&#8221; would mean, in the case of a study on a new medicine, that the medicine doesn&#8217;t do anything.</p>



<p>To put it in terms of coin flips: suppose you&#8217;re trying to decide if a coin is fair (i.e., if it has an equal chance of landing on heads and tails &#8211; so that&#8217;s your &#8220;null hypothesis&#8221; in this context). You flip the coin 100 times and get 60 heads. You calculate the p-value (p=0.06).</p>



<p>This p-value tells you there&#8217;s a 6% chance you&#8217;d get 60 or more heads OR 60 or more tails out of 100 flips if the coin were actually fair.</p>



<p>What makes p-values useful is that when they are high, you usually can&#8217;t rule out your effect being due to random chance or luck. And, when they are very low, random chance is (in most cases) unlikely to be the explanation for your result.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><strong>— What&#8217;s the problem with p-values? —</strong></p>



<p>In social science, p&lt;0.05 is often used as the cutoff for a &#8220;successful&#8221; result (i.e., they treat the effect as real and potentially publishable). This is an arbitrary cutoff; there&#8217;s nothing special about 0.05. The phrase &#8220;statistically significant&#8221; is defined simply to mean that p&lt;0.05.</p>



<p>There are many ways that p-values get commonly misused, creating lots of problems. For instance:</p>



<p>• p-values often get misinterpreted as the probability that an effect is not real (recall: p-values are actually the probability of getting a result this extreme if there is no effect, which is not the same thing)</p>



<p>• If you see one study where the main finding&#8217;s p-value is, say, 0.05, and another study where the main finding&#8217;s p-value is, say, 0.01, it&#8217;s tempting to conclude that the finding of the 2nd study is much less likely to be the result of chance (e.g., 1/5th as likely) than the 1st study&#8217;s finding. Unfortunately, we can&#8217;t draw this conclusion. The probability that a study&#8217;s finding is the result of chance is not the same as the p-value, and in fact, it can&#8217;t even be calculated just by knowing the p-value.</p>



<p>• Because a p-value threshold is often used for a result to be publishable (p&lt;0.05 in social science), researchers sometimes engage in fishy methods to get their p-values below the threshold. This is known as &#8220;p-hacking.:</p>



<p>• A result&#8217;s p-value (or &#8220;statistical significance&#8221;) is sometimes focused on instead of focusing on other factors that are also important. For instance, a result may have a low p-value but be such a weak effect that it&#8217;s totally useless or uninteresting.</p>



<p>• While a low p-value helps you rule out the possibility that your effect is merely due to random chance, unfortunately, that&#8217;s all it helps you with. But researchers sometimes act as though it tells them more than that. Even an extremely low p-value doesn&#8217;t mean an effect is &#8220;real&#8221; or that the effect means what you think. Low p-values can result from a variety of causes, including mistakes in experimental design or confounds.</p>



<p>Here&#8217;s another way to think about what a p-value is and isn&#8217;t that some people find helpful: a p-value does not tell you the probability that your result is due to chance. It tells you how consistent your results are with being due to chance. (I&#8217;m paraphrasing from <a href="https://statmodeling.stat.columbia.edu/2013/03/12/misunderstanding-the-p-value/#comment-143473">here</a>.) So, the lower the p-value, the less consistent your results are with them being due to chance.</p>



<p>It&#8217;s interesting to note that, empirically, results with lower p-values are more likely to be genuine effects (i.e., not false positives). I looked at results for 325 psychology study replications, and when the original study p-value was at most 0.01, about 72% replicated. When p&gt;0.01, only 48% did.</p>



<p>Ultimately, p-values are a useful (though often abused) statistical tool.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><strong>— BONUS APPENDIX: what&#8217;s the chance of a hypothesis being &#8220;true&#8221; if p&lt;0.05?  —</strong></p>



<p>One annoying thing about p-values is that they don&#8217;t answer the question we are usually interested in. Usually, we want to know something like &#8220;What&#8217;s the probability that my hypothesis is true?&#8221; or &#8220;What&#8217;s the probability that the effect of this drug is bigger than X?&#8221; but p-values don&#8217;t tell us those things.</p>



<p>However, we can put a different spin on p-values to get them to answer questions that are closer to what we&#8217;re really interested in. Let&#8217;s think of p-values as giving us a decision procedure (in an overly simplified world where you either &#8220;believe&#8221; in an effect or you fail to believe in it).&nbsp;</p>



<p>Suppose you test 100 totally separate, previously unexplored hypotheses about humans, and suppose that you commit to &#8220;believe&#8221; a hypothesis is true if and only if you get p&lt;0.05 (and otherwise, you don&#8217;t believe it).</p>



<p>I think it&#8217;s realistic that in a social science context, most hypotheses studied will be false since discovering novel, publishable hypotheses about humans is hard. So let&#8217;s suppose that 80% of the hypotheses you test are *not* true.&nbsp;</p>



<p>Finally, suppose that you use a large enough number of participants in your studies so that if you are testing for the presence of a real effect, there is an 80% chance you&#8217;ll be able to find it (this 80% figure is a common recommendation for &#8220;statistical power&#8221;).&nbsp;</p>



<p>Under these assumptions, if you test 100 hypotheses, then you will end up believing in 20 hypotheses, and 80% of those you believe will be true (with the other 20% being false positives). That means that of the results you believe in, 80% will be correct! Of course, this assumes no mistakes are made in the process of designing the experiment, running the statistics, and so on.</p>



<p>Here&#8217;s how the math works out if you&#8217;re curious:</p>



<p>• Out of the 100 hypotheses, 20 will be true, and of those, you&#8217;ll believe 16 = 0.80 * 20 (these are the true positives) and fail to believe 4 (these are the false negatives).</p>



<p>• Out of the 100 hypotheses, 80 will be false, and of those, you&#8217;ll believe 4 = 0.05 * 80 (these are the false positives), and you&#8217;ll reject 76 (these are the true negatives).</p>



<p>Of course, if the numbers here had been different, the conclusions would be different as well. For instance, imagine if you started with 2000 hypotheses, and this time, imagine that only 1% of them were true. If the power was still 80%, then:</p>



<p>&nbsp;• Out of the 2000 hypotheses, 20 of them would be true, and of those, you&#8217;d believe 16 (0.80 * 20) of them (these are true positives) and fail to believe 4 of them (these are false negatives).</p>



<p>• Out of the 2000 hypotheses, 1980 would be false, and of those, you&#8217;d believe 99 (0.05*1980) of them (these are false positives), and you&#8217;d reject the other 1881 of them (these are true negatives).</p>



<p>• So, altogether, you&#8217;d believe 115 (16 + 99) hypotheses, of which only 16 would&#8217;ve actually been true, so of the results you believe in, less than 14% would be correct!&nbsp;</p>



<p>From analyses like these, we can see that the probability that a specific hypothesis is true, given that we&#8217;ve found p&lt;0.05, depends on a variety of factors, including the sample size, the true effect size, the base rate probability that a new hypothesis tested by that researcher is true, the probability of errors being made in the experimental design or statistical analysis, and so on.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p>In real life:</p>



<p>(1) Studies often don&#8217;t use large enough numbers of participants (and so are underpowered).</p>



<p>(2) Researchers sometimes engage in p-hacking to artificially lower their p-values to help their papers get published.</p>



<p>(3) Researchers often don&#8217;t carefully track how many hypotheses they&#8217;ve really tested.</p>



<p>(4) The decision procedure described above is often not adhered to so strictly (e.g., a result of p=0.08 might be treated as suggestive evidence for the hypothesis, and hence the hypothesis is not rejected).</p>



<p>(5) Real hypotheses often have auxiliary assumptions beyond what the p-value accounts for (such as an assumption that there is a lack of confounders, a lack of serious errors in the experimental setup, and so on).</p>



<p>I personally don&#8217;t like thinking in terms of this decision procedure for p-values because I believe that modeling hypotheses as &#8220;true&#8221; or &#8220;false&#8221; is not a good approach to thinking clearly. This is because I believe it&#8217;s usually much better to think in terms of probabilities rather than a &#8220;true&#8221;/&#8221;false&#8221; dichotomy when trying to understand the answers to complex questions.</p>



<p>Some people have argued that we should switch to a Bayesian approach to hypothesis testing since such an approach avoids many of the issues of p-values (including avoiding the problematic &#8220;true&#8221;/&#8221;false&#8221; dichotomy). But it also introduces other challenges, such as how to come up with an appropriate &#8220;prior&#8221; (which represents one&#8217;s belief about the probability of the hypothesis having different strengths of effects prior to seeing the study results).</p>



<p></p>



<p><em>This piece was first written on December 31, 2022, and first appeared on this site on April 2, 2023.</em></p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p><a href="https://www.guidedtrack.com/programs/4zle8q9/run?essaySpecifier=%3A+Demystifying%20p-values" target="_blank" rel="noreferrer noopener">If you read this line, please do us a favor and click here to answer one quick question.</a></p>



<p></p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.spencergreenberg.com/2022/12/demystifying-p-values/feed/</wfw:commentRss>
			<slash:comments>1</slash:comments>
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">3382</post-id>	</item>
		<item>
		<title>Bias based on facial attractiveness</title>
		<link>https://www.spencergreenberg.com/2020/07/bias-based-on-facial-attractiveness/</link>
					<comments>https://www.spencergreenberg.com/2020/07/bias-based-on-facial-attractiveness/#respond</comments>
		
		<dc:creator><![CDATA[admin]]></dc:creator>
		<pubDate>Fri, 03 Jul 2020 03:15:00 +0000</pubDate>
				<category><![CDATA[Essays]]></category>
		<category><![CDATA[cultural norms]]></category>
		<category><![CDATA[cultural values]]></category>
		<category><![CDATA[discrimination based on appearance]]></category>
		<category><![CDATA[discrimination based on faces]]></category>
		<category><![CDATA[evolution]]></category>
		<category><![CDATA[false positives]]></category>
		<category><![CDATA[gender]]></category>
		<category><![CDATA[health]]></category>
		<category><![CDATA[individual variation]]></category>
		<category><![CDATA[injustice]]></category>
		<category><![CDATA[lookism]]></category>
		<category><![CDATA[prediction]]></category>
		<category><![CDATA[selection pressures]]></category>
		<category><![CDATA[sexual attraction]]></category>
		<category><![CDATA[testosterone]]></category>
		<guid isPermaLink="false">https://www.spencergreenberg.com/?p=2542</guid>

					<description><![CDATA[There&#8217;s a deeply-rooted, incredibly superficial aspect of human nature that is rarely discussed: our obsession with small variations in bone structure/skin smoothness on a person&#8217;s face. At extremes, people are desired or shunned due to tiny, otherwise almost meaningless facial details. In the attached image, there are two non-existent women (generated by a face generation [&#8230;]]]></description>
										<content:encoded><![CDATA[
<p>There&#8217;s a deeply-rooted, incredibly superficial aspect of human nature that is rarely discussed: our obsession with small variations in bone structure/skin smoothness on a person&#8217;s face. At extremes, people are desired or shunned due to tiny, otherwise almost meaningless facial details.</p>



<p>In the attached image, there are two non-existent women (generated by a face generation AI set to generate &#8220;brown hair white adult female&#8221;). If these were real people, they would likely be treated differently throughout their lives due to very minor differences in facial structure and skin smoothness.</p>



<p>Based on their faces alone, there&#8217;s no way to know with non-negligible accuracy which of these people (if they existed) would be more hard-working, more moral, wiser, or otherwise in possession of personal traits that we actually might care about. So why are humans so obsessed with faces? It seems likely to be caused by a combination of two factors:</p>



<hr class="wp-block-separator"/>



<p>(<strong>1) Runaway Sexual Selection</strong></p>



<p>If peacocks find large tail plumage sexually attractive, then even if those feathers are not useful for anything else, that still creates an evolutionary selection pressure where those with larger tail plumage are more likely to pass on their genes (due to improved chances of mating). Similarly, if certain humans are found to be more attractive based on their faces, that creates an evolutionary selection pressure in favor of mating with those people because then their children have a higher probability of finding mating success themselves (and hence passing on their genes). This phenomenon reinforces faces being attractive (because those attracted to &#8220;good-looking&#8221; faces mate with &#8220;good-looking&#8221; people more often, therefore their children are more good-looking and so have an easier time mating, plus have a preference for &#8220;good-looking&#8221; faces).</p>



<p>Today, this selection pressure is likely much weaker than it once was since most people now end up having children. For instance, now the vast majority of people in the US live to be at least 50, and only about 15% of women and 25% of men in the 40-50-year age bracket are childless. In contrast, tens of thousands of years ago, far fewer would make it to the point where they would have children.</p>



<hr class="wp-block-separator"/>



<p></p>



<p><strong>(2) Health Correlations</strong></p>



<p>In the environment we lived in tens of thousands of years ago, some aspects of a person&#8217;s face correlated with the likelihood of the survival of their genes, in particular ones related to disease (some diseases impact the face), genetic disorders (some of them cause facial changes), and development in the womb (where abnormal development can cause facial changes). </p>



<p>The correlation between health and facial features is likely to be lower now than it used to be back then. Today, a person&#8217;s facial features might still help to predict someone&#8217;s age, their most probable gender identity, and whether they have certain health conditions &#8211; but, of course, none of these give us any legitimate justification for treating some people better and others worse based just on their face.</p>



<p>It has been found that certain facial features do correlate with hormone levels (like testosterone). While testosterone levels may play a role in aggression (they may be part of the explanation for why men are violent so much more often than women), using these small correlations to make judgments about any one person is going to be both highly inaccurate and highly unjust. Some other personality traits may also be very weakly correlated with a person&#8217;s facial features, but talking to the person for 20 minutes will, of course, give you dramatically more information about what that person is like. Yet, we are prone to read so much into the way a person looks.</p>



<hr class="wp-block-separator"/>



<p><strong>Note: </strong>there is an additional effect when it comes to faces, which is that we are sometimes taught by our culture to value certain facial attributes more than others. This can act above and beyond the previously mentioned two factors.</p>



<hr class="wp-block-separator"/>



<p>We humans act as though faces are incredibly important despite them being a substantially arbitrary mask our genes have programmed for us. And they often impact how we humans treat each other, despite this unequal treatment being both unjust and unjustified. If you ever notice yourself treating someone less well because of their face, take note and adjust your behavior.</p>



<p>I am not saying that people should, for example, date people they are not attracted to. Obviously, attraction is an important part of relationships for most people, and the face is one part of what determines attraction. (You may also care about your children one day having attractive faces, so they can more easily find life partners they like.) Rather, what I&#8217;m saying is that we should be very wary about making negative inferences about any individual person based on their face (which is something that, unfortunately, the human mind seems to do often). The face says too little about a person&#8217;s character to be useful for predicting at the level of any individual.</p>



<hr class="wp-block-separator"/>



<p><em>This essay was first written on July 2, 2020, and first appeared on this site on December 17, 2021.</em></p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.spencergreenberg.com/2020/07/bias-based-on-facial-attractiveness/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">2542</post-id>	</item>
	</channel>
</rss>
