{"id":432,"date":"2012-05-16T16:26:25","date_gmt":"2012-05-16T21:26:25","guid":{"rendered":"http:\/\/blog.richmond.edu\/physicsbunn\/?p=432"},"modified":"2012-05-17T10:03:09","modified_gmt":"2012-05-17T15:03:09","slug":"nature-really-hates-psychology","status":"publish","type":"post","link":"https:\/\/blog.richmond.edu\/physicsbunn\/2012\/05\/16\/nature-really-hates-psychology\/","title":{"rendered":"Nature really hates psychology"},"content":{"rendered":"<p>That&#8217;s <em>Nature<\/em>\u00a0the premier science journal, not the nature that abhors a vacuum. Check out this <a href=\"http:\/\/www.nature.com\/news\/replication-studies-bad-copy-1.10634\">news feature<\/a> on problems in the field.<\/p>\n<blockquote><p>These problems occur throughout the sciences, but psychology has a number of deeply entrenched cultural norms that exacerbate them. It has become common practice, for example, to tweak experimental designs in ways that practically guarantee positive results. And once positive results are published, few researchers replicate the experiment exactly, instead carrying out &#8216;conceptual replications&#8217; that test similar hypotheses using different methods. This practice, say critics, builds a house of cards on potentially shaky foundations.<\/p><\/blockquote>\n<p>I&#8217;m not a psychologist, so I can&#8217;t be sure how much merit there is in the article&#8217;s indictments. But don&#8217;t worry &#8212; I won&#8217;t let that stop me from commenting!<\/p>\n<p>The article seems to me to make three main claims:<\/p>\n<ol>\n<li>&#8220;Publication bias&#8221; is a serious problem &#8212; results that show something positive and surprising are much more likely to be published.<\/li>\n<li>Psychologists often use dishonest statistical methods (perhaps unintentionally).<\/li>\n<li>People don&#8217;t replicate previous results exactly.<\/li>\n<\/ol>\n<p>Let&#8217;s take each in turn.<\/p>\n<p><strong>1. Publication bias.<\/strong><\/p>\n<p>There&#8217;s no doubt that experiments that show a positive result are more highly valued than experiments that don&#8217;t show one. It&#8217;s much better for your career to find something than to find nothing. That&#8217;s bound to lead to a certain amount of &#8220;publication bias&#8221;: the published literature will contain many more false positives than you&#8217;d naively expect, because the false positives are more likely to get published.<\/p>\n<p>This problem is exacerbated by the fact that people often use the ridiculously low threshold of 95% confidence to decide whether a result is statistically significant. This means that you expect one out of every 20 tests you do to yield a false positive result. There are lots of people doing lots of tests, so of course there are lots of false positives. That&#8217;s why, according to John Ioannidis, &#8220;<a href=\"http:\/\/www.plosmedicine.org\/article\/info:doi\/10.1371\/journal.pmed.0020124\">most published [medical] research findings are false<\/a>.&#8221;<\/p>\n<p>The <em>Nature<\/em>\u00a0piece claims that these problems are worse in psychology than in other fields. I don&#8217;t know if that&#8217;s true or not. Some evidence from the article:<\/p>\n<blockquote><p>Psychology and psychiatry, according to other work by Fanelli, are the worst offenders: they are five times more likely to report a positive result than are the space sciences, which are at the other end of the spectrum. The situation is not improving. In 1959, statistician Theodore Sterling found that 97% of the studies in four major psychology journals had reported statistically significant positive results5. When he repeated the analysis in 1995, nothing had changed.<\/p>\n<p>One reason for the excess in positive results for psychology is an emphasis on \u201cslightly freak-show-ish\u201d results, says Chris Chambers, an experimental psychologist at Cardiff University, UK. \u201cHigh-impact journals often regard psychology as a sort of parlour-trick area,\u201d he says. Results need to be exciting, eye-catching, even implausible. Simmons says that the blame lies partly in the review process. \u201cWhen we review papers, we&#8217;re often making authors prove that their findings are novel or interesting,\u201d he says. \u201cWe&#8217;re not often making them prove that their findings are true.\u201d<\/p><\/blockquote>\n<p>Incidentally, the article has a graphic illustrating the fraction of times that papers in different disciplines quote positive results, but it doesn&#8217;t seem to support the claim that psychology is five times worse than space science:<\/p>\n<p>&nbsp;<\/p>\n<p><a href=\"http:\/\/blog.richmond.edu\/physicsbunn\/files\/2012\/05\/replication_graph.jpg\"><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-medium wp-image-434\" src=\"http:\/\/blog.richmond.edu\/physicsbunn\/files\/2012\/05\/replication_graph-236x300.jpg\" alt=\"\" width=\"236\" height=\"300\" srcset=\"https:\/\/blog.richmond.edu\/physicsbunn\/files\/2012\/05\/replication_graph-236x300.jpg 236w, https:\/\/blog.richmond.edu\/physicsbunn\/files\/2012\/05\/replication_graph.jpg 300w\" sizes=\"auto, (max-width: 236px) 100vw, 236px\" \/><\/a><\/p>\n<p>I estimate psychology is at about 93% and space science is at about 70%. That&#8217;s not a factor of 5.<\/p>\n<p><strong>2. Dishonest statistics.<\/strong><\/p>\n<p><em>Nature <\/em>doesn&#8217;t use a loaded word like &#8220;dishonest,&#8221; but here&#8217;s what they claim:<\/p>\n<blockquote><p>Many psychologists make on-the-fly decisions about key aspects of their studies, including how many volunteers to recruit, which variables to measure and how to analyse the results. These choices could be innocently made, but they give researchers the freedom to torture experiments and data until they produce positive results.<\/p>\n<p>In a survey of more than 2,000 psychologists, Leslie John, a consumer psychologist from Harvard Business School in Boston, Massachusetts, showed that more than 50% had waited to decide whether to collect more data until they had checked the significance of their results, thereby allowing them to hold out until positive results materialize. More than 40% had selectively reported studies that \u201cworked\u201d. On average, most respondents felt that these practices were defensible. \u201cMany people continue to use these approaches because that is how they were taught,\u201d says Brent Roberts, a psychologist at the University of Illinois at Urbana\u2013Champaign.<\/p><\/blockquote>\n<p>If you deliberately choose to report data that lead to positive results and not data that lead to negative results, then you&#8217;re just plain lying about your data. <a href=\"http:\/\/allendowney.blogspot.com\/2011\/10\/repeated-tests-how-bad-can-it-be.html\">The same goes <\/a>for continuing to gather data, only stopping when &#8220;positive results materialize.&#8221;<\/p>\n<p>Once again, I am not claiming that psychologists do this. I am claiming that <em>if<\/em> they do this, as\u00a0<em>Nature<\/em>\u00a0claims, it&#8217;s a very serious problem.<\/p>\n<p><strong>3. People don&#8217;t replicate previous studies.<\/strong><\/p>\n<p>The <em>Nature<\/em>\u00a0article tells the story of the people who tried to replicate a previous study claiming to show precognition. Journals wouldn&#8217;t accept the paper detailing the failed replication. I <a href=\"http:\/\/blog.richmond.edu\/physicsbunn\/2011\/04\/29\/science-is-all-about-replication-of-results-right\/\">agree<\/a> with the author that this is a shame.<\/p>\n<p>But the article&#8217;s concerns on this seem quite overwrought to me. They point out that most of the time people don&#8217;t try to replicate studies exactly, instead performing &#8220;conceptual replications,&#8221; in which they do something similar but not identical to what&#8217;s been done before. The author seems to think that this is a problem, but I don&#8217;t really see why. Here&#8217;s his argument:<\/p>\n<blockquote><p>But to other psychologists, reliance on conceptual replication is problematic. \u201cYou can&#8217;t replicate a concept,\u201d says Chambers. \u201cIt&#8217;s so subjective. It&#8217;s anybody&#8217;s guess as to how similar something needs to be to count as a conceptual replication.\u201d The practice also produces a \u201clogical double-standard\u201d, he says. For example, if a heavy clipboard unconsciously influences people&#8217;s judgements, that could be taken to conceptually replicate the slow-walking effect. But if the weight of the clipboard had no influence, no one would argue that priming had been conceptually falsified. With its ability to verify but not falsify, conceptual replication allows weak results to support one another. \u201cIt is the scientific embodiment of confirmation bias,\u201d says Brian Nosek, a social psychologist from the University of Virginia in Charlottesville. \u201cPsychology would suffer if it wasn&#8217;t practised but it doesn&#8217;t replace direct replication. To show that &#8216;A&#8217; is true, you don&#8217;t do &#8216;B&#8217;. You do &#8216;A&#8217; again.\u201d<\/p><\/blockquote>\n<p>It&#8217;s true that a &#8220;conceptual replication&#8221; doesn&#8217;t directly falsify a particular result in the same way that a (failed) exact replication would. But I can&#8217;t bring myself to care that much. If a result is incorrect, it will gradually become clear that it doesn&#8217;t fit in with the pattern built up by many subsequent similar-but-not-identical experiments. The incorrect result will gradually atrophy from lack of attention, even if there&#8217;s not a single definitive refutation.<\/p>\n<p>At least, <a href=\"http:\/\/blog.richmond.edu\/physicsbunn\/2011\/06\/27\/error-correcting-in-science\/\">that&#8217;s the way I see things working<\/a> in physics, where direct replications of previous experiments are extremely uncommon.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>That&#8217;s Nature\u00a0the premier science journal, not the nature that abhors a vacuum. Check out this news feature on problems in the field. These problems occur throughout the sciences, but psychology has a number of deeply entrenched cultural norms that exacerbate them. It has become common practice, for example, to tweak experimental designs in ways that &hellip; <a href=\"https:\/\/blog.richmond.edu\/physicsbunn\/2012\/05\/16\/nature-really-hates-psychology\/\" class=\"more-link\">Continue reading <span class=\"screen-reader-text\">Nature really hates psychology<\/span><\/a><\/p>\n","protected":false},"author":12,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[1],"tags":[],"class_list":["post-432","post","type-post","status-publish","format-standard","hentry","category-uncategorized"],"jetpack_featured_media_url":"","_links":{"self":[{"href":"https:\/\/blog.richmond.edu\/physicsbunn\/wp-json\/wp\/v2\/posts\/432","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/blog.richmond.edu\/physicsbunn\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/blog.richmond.edu\/physicsbunn\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/blog.richmond.edu\/physicsbunn\/wp-json\/wp\/v2\/users\/12"}],"replies":[{"embeddable":true,"href":"https:\/\/blog.richmond.edu\/physicsbunn\/wp-json\/wp\/v2\/comments?post=432"}],"version-history":[{"count":0,"href":"https:\/\/blog.richmond.edu\/physicsbunn\/wp-json\/wp\/v2\/posts\/432\/revisions"}],"wp:attachment":[{"href":"https:\/\/blog.richmond.edu\/physicsbunn\/wp-json\/wp\/v2\/media?parent=432"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/blog.richmond.edu\/physicsbunn\/wp-json\/wp\/v2\/categories?post=432"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/blog.richmond.edu\/physicsbunn\/wp-json\/wp\/v2\/tags?post=432"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}