<?xml version="1.0" encoding="ISO-8859-1"?><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
<front>
<journal-meta>
<journal-id>0188-2198</journal-id>
<journal-title><![CDATA[Revista mexicana de cardiología]]></journal-title>
<abbrev-journal-title><![CDATA[Rev. Mex. Cardiol]]></abbrev-journal-title>
<issn>0188-2198</issn>
<publisher>
<publisher-name><![CDATA[Asociación Nacional de Cardiólogos de México, Sociedad de Cardiología Intervencionista de México]]></publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id>S0188-21982015000100008</article-id>
<title-group>
<article-title xml:lang="en"><![CDATA[Demystification of the significance of p in statistical tests]]></article-title>
<article-title xml:lang="es"><![CDATA[Desmitificación de la significancia de p en las pruebas estadísticas]]></article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname><![CDATA[Sánchez Turcios]]></surname>
<given-names><![CDATA[Reinaldo Alberto]]></given-names>
</name>
<xref ref-type="aff" rid="A01"/>
</contrib>
</contrib-group>
<aff id="A01">
<institution><![CDATA[,Hospital General Milpa Alta  ]]></institution>
<addr-line><![CDATA[Mexico City ]]></addr-line>
<country>Mexico</country>
</aff>
<pub-date pub-type="pub">
<day>00</day>
<month>03</month>
<year>2015</year>
</pub-date>
<pub-date pub-type="epub">
<day>00</day>
<month>03</month>
<year>2015</year>
</pub-date>
<volume>26</volume>
<numero>1</numero>
<fpage>56</fpage>
<lpage>58</lpage>
<copyright-statement/>
<copyright-year/>
<self-uri xlink:href="http://www.scielo.org.mx/scielo.php?script=sci_arttext&amp;pid=S0188-21982015000100008&amp;lng=en&amp;nrm=iso"></self-uri><self-uri xlink:href="http://www.scielo.org.mx/scielo.php?script=sci_abstract&amp;pid=S0188-21982015000100008&amp;lng=en&amp;nrm=iso"></self-uri><self-uri xlink:href="http://www.scielo.org.mx/scielo.php?script=sci_pdf&amp;pid=S0188-21982015000100008&amp;lng=en&amp;nrm=iso"></self-uri><abstract abstract-type="short" xml:lang="en"><p><![CDATA[All statistical tests have a p value that is significant when < 0.050. This value was arbitrarily determined by RA Fisher and accepted consensually over time. Since its genesis, this value has been questioned, and nowadays it is under the careful eye of many statisticians. This issue has led to a debate among the scientific community: obtaining p significance was considered as a guarantee that the research project would be an appropriate contrast between the hypothesis and the acceptance, or rejection, of it. The purpose of this paper is to construct a discussion about p significance.]]></p></abstract>
<abstract abstract-type="short" xml:lang="es"><p><![CDATA[Todas la pruebas estadísticas tienen un valor de p significativo a partir de < de 0.050, el cual fue arbitrariamente determinado por RA Fisher y aceptado por consenso a través del tiempo. Desde su génesis, este valor ha sido cuestionado y actualmente está bajo la mirada escrupulosa de muchos estadígrafos, por lo que se establece un debate en la comunidad científica donde clásicamente se consideraba obtener la significancia de p un sello de garantía, que el proyecto de investigación era capaz de aceptar o rechazar la hipótesis. El objetivo de este artículo es discutir los cuestionamientos de la significancia de p.]]></p></abstract>
<kwd-group>
<kwd lng="en"><![CDATA[Statistic test]]></kwd>
<kwd lng="en"><![CDATA[p value]]></kwd>
<kwd lng="en"><![CDATA[p significance]]></kwd>
<kwd lng="es"><![CDATA[Pruebas estadísticas]]></kwd>
<kwd lng="es"><![CDATA[valor de p]]></kwd>
<kwd lng="es"><![CDATA[significancia de p]]></kwd>
</kwd-group>
</article-meta>
</front><body><![CDATA[ <p align="justify"><font face="verdana" size="4">Comunicado breve</font></p>     <p align="justify"><font face="verdana" size="2">&nbsp;</font></p>     <p align="center"><font face="verdana" size="4"><b>Demystification of the significance of p in statistical tests</b></font></p>    <p align="justify"><font face="verdana" size="2">&nbsp;</font></p>     <p align="center"><font face="verdana" size="3"><b>Desmitificaci&oacute;n de la significancia de p en las pruebas estad&iacute;sticas</b></font></p>    <p align="justify"><font face="verdana" size="2">&nbsp;</font></p>     <p align="center"><font face="verdana" size="2"><b>Reinaldo Alberto S&aacute;nchez Turcios&#42;</b></font></p>    <p align="justify"><font face="verdana" size="2">&nbsp;</font></p>     <p align="justify"><font face="verdana" size="2">&#42; MD Endocrinology, MSc Pharmacology. Hospital General Milpa Alta. Mexico City, Mexico.</font></p>    <p align="justify"><font face="verdana" size="2">&nbsp;</font></p>     ]]></body>
<body><![CDATA[<p align="justify"><font face="verdana" size="2"><i>Address for correspondence:</i>    <br><b>Reinaldo Alberto S&aacute;nchez Turcios</b>    <br>Tepic N&uacute;m. 113-610,    <br>Col. Roma Sur, 06760, M&eacute;xico, Distrito Federal.     <br>Tel. 015552648061,    <br>Cel. 0445543508824,    <br>E-mail: <a href="mailto:rturcios@live.com.mx" target="_blank">rturcios@live.com.mx</a></font></p>    <p align="justify"><font face="verdana" size="2">&nbsp;</font></p>     <p align="justify"><font face="verdana" size="2"><b>ABSTRACT</b></font></p>     <p align="justify"><font face="verdana" size="2">All statistical tests have a p value that is significant when &lt; 0.050. This value was arbitrarily determined by RA Fisher and accepted consensually over time. Since its genesis, this value has been questioned, and nowadays it is under the careful eye of many statisticians. This issue has led to a debate among the scientific community: obtaining p significance was considered as a guarantee that the research project would be an appropriate contrast between the hypothesis and the acceptance, or rejection, of it. The purpose of this paper is to construct a discussion about p significance.</font></p>     ]]></body>
<body><![CDATA[<p align="justify"><font face="verdana" size="2"><b>Key words:</b> Statistic test, p value, p significance.</font></p>     <p align="justify"><font face="verdana" size="2">&nbsp;</font></p>     <p align="justify"><font face="verdana" size="2"><b>RESUMEN</b></font></p>     <p align="justify"><font face="verdana" size="2">Todas la pruebas estad&iacute;sticas tienen un valor de p significativo a partir de &lt; de 0.050, el cual fue arbitrariamente determinado por RA Fisher y aceptado por consenso a trav&eacute;s del tiempo. Desde su g&eacute;nesis, este valor ha sido cuestionado y actualmente est&aacute; bajo la mirada escrupulosa de muchos estad&iacute;grafos, por lo que se establece un debate en la comunidad cient&iacute;fica donde cl&aacute;sicamente se consideraba obtener la significancia de p un sello de garant&iacute;a, que el proyecto de investigaci&oacute;n era capaz de aceptar o rechazar la hip&oacute;tesis. El objetivo de este art&iacute;culo es discutir los cuestionamientos de la significancia de p.</font></p>     <p align="justify"><font face="verdana" size="2"><b>Palabras clave: </b>Pruebas estad&iacute;sticas, valor de p, significancia de p.</font></p>     <p align="justify"><font face="verdana" size="2">&nbsp;</font></p>     <p align="justify"><font face="verdana" size="2"><b>INTRODUCTION</b></font></p>     <p align="justify"><font face="verdana" size="2">It was RA Fisher, a British statistician and geneticist, who first proposed p significance formally in 1925 (which was arbitrarily determined and consensually accepted over time), it was published in Statistical Methods for Research Workers<sup>1</sup>. This p significance was not founded on scientific basis, but on the assumption that p was &lt; 0.050.<sup>2</sup> The significance of p received strong criticism from Jerzy Neyman, a Polish Mathematician, since its inception and from the British statistician Egon Person<sup>3</sup> in our time. Sthepen Ziliak,<sup>3</sup> an economist with scientific thought, criticizes the statistic tests used in research which frequently are inadequately used. However, there are defenders of "p significance".</font></p>     <p align="justify"><font face="verdana" size="2">P significance is 5%, but it is probably the most misunderstood and misinterpreted index calculated for research in different disciplines of knowledge.<sup>2-4</sup> In a recent survey to medical residents published by JAMA,<sup>4</sup> 88% of the people surveyed could not interpret correctly p value; it is for this reason that it was considered necessary a careful look at p significance.</font></p>     <p align="justify"><font face="verdana" size="2"></font></p>    ]]></body>
<body><![CDATA[<p align="justify"><font face="verdana" size="2">&nbsp;</font></p>    <p align="justify"><font face="verdana" size="2">  <b>METHODOLOGY</b></font></p>     <p align="justify"><font face="verdana" size="2">In 12 papers and research reports that appeared in Science (December 14<sup>th</sup> 2014) and in 20 out of 22 papers published in the Journal of Pharmacology and Experimental Therapeutics (December 2012), p values are mentioned when describing experimental results. These values are considered valid even when several of them are at the threshold of their significance in those studies.</font></p>     <p align="justify"><font face="verdana" size="2"><b>Nature of "p significance" in statistic tests</b></font></p>     <p align="justify"><font face="verdana" size="2">The significance of p is defined as the probability that Ho is true according to the research methodology whose hypothesis is proposed to demonstrate.</font></p>     <p align="justify"><font face="verdana" size="2">One of the objectives of this paper is to clarify the steps that R.A. Fisher described originally: 1. "p significance" is not definite; 2. RA Fisher tried to simplify the procedure of the test to make a judgment on the evidence, and that it would be true; 3. He stated that a set of experiments should be carried out to observe whether the results were consistent with randomness; 4. The researchers must first postulate a Ho that must be rejected, such as the existence of correlation, or difference between the two groups being analyzed; 5. The researchers must assume that Ho is in fact true and calculate the observed value with the data obtained, and they have to see how distant it is from a p value previously established and 6. Fisher suggested that if p had a smaller magnitude than the set value, then Ho would be false.</font></p>     <p align="justify"><font face="verdana" size="2">Once the Ho and the Ha had been specified, research should define the level of significance which is usually known as &alpha; &lt; 0.050. This rejects Ho, that is to say if the p value calculated with the data from the two samples is less than &alpha; value previously established, then Ho is rejected. In this step is where the problem of p interpretation resides.</font></p>     <p align="justify"><font face="verdana" size="2"></font></p>    <p align="justify"><font face="verdana" size="2">&nbsp;</font></p>    <p align="justify"><font face="verdana" size="2">  <b>DISCUSSION</b></font></p>     ]]></body>
<body><![CDATA[<p align="justify"><font face="verdana" size="2">The main restriction of p value is that the behavior of the results cannot be visualized simultaneously, in the long or short term (over time), of the treatments whose data will be subjected to statistic tests.</font></p>     <p align="justify"><font face="verdana" size="2">The deductive method is used, in the long term, to obtain these data. If we hypothetically carried out a repetition of the experiment, the results should be consistent with the p value obtained. When designing a new drug, many samples are analyzed for a given pathology, and the results are interpreted in a global fashion.</font></p>     <p align="justify"><font face="verdana" size="2">One of the problems is that the repetition of the drug usage in different samples is not the same; it is assumed that these data would be consistent with the p value found in preliminary studies. Other way to look at the problem is when you do a single experiment (proving Ho); this study is inductive and establishes only one result as evidence, expecting to generalize the obtained result and replicate it in other samples.<sup>5,6</sup> The latter proposition comes true when pretending to describe, with one short term study, the behavior in the long term of a new drug, which will replace the one in use.</font></p>     <p align="justify"><font face="verdana" size="2"><b>Main criticisms of p value</b></font></p>     <p align="justify"><font face="verdana" size="2">The main criticisms of p value are:</font></p>     <p align="justify">    <blockquote><font face="verdana" size="2"> 1) P values is not a usual measurement for inference.<sup>7</sup></font></p>     <p align="justify"><font face="verdana" size="2">2) There are at least 12 wrong interpretations of p significance.<sup>5,6,8,9</sup></font></p>     <p align="justify"><font face="verdana" size="2">3) P values do not offer exact evidence of sample differences. P values are affected by the size of the samples; if you increase the size of a sample, p value will change from significant to non-significant.<sup>9</sup> </font></blockquote></p>     <p align="justify"><font face="verdana" size="2">Other factors that can be observed of p values misinterpretation in research papers are:<sup>10</sup></font></p>     ]]></body>
<body><![CDATA[<p align="justify">    <blockquote><font face="verdana" size="2"> a) The smaller the size of the sample, the more likely p values are wrong. </font></p>     <p align="justify"><font face="verdana" size="2">b) The smaller the sample effect size, the less likely the obtained results are truthful. (Size effect analysis).</font></p>     <p align="justify"><font face="verdana" size="2">c) If the preliminary studies are scarce and demonstrative of the studied effect association, it is more likely that the findings will be truthful.</font></p>     <p align="justify"><font face="verdana" size="2">d) The flexibility, design, definitions, favorable events, and analytic methods used in the research may lead us to false results.</font></p>     <p align="justify"><font face="verdana" size="2">e) The more controversial the study (with a larger team of researchers involved), the more likely the findings are false. </font></blockquote></p>     <p align="justify"><font face="verdana" size="2"><b>Effect size</b><sup>10</sup></font></p>     <p align="justify"><font face="verdana" size="2">The effect size, that a researcher hypothesizes to exist between two patients subjected to two different treatments, is a value that can be referred as the heterogeneous difference in the effects of both treatments.</font></p>     <p align="justify"><font face="verdana" size="2">The effect size is the difference that the researcher proposes; whether there is a/or several effects between two treatments in a set of patients with a common pathology. Its importance in the statistical analysis is that with this value (effect size), significance and a pre-established sample size, the researcher is able to calculate the power of its statistical test, that is to say: what probability the result has to reject Ho with the available data if the statistical test has a very small power (usually less than 80%). In order to improve statistical power, the size of the sample has to be increased due to the significance level previously assigned and the effect size proposed. This latter premise is relevant since in most protocols the obtained results when rejecting Ho are not important, because the statistical analysis was carried without an acceptable power. The greater the power in a statistical design, the bigger the likelihood of Ho rejection when Ho is false.</font></p>     <p align="justify"><font face="verdana" size="2"><b>Statistical power depends on:</b></font></p>     ]]></body>
<body><![CDATA[<p align="justify">    <blockquote><font face="verdana" size="2"> 1. Sample size.</font></p>     <p align="justify"><font face="verdana" size="2">2. The previously determined &alpha; significance level; it is usually from 0.050 to 0.01. It is important to point out that when the effect size is very small (the researchers expect that the investigated drug will not modify the previously established treatment), the sample sizes for the fixed &alpha; significance level will be bigger.</font></p>     <p align="justify"><font face="verdana" size="2">Conversely, if one knows that the effect size is big, the sample size is relatively small: 1) &lt; 0.20: very small, 2) 0.030-0.050: small, 3) 0.050-0.70: medium, 4) &gt; 0.80: big. It is common to assign an effect size of 1.0 to 2.0.</font></p>     <p align="justify"><font face="verdana" size="2">Cohen,<sup>11</sup> the creator of this concept, makes clear that the terms: small, medium, and big have to be interpreted within the concept of the statistical analysis being concreted. </font></blockquote></p>     <p align="justify"><font face="verdana" size="2">Conclusions</font></p>     <p align="justify"><font face="verdana" size="2">P significance is still valid if the originally proposed procedures by RA Fisher are executed with a strict scientific discipline when obtaining a set of p values.</font></p>     <p align="justify"><font face="verdana" size="2"><b>Outlook.</b> An adequate knowledge of the effect size is proposed to be able to raise an appropriate sample size with proper significance level, and statistical power of a correct analysis. It is likely that Bayesian statistics could solve the problem of the different sizes.</font></p>     <p align="justify"><font face="verdana" size="2">&nbsp;</font></p>    <p align="justify"><font face="verdana" size="2"><b>REFERENCIAS</b></font></p>    ]]></body>
<body><![CDATA[<!-- ref --><p align="justify"><font face="verdana" size="2">1. Fisher RA. <i>Statistical methods for research workers</i>. 5 de Biological monographs and manuals. La Universidad de California, 1932, pp. 1-307.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=7720963&pid=S0188-2198201500010000800001&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>     <!-- ref --><p align="justify"><font face="verdana" size="2">2. Lew MJ. To P or not to P: on the evidential nature of P-values and their place in scientific inference stat. <i>ME</i>. 2013; 3: 27-46.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=7720965&pid=S0188-2198201500010000800002&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>     <!-- ref --><p align="justify"><font face="verdana" size="2">3. Nuzzo R. P values, the "gold standard" of statistical validity, are not as reliable as many scientists assume. <i>Nature</i>. 2014; 506: 150-152.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=7720967&pid=S0188-2198201500010000800003&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>     <!-- ref --><p align="justify"><font face="verdana" size="2">4. Donna M, Windish MD, Stephen J, Huot SJ, Green ML. Medicine residents' understanding of the biostatistics and results in the medical literature. <i>JAMA</i>. 2007; 298: 1010-1022.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=7720969&pid=S0188-2198201500010000800004&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>     <!-- ref --><p align="justify"><font face="verdana" size="2">5. Lecoutre MP, Poitevineau J, Lecoutre B. Even statisticians are not immune to misinterpretations of null hypothesis significance tests. <i>Int J Psychol</i>. 2003; 38: 37-45.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=7720971&pid=S0188-2198201500010000800005&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>     ]]></body>
<body><![CDATA[<!-- ref --><p align="justify"><font face="verdana" size="2">6. Hubbard R, Lindsay MR. Why p values are not a useful measure of evidence in statistical significance testing. <i>Theory Psychol</i>. 2008; 18: 69-88.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=7720973&pid=S0188-2198201500010000800006&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>     <!-- ref --><p align="justify"><font face="verdana" size="2">7. Lew MJ. Bad statistical practice in pharmacology (and other basic biomedical disciplines): you probably don't know P. <i>Br J Pharmacol</i>. 2012; 166: 1559-1567.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=7720975&pid=S0188-2198201500010000800007&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>     <!-- ref --><p align="justify"><font face="verdana" size="2">8. Goodman S. A dirty dozen: twelve p-value misconceptions. <i>Semin Hematol</i>. 2008; 45: 135-140.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=7720977&pid=S0188-2198201500010000800008&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>     <!-- ref --><p align="justify"><font face="verdana" size="2">9. Hurlbert SH, Lombardi CM. Final collapse of the Neyman-Pearson decision theoretic framework and rise of the neo Fisherian. <i>Ann Zool Fennici</i>. 2009; 46: 311-349.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=7720979&pid=S0188-2198201500010000800009&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>     <!-- ref --><p align="justify"><font face="verdana" size="2">10. Coe R. <i>It's the effect size, stupid. What effect size is and why it is important. School of education, University of Durham</i>. Paper presented at the British Educational Research Association, Annual Conference. Exter, 12-14 September, 2002.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=7720981&pid=S0188-2198201500010000800010&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>     ]]></body>
<body><![CDATA[<!-- ref --><p align="justify"><font face="verdana" size="2">11. Cohen J. Statistical power analysis for the behavioral sciences. New York, Academic Press. 1977, pp. 216-380.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=7720983&pid=S0188-2198201500010000800011&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>     <p align="justify"><font face="verdana" size="2"></font></p>    <p align="justify"><font face="verdana" size="2">&nbsp;</font></p>    <p align="justify"><font face="verdana" size="2">      <br> <b>Nota</b>     <br>      <br> Este art&iacute;culo puede ser consultado en versi&oacute;n completa en: <a href="http://www.medigraphic.com/revmexcardiol" target="_blank">http://<b>www.medigraphic.com/revmexcardiol</b></a></font></p>       ]]></body><back>
<ref-list>
<ref id="B1">
<label>1</label><nlm-citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Fisher]]></surname>
<given-names><![CDATA[RA.]]></given-names>
</name>
</person-group>
<source><![CDATA[Statistical methods for research workers]]></source>
<year>1932</year>
<page-range>1-307</page-range><publisher-name><![CDATA[La Universidad de California]]></publisher-name>
</nlm-citation>
</ref>
<ref id="B2">
<label>2</label><nlm-citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Lew]]></surname>
<given-names><![CDATA[MJ.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[To P or not to P: on the evidential nature of P-values and their place in scientific inference stat]]></article-title>
<source><![CDATA[ME]]></source>
<year>2013</year>
<volume>3</volume>
<page-range>27-46</page-range></nlm-citation>
</ref>
<ref id="B3">
<label>3</label><nlm-citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Nuzzo]]></surname>
<given-names><![CDATA[R.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[P values, the "gold standard" of statistical validity, are not as reliable as many scientists assume]]></article-title>
<source><![CDATA[Nature]]></source>
<year>2014</year>
<volume>506</volume>
<page-range>150-152</page-range></nlm-citation>
</ref>
<ref id="B4">
<label>4</label><nlm-citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Donna]]></surname>
<given-names><![CDATA[M]]></given-names>
</name>
<name>
<surname><![CDATA[Windish]]></surname>
<given-names><![CDATA[MD]]></given-names>
</name>
<name>
<surname><![CDATA[Stephen]]></surname>
<given-names><![CDATA[J]]></given-names>
</name>
<name>
<surname><![CDATA[Huot]]></surname>
<given-names><![CDATA[SJ]]></given-names>
</name>
<name>
<surname><![CDATA[Green]]></surname>
<given-names><![CDATA[ML.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Medicine residents' understanding of the biostatistics and results in the medical literature]]></article-title>
<source><![CDATA[JAMA]]></source>
<year>2007</year>
<volume>298</volume>
<page-range>1010-1022</page-range></nlm-citation>
</ref>
<ref id="B5">
<label>5</label><nlm-citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Lecoutre]]></surname>
<given-names><![CDATA[MP]]></given-names>
</name>
<name>
<surname><![CDATA[Poitevineau]]></surname>
<given-names><![CDATA[J]]></given-names>
</name>
<name>
<surname><![CDATA[Lecoutre]]></surname>
<given-names><![CDATA[B.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Even statisticians are not immune to misinterpretations of null hypothesis significance tests]]></article-title>
<source><![CDATA[Int J Psychol]]></source>
<year>2003</year>
<volume>38</volume>
<page-range>37-45</page-range></nlm-citation>
</ref>
<ref id="B6">
<label>6</label><nlm-citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Hubbard]]></surname>
<given-names><![CDATA[R]]></given-names>
</name>
<name>
<surname><![CDATA[Lindsay]]></surname>
<given-names><![CDATA[MR.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Why p values are not a useful measure of evidence in statistical significance testing]]></article-title>
<source><![CDATA[Theory Psychol]]></source>
<year>2008</year>
<volume>18</volume>
<page-range>69-88</page-range></nlm-citation>
</ref>
<ref id="B7">
<label>7</label><nlm-citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Lew]]></surname>
<given-names><![CDATA[MJ.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Bad statistical practice in pharmacology (and other basic biomedical disciplines): you probably don't know P]]></article-title>
<source><![CDATA[Br J Pharmacol]]></source>
<year>2012</year>
<volume>166</volume>
<page-range>1559-1567</page-range></nlm-citation>
</ref>
<ref id="B8">
<label>8</label><nlm-citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Goodman]]></surname>
<given-names><![CDATA[S.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[A dirty dozen: twelve p-value misconceptions]]></article-title>
<source><![CDATA[Semin Hematol]]></source>
<year>2008</year>
<volume>45</volume>
<page-range>135-140</page-range></nlm-citation>
</ref>
<ref id="B9">
<label>9</label><nlm-citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Hurlbert]]></surname>
<given-names><![CDATA[SH]]></given-names>
</name>
<name>
<surname><![CDATA[Lombardi]]></surname>
<given-names><![CDATA[CM.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Final collapse of the Neyman-Pearson decision theoretic framework and rise of the neo Fisherian]]></article-title>
<source><![CDATA[Ann Zool Fennici]]></source>
<year>2009</year>
<volume>46</volume>
<page-range>311-349</page-range></nlm-citation>
</ref>
<ref id="B10">
<label>10</label><nlm-citation citation-type="">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Coe]]></surname>
<given-names><![CDATA[R.]]></given-names>
</name>
</person-group>
<source><![CDATA[It's the effect size, stupid: What effect size is and why it is important. School of education, University of Durham]]></source>
<year>12-1</year>
<month>4 </month>
<day>Se</day>
</nlm-citation>
</ref>
<ref id="B11">
<label>11</label><nlm-citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Cohen]]></surname>
<given-names><![CDATA[J.]]></given-names>
</name>
</person-group>
<source><![CDATA[Statistical power analysis for the behavioral sciences]]></source>
<year>1977</year>
<page-range>216-380</page-range><publisher-loc><![CDATA[New York ]]></publisher-loc>
<publisher-name><![CDATA[Academic Press]]></publisher-name>
</nlm-citation>
</ref>
</ref-list>
</back>
</article>
