My results were not significant now what? - Statistics Solutions statistical significance - Reporting non-significant regression [1] Comondore VR, Devereaux PJ, Zhou Q, et al. For r-values the adjusted effect sizes were computed as (Ivarsson, Andersen, Johnson, & Lindwall, 2013), Where v is the number of predictors. significant. At the risk of error, we interpret this rather intriguing Within the theoretical framework of scientific hypothesis testing, accepting or rejecting a hypothesis is unequivocal, because the hypothesis is either true or false. This indicates the presence of false negatives, which is confirmed by the Kolmogorov-Smirnov test, D = 0.3, p < .000000000000001. For example: t(28) = 2.99, SEM = 10.50, p = .0057.2 If you report the a posteriori probability and the value is less than .001, it is customary to report p < .001. Although my results are significants, when I run the command the significance level is never below 0.1, and of course the point estimate is outside the confidence interval since the beginning. For example, in the James Bond Case Study, suppose Mr. 0. You should cover any literature supporting your interpretation of significance. Power of Fisher test to detect false negatives for small- and medium effect sizes (i.e., = .1 and = .25), for different sample sizes (i.e., N) and number of test results (i.e., k). Example 11.6. Summary table of Fisher test results applied to the nonsignificant results (k) of each article separately, overall and specified per journal. [Non-significant in univariate but significant in multivariate analysis: a discussion with examples] Changgeng Yi Xue Za Zhi. }, author={S. Lo and I. T. Li and T. Tsou and L. Suppose a researcher recruits 30 students to participate in a study. This is also a place to talk about your own psychology research, methods, and career in order to gain input from our vast psychology community. However, the difference is not significant. Peter Dudek was one of the people who responded on Twitter: "If I chronicled all my negative results during my studies, the thesis would have been 20,000 pages instead of 200." one should state that these results favour both types of facilities Statistical significance does not tell you if there is a strong or interesting relationship between variables. All rights reserved. P50 = 50th percentile (i.e., median). Assume he has a \(0.51\) probability of being correct on a given trial \(\pi=0.51\).
For the discussion, there are a million reasons you might not have replicated a published or even just expected result. I usually follow some sort of formula like "Contrary to my hypothesis, there was no significant difference in aggression scores between men (M = 7.56) and women (M = 7.22), t(df) = 1.2, p = .50." The Fisher test proved a powerful test to inspect for false negatives in our simulation study, where three nonsignificant results already results in high power to detect evidence of a false negative if sample size is at least 33 per result and the population effect is medium. In this short paper, we present the study design and provide a discussion of (i) preliminary results obtained from a sample, and (ii) current issues related to the design. BMJ 2009;339:b2732. How would the significance test come out? Table 4 also shows evidence of false negatives for each of the eight journals. For example, a large but statistically nonsignificant study might yield a confidence interval (CI) of the effect size of [0.01; 0.05], whereas a small but significant study might yield a CI of [0.01; 1.30]. These decisions are based on the p-value; the probability of the sample data, or more extreme data, given H0 is true.
Non significant result but why? | ResearchGate Instead, we promote reporting the much more . For example, the number of participants in a study should be reported as N = 5, not N = 5.0. were reported. Libby Funeral Home Beacon, Ny. A study is conducted to test the relative effectiveness of the two treatments: \(20\) subjects are randomly divided into two groups of 10. The results indicate that the Fisher test is a powerful method to test for a false negative among nonsignificant results. Null findings can, however, bear important insights about the validity of theories and hypotheses. They concluded that 64% of individual studies did not provide strong evidence for either the null or the alternative hypothesis in either the original of the replication study. Two erroneously reported test statistics were eliminated, such that these did not confound results. Cells printed in bold had sufficient results to inspect for evidential value. When considering non-significant results, sample size is partic-ularly important for subgroup analyses, which have smaller num-bers than the overall study. the results associated with the second definition (the mathematically Concluding that the null hypothesis is true is called accepting the null hypothesis. We computed three confidence intervals of X: one for the number of weak, medium, and large effects. The levels for sample size were determined based on the 25th, 50th, and 75th percentile for the degrees of freedom (df2) in the observed dataset for Application 1. Journals differed in the proportion of papers that showed evidence of false negatives, but this was largely due to differences in the number of nonsignificant results reported in these papers. tbh I dont even understand what my TA was saying to me, but she said that there was no significance in my results. In laymen's terms, this usually means that we do not have statistical evidence that the difference in groups is. Although there is never a statistical basis for concluding that an effect is exactly zero, a statistical analysis can demonstrate that an effect is most likely small. biomedical research community. We investigated whether cardiorespiratory fitness (CRF) mediates the association between moderate-to-vigorous physical activity (MVPA) and lung function in asymptomatic adults. First, just know that this situation is not uncommon. One group receives the new treatment and the other receives the traditional treatment.
Guide to Writing the Results and Discussion Sections of a - GoldBio To conclude, our three applications indicate that false negatives remain a problem in the psychology literature, despite the decreased attention and that we should be wary to interpret statistically nonsignificant results as there being no effect in reality.
You didnt get significant results. ive spoken to my ta and told her i dont understand. Bond and found he was correct \(49\) times out of \(100\) tries. Summary table of articles downloaded per journal, their mean number of results, and proportion of (non)significant results. By accepting all cookies, you agree to our use of cookies to deliver and maintain our services and site, improve the quality of Reddit, personalize Reddit content and advertising, and measure the effectiveness of advertising. However, once again the effect was not significant and this time the probability value was \(0.07\). If you didn't run one, you can run a sensitivity analysis.Note: you cannot run a power analysis after you run your study and base it on observed effect sizes in your data; that is just a mathematical rephrasing of your p-values. Published on 21 March 2019 by Shona McCombes. so i did, but now from my own study i didnt find any correlations.
For example, for small true effect sizes ( = .1), 25 nonsignificant results from medium samples result in 85% power (7 nonsignificant results from large samples yield 83% power). Results of the present study suggested that there may not be a significant benefit to the use of silver-coated silicone urinary catheters for short-term (median of 48 hours) urinary bladder catheterization in dogs. Hence, most researchers overlook that the outcome of hypothesis testing is probabilistic (if the null-hypothesis is true, or the alternative hypothesis is true and power is less than 1) and interpret outcomes of hypothesis testing as reflecting the absolute truth. Therefore, these two non-significant findings taken together result in a significant finding. The proportion of reported nonsignificant results showed an upward trend, as depicted in Figure 2, from approximately 20% in the eighties to approximately 30% of all reported APA results in 2015. All research files, data, and analyses scripts are preserved and made available for download at http://doi.org/10.5281/zenodo.250492.
Reporting Research Results in APA Style | Tips & Examples - Scribbr Figure1.Powerofanindependentsamplest-testwithn=50per All. First, we determined the critical value under the null distribution. The forest plot in Figure 1 shows that research results have been ^contradictory _ or ^ambiguous. Consider the following hypothetical example. analysis. Talk about power and effect size to help explain why you might not have found something.
non significant results discussion example - lindoncpas.com 2016). Strikingly, though since its inception in 1956 compared to only 3 for Manchester United; Discussion. Simulations indicated the adapted Fisher test to be a powerful method for that purpose. One (at least partial) explanation of this surprising result is that in the early days researchers primarily reported fewer APA results and used to report relatively more APA results with marginally significant p-values (i.e., p-values slightly larger than .05), compared to nowadays. Both one-tailed and two-tailed tests can be included in this way. If = .1, the power of a regular t-test equals 0.17, 0.255, 0.467 for sample sizes of 33, 62, 119, respectively; if = .25, power values equal 0.813, 0.998, 1 for these sample sizes. More technically, we inspected whether p-values within a paper deviate from what can be expected under the H0 (i.e., uniformity). The reanalysis of the nonsignificant RPP results using the Fisher method demonstrates that any conclusions on the validity of individual effects based on failed replications, as determined by statistical significance, is unwarranted. The repeated concern about power and false negatives throughout the last decades seems not to have trickled down into substantial change in psychology research practice. Maybe there are characteristics of your population that caused your results to turn out differently than expected. The debate about false positives is driven by the current overemphasis on statistical significance of research results (Giner-Sorolla, 2012). 178 valid results remained for analysis. <- for each variable. If your p-value is over .10, you can say your results revealed a non-significant trend in the predicted direction. The power of the Fisher test for one condition was calculated as the proportion of significant Fisher test results given Fisher = 0.10. those two pesky statistically non-significant P values and their equally Sounds ilke an interesting project! Note that this application only investigates the evidence of false negatives in articles, not how authors might interpret these findings (i.e., we do not assume all these nonsignificant results are interpreted as evidence for the null). For example, the number of participants in a study should be reported as N = 5, not N = 5.0. If the \(95\%\) confidence interval ranged from \(-4\) to \(8\) minutes, then the researcher would be justified in concluding that the benefit is eight minutes or less. Besides in psychology, reproducibility problems have also been indicated in economics (Camerer, et al., 2016) and medicine (Begley, & Ellis, 2012). [2], there are two dictionary definitions of statistics: 1) a collection Table 2 summarizes the results for the simulations of the Fisher test when the nonsignificant p-values are generated by either small- or medium population effect sizes. relevance of non-significant results in psychological research and ways to render these results more . Our study demonstrates the importance of paying attention to false negatives alongside false positives. Expectations were specified as H1 expected, H0 expected, or no expectation. When researchers fail to find a statistically significant result, it's often treated as exactly that - a failure. [Article in Chinese] . Another potential explanation is that the effect sizes being studied have become smaller over time (mean correlation effect r = 0.257 in 1985, 0.187 in 2013), which results in both higher p-values over time and lower power of the Fisher test. Therefore we examined the specificity and sensitivity of the Fisher test to test for false negatives, with a simulation study of the one sample t-test. As opposed to Etz and Vandekerckhove (2016), Van Aert and Van Assen (2017; 2017) use a statistically significant original and a replication study to evaluate the common true underlying effect size, adjusting for publication bias. Finally, the Fisher test may and is also used to meta-analyze effect sizes of different studies. A researcher develops a treatment for anxiety that he or she believes is better than the traditional treatment. We also propose an adapted Fisher method to test whether nonsignificant results deviate from H0 within a paper. Application 1: Evidence of false negatives in articles across eight major psychology journals, Application 2: Evidence of false negative gender effects in eight major psychology journals, Application 3: Reproducibility Project Psychology, Section: Methodology and Research Practice, Nuijten, Hartgerink, van Assen, Epskamp, & Wicherts, 2015, Marszalek, Barber, Kohlhart, & Holmes, 2011, Borenstein, Hedges, Higgins, & Rothstein, 2009, Hartgerink, van Aert, Nuijten, Wicherts, & van Assen, 2016, Wagenmakers, Wetzels, Borsboom, van der Maas, & Kievit, 2012, Bakker, Hartgerink, Wicherts, & van der Maas, 2016, Nuijten, van Assen, Veldkamp, & Wicherts, 2015, Ivarsson, Andersen, Johnson, & Lindwall, 2013, http://science.sciencemag.org/content/351/6277/1037.3.abstract, http://pss.sagepub.com/content/early/2016/06/28/0956797616647519.abstract, http://pps.sagepub.com/content/7/6/543.abstract, https://doi.org/10.3758/s13428-011-0089-5, http://books.google.nl/books/about/Introduction_to_Meta_Analysis.html?hl=&id=JQg9jdrq26wC, https://cran.r-project.org/web/packages/statcheck/index.html, https://doi.org/10.1371/journal.pone.0149794, https://doi.org/10.1007/s11192-011-0494-7, http://link.springer.com/article/10.1007/s11192-011-0494-7, https://doi.org/10.1371/journal.pone.0109019, https://doi.org/10.3758/s13423-012-0227-9, https://doi.org/10.1016/j.paid.2016.06.069, http://www.sciencedirect.com/science/article/pii/S0191886916308194, https://doi.org/10.1053/j.seminhematol.2008.04.003, http://www.sciencedirect.com/science/article/pii/S0037196308000620, http://psycnet.apa.org/journals/bul/82/1/1, https://doi.org/10.1037/0003-066X.60.6.581, https://doi.org/10.1371/journal.pmed.0020124, http://journals.plos.org/plosmedicine/article/asset?id=10.1371/journal.pmed.0020124.PDF, https://doi.org/10.1016/j.psychsport.2012.07.007, http://www.sciencedirect.com/science/article/pii/S1469029212000945, https://doi.org/10.1080/01621459.2016.1240079, https://doi.org/10.1027/1864-9335/a000178, https://doi.org/10.1111/j.2044-8317.1978.tb00578.x, https://doi.org/10.2466/03.11.PMS.112.2.331-348, https://doi.org/10.1080/01621459.1951.10500769, https://doi.org/10.1037/0022-006X.46.4.806, https://doi.org/10.3758/s13428-015-0664-2, http://doi.apa.org/getdoi.cfm?doi=10.1037/gpr0000034, https://doi.org/10.1037/0033-2909.86.3.638, http://psycnet.apa.org/journals/bul/86/3/638, https://doi.org/10.1037/0033-2909.105.2.309, https://doi.org/10.1177/00131640121971392, http://epm.sagepub.com/content/61/4/605.abstract, https://books.google.com/books?hl=en&lr=&id=5cLeAQAAQBAJ&oi=fnd&pg=PA221&dq=Steiger+%26+Fouladi,+1997&ots=oLcsJBxNuP&sig=iaMsFz0slBW2FG198jWnB4T9g0c, https://doi.org/10.1080/01621459.1959.10501497, https://doi.org/10.1080/00031305.1995.10476125, https://doi.org/10.1016/S0895-4356(00)00242-0, http://www.ncbi.nlm.nih.gov/pubmed/11106885, https://doi.org/10.1037/0003-066X.54.8.594, https://www.apa.org/pubs/journals/releases/amp-54-8-594.pdf, http://creativecommons.org/licenses/by/4.0/, What Diverse Samples Can Teach Us About Cognitive Vulnerability to Depression, Disentangling the Contributions of Repeating Targets, Distractors, and Stimulus Positions to Practice Benefits in D2-Like Tests of Attention, Prespecification of Structure for the Optimization of Data Collection and Analysis, Binge Eating and Health Behaviors During Times of High and Low Stress Among First-year University Students, Psychometric Properties of the Spanish Version of the Complex Postformal Thought Questionnaire: Developmental Pattern and Significance and Its Relationship With Cognitive and Personality Measures, Journal of Consulting and Clinical Psychology (JCCP), Journal of Experimental Psychology: General (JEPG), Journal of Personality and Social Psychology (JPSP). Use the same order as the subheadings of the methods section. The data support the thesis that the new treatment is better than the traditional one even though the effect is not statistically significant. { "11.01:_Introduction_to_Hypothesis_Testing" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.
b__1]()", "11.02:_Significance_Testing" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "11.03:_Type_I_and_II_Errors" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "11.04:_One-_and_Two-Tailed_Tests" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "11.05:_Significant_Results" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "11.06:_Non-Significant_Results" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "11.07:_Steps_in_Hypothesis_Testing" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "11.08:_Significance_Testing_and_Confidence_Intervals" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "11.09:_Misconceptions_of_Hypothesis_Testing" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "11.10:_Statistical_Literacy" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "11.E:_Logic_of_Hypothesis_Testing_(Exercises)" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()" }, { "00:_Front_Matter" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "01:_Introduction_to_Statistics" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "02:_Graphing_Distributions" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "03:_Summarizing_Distributions" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "04:_Describing_Bivariate_Data" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "05:_Probability" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "06:_Research_Design" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "07:_Normal_Distribution" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "08:_Advanced_Graphs" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "09:_Sampling_Distributions" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "10:_Estimation" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "11:_Logic_of_Hypothesis_Testing" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "12:_Tests_of_Means" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "13:_Power" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "14:_Regression" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "15:_Analysis_of_Variance" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "16:_Transformations" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "17:_Chi_Square" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "18:_Distribution-Free_Tests" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "19:_Effect_Size" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "20:_Case_Studies" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "21:_Calculators" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()", "zz:_Back_Matter" : "property get [Map MindTouch.Deki.Logic.ExtensionProcessorQueryProvider+<>c__DisplayClass228_0.b__1]()" }, [ "article:topic", "authorname:laned", "showtoc:no", "license:publicdomain", "source@https://onlinestatbook.com" ], https://stats.libretexts.org/@app/auth/3/login?returnto=https%3A%2F%2Fstats.libretexts.org%2FBookshelves%2FIntroductory_Statistics%2FBook%253A_Introductory_Statistics_(Lane)%2F11%253A_Logic_of_Hypothesis_Testing%2F11.06%253A_Non-Significant_Results, \( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}}}\) \( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash{#1}}} \)\(\newcommand{\id}{\mathrm{id}}\) \( \newcommand{\Span}{\mathrm{span}}\) \( \newcommand{\kernel}{\mathrm{null}\,}\) \( \newcommand{\range}{\mathrm{range}\,}\) \( \newcommand{\RealPart}{\mathrm{Re}}\) \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\) \( \newcommand{\Argument}{\mathrm{Arg}}\) \( \newcommand{\norm}[1]{\| #1 \|}\) \( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\) \( \newcommand{\Span}{\mathrm{span}}\) \(\newcommand{\id}{\mathrm{id}}\) \( \newcommand{\Span}{\mathrm{span}}\) \( \newcommand{\kernel}{\mathrm{null}\,}\) \( \newcommand{\range}{\mathrm{range}\,}\) \( \newcommand{\RealPart}{\mathrm{Re}}\) \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\) \( \newcommand{\Argument}{\mathrm{Arg}}\) \( \newcommand{\norm}[1]{\| #1 \|}\) \( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\) \( \newcommand{\Span}{\mathrm{span}}\)\(\newcommand{\AA}{\unicode[.8,0]{x212B}}\). By rejecting non-essential cookies, Reddit may still use certain cookies to ensure the proper functionality of our platform. See osf.io/egnh9 for the analysis script to compute the confidence intervals of X. As Albert points out in his book Teaching Statistics Using Baseball A uniform density distribution indicates the absence of a true effect. A larger 2 value indicates more evidence for at least one false negative in the set of p-values. In order to compute the result of the Fisher test, we applied equations 1 and 2 to the recalculated nonsignificant p-values in each paper ( = .05). When you need results, we are here to help! For instance, 84% of all papers that report more than 20 nonsignificant results show evidence for false negatives, whereas 57.7% of all papers with only 1 nonsignificant result show evidence for false negatives. Additionally, in applications 1 and 2 we focused on results reported in eight psychology journals; extrapolating the results to other journals might not be warranted given that there might be substantial differences in the type of results reported in other journals or fields. All it tells you is whether you have enough information to say that your results were very unlikely to happen by chance. The remaining journals show higher proportions, with a maximum of 81.3% (Journal of Personality and Social Psychology). status page at https://status.libretexts.org, Explain why the null hypothesis should not be accepted, Discuss the problems of affirming a negative conclusion. evidence that there is insufficient quantitative support to reject the For example, suppose an experiment tested the effectiveness of a treatment for insomnia. Hence, we expect little p-hacking and substantial evidence of false negatives in reported gender effects in psychology. AppreciatingtheSignificanceofNon-Significant FindingsinPsychology It does depend on the sample size (the study may be underpowered), type of analysis used (for example in regression the other variable may overlap with the one that was non-significant),. Sample size development in psychology throughout 19852013, based on degrees of freedom across 258,050 test results. Similar Corpus ID: 20634485 [Non-significant in univariate but significant in multivariate analysis: a discussion with examples]. This might be unwarranted, since reported statistically nonsignificant findings may just be too good to be false. Secondly, regression models were fitted separately for contraceptive users and non-users using the same explanatory variables, and the results were compared. They also argued that, because of the focus on statistically significant results, negative results are less likely to be the subject of replications than positive results, decreasing the probability of detecting a false negative. It is generally impossible to prove a negative. An example of statistical power for a commonlyusedstatisticaltest,andhowitrelatesto effectsizes,isdepictedinFigure1. Avoid using a repetitive sentence structure to explain a new set of data. I understand when you write a report where you write your hypotheses are supported, you can pull on the studies you mentioned in your introduction in your discussion section, which i do and have done in past courseworks, but i am at a loss for what to do over a piece of coursework where my hypotheses aren't supported, because my claims in my introduction are essentially me calling on past studies which are lending support to why i chose my hypotheses and in my analysis i find non significance, which is fine, i get that some studies won't be significant, my question is how do you go about writing the discussion section when it is going to basically contradict what you said in your introduction section?, do you just find studies that support non significance?, so essentially write a reverse of your intro, I get discussing findings, why you might have found them, problems with your study etc my only concern was the literature review part of the discussion because it goes against what i said in my introduction, Sorry if that was confusing, thanks everyone, The evidence did not support the hypothesis. The concern for false positives has overshadowed the concern for false negatives in the recent debate, which seems unwarranted. To the contrary, the data indicate that average sample sizes have been remarkably stable since 1985, despite the improved ease of collecting participants with data collection tools such as online services.
Blue Streak Roller Coaster Death,
What Does Alt Points Mean Fanduel,
Accident In Brighton, Co Today,
Keto Pee Smells Like Asparagus,
Trader Joe's Distribution Center Jobs,
Articles N