Skip to main content

Bestimmung von Teststärke, Effektgröße und optimalem Stichprobenumfang

  • Chapter
  • First Online:
Forschungsmethoden und Evaluation in den Sozial- und Humanwissenschaften

Part of the book series: Springer-Lehrbuch ((SLB))

Zusammenfassung

Dieses Kapitel vermittelt folgende Lernziele: Die Teststärke definieren und Post-hoc- sowie A-priori-Teststärkeanalysen voneinander abgrenzen können. Wissen, was man unter der Effektgröße versteht und wie man sie berechnet. Verschiedene standardisierte Effektgrößenmaße unterscheiden und hinsichtlich ihrer Ausprägung als kleine, mittlere oder große Effekte einordnen können. Das Konzept des optimalen Stichprobenumfangs erläutern können. Wissen, wie man den optimalen Stichprobenumfang für Studien mit unterschiedlichen Signifikanztests im Zuge der Untersuchungsplanung festlegt.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 49.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Literatur

  • Abelson, R. P. & Prentice, D. A. (1997). Contrast tests of interaction hypothesis. Psychological Methods, 2(4), 315–328.

    Google Scholar 

  • Aberson, C. L. (2010). Applied power analysis for the behavioral sciences. New York: Psychology Press.

    Google Scholar 

  • Algina, J., Keselman, H. J., & Penfield, R. D. (2005). An alternative to Cohen’s standardized mean difference effect size: A robust parameter and confidence interval in the two independent group case. Psychological Methods, 10(3), 317–328.

    Google Scholar 

  • APA, American Psychological Association. (2001). Publication manual of the American Psychological Association (5th ed.). Washington DC: American Psychological Association.

    Google Scholar 

  • APA, American Psychological Association. (2009). Publication manual of the American Psychological Association (6th ed.). Washington DC: American Psychological Association.

    Google Scholar 

  • Bausell, B. R. & Li, Y.–F. (2003) Power analysis for experimental research: a practical guide for the biological, medical and social sciences. Cambridge: Cambridge University Press.

    Google Scholar 

  • Bird, K. D. (2002). Confidence intervalls for effect sizes in analysis of variance. Educational Psychological Measurement, 62(2), 197–226.

    Google Scholar 

  • Bortz, J. (2005). Statistik (6. Aufl.). Berlin: Springer.

    Google Scholar 

  • Bortz, J. & Lienert, G. A. (2003). Kurzgefasste Statistik für die klinische Forschung (2. Aufl.). Berlin: Springer.

    Google Scholar 

  • Bortz, J. & Lienert, G. A. (2008). Kurzgefasste Statistik für die klinische Forschung (3. Aufl.). Heidelberg: Springer.

    Google Scholar 

  • Bortz, J., Lienert, G. A., & Boehnke, K. (2008). Verteilungsfreie Methoden in der Biostatistik (3. Aufl.). Berlin: Springer.

    Google Scholar 

  • Bortz, J. & Schuster, C. (2010). Statistik für Human- und Sozialwissenschaftler. Berlin: Springer.

    Google Scholar 

  • Brizendine, L. (2006). The female brain. New York: Broadway Books.

    Google Scholar 

  • Chen, H., Cohen, P., & Chen, S. (2010). How big is a big odds ratio? Interpreting the magnitudes of odds ratios in epidemiological studies. Communications in Statistics – Simulation and Computation 39(4), 860–864.

    Google Scholar 

  • Cohen, J. (1962). The statistical power of abnormal-social psychological research: A review. Journal of Abnormal and Social Research, 65(3), 145–153.

    Google Scholar 

  • Cohen, J. (1988). Statistical power analysis for the behavioral sciences. New York: Erlbaum.

    Google Scholar 

  • Cohen, J. (1990). Things I have learned (so far). American Psychologist, 45(12), 1304–1312.

    Google Scholar 

  • Cohen, J. (1992). A power primer. Psychological Bulletin, 112(1), 155–159.

    Google Scholar 

  • Cohen, J. (1994). The earth is round (p \(<\) 0.05). American Psychologist, 49(12), 997–1003.

    Google Scholar 

  • Cramer, E. M. & Nicewander, W. A. (1979). Some symmetric, invariant measures of multivariate association. Psychometrika, 44(1), 43–54.

    Google Scholar 

  • Cumming, G. (2012). Understanding the new statistics: Effect sizes, confidence intervals, and meta-analysis. New York: Routledge.

    Google Scholar 

  • Cumming, G. & Finch, S. (2001). A primer on the understanding, Use, and calculation of confidence intervals that are based on central and noncentral distribution. Educational Psychological Measurement, 61(4), 532–574.

    Google Scholar 

  • Dunlop, W. P., Cortina, J. M., Vaslow, J. B., & Burke, M. J. (1996). Meta-analysis of experiments with matched groups or repeated measure designs. Psychological Methods, 1(2), 170–177.

    Google Scholar 

  • Ellis, P. D. (2010). The essential guide to effect sizes: an introduction to statistical power, meta-analysis and the interpretation of research results. Cambridge Cambridge University Press.

    Google Scholar 

  • Erdfelder, E. (1984). Zur Bedeutung und Kontrolle des beta-Fehlers bei der inferenzstatistischen Prüfung log-linearer Modelle. Zeitschrift für Sozialpsychologie, 15(1), 18–32.

    Google Scholar 

  • Erdfelder, E., Faul, F., Buchner, A., & Cüpper, L. (2010). Effektgröße und Teststärke. In H. Holling & B. Schmitz (Hrsg.), Handbuch der Psychologischen Methoden und Evaluation (S. 358–369). Göttingen: Hogrefe.

    Google Scholar 

  • Faul, F., Erdfelder, E., Lang, A.-G., & Buchner, A. (2007). G*Power 3: A flexible statistical power analysis program for the social, behavioral, and biomedical sciences. Behavior Research Methods, 39(2), 175–191.

    Google Scholar 

  • Faul, F., Erdfelder, E., Buchner, A., & Lang, A.-G. (2009). Statistical power analyses using G*Power 3.1: Tests for correlation and regression analyses. Behavior Research Methods, 41(4), 1149–1160.

    Google Scholar 

  • Ferguson, C. J. (2009). An effect size primer: A guide for clinicians and researchers. Professional Psychology: Research and Practice, 40(5), 532–538.

    Google Scholar 

  • Fleiss, J. L. (1994). Measures of effect size for categorical data. In H. Cooper & L. V. Hedges (Eds.), The Handbook of Research Synthesis (pp. 246–259). New York: Sage.

    Google Scholar 

  • Fowler, R. L. (1987). A general method for comparing effect-magnitudes in ANOVA designs. Educational and Psychological Measurement, 47, 361–367.

    Google Scholar 

  • Fritz, C. O., Morris, P. E., & Richler, J. J. (2012). Effect size estimates: Current use, calculations, and interpretation. Journal of Experimental Psychology: General, 141(1), 2–18.

    Google Scholar 

  • Gatsonis, C. & Sampson, A. R. (1989). Multiple correlation: Exact power and sample size calculations. Psychological Bulletin, 106(3), 519–524.

    Google Scholar 

  • Gigerenzer, G. (1993). The superego, the ego, and the ID in statistical reasoning. In G. Keren & C. Lewis (Eds.), A handbook for data analysis in the behavioural sciences. Methodological Issues (pp. 311–339). Hillsdale: Erlbaum.

    Google Scholar 

  • Gillett, R. (1994). The average power criterion for sample size estimation. Statistician, 43, 389–394.

    Google Scholar 

  • Gillett, R. (2003). The metric comparability of meta-analytic effectsize estimators from factorial designs. Psychological Methods, 8(4), 419–433.

    Google Scholar 

  • Glass, G. V. (1976). Primary, secondary, and meta-analysis of research. Educational Researcher, 5(10), 3–8.

    Google Scholar 

  • Grissom, R. J. & Kim, J. J. (2011). Effect sizes for research: Univariate and multivariate applications (2nd ed.). Milton Park: Routledge.

    Google Scholar 

  • Haddock, C. K., Rindskopf, D., & Shadish, W. R. (1998). Using odds ratios as effect sizes for meta-analysis of dichotomous data. A primer on methods and issues. Psychological Methods, 3(3), 339–353.

    Google Scholar 

  • Hager, W. (2004). Testplanung zur statistischen Prüfung psychologischer Hypothesen. Göttingen: Hogrefe.

    Google Scholar 

  • Halpern, S. D., Karlawish, J. H. T., & Berlin, J. A. (2002). The continuing unethical conduct of underpowered clinical trials. Journal of the American Medical Association, 288(3), 358–362.

    Google Scholar 

  • Hays, W. L. (1994). Statistics (5th ed.). New York: Harcourt College Publishers.

    Google Scholar 

  • Hedges, L. V. (1982). Estimation of effect size from a series of independent experiments. Psychological Bulletin, 92(2), 490–499.

    Google Scholar 

  • Held, U. (2010). Was ist eine „Odds Ratio“ und wann wird sie verwendet? Schweiz Med Forum, 10(37), 634–645.

    Google Scholar 

  • Hoenig, J. M. & Heisey, D. M. (2001). The abuse of power: The pervasive fallacy of power calculations for data analysis. The American Statistician, 55(1), 19–24.

    Google Scholar 

  • Hsu, L. M. (2004). Biases of success rate differences shown in binomial effect size displays. Psychological Methods, 9(2), 183–197.

    Google Scholar 

  • James, D. & Drakich, J. (1993). Understanding gender differences in amount of talk. In D. Tannen (Ed.), Gender and conversational interaction (pp. 281–312). Oxford: Oxford University Press.

    Google Scholar 

  • Janosky, J. E. (2002). The ethics of underpowered clinical trials. Journal of the American Medical Association, 288(17), 2118.

    Google Scholar 

  • Kelley, K. & Maxwell, S. E. (2003). Sample Size for multiple regression. Obtaining regression coefficients that are accurate, not simply significant. Psychological Methods, 8(3), 305–321.

    Google Scholar 

  • Keren, G. & Lewis, C. (1979). Partial omega squared for ANOVA designs. Educational and Psychological Measurement, 39, 119–128.

    Google Scholar 

  • Kirk, R. E. (1996). Practical significance: A concept whose time has come. Educational and Psychological Measurement, 56(5), 746–759.

    Google Scholar 

  • Kline, R. B. (2004). Beyond significance testing. Washington: American Psychological Association.

    Google Scholar 

  • Kraemer, H. C. & Thiemann, S. (1987). How many subjects? Statistical power analysis in research. Thousand Oaks: Sage.

    Google Scholar 

  • Kshirsagar, A. M. (1972). Multivariate analysis. New York: Dekker.

    Google Scholar 

  • Küchler, M. (1980). The analysis of nonmetric data. Sociological Methods and Research, 8(4), 369–388.

    Google Scholar 

  • Leonhart, R. (2009). Statistik – Einstieg und Vertiefung (2. Aufl.). Bern: Huber.

    Google Scholar 

  • Lilford, R., & Stevens, A. J. (2002). Underpowered studies. British Journal of Surgery, 89(2), 129–131.

    Google Scholar 

  • Lipsey, M. W. (1997). Design sensitivity: Statistical power for applied experimental research. In L. Bickman & D. Rog (Eds.), Handbook of applied social research methods (pp. 39–68). Thousand Oaks: Sage.

    Google Scholar 

  • Lipsey, M. W. & Wilson, D. B. (2001). Practical meta-analysis. Thousand Oaks: Sage.

    Google Scholar 

  • Maxwell, S. E. (2000). Sample size and multiple regression analysis. Psychological Methods, 5(4), 434–458.

    Google Scholar 

  • Maxwell, S. E. (2004). The persistence of underpowered studies in psychological research: Causes, consequences, and remedies. Psychological Methods, 9(2), 147–163.

    Google Scholar 

  • Mehl, M. R., Vazire, S., Ramirez–Esparza, N., Slatcher, R. B., & Pennebaker, J. W. (2007). Are women really more talkative than men? Science, 317(5834), 82.

    Google Scholar 

  • Mendoza, J. L. & Stafford, K. L. (2001). Confidence Intervals, power calculation, and sample size estimation for the squared multiple correlation coefficient under the fixed and random regression models: A computer program and useful standard tables. Educational Psychological Measurement, 61(4), 650–667.

    Google Scholar 

  • Murphy, K. R., Myors, B., & Wolach, A. (2008). Statistical power analysis: 2. Milton Park: Routledge.

    Google Scholar 

  • Nickerson, R. S. (2000). Null hypothesis significance testing: A review of an old and continuing controversy. Psychological Methods, 5(2), 241–301.

    Google Scholar 

  • Olejnik, S. & Algina, J. (2003). Generalized eta and omega squared statistics: Measures of effect size for some common research designs. Psychological Methods, 8(4), 434–477.

    Google Scholar 

  • Peng, C.-Y. J., Long, H., & Abaci, S. (2012). Power analysis software for educational researchers. Journal of Experimental Education, 80(2), 113–136.

    Google Scholar 

  • Reedera, H. M. (1996). A critical look at gender difference in communication research. Communication Studies, 47(4), 318–330.

    Google Scholar 

  • Rosenthal, M. C. (1994). The fugitive literature. In H. Cooper & L. V. Hedges (Eds.), The handbook of research synthesis (pp. 85–94). Thousand Oaks: Sage.

    Google Scholar 

  • Rosenthal, R. & Rubin, D. B. (1982). A simple, general purpose display of magnitudes of experimental effect. Journal of Educational Psychology, 74(2), 166–169.

    Google Scholar 

  • Sachs, L. (2002). Statistische Auswertungsmethoden (10. Aufl.). Berlin: Springer.

    Google Scholar 

  • Sedlmeier, P. & Gigerenzer, G. (1989). Do studies of statistical power have an effect on the power of studies? Psychological Bulletin, 105(2), 309–316.

    Google Scholar 

  • Shadish, W. R., Cook, T. D., & Campbell, D. T. (2002). Experimental and quasi-experimental designs for generalized causal inference. Boston: Houghton Mifflin.

    Google Scholar 

  • Smithson, M. J. (2003). Confidence intervals. Thousand Oaks: Sage.

    Google Scholar 

  • Steiger, J. H. (2004). Beyond the F-Test: Effect size confidence intervals and tests of close fit in the analysis of variance and contrast analysis. Psychological Methods, 9(2), 164–182.

    Google Scholar 

  • Thompson, B. (1994). Guidelines for authors. Educational and Psychological Measurement, 54, 837–847.

    Google Scholar 

  • Thompson, B. (2002). „Statistical“, „practical“, and „clinical“: How many kinds of significance do counselors need to consider? Journal of Counseling and Development, 80(1), 64–71.

    Google Scholar 

  • Westermann, R. (2000). Wissenschaftstheorie und Experimentalmethodik. Ein Lehrbuch zur Psychologischen Methodenlehre. Göttingen: Hogrefe.

    Google Scholar 

  • Wilkinson, L. & Inference, T. T. F. o. S. (1999). Statistical methods in psychological journals: Guidelines and explanations. American Psychologist, 54(8), 594–604.

    Google Scholar 

  • Winer, B. J., Brown, D. R., & Michels, K. M. (1991). Statistical principles in experimental design. New York: Mc–Graw Hill.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Rights and permissions

Reprints and permissions

Copyright information

© 2016 Springer-Verlag Berlin Heidelberg

About this chapter

Cite this chapter

Döring, N., Bortz, J. (2016). Bestimmung von Teststärke, Effektgröße und optimalem Stichprobenumfang. In: Forschungsmethoden und Evaluation in den Sozial- und Humanwissenschaften. Springer-Lehrbuch. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-41089-5_14

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-41089-5_14

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-41088-8

  • Online ISBN: 978-3-642-41089-5

  • eBook Packages: Psychology (German Language)

Publish with us

Policies and ethics