CHF62.00
Download steht sofort bereit
Meta Analysis: A Guide to Calibrating and Combining Statistical
Evidence acts as a source of basic methods for scientists
wanting to combine evidence from different experiments. The authors
aim to promote a deeper understanding of the notion of statistical
evidence.
The book is comprised of two parts - The Handbook, and
The Theory. The Handbook is a guide for combining and
interpreting experimental evidence to solve standard statistical
problems. This section allows someone with a rudimentary knowledge
in general statistics to apply the methods. The Theory
provides the motivation, theory and results of simulation
experiments to justify the methodology.
This is a coherent introduction to the statistical concepts
required to understand the authors' thesis that evidence in a
test statistic can often be calibrated when transformed to the
right scale.
Autorentext
Dr. E. Kulinskaya - Director, Statistical Advisory
Service, Imperial College, London.
Professor S. Morgenthaler - Chair of Applied
Statistics, Ecole Polytechnique Fédérale de Lausanne,
Switzerland. Professor Morgenthaler was Assistant Professor at Yale
University prior to moving to EPFL and has chaired various ISI
committees.
Professor R. G. Staudte - Department of Statistical
Science, La Trobe University, Melbourne. During his career at La
Trobe he has served as Head of the Department of Statistical
Science for five years and Head of the School of Mathematical
and Statistical Sciences for two years. He was an Associate Editor
for the Journal of Statistical Planning & Inference for 4
years, and is a member of the American Statistical
Association, the Sigma Xi Scientific Research Society and the
Statistical Society of Australia.
Zusammenfassung
Meta Analysis: A Guide to Calibrating and Combining Statistical Evidence acts as a source of basic methods for scientists wanting to combine evidence from different experiments. The authors aim to promote a deeper understanding of the notion of statistical evidence.
The book is comprised of two parts The Handbook, and The Theory. The Handbook is a guide for combining and interpreting experimental evidence to solve standard statistical problems. This section allows someone with a rudimentary knowledge in general statistics to apply the methods. The Theory provides the motivation, theory and results of simulation experiments to justify the methodology.
This is a coherent introduction to the statistical concepts required to understand the authors' thesis that evidence in a test statistic can often be calibrated when transformed to the right scale.
Inhalt
Preface.
Part I The Methods.
1 What can the reader expect from this book?
1.1 A calibration scale for evidence.
1.2 The efficacy of glass ionomer versus resin sealants for prevention of caries.
1.3 Measures of effect size for two populations.
1.4 Summary.
2 Independent measurements with known precision.
2.1 Evidence for one-sided alternatives.
2.2 Evidence for two-sided alternatives.
2.3 Examples.
3 Independent measurements with unknown precision.
3.1 Effects and standardized effects.
3.2 Paired comparisons.
3.3 Examples.
4 Comparing treatment to control.
4.1 Equal unknown precision.
4.2 Differing unknown precision.
4.3 Examples.
5 Comparing K treatments.
5.1 Methodology.
5.2 Examples.
6 Evaluating risks.
6.1 Methodology.
6.2 Examples.
7 Comparing risks.
7.1 Methodology.
7.2 Examples.
8 Evaluating Poisson rates.
8.1 Methodology.
8.2 Example.
9 Comparing Poisson rates.
9.1 Methodology.
9.2 Example.
10 Goodness-of-fit testing.
10.1 Methodology.
10.2 Example.
11 Evidence for heterogeneity of effects and transformed effects.
11.1 Methodology.
11.2 Examples.
12 Combining evidence: fixed standardized effects model.
12.1 Methodology.
12.2 Examples.
13 Combining evidence: random standardized effects mode.
13.1 Methodology.
13.2 Example..
14 Meta-regression.
14.1 Methodology.
14.2 Commonly encountered situations.
14.3 Examples.
15 Accounting for publication bias.
15.1 The downside of publishing.
15.2 Examples.
Part II The Theory.
16 Calibrating evidence in a test.
16.1 Evidence for one-sided alternatives.
16.2 Random p-value behavior.
16.3 Publication bias.
16.4 Comparison with a Bayesian calibration.
16.5 Summary.
17 The basics of variance stabilizing transformations.
17.1 Standardizing the sample mean.
17.2 Variance stabilizing transformations.
17.3 Poisson model example.
17.4 Two-sided evidence from one-sided evidence.
17.5 Summary.
18 One-sample binomial tests.
18.1 Variance stabilizing the risk estimator.
18.2 Confidence intervals for p.
18.3 Relative risk and odds ratio.
18.4 Confidence intervals for small risks p.
18.5 Summary.
19 Two-sample binomial tests.
19.1 Evidence for a positive effect.
19.2 Confidence intervals for effect sizes.
19.3 Estimating the risk difference.
19.4 Relative risk and odds ratio.
19.5 Recurrent urinary tract infections.
19.6 Summary.
20 Defining evidence in t-statistics.
20.1 Example.
20.2 Evidence in the Student t-statistic.
20.3 The Key Inferential Function for Student's model.
20.4 Corrected evidence.
20.5 A confidence interval for the standardized effect.
20.6 Comparing evidence in t- and z-tests.
20.7 Summary.
21 Two-sample comparisons.
21.1 Drop in systolic blood pressure.
21.2 Defining the standardized effect.
21.3 Evidence in the Welch statistic.
21.4 Confidence intervals for d.
21.5 Summary.
22 Evidence in the chi-squared statistic.
22.1 The noncentral chi-squared distribution.
22.2 A vst for the noncentral chi-squared statistic.
22.3 Simulation studies.
22.4 Choosing the sample size.
22.5 Evidence for l > l0.
22.6 Summary.
23 Evidence in F-tests. 23.1 Variance stabilizing tr...