Statistical Tools for the Comprehensive Practice of Industrial Hygiene and Environmental Health Sciences
John Wiley & Sons Inc (Verlag)
978-1-119-14301-7 (ISBN)
- Titel z.Zt. nicht lieferbar
- Versandkostenfrei
- Auch auf Rechnung
- Artikel merken
Includes both parametric and non-parametric techniques described and illustrated in a worker health and environmental protection practice context
Illustrated through numerous examples presented in the context of IH/EHS field practice and research, using the statistical analysis tools available in Excel® wherever possible
Emphasizes the application of statistical tools to IH/EHS-type data in order to answer IH/EHS-relevant questions
Includes an instructor’s manual that follows in parallel with the textbook, including PowerPoints to help prepare lectures and answers in the text as for the Exercises section of each chapter.
David L. Johnson has over 40 years of experience in environmental engineering and occupational safety and health practice, research, and teaching. Dr. Johnson was a practicing environmental engineer and industrial hygienist with the United States Army for 20 years, serving in a variety of positions in the United States, Europe, and the Middle East. He joined the faculty of the University of Oklahoma’s College of Public Health, Department of Occupational and Environmental Health in 1991.
Preface xv
Acknowledgments xvii
About the Author xix
About the Companion Website xxi
1 Some Basic Concepts 1
1.1 Introduction 1
1.2 Physical versus Statistical Sampling 2
1.3 Representative Measures 3
1.4 Strategies for Representative Sampling 3
1.5 Measurement Precision 4
1.6 Probability Concepts 6
1.6.1 The Relative Frequency Approach 7
1.6.2 The Classical Approach – Probability Based on Deductive Reasoning 7
1.6.3 Subjective Probability 7
1.6.4 Complement of a Probability 7
1.6.5 Mutually Exclusive Events 8
1.6.6 Independent Events 8
1.6.7 Events that Are Not Mutually Exclusive 9
1.6.8 Marginal and Conditional Probabilities 9
1.6.9 Testing for Independence 11
1.7 Permutations and Combinations 12
1.7.1 Permutations for Sampling without Replacement 12
1.7.2 Permutations for Sampling with Replacement 13
1.7.3 Combinations 13
1.8 Introduction to Frequency Distributions 14
1.8.1 The Binomial Distribution 14
1.8.2 The Normal Distribution 16
1.8.3 The Chi-Square Distribution 20
1.9 Confidence Intervals and Hypothesis Testing 22
1.10 Summary 23
1.11 Addendum: Glossary of Some Useful Excel Functions 23
1.12 Exercises 26
References 28
2 Descriptive Statistics and Methods of Presenting Data 29
2.1 Introduction 29
2.2 Quantitative Descriptors of Data and Data Distributions 29
2.3 Displaying Data with Frequency Tables 33
2.4 Displaying Data with Histograms and Frequency Polygons 34
2.5 Displaying Data Frequency Distributions with Cumulative Probability Plots 35
2.6 Displaying Data with NED and Q–Q Plots 38
2.7 Displaying Data with Box-and-Whisker Plots 41
2.8 Data Transformations to Achieve Normality 42
2.9 Identifying Outliers 43
2.10 What to Do with Censored Values? 45
2.11 Summary 45
2.12 Exercises 46
References 48
3 Analysis of Frequency Data 49
3.1 Introduction 49
3.2 Tests for Association and Goodness-of-Fit 50
3.2.1 r × c Contingency Tables and the Chi-Square Test 50
3.2.2 Fisher’s Exact Test 54
3.3 Binomial Proportions 55
3.4 Rare Events and the Poisson Distribution 57
3.4.1 Poisson Probabilities 57
3.4.2 Confidence Interval on a Poisson Count 60
3.4.3 Testing for Fit with the Poisson Distribution 61
3.4.4 Comparing Two Poisson Rates 62
3.4.5 Type I Error, Type II Error, and Power 64
3.4.6 Power and Sample Size in Comparing Two Poisson Rates 64
3.5 Summary 65
3.6 Exercises 66
References 69
4 Comparing Two Conditions 71
4.1 Introduction 71
4.2 Standard Error of the Mean 71
4.3 Confidence Interval on a Mean 72
4.4 The t-Distribution 73
4.5 Parametric One-Sample Test – Student’s t-Test 74
4.6 Two-Tailed versus One-Tailed Hypothesis Tests 76
4.7 Confidence Interval on a Variance 77
4.8 Other Applications of the Confidence Interval Concept in IH/EHS Work 79
4.8.1 OSHA Compliance Determinations 79
4.8.2 Laboratory Analyses – LOB, LOD, and LOQ 80
4.9 Precision, Power, and Sample Size for One Mean 81
4.9.1 Sample Size Required to Estimate a Mean with a Stated Precision 81
4.9.2 Sample Size Required to Detect a Specified Difference in Student’s t-Test 81
4.10 Iterative Solutions Using the Excel Goal Seek Utility 82
4.11 Parametric Two-Sample Tests 83
4.11.1 Confidence Interval for a Difference in Means: The Two-Sample t-Test 83
4.11.2 Two-Sample t-Test When Variances Are Equal 84
4.11.3 Verifying the Assumptions of the Two-Sample t-Test 85
4.11.3.1 Lilliefors Test for Normality 86
4.11.3.2 Shapiro–Wilk W-Test for Normality 87
4.11.3.3 Testing for Homogeneity of Variance 91
4.11.3.4 Transformations to Stabilize Variance 93
4.11.4 Two-Sample t-Test with Unequal Variances – Welch’s Test 93
4.11.5 Paired Sample t-Test 95
4.11.6 Precision, Power, and Sample Size for Comparing Two Means 96
4.12 Testing for Difference in Two Binomial Proportions 99
4.12.1 Testing a Binomial Proportion for Difference from a Known Value 100
4.12.2 Testing Two Binomial Proportions for Difference 100
4.13 Nonparametric Two-Sample Tests 102
4.13.1 Mann–Whitney U Test 102
4.13.2 Wilcoxon Matched Pairs Test 104
4.13.3 McNemar and Binomial Tests for Paired Nominal Data 105
4.14 Summary 107
4.15 Exercises 107
References 111
5 Characterizing the Upper Tail of the Exposure Distribution 113
5.1 Introduction 113
5.2 Upper Tolerance Limits 113
5.3 Exceedance Fractions 115
5.4 Distribution Free Tolerance Limits 117
5.5 Summary 119
5.6 Exercises 119
References 121
6 One-Way Analysis of Variance 123
6.1 Introduction 123
6.2 Parametric One-Way ANOVA 123
6.2.1 How the Parametric ANOVA Works – Sums of Squares and the F-Test 124
6.2.2 Post hoc Multiple Pairwise Comparisons in Parametric ANOVA 127
6.2.2.1 Tukey’s Test 127
6.2.2.2 Tukey–Kramer Test 128
6.2.2.3 Dunnett’s Test for Comparing Means to a Control Mean 130
6.2.2.4 Planned Contrasts Using the Scheffé S Test 132
6.2.3 Checking the ANOVA Model Assumptions – NED Plots and Variance Tests 134
6.2.3.1 Levene’s Test 134
6.2.3.2 Bartlett’s Test 135
6.3 Nonparametric Analysis of Variance 136
6.3.1 Kruskal–Wallis Nonparametric One-Way ANOVA 137
6.3.2 Post hoc Multiple Pairwise Comparisons in Nonparametric ANOVA 139
6.3.2.1 Nemenyi’s Test 139
6.3.2.2 Bonferroni–Dunn Test 140
6.4 ANOVA Disconnects 142
6.5 Summary 144
6.6 Exercises 145
References 149
7 Two-Way Analysis of Variance 151
7.1 Introduction 151
7.2 Parametric Two-Way ANOVA 151
7.2.1 Two-Way ANOVA without Interaction 154
7.2.2 Checking for Homogeneity of Variance 154
7.2.3 Multiple Pairwise Comparisons When There Is No Interaction Term 154
7.2.4 Two-Way ANOVA with Interaction 156
7.2.5 Multiple Pairwise Comparisons with Interaction 158
7.2.6 Two-Way ANOVA without Replication 160
7.2.7 Repeated-Measures ANOVA 160
7.2.8 Two-Way ANOVA with Unequal Sample Sizes 162
7.3 Nonparametric Two-Way ANOVA 162
7.3.1 Rank Tests 162
7.3.1.1 The Rank Test 162
7.3.1.2 The Rank Transform Test 166
7.3.1.3 Other Options – Aligned Rank Tests 166
7.3.2 Repeated-Measures Nonparametric ANOVA – Friedman’s Test 166
7.3.2.1 Friedman’s Test without Replication 167
7.3.2.2 Multiple Comparisons for Friedman’s Test without Replication 169
7.3.2.3 Friedman’s Test with Replication 170
7.3.2.4 Multiple Comparisons for Friedman’s Test with Replication 172
7.4 More Powerful Non-ANOVA Approaches: Linear Modeling 172
7.5 Summary 172
7.6 Exercises 172
References 178
8 Correlation Analysis 181
8.1 Introduction 181
8.2 Simple Parametric Correlation Analysis 181
8.2.1 Testing the Correlation Coefficient for Significance 184
8.2.1.1 t-Test for Significance 185
8.2.1.2 F-Test for Significance 186
8.2.2 Confidence Limits on the Correlation Coefficient 186
8.2.3 Power in Simple Correlation Analysis 187
8.2.4 Comparing Two Correlation Coefficients for Difference 188
8.2.5 Comparing More Than Two Correlation Coefficients for Difference 189
8.2.6 Multiple Pairwise Comparisons of Correlation Coefficients 190
8.3 Simple Nonparametric Correlation Analysis 190
8.3.1 Spearman Rank Correlation Coefficient 190
8.3.2 Testing Spearman’s Rank Correlation Coefficient for Statistical Significance 191
8.3.3 Correction to Spearman’s Rank Correlation Coefficient When There Are Tied Ranks 193
8.4 Multiple Correlation Analysis 195
8.4.1 Parametric Multiple Correlation 195
8.4.2 Nonparametric Multiple Correlation: Kendall’s Coefficient of Concordance 195
8.5 Determining Causation 198
8.6 Summary 198
8.7 Exercises 198
References 204
9 Regression Analysis 205
9.1 Introduction 205
9.2 Linear Regression 205
9.2.1 Simple Linear Regression 207
9.2.2 Nonconstant Variance – Transformations and Weighted Least Squares Regression 209
9.2.3 Multiple Linear Regression 213
9.2.3.1 Multiple Regression in Excel 215
9.2.3.2 Multiple Regression Using the Excel Solver Utility 218
9.2.3.3 Multiple Regression Using Advanced Software Packages 221
9.2.4 Using Regression for Factorial ANOVA with Unequal Sample Sizes 222
9.2.5 Multiple Correlation Analysis Using Multiple Regression 227
9.2.5.1 Assumptions of Parametric Multiple Correlation 233
9.2.5.2 Options When Collinearity Is a Problem 233
9.2.6 Polynomial Regression 234
9.2.7 Interpreting Linear Regression Results 234
9.2.8 Linear Regression versus ANOVA 235
9.3 Logistic Regression 235
9.3.1 Odds and Odds Ratios 236
9.3.2 The Logit Transformation 238
9.3.3 The Likelihood Function 240
9.3.4 Logistic Regression in Excel 240
9.3.5 Likelihood Ratio Test for Significance of MLE Coefficients 241
9.3.6 Odds Ratio Confidence Limits in Multivariate Models 243
9.4 Poisson Regression 243
9.4.1 Poisson Regression Model 243
9.4.2 Poisson Regression in Excel 244
9.5 Regression with Excel Add-ons 245
9.6 Summary 246
9.7 Exercises 246
References 252
10 Analysis of Covariance 253
10.1 Introduction 253
10.2 The Simple ANCOVA Model and Its Assumptions 253
10.2.1 Required Regressions 255
10.2.2 Checking the ANCOVA Assumptions 258
10.2.2.1 Linearity, Independence, and Normality 258
10.2.2.2 Similar Variances 258
10.2.2.3 Equal Regression Slopes 258
10.2.3 Testing and Estimating the Treatment Effects 259
10.3 The Two-Factor Covariance Model 261
10.4 Summary 261
10.5 Exercises 261
Reference 263
11 Experimental Design 265
11.1 Introduction 265
11.2 Randomization 266
11.3 Simple Randomized Experiments 266
11.4 Experimental Designs Blocking on Categorical Factors 267
11.5 Randomized Full Factorial Experimental Design 270
11.6 Randomized Full Factorial Design with Blocking 271
11.7 Split Plot Experimental Designs 272
11.8 Balanced Experimental Designs – Latin Square 273
11.9 Two-Level Factorial Experimental Designs with Quantitative Factors 274
11.9.1 Two-Level Factorial Designs for Exploratory Studies 274
11.9.2 The Standard Order 275
11.9.3 Calculating Main Effects 276
11.9.4 Calculating Interactions 278
11.9.5 Estimating Standard Errors 278
11.9.6 Estimating Effects with REGRESSION in Excel 279
11.9.7 Interpretation 280
11.9.8 Cube, Surface, and NED Plots as an Aid to Interpretation 280
11.9.9 Fractional Factorial Two-Level Experiments 282
11.10 Summary 282
11.11 Exercises 283
References 284
12 Uncertainty and Sensitivity Analysis 285
12.1 Introduction 285
12.2 Simulation Modeling 285
12.2.1 Propagation of Errors 286
12.2.2 Simple Bounding 287
12.2.2.1 Sums and Differences 287
12.2.2.2 Products and Ratios 287
12.2.2.3 Powers 289
12.2.3 Addition in Quadrature 289
12.2.3.1 Sums and Differences 289
12.2.3.2 Products and Ratios 290
12.2.3.3 Powers 292
12.2.4 LOD and LOQ Revisited – Dust Sample Gravimetric Analysis 292
12.3 Uncertainty Analysis 295
12.4 Sensitivity Analysis 296
12.4.1 One-at-a-Time (OAT) Analysis 296
12.4.2 Variance-Based Analysis 297
12.5 Further Reading on Uncertainty and Sensitivity Analysis 297
12.6 Monte Carlo Simulation 297
12.7 Monte Carlo Simulation in Excel 298
12.7.1 Generating Random Numbers in Excel 298
12.7.2 The Populated Spreadsheet Approach 299
12.7.3 Monte Carlo Simulation Using VBA Macros 299
12.8 Summary 303
12.9 Exercises 303
References 307
13 Bayes’ Theorem and Bayesian Decision Analysis 309
13.1 Introduction 309
13.2 Bayes’ Theorem 310
13.3 Sensitivity, Specificity, and Positive and Negative Predictive Value in Screening
Tests 310
13.4 Bayesian Decision Analysis in Exposure Control Banding 312
13.4.1 Introduction to BDA 312
13.4.2 The Prior Distribution and the Parameter Space 314
13.4.3 The Posterior Distribution and Likelihood Function 314
13.4.4 Relative Influences of the Prior and the Data 315
13.4.5 Frequentist versus Bayesian Perspectives 316
13.5 Exercises 316
References 318
A z-Tables of the Standard Normal Distribution 321
B Critical Values of the Chi-Square Distribution 327
C Critical Values for the t-Distribution 329
D Critical Values for Lilliefors Test 331
Reference 332
E Shapiro–Wilk W Test 𝜶 Coefficients and Critical Values 333
Reference 336
F Critical Values of the F Distribution for 𝜶 = 0.05 337
G Critical U Values for the Mann–Whitney U Test 341
Reference 342
H Critical Wilcoxon Matched Pairs Test t Values 343
Reference 344
I K Values for Upper Tolerance Limits 345
Reference 346
J Exceedance Fraction 95% Lower Confidence Limit versus Z 347
Reference 347
K q Values for Tukey’s, Tukey–Kramer, and Nemenyi’s MSD Tests 349
L q′ Values for Dunnett’s Test 351
Reference 353
M Q Values for the Bonferroni–Dunn MSD Test 355
N Critical Spearman Rank Correlation Test Values 357
O Critical Values of Kendall’s W 359
Reference 361
Index 363
Erscheinungsdatum | 11.02.2017 |
---|---|
Verlagsort | New York |
Sprache | englisch |
Maße | 183 x 257 mm |
Gewicht | 839 g |
Themenwelt | Mathematik / Informatik ► Mathematik |
Studium ► Querschnittsbereiche ► Prävention / Gesundheitsförderung | |
Naturwissenschaften ► Chemie | |
Sozialwissenschaften ► Soziologie | |
Technik ► Bauwesen | |
Wirtschaft | |
ISBN-10 | 1-119-14301-2 / 1119143012 |
ISBN-13 | 978-1-119-14301-7 / 9781119143017 |
Zustand | Neuware |
Haben Sie eine Frage zum Produkt? |
aus dem Bereich