You are on page 1of 11

I.

Introduction
Prerequisites none A. B. C. D. E. F. G. H. What are Statistics? Importance of Statistics Descriptive Statistics Inferential Statistics Sampling Demonstration Variables Percentiles Measurement 1. Levels of Measurement 2. Measurement Demonstration 3. Basics of Data Collection I. Distributions J. Summation Notation K. Linear Transformations L. Exercises M. PDF Files (in .zip archive)

II. Graphing Distributions


Prerequisites none A. Introduction B. Qualitative Variables C. Quantitative Variables 1. Stem and Leaf Displays 2. Histograms 3. Frequency Polygons 4. Box Plots 5. Box Plot Demonstration 6. Bar Charts 7. Line Graphs D. Exercises E. PDF Files (in .zip archive)

III. Summarizing Distributions


Prerequisites none A. Central Tendency 1. What is Central Tendency 2. Measures of Central Tendency 3. Balance Scale Simulation 4. Absolute Difference Simulation 5. Squared Differences Simulation 6. Median and Mean 7. Mean and Median Simulation 8. Additional Measures 9. Comparing measures B. Variability 1. Measures of Variability 2. Variability Demo 3. Estimating Variance Simulation C. Shape D. Comparing Distributions Demo E. Effects of Transformations F. Variance Sum Law I G. Exercises H. PDF Files (in .zip archive) Descriptive statistics often involves using a few numbers to summarize a distribution. One important aspect of a distribution is where its center is located. Measures of central tendency are discussed first. A second aspect of a distribution is how spread out it is. In other words, how much the numbers in the distribution vary from one another. The second section describes measures of variability. Distributions can differ in shape. Some distributions are symmetric whereas others have long tails in just one direction. The third section describes measures of the shape of distributions. The final two sections concern (1) how transformations affect measures summarizing distributions and (2) the variance sum law, an important relationship involving a measure of variability.

IV. Describing Bivariate Data


Prerequisites none

A. B. C. D. E. F. G. H. I.

Introduction to Bivariate Data Values of the Pearson Correlation Guessing Correlations Simulation Properties of Pearson's r Computing Pearson's r Restriction of Range Demo Variance Sum Law II Exercises PDF Files (in .zip archive)

A dataset with two variables contains what is called bivariate data. This chapter discusses ways to describe the relationship between two variables. For example, you may wish to describe the relationship between the heights and weights of people to determine the extent to which taller people weigh more. The introductory section gives more examples of bivariate relationships and presents the most common way of portraying these relationships graphically. The next five sections discuss Pearson's correlation, the most common index of the relationship between two variables. The final section, "Variance Sum Law II" makes use of Pearson's correlation to generalize this law to bivariate data.

V. Probability
Prerequisites none A. B. C. D. E. F. G. H. I. J. K. L. Introduction Basic Concepts Conditional Probability Demo Gamblers Fallacy Simulation Birthday Demonstration Binomial Distribution Binomial Demonstration Base Rates Bayes' Theorem Demonstration Monty Hall Problem Demonstration Exercises PDF Files (in .zip archive)

Probability is an important and complex field of study. Fortunately, only a few basic issues in probability theory are essential for understanding

statistics at the level covered in this book. These basic issues are covered in this chapter. The introductory section discusses the definitions of probability. This is not as simple as it may seem. The section on basic concepts covers how to compute probabilities in a variety of simple situations. The Gambler's Fallacy Simulation provides an opportunity to explore this fallacy by simulation. The Birthday Demonstration illustrates the probability of finding two or more people with the same birthday. The Binomial Demonstration shows the binomial distribution for different parameters. The section on base rates discusses an important but often-ignored factor in determining probabilities. It also presents Bays' Theorem. The Bays' Theorem Demonstration shows how a tree diagram and Bays' Theorem result in the same answer. Finally, the Monty Hall Demonstration lets you play a game with a very counterintuitive result.

Normal Distributions
Prerequisites none A. B. C. D. E. F. G. H. I. Introduction History Areas of Normal Distributions Varieties of Normal Distribution Demo Standard Normal Normal Approximation to the Binomial Normal Approximation Demo Exercises PDF Files (in .zip archive)

Most of the statistical analyses presented in this book are based on the bellshaped or normal distribution. The introductory section defines what it means for a distribution to be normal and presents some important properties of normal distributions. The interesting history of the discovery of the normal distribution is described in the second section. Methods for calculating probabilities based on the normal distribution are described in Areas of Normal Distributions. The Varieties of Normal Distribution Demo

allows you to enter values for the mean and standard deviation of a normal distribution and see a graph of the resulting distribution. A frequently used normal distribution is called the Standard Normal distribution and is described in the section with that name. The binomial distribution can be approximated by a normal distribution. The section Normal Approximation to the Binomial shows this approximation. The Normal Approximation Demo allows you to explore the accuracy of this approximation.

VII. Sampling Distributions


Prerequisites none A. B. C. D. E. F. G. H. I. J. Introduction Basic Demo Sample Size Demo Central Limit Theorem Demo Sampling Distribution of the Mean Sampling Distribution of Difference Between Means Sampling Distribution of Pearson's r Sampling Distribution of a Proportion Exercises PDF Files (in .zip archive)

The concept of a sampling distribution is perhaps the most basic concept in inferential statistics. It is also a difficult concept to teach because a sampling distribution is a theoretical distribution rather than an empirical distribution. The introductory section defines the concept and gives an example for both a discrete and a continuous distribution. It also discusses how sampling distributions are used in inferential statistics. The Basic Demo is an interactive demonstration of sampling distributions. It is designed to make the abstact concept of sampling distributions more concrete. The Sample Size Demo allows you to investigate the effect of sample size on the sampling distribution of the mean. The Central Limit Theorem (CLT) Demo is an interactive illustration of

a very important and counter-intuitive characteristic of the sampling distribution of the mean. The remaining sections of the chapter concern the sampling distributions of important statistics: the Sampling Distribution of the Mean, the Sampling Distribution of the Difference Between Means, the Sampling Distribution of r, and the Sampling Distribution of a Proportion.

Estimation
Prerequisites See individual sections A. B. C. D. E. Introduction Degrees of Freedom Characteristics of Estimators Bias and Variability Simulation Confidence Intervals 1. Introduction 2. Confidence Interval for the Mean 3. t distribution 4. Confidence Interval Simulation 5. Confidence Interval for the Difference Between Means 6. Confidence Interval for Pearson's Correlation 7. Confidence Interval for a Proportion F. Exercises G. PDF Files (in .zip archive)

One of the major applications of statistics is estimating population parameters from sample statistics . For example, a poll may seek to estimate the proportion of adult residents of a city that support a proposition to build a new sports stadium. Out of a random sample of 200 people, 106 say they support the proposition. Thus in the sample, 0.53 of the people supported the proposition. This value of 0.53 is called a point estimate of the population proportion. It is called a point estimate because the estimate consists of a single value or point.

The concept of degrees of freedom and its relationship to estimation is discussed in Section B. "Characteristics of Estimators" discusses two important concepts: bias and precision. Point estimates are usually supplemented by interval estimates called confidence intervals . Confidence intervals are intervals constructed using a method that contains the population parameter a specified proportion of the time. For example, if the pollster used a method that contains the parameter 95% of the time it is used, he or she would arrive at the following 95% confidence interval: 0.46 < < 0.60. The pollster would then conclude that somewhere between 0.46 and 0.60 of the population supports the proposal. The media usually reports this type of result by saying that 53% favor the proposition with a margin of error of 7%. The sections on confidence interval show how to compute confidence intervals for a variety of parameters.

IX. Logic of Hypothesis Testing


Prerequisites none A. B. C. D. E. F. G. H. I. J. K. Introduction Significance Testing Type I and Type II Errors One- and Two-Tailed Tests Interpreting Significant Results Interpreting Non-Significant Results Steps in Hypothesis Testing Significance Testing and Confidence Intervals Misconceptions Exercises PDF Files (in .zip archive)

When interpreting an experimental finding, a natural question arises as to whether the finding could have occurred by chance. Hypothesis testing is a statistical procedure for testing whether chance is a plausible explanation of an experimental finding. Misconceptions about hypothesis testing are common among practitioners as well as students. To help prevent these misconception, this chapter goes into more detail about the logic of hypothesis testing than is typical for an introductory-level text.

Testing Means
Prerequisites Specified in individual sections A. B. C. D. E. F. G. H. I. J. K. L. Single Mean t Distribution Demo Difference between Two Means (Independent Groups) Robustness Simulation All Pairwise Comparisons Among Means Specific Comparisons Difference between Two Means (Correlated Pairs) Correlated t Simulation Specific Comparisons (Correlated Observations) Pairwise Comparisons (Correlated Observations) Exercises PDF Files (in .zip archive)

Many if not most experiments are designed to compare means. The experiment may involve only one sample mean that is to be compared to a specific value. Or the experiment could be testing differences among many different experimental conditions, and the experimenter could be interested in comparing each mean with each other mean. This chapter covers methods of comparing means in many different experimental situations. The topics covered here in sections D, E, G, and H are typically covered in other texts in a chapter on Analysis of Variance. We prefer to cover them here since they bear no necessary relationship to analysis of variance. As has been pointed out elsewhere, it is not logical to consider the procedures in this chapter tests to be performed subsequent to an analysis of variance. Nor is it logical to call them post-hoc tests as some computer programs do.

Power
Prerequisites Logic of Hypothesis Testing, Normal Distribution, Testing Single Means A. Introduction B. Example Calculations

C. D. E. F. G.

Power Demo 1 Power Demo 2 Factors Affecting Power Exercises PDF Files (in .zip archive)

Prediction
Prerequisites Specified in individual sections A. B. C. D. E. F. G. H. Introduction to Simple Linear Regression Linear Fit Demo Partitioning Sums of Squares Standard Error of the Estimate Prediction Line Demo Inferential Statistics for b and r Exercises PDF Files (in zip format)

Statisticians are often called upon to develop methods to predict one variable from other variables. For example, one might want to predict college grade point average from high school grade point average. Or, one might want to predict income from the number of years of education.

Analysis of Variance
Prerequisites Specified in individual sections A. B. C. D. E. F. G. H. I. J. K. Introduction ANOVA Designs One-Factor ANOVA (Between-Subjects) One-Way Demo Multi-Factor ANOVA (Between-Subjects) Unequal Sample Sizes Tests Supplementing ANOVA Within-Subjects ANOVA Power of Within-Subjects Designs Demo PDF Files (in .zip archive) Exercises

Chi Square

Prerequisites Distributions, Areas of Normal Distributions, Standard Normal Distribution A. B. C. D. E. F. G. Chi Square Distribution One-Way Tables Testing Distributions Demo Contingency Tables 2 x 2 Table Simulation Exercises PDF Files (in .zip archive)

Chi Square is a distribution that has proven to be particularly useful in statistics. The first section describes the basics of this distribution. The following two sections cover the most common statistical tests that make use of the Chi Square distribution. The section "One-Way Tables" shows how to use the Chi Square Distribution to test the difference between theoretically expected and observed frequencies. The section "Contingency Tables" shows how to use Chi Square to test the association between two nominal variables. This use of Chi Square is so common that it is often referred to as the "Chi Square Test." Prerequisites none Analysis Lab Binomial Distribution Chi Square Distribution F Distribution Inverse Normal Distribution Inverse t Distribution Normal Distribution Power Calculator r to z' t Distribution Studentized Range Distribution Does a wide variety of statistical analyses. Computes probabilities of the binomial. Computes area of the Chi Square distribution. Computes areas of the F distribution. Use for confidence intervals. Use for confidence intervals. Computes areas of the normal distribution. Computes power for a two-sample t-test. Computes transformations in both directions. Computes areas of the t distribution. Use for range tests such as the Tukey hsd test.

http://onlinestatbook.com/version_1.html

You might also like