STATISTICS EXPLAINED: Module descriptions


Description of each module in STATISTICS EXPLAINED


Module Descriptions for STATISTICS EXPLAINED
Contents of STATISTICS EXPLAINED:
(There are over 90 modules!)
    Introduction
Group #1: SOME DESCRIPTIVE STATISTICS
  • Describing collections of numbers. (Modules 1-6)
  • The Normal Curve. (Modules 7-13)
Group #2: SOME INFERENTIAL STATISTICS
  • Sampling Distribution of the Mean. (Modules 14-17)
  • Tests of a hypothesis about a population mean. (Modules 18-23)
  • Sampling distributions based on S-square. (Modules 24-30)
  • Some comparisons among Z, t and F. (Modules 31-38)
  • Tests of the differences between two means. (Modules 39-47)
  • The concept of beta and the power of a test. (Module 48)
Group #3: RELATED MEASURES
  • Introduction to correlation and regression. (Modules 49-53)
  • More on correlation and regression. (Modules 54-62)
  • Tests of related measures. (Modules 63-68)
Group #4: ENUMERATION STATISTICS
  • Chi-Square (goodness of fit). (Modules 69-71)
Group #5: ANALYSIS OF VARIANCE (ANOVA)
  • Introduction to one-way analysis of variance. (Modules 72-74)
  • More on the one-way analysis of variance. (Modules 75-80)
  • The two-way analysis of variance. (Modules 81-86)
  • More on the two-way analysis of variance. (Modules 87-92)
Final sections:
  • Commentary, statistics quiz, tables & glossary (Modules 93-94)
  • About the program and its creators. (Modules 95-100)
  • Other Animated Software Company products. (Modules 101-107)


Here are the topics for each of the modules in Statistics Explained:
(1) On the need for algebra
    This introductory module provides an example of the surprising insights that are possible when we apply some basic algebra to problems that at first glance seem to have obvious answers.
(2) Techniques for visualizing data
    Statisticians need to be able to visualize collections of data. This module describes three of the techniques that have proven especially useful:
(3) Measures of Central Tendency
    Statisticians need a way to indicate where, along a scale of measures, a given distribution is centered. This module describes three measures of central tendency.
(4) Measures of Variability
    Statisticians have developed ways to indicate the degree to which the items in a distribution are separated from each other. The measures of variability described in this module are:
(5) The effects of uniform transformations of X
    This module shows what happens to the mean and variance of a distribution when every item in the distribution is transformed by either adding a given amount to it or by multiplying it by a given amount.
(6) The z score conversion
    When the items in a distribution have been converted to z scores, it is possible to compare the relative standings of items from distributions that initially had different means and variances. This module shows how and why this happens.
(7) Introduction to the Normal Curve
    When one measures a very large number of items and then graphically represents the data using very fine class intervals, the histogram that results is often somewhat bell shaped.
(8) The details of the equation for the Normal Curve
    Re-examining the equation that specifies a normal curve.
(9) Specifying a given normal curve
    Gives an example of normal curve with the same area, mean and standard deviation as a histogram of 100 items.
(10) Normal curves versus curves that are not normal
    This module illustrates several kinds of curves that are not normal, including two bell shaped curves.
(11) Some examples of Normal Distributions
    Because of the nature of our universe, many kinds of measures are distributed normally, and hence we can use the equation for a normal curve to completely describe their distribution.
(12) The arrangement of areas in a normal distribution
    We saw in the previous module that the ordinates in a normal curve are distributed in a special way. In this module we learn that the areas under a normal curve are also distributed in a special way.
(13) The z score conversion of a normal curve
    It would be tremendously burdensome, indeed impossible, to specify the areas under various portions of all possible normal curves, because N, i, and can each have any of an infinite number of possible
(14) The Concept of a Random Sample
    Any set of items with a common observable characteristic can be regarded as a population (or universe). Once we have specified a given population we regard any set of items drawn from that population as a sample.
(15) The Sampling Distribution of the Mean and The Central Limit Theorem
    Introduces the concept of a sampling distribution and it indicates how The Central Limit Theorem can describe the characteristics of a particular Sampling Distribution of the Mean.
(16) The rationale for The Central Limit Theorem
    If we are to understand why the Central Limit Theorm is true, we will have to look very closely at what happens when we take random samples.
(17) The z score conversion of The Sampling Distribution of the Mean
    Explains how to assess the probability of obtaining various values of a sample mean when we take a random sample from a normal population with a known mean and variance.
(18) A test of a hypothesis about a population's mean when the population's variance is known
    Introduces the z test of a statistical hypothesis about the mean of a population from which a random sample has been drawn.
(19) A Review and Summary of the z test
    Reviews and summarizes the ideas presented in the previous module.
(20) An example of the z test
    A practical example to help clarify the strategy described in the previous module.
(21) An introduction to the t test
    Introduces a test of a hypothesis about the mean of a population that can be used when the variance of the population is unknown and must be estimated from the information in the sample.
(22) The rational for using S-squared x to estimate Sigma-squared x
    This is because the expression sigma(X-Mu x) squared is algebriacally equal to sigma(X-Xbar+Xbar-Mu x) squared.
(23) Some notes on the use of n-1 and the concept of degrees of freedom
    This module summarizes the concepts introduced in the previous two modules.
(24) The Sampling Distribution of s squared
    Describes the sampling distribution of the statistic s squared and shows how its form is determined by the statistic's degrees of freedom.
(25) The Sampling Distribution of the statistic (s squared x/sigma-squared x)
    Introduces the statistic and describes some of the situations where it can be used.
(26) Testing a Statistical Hypothesis about the Variance of a Population
    Describes an example that summarizes the procedure to test a statistical hypothesis about the value of sigma-squared x.
(27) The Sampling Distribution of F
    Provides a description of the statistic F which was named after the great statistician Ronald Fisher. F is the ratio of two s squared xs.
(28) The table of F values
    Introduces the table of F values and describes the basic features of its entries.
(29) Two-Tailed tests with the F Table
    Provides further insight into the F table by describing how to use it to carry out a Two-Tailed test.
(30) Two-Tailed tests with the F Table -- A summary
    Provides a short account of the sequence of steps required to carry out a Two-Tailed test using the F tables.
(31) The t-statistic
    Provides an introduction to the t test.
(32) More on the t test
    Provides additional details on the differences between t and z.
(33) The Sampling Distribution of t
    Compares the t distribution to a normal distribution.
(34) The effect on t of the s square in its denominator
    Provides an account of how and why the t distribution differs from a normal distribution.
(35) An example of a t test
    The use of t as a test of a hypothesis about the mean when the variance of the population is not known.
(36) The relationship between t and F
    Indicates the nature of the relationship between t and F.
(37) More on the relationship between t and F
    Indicates the nature of the relationship between t and F.
(38) Uncovering the value of t within an F table
    Provides a quantitative example of how values of t are contained within the F table.
(39) Tests of hypotheses about the means of two populations
    Introduces the Z test of a hypothesis about the means of two populations when the population variances are known.
(40) More on the Z test of the difference between two means
    Examines the details of the Z test of the difference between two means.
(41) An example of a test of a hypothesis about the means from two populations
    Summarizes the Z test of a difference between two sample means with known variances.
(42) The t test of the difference between two sample means
    Introduces the t test of a hypotheses about the means of two populations when the variances are not known.
(43) Some constraints on the use of the t test
    Describes some of the assumptions underlying the t test and discusses how to use t to best advantage.
(44) Some computing formulae for the t test
    Introduces several of the computational formulae for the t test and it shows their derivations.
(45) More on the assumptions that underlie the t test
    Discusses the use of t when its assumptions may not have been met.
(46) More on the relationship between t and F
    Explains the relationship between t and F when dealing with the t test of the difference between two sample means.
(47) An example of a t test of the difference between two means when the population variances are not known
    Provides a summary of the steps in performing a t test of the observed difference between the means of two samples.
(48) The concept of beta and the power of a test
    This somewhat complex module provides an explanation of the important concepts of alpha, beta and Power. Because of the central role of these concepts to all of statistical inference, this module is worthy of especially careful study.
(49) Correlation
    Introduces the concept of a correlation between two measures on each of a set of items.
(50) Quantifying the relationship (the coefficient of correlation)
    This module shows how to quantify a measure of correlation (r).
(51) A Computational Formula for r
    There are times when it is convenient to express r in a form that is designed to make its numerical calculation as convenient as possible. This module shows the derivation of one of the various ways that r can be conveniently calculated.
(52) The concept of regression
    Introduces the concept of regression and relates it to the concept of correlation.
(53) Specifying a regression line
    Shows the derivation of the regression equations that are used to specify the parameters of the line that specifies the best estimate of the value of Y given a value of X.
(54) The z score transformation in the context of correlation
    This module provides a graphic illustration of the relationship between regression and correlation by showing what happens when z score transformations are applied to a set of related measures.
(55) The regression equations after z score transformations
    This module explains why, after the Z score conversion, the value of b is equal to r and the value of a is equal to zero.
(56) Partitioning the variance in the context of correlation
    Shows why the variance of the Y values in the context of correlation can be partitioned into two variances.
(57) More on the partitioning of variance of Y
    An algebraic proof of the equation describing the partitioning of the variance of Y in the context of correlation.
(58) Some graphic examples of the key ideas behind correlation and regression
    Provides a graphic illustration of the deviations that make up:
(59) The second regression line
    Introduces the second regression line, namely the regression of X on Y.
(60) How the regression lines change with changes in strength of a relationship
    Provides a graphic illustration of the scatterplot's characteristic of both weak and strong positive and negative correlations.
(61) Regression toward the mean
    Provides a detailed account of the concept of regression towards the mean.
(62) Some final considerations with respect to correlation and regression
    Discusses some of the factors that must be considered to properly interpret the implications of an obtained correlation coefficient.
(63) A test of a hypothesis about the means of a set of related measures
    Introduces the formula for a z test of a set of related measures.
(64) Some algebraic relationships in the z test of related measures
    Introduces a second formula for the z test of related measures and shows how it is derived.
(65) Tests involving independent measures versus tests involving related measures
    Highlights some of the special features of tests of related measures.
(66) Test of hypotheses about population means using related measures when the population variances are unknown
    Introduces the t test for related measures.
(67) Comparing t and z tests
    Illustrates the differences between the z and t tests for related versus independent measures.
(68) An example of a t test of related measures
    Summarizes the steps in performing a t test of related measures.
(69) Chi-square: An enumeration statistic
    Each of the statistical procedures that we have discussed thus far can be conceptualized as being based on the circumstances that arise when we have a set of n or more items where each item yields a measure of a given magnitude.
(70) The logic of the Chi-square test
    We can gain some appreciation of the logic of Chi Sq. if we consider an "real life" example.
(71) Chi-square: A test for independence
    There are times when each of two aspects of an item are categorized and we wish to determine if the categorizations across items are independent.
(72) An introduction to the analysis of variance
    Up to now we have examined a variety of ways to decide whether or not an observed difference between two means is statistically reliable. At this point, we will examine a technique for deciding whether or not the observed differences among a set of K means is statistically reliable.
(73) The analysis of variance and the F distribution
    Shows how, by using the F distribution, the analysis of variance can be used to test a statistical hypothesis about the means of a set of samples.
(74) A preferred format for reporting the results of an analysis of variance
    Shows both the conceptual and computing formula for an analysis of variance arranged in a preferred format.
(75) Derivation of the computing formula for the between sum of squares
    Demonstrates the algebraic equivalence of the conceptual and computing formula for the between sum of squares.
(76) Derivation of the computing formula for the within sum of squares
    Demonstrates the algebraic equivalence of the conceptual and computing formula for the within sum of squares.
(77) Derivation of the computing formula for the total sum of squares
    Demonstrates the algebraic equivalence of the conceptual and computing formula for the total sum of squares.
(78) An example of a problem requiring a One-Way Analysis of Variance
    Describes an experiment for which a One-Way Analysis of Variance is needed to determine if the results are statistically significant.
(79) The calculations for the example of a One-Way Analysis of Variance
    Illustrates both the conceptual and computing formula approach to the calculations for the example of a One-Way Analysis of Variance.
(80) An additional comment on the analysis of variance
    Describes an experiment for which a One-Way Analysis of Variance is needed to determine if the results are statistically significant.
(81) An introduction to the Two-Way Analysis of Variance
    Provides an example of a Two-Way Analysis of Variance and shows how to calculate the grand mean.
(82) The calculation of the cell means
    This module shows how to calculate the cell means.
(83) The calculation of the row and column means
    This module shows the data used to calculate the means of the several (R) rows and (C) columns.
(84) The calculation of row and column effects
    Introduces a procedure for assessing the effects of the row and the column treatments.
(85) Introduction to the concept of interaction
    Introduces a procedure to test the statistical hypothesis that the row and column treatments are producing independent effects. If we can reject that statistical hypothesis, we will have reason to conclude that the effects of the row and column treatments are interacting with each other.
(86) Variance estimates in the Two-Way ANOVA when the null hypothesis is true
    Discusses the several estimates of the population variance in the Two-Way ANOVA when the null hypothesis is true in each case.
(87) Variance estimates in the Two-Way Analysis of Variance when the null hypothesis is false
    Discusses the several estimates of the population variance in the Two-Way ANOVA when the null hypothesis is false in each case.
(88) Testing for significant effects in the Two-Way Analysis of Variance
    Describes the tests for the statistical significance of the row, column and interaction effects.
(89) A preferred format for reporting the results of a Two-Way ANOVA
    Shows both the conceptual and computing formula for a two-way ANOVA, arranged in a preferred format.
(90) An example of a problem requiring a two-way ANOVA
    Describes an experiment for which a Two-Way ANOVA is needed to determine if the numerical results are statistically significant.
(91) Numerical results for our example of a Two-Way ANOVA
    Shows the calculations necessary for the Two-Way ANOVA of the data presented in the previous module. It also discusses the relationships among the various results of those calculations.
(92) A General Comment about the Two-Way ANOVA
    By indicating that the observed differences among the row and column means are larger than expected by chance, we are led to conclude that those differences are due to the treatments.
(93) A final comment
    Some notes on the proper use of statistics.
(94) Statistics Quiz
    25 Questions presented in random order.
(95) Selected further reading
    Here are some classic as well as some obscure books for those interested in additional concepts related to the material covered in this program.
(96) Thanks to the following people
    (Listed in program)
(97) About Statistics Explained
    This program was originally a book plus a diskette, written by Howard S. Hoffman.
(98) About the authors
    Howard Hoffman is a professor emeritus at Bryn Mawr College, Bryn Mawr, Pennsylvania. He taught psychology and statistics for nearly 50 years. Russell Hoffman, Howard's eldest living son, is a computer programmer and animator, and owner of The Animated Software Company. Sharon Hoffman, Russell's wife, is an editor with experience in technical reports, newsletters and magazines. She has nearly two decades' experience in the field of IBM AS/400 computers.
(99) About The Animated Software Company
    The Animated Software Company was founded in 1984 as "P11 Enterprises".
(100) About Our Web Site
    We began offering our software through online services in 1990. We have offered most of our software via the Internet since then.
(101) Why do we use P11?
    We love the Internet but don't believe its "one size fits all" approach is yet (1999) capable of the animations or smooth user interaction we are able to achieve by using our own development tools.
(102) All About Pumps
    Originally released in 1995, All About Pumps is a complete introduction to the second-most common machine on Earth, after electric motors. YOU need to know about pumps!
(103) The Heart: The Engine of Life
    The Heart: The Engine of Life is an animated tutorial about the human heart.
(104) Russell's "P11" Animation Machine
    Russell's "P11" Animation Machine was originally created to produce THE HEART: THE ENGINE OF LIFE. Work on what eventually became "P11" began in 1984. The current release is 8.68, compiled November 7th, 1994
(105) Global Energy Network International
    Global Energy Network International is a U.S. tax exempt 501(c)(3) corporation committed to improving the quality of life for everyone without damage to the planet. In 1995 the Animated Software Company created an educational tutorial describing the initiative, which is available for free download and distribution via the Internet (www.geni.org).
(106) Assembler Language Source Code Available
    This and all our products are created with a 100% Assembler Language computer program known as "P11".
(107) What's next at The Animated Software Company?
    We will continue to upgrade the products we already have on the market and develop new and exciting interactive educational products along the way.
Statistics Cartoon by Howard S. Hoffman 2005
 FAQ's
Installation or other problems?

Please visit our Q-and-A section first!
Our software
is hand-crafted
in the U.S.A.

banner banner

Catalog / Site Map / Company Home page / Contact / 321...
statistics glossary pump glossary heart glossary list of U.S. Nuclear Power Plants The Demon Hot Atom Poison Fire USA

Last modified MMXIX
Webmaster: Ace Hoffman
Copyright (c) Ace Hoffman