Statistics theory

From Citizendium
Revision as of 22:08, 18 August 2008 by imported>Robert Badgett (→‎Frequentist method: Added a para I helped with at WP)
Jump to navigation Jump to search
This article is developing and not approved.
Main Article
Discussion
Related Articles  [?]
Bibliography  [?]
External Links  [?]
Citable Version  [?]
Advanced [?]
 
This editable Main Article is under development and subject to a disclaimer.

Statistics refers primarily to a branch of mathematics that specializes in enumeration, or counted, data and their relation to measured data.[1] It may also refer to a fact of classification, which is the chief source of all statistics, and has a relationship to psychometric applications in the social sciences.

An individual statistic refers to a derived numerical value, such as a mean, a coefficient of correlation, or some other single concept of descriptive statistics . It may also refer to an idea associated with an average, such as a median, or standard deviation, or some value computed from a set of data. [2]

More precisely, in mathematical statistics, and in general usage, a statistic is defined as any measurable function of a data sample [3]. A data sample is described by instances of a random variable of interest, such as a height, weight, polling results, test performance, etc., obtained by random sampling of a population.

Simple illustration

Suppose one wishes to embark on a quantitative study of the height of adult males in some country C. How should one go about doing this and how can the data be summarized? In statistics, the approach taken is to assume/model the quantity of interest, i.e., "height of adult men from the country C" as a random variable X, say, taking on values in [0,5] (measured in metres) and distributed according to some unknown probability distribution[4] F on [0,5] . One important theme studied in statistics is to develop theoretically sound methods (firmly grounded in probability theory) to learn something about the postulated random variable X and also its distribution F by collecting samples, for this particular example, of the height of a number of men randomly drawn from the adult male population of C.

Suppose that N men labeled have been randomly drawn by simple random sampling (this means that each man in the population is equally likely to be selected in the sampling process) whose heights are , respectively. An important yet subtle point to note here is that, due to random sampling, the data sample obtained is actually an instance or realization of a sequence of independent random variables with each random variable being distributed identically according to the distribution of (that is, each has the distribution F). Such a sequence is referred to in statistics as independent and identically distributed (i.i.d) random variables. To further clarify this point, suppose that there are two other investigators, Tim and Allen, who are also interested in the same quantitative study and they in turn also randomly sample N adult males from the population of C. Let Tim's height data sample be and Allen's be , then both samples are also realizations of the i.i.d sequence , just as the first sample was.

From a data sample one may define a statistic T as for some real-valued function f which is measurable (here with respect to the Borel sets of ). Two examples of commonly used statistics are:

  1. . This statistic is known as the sample mean
  2. . This statistic is known as the sample variance. Often the alternative definition of sample variance is preferred because it is an unbiased estimator of the variance of X, while the former is a biased estimator.

Summary statistics

Inferential statistics and hypothesis testing

Frequentist method

This approach uses mathematical formulas to calculate deductive probabilities (p-value) of an experimental result.[5] This approach can generate confidence intervals.

A problem with the frequentist analyses of p-values is that they may overstate "statistical significance".[6][7] See Bayes factor for details.

Likelihood or Bayesian method

Some argue that the P-value should be interpreted in light of how plausible is the hypothesis based on the totality of prior research and physiologic knowledge.[8][5][9] This approach can generate Bayesian 95% credibility intervals.[10]

Classification

See also

References

  1. Trapp, Robert; Beth Dawson (2004). Basic & clinical biostatistics. New York: Lange Medical Books/McGraw-Hill. LCC QH323.5 .D38LCCN 2005-263. ISBN 0-07-141017-1. 
  2. Guilford, J.P., Fruchter, B. (1978). Fundamental statistics in psychology and education. New York: McGraw-Hill.
  3. Shao, J. (2003). Mathematical Statistics (2 ed.). ser. Springer Texts in Statistics, New York: Springer-Verlag, p. 100.
  4. This is the case in non-parametric statistics. On the other hand, in parametric statistics the underlying distribution is assumed to be of some particular type, say a normal or exponential distribution, but with unknown parameters that are to be estimated.
  5. 5.0 5.1 Goodman SN (1999). "Toward evidence-based medical statistics. 1: The P value fallacy". Ann Intern Med 130: 995–1004. PMID 10383371[e]
  6. Goodman S (1999). "Toward evidence-based medical statistics. 1: The P value fallacy.". Ann Intern Med 130 (12): 995–1004. PMID 10383371.
  7. Goodman S (1999). "Toward evidence-based medical statistics. 2: The Bayes factor.". Ann Intern Med 130 (12): 1005–13. PMID 10383350.
  8. Browner WS, Newman TB (1987). "Are all significant P values created equal? The analogy between diagnostic tests and clinical research". JAMA 257: 2459–63. PMID 3573245[e]
  9. Goodman SN (1999). "Toward evidence-based medical statistics. 2: The Bayes factor". Ann Intern Med 130: 1005–13. PMID 10383350[e]
  10. Gelfand, Alan E.; Sudipto Banerjee; Carlin, Bradley P. (2003). Hierarchical Modeling and Analysis for Spatial Data (Monographs on Statistics and Applied Probability). Boca Raton: Chapman & Hall/CRC. LCC QA278.2 .B36. ISBN 1-58488-410-X.