Jump to content

Philosophy of statistics

From Wikipedia, the free encyclopedia

The philosophy of statistics is the study of the mathematical, conceptual, and philosophical foundations and analyses of statistics and statistical inference. For example, Dennis Lindely argues for the more general analysis of statistics as the study of uncertainty.[1] The subject involves the meaning, justification, utility, use and abuse of statistics and its methodology, and ethical and epistemological issues involved in the consideration of choice and interpretation of data and methods of statistics.[2]

Topics of interest

[edit]
  • Foundations of statistics involves issues in theoretical statistics, its goals and optimization methods to meet these goals, parametric assumptions or lack thereof considered in nonparametric statistics, model selection for the underlying probability distribution, and interpretation of the meaning of inferences made using statistics, related to the philosophy of probability and the philosophy of science. Discussion of the selection of the goals and the meaning of optimization, in foundations of statistics, are the subject of the philosophy of statistics. Selection of distribution models, and of the means of selection, is the subject of the philosophy of statistics, whereas the mathematics of optimization is the subject of nonparametric statistics.
  • The relevance and capacity of statistics to make inductive inferences and resolve (in some respect) the problem of induction, first formally raised by David Hume in A Treatise of Human Nature.[2]
  • Issues arise involving sample size, such as cost and efficiency, are common, such as in polling and pharmaceutical research.
  • Extra-mathematical considerations in the design of experiments and accommodating these issues arise in most actual experiments.[further explanation needed]
  • The motivation and justification of data analysis and experimental design, as part of the scientific method are considered.
  • Distinctions between induction and logical deduction relevant to inferences from data and evidence arise, such as when frequentist interpretations are compared with degrees of certainty derived from Bayesian inference. However, the difference between induction and ordinary reasoning is not generally appreciated.[3]
  • Leo Breiman exposed the diversity of thinking in his article on 'The Two Cultures', making the point that statistics has several kinds of inference to make, modelling and prediction amongst them.[4]
  • Issues in the philosophy of statistics arise throughout the history of statistics. Causality considerations arise with interpretations of, and definitions of, correlation, and in the theory of measurement.
  • Objectivity in statistics is often confused with truth whereas it is better understood as replicability, which then needs to be defined in the particular case. Theodore Porter develops this as being the path pursued when trust has evaporated, being replaced with criteria.[5]
  • Ethics associated with epistemology and medical applications arise from potential abuse of statistics, such as selection of method or transformations of the data to arrive at different probability conclusions for the same data set. For example, the meaning of applications of a statistical inference to a single person, such as one single cancer patient, when there is no frequentist interpretation for that patient to adopt.
  • Campaigns for statistical literacy must wrestle with the problem that most interesting questions around individual risk are very difficult to determine or interpret, even with the computer power currently available.

Notes

[edit]
  1. ^ Lindley, Dennis (September 2000). "The philosophy of statistics". Royal Statistical Society. 49 (3): 293–337. doi:10.1111/1467-9884.00238. Retrieved June 21, 2024.
  2. ^ a b Romijn, Jan-Willem (2014). "Philosophy of statistics". Stanford Encyclopedia of Philosophy.
  3. ^ Hacking 2006.
  4. ^ Breiman 2001.
  5. ^ Porter 1995.

Further reading

[edit]