161:
29:
229:
or simply through natural deviations in populations. A sample may have been contaminated with elements from outside the population being examined. Alternatively, an outlier could be the result of a flaw in the assumed theory, calling for further investigation by the researcher. Additionally, the pathological appearance of outliers of a certain form appears in a variety of datasets, indicating that the causative mechanism for the data might differ at the extreme end (
1818:
1689:
rejection, they do not make the practice more scientifically or methodologically sound, especially in small sets or where a normal distribution cannot be assumed. Rejection of outliers is more acceptable in areas of practice where the underlying model of the process being measured and the usual distribution of measurement error are confidently known.
1738:, the sample variance increases with the sample size, the sample mean fails to converge as the sample size increases, and outliers are expected at far larger rates than for a normal distribution. Even a slight difference in the fatness of the tails can make a large difference in the expected number of extreme values.
1323:
2013:, p. 1 stating "An outlying observation may be merely an extreme manifestation of the random variability inherent in the data. ... On the other hand, an outlying observation may be the result of gross deviation from prescribed experimental procedure or an error in calculating or recording the numerical value."
107:, or it may be that some observations are far from the center of the data. Outlier points can therefore indicate faulty data, erroneous procedures, or areas where a certain theory might not be valid. However, in large samples, a small number of outliers is to be expected (and not due to any anomalous condition).
1684:
When deciding whether to remove an outlier, the cause has to be considered. As mentioned earlier, if the outlier's origin can be attributed to an experimental error, or if it can be otherwise determined that the outlying data point is erroneous, it is generally recommended to remove it. However, it
699:
The modified
Thompson Tau test is a method used to determine if an outlier exists in a data set. The strength of this method lies in the fact that it takes into account a data set's standard deviation, average and provides a statistically determined rejection zone; thus providing an objective method
180:
or more from the mean, and 1 in 370 will deviate by three times the standard deviation. In a sample of 1000 observations, the presence of up to five observations deviating from the mean by more than three times the standard deviation is within the range of what can be expected, being less than twice
1675:
Even when a normal distribution model is appropriate to the data being analyzed, outliers are expected for large sample sizes and should not automatically be discarded if that is the case. Instead, one should use a method that is robust to outliers to model or analyze data with naturally occurring
228:
Outliers can have many anomalous causes. A physical apparatus for taking measurements may have suffered a transient malfunction. There may have been an error in data transmission or transcription. Outliers arise due to changes in system behaviour, fraudulent behaviour, human error, instrument error
133:
of the data will be between 20 and 25 °C but the mean temperature will be between 35.5 and 40 °C. In this case, the median better reflects the temperature of a randomly sampled object (but not the temperature in the room) than the mean; naively interpreting the mean as "a typical sample",
241:
There is no rigid mathematical definition of what constitutes an outlier; determining whether or not an observation is an outlier is ultimately a subjective exercise. There are various methods of outlier detection, some of which are treated as synonymous with novelty detection. Some are graphical
1688:
Removing a data point solely because it is an outlier, on the other hand, is a controversial practice, often frowned upon by many scientists and science instructors, as it typically invalidates statistical results. While mathematical criteria provide an objective and quantitative method for data
349:
such observations. The principle upon which it is proposed to solve this problem is, that the proposed observations should be rejected when the probability of the system of errors obtained by retaining them is less than that of the system of errors obtained by their rejection multiplied by the
825:
1700:. Trimming discards the outliers whereas Winsorising replaces the outliers with the nearest "nonsuspect" data. Exclusion can also be a consequence of the measurement process, such as when an experiment is not entirely capable of measuring such extreme values, resulting in
973:
is removed if it is an outlier). Meaning, if a data point is found to be an outlier, it is removed from the data set and the test is applied again with a new average and rejection region. This process is continued until no outliers remain in a data set.
700:
to determine if a data point is an outlier. How it works: First, a data set's average is determined. Next the absolute deviation between each data point and the average are determined. Thirdly, a rejection region is determined using the formula:
1557:
1043:
56:
that differs significantly from other observations. An outlier may be due to a variability in the measurement, an indication of novel data, or it may be the result of experimental error; the latter are sometimes excluded from the
977:
Some work has also examined outliers for nominal (or categorical) data. In the context of a set of examples (or instances) in a data set, instance hardness measures the probability that an instance will be misclassified (
253:
Model-based methods which are commonly used for identification assume that the data are from a normal distribution, and identify observations which are deemed "unlikely" based on mean and standard deviation:
706:
2708:
Bakker, Marjan; Wicherts, Jelte M. (2014). "Outlier removal, sum scores, and the inflation of the type I error rate in independent samples t tests: The power of alternatives and recommendations".
956:
1048:
118:, or both, depending on whether they are extremely high or low. However, the sample maximum and minimum are not always outliers because they may not be unusually far from other observations.
84:. A frequent cause of outliers is a mixture of two distributions, which may be two distinct sub-populations, or may indicate 'correct trial' versus 'measurement error'; this is modeled by a
868:
1602:
185:– and not indicate an anomaly. If the sample size is only 100, however, just three such outliers are already reason for concern, being more than 11 times the expected number.
1318:{\displaystyle {\begin{aligned}IH(\langle x,y\rangle )&=\sum _{H}(1-p(y,x,h))p(h|t)\\&=\sum _{H}p(h|t)-p(y,x,h)p(h|t)\\&=1-\sum _{H}p(y,x,h)p(h|t).\end{aligned}}}
1401:
1393:
1019:
1653:
550:
1367:
1629:
603:
419:
392:
629:
1711:
problems, an alternative approach may be to only exclude points which exhibit a large degree of influence on the estimated coefficients, using a measure such as
220:
is approximately 0.3%, and thus for 1000 trials one can approximate the number of samples whose deviation exceeds 3 sigmas by a
Poisson distribution with λ = 3.
573:
347:
327:
2090:
2800:
Cook, R. Dennis (Feb 1977). "Detection of
Influential Observations in Linear Regression". Technometrics (American Statistical Association) 19 (1): 15–18.
2590:(2012). "Local outlier detection reconsidered: A generalized view on locality with applications to spatial, video, and network outlier detection".
1850:
In cases where the cause of the outliers is known, it may be possible to incorporate this effect into the model structure, for example by using a
350:
probability of making so many, and no more, abnormal observations. (Quoted in the editorial note on page 516 to Peirce (1982 edition) from
329:
observations the limit of error, beyond which all observations involving so great an error may be rejected, provided there are as many as
121:
Naive interpretation of statistics derived from data sets that include outliers may be misleading. For example, if one is calculating the
2152:
Pimentel, M. A., Clifton, D. A., Clifton, L., & Tarassenko, L. (2014). A review of novelty detection. Signal
Processing, 99, 215-249.
2001:
Pimentel, M. A., Clifton, D. A., Clifton, L., & Tarassenko, L. (2014). A review of novelty detection. Signal
Processing, 99, 215-249.
2863:
Roberts, S. and
Tarassenko, L.: 1995, A probabilistic resource allocating network for novelty detection. Neural Computation 6, 270–284.
2314:
2306:
2042:
820:{\displaystyle {\text{Rejection Region}}{=}{\frac {{t_{\alpha /2}}{\left(n-1\right)}}{{\sqrt {n}}{\sqrt {n-2+{t_{\alpha /2}^{2}}}}}}}
2953:
1949:
An outlying observation, or "outlier," is one that appears to deviate markedly from other members of the sample in which it occurs.
134:
equivalent to the median, is incorrect. As illustrated in this case, outliers may indicate data points that belong to a different
2166:
1663:
The choice of how to deal with an outlier should depend on the cause. Some estimators are highly sensitive to outliers, notably
2026:
2434:
2406:
2235:
2073:
1983:
885:
61:. An outlier can be an indication of exciting possibility, but can also cause serious problems in statistical analyses.
1730:
The possibility should be considered that the underlying distribution of the data is not approximately normal, having "
64:
Outliers can occur by chance in any distribution, but they can indicate novel behaviour or structures in the data-set,
2370:
Report of the
Superintendent of the United States Coast Survey Showing the Progress of the Survey During the Year 1870
2752:
2570:
1664:
2041:
Chandan
Mukherjee, Howard White, Marc Wuyts, 1998, "Econometrics and Data Analysis for Developing Countries Vol. 1"
1882:
834:
2129:
37:
2520:
2507:. Proceedings of the 2000 ACM SIGMOD international conference on Management of data - SIGMOD '00. p. 427.
2963:
2931:
882:
is the sample size, and s is the sample standard deviation. To determine if a value is an outlier: Calculate
263:
76:
to outliers, while in the case of heavy-tailed distributions, they indicate that the distribution has high
1776:'s. When outliers occur, this intersection could be empty, and we should relax a small number of the sets
2926:
694:
2958:
1867:
1655:. Instance hardness provides a continuous value for determining if an instance is an outlier instance.
1565:
1552:{\displaystyle IH_{L}(\langle x,y\rangle )=1-{\frac {1}{|L|}}\sum _{j=1}^{|L|}p(y|x,g_{j}(t,\alpha ))}
2458:
Knorr, E. M.; Ng, R. T.; Tucakov, V. (2000). "Distance-based outliers: Algorithms and applications".
1033:). Ideally, instance hardness would be calculated by summing over the set of all possible hypotheses
69:
2625:
2391:(1986) . "On the Theory of Errors of Observation". In Kloesel, Christian J. W.; et al. (eds.).
2375:
2196:
2472:
2065:
1851:
1369:
is unknown for many algorithms. Thus, instance hardness can be approximated using a diverse subset
258:
193:
104:
20:
2937:
2091:"There and back again: Outlier detection between statistical reasoning and data mining algorithms"
196:
from what can be expected: for a given cutoff (so samples fall beyond the cutoff with probability
2769:
1897:
1892:
1783:(as small as possible) in order to avoid any inconsistency. This can be done using the notion of
1693:
243:
2638:
2628:".Journal of the Royal Statistical Society. Series B (Methodological), Vol. 47, No. 1, pp. 53-55
2398:
2392:
1372:
2467:
2388:
2365:
2191:
1701:
135:
981:
2921:
1975:
1969:
1965:
1877:
1638:
431:
303:
290:
are often used to detect outliers, especially in the development of linear regression models.
287:
201:
2057:
1931:
Grubbs, F. E. (February 1969). "Procedures for detecting outlying observations in samples".
1335:
1907:
1887:
1788:
1607:
678:
582:
397:
370:
283:
216:. Thus if one takes a normal distribution with cutoff 3 standard deviations from the mean,
209:
182:
148:
capable of coping with outliers are said to be robust: the median is a robust statistic of
2823:
72:. In the case of measurement error, one wishes to discard them or use statistics that are
40:
displaying four outliers in the middle column, as well as one outlier in the first column.
8:
2541:
2257:(2012). "A survey on unsupervised outlier detection in high-dimensional numerical data".
2205:
2058:
1735:
1708:
677:
may take other approaches. Some of these may be distance-based and density-based such as
608:
364:
169:
139:
81:
425:
respectively, then one could define an outlier to be any observation outside the range:
2846:
2810:
2687:
2607:
2587:
2537:
2485:
2424:
2347:
2274:
2254:
2209:
2182:
Hodge, Victoria J.; Austin, Jim (2004), "A Survey of
Outlier Detection Methodologies",
2121:
1712:
969:
The modified
Thompson Tau test is used to find one outlier at a time (largest value of
682:
558:
332:
312:
181:
the expected number and hence within 1 standard deviation of the expected number – see
177:
2748:
2725:
2691:
2679:
2566:
2516:
2440:
2430:
2402:
2231:
2113:
2069:
1979:
1902:
1872:
650:
640:
268:
73:
65:
2850:
2626:
A Note on Restricted Maximum Likelihood Estimation with an Alternative Outlier Model
2611:
2489:
2125:
2023:
2882:
2842:
2838:
2781:
2717:
2671:
2599:
2558:
2508:
2477:
2397:. Vol. 3, 1872–1878. Bloomington, Indiana: Indiana University Press. pp.
2339:
2278:
2266:
2213:
2201:
2105:
1990:
An outlier is an observation that is far removed from the rest of the observations.
1944:
1940:
173:
160:
149:
96:
2291:
152:, while the mean is not. However, the mean is generally a more precise estimator.
2551:
Proceedings of the 2000 ACM SIGMOD International Conference on Management of Data
2327:
2302:
2162:
2030:
666:
126:
92:
91:
In most larger samplings of data, some data points will be further away from the
2873:
Bishop, C. M. (August 1994). "Novelty detection and Neural Network validation".
80:
and that one should be very cautious in using tools or intuitions that assume a
1747:
115:
111:
2886:
2786:
2603:
2947:
2905:
2117:
1961:
1855:
1719:
662:
293:
Subspace and correlation based techniques for high-dimensional numerical data
85:
125:
temperature of 10 objects in a room, and nine of them are between 20 and 25
2729:
2683:
2444:
658:
2562:
2512:
2481:
28:
1912:
1697:
670:
230:
2351:
1029:
represent the input attribute value for an instance in the training set
2721:
2675:
646:
576:
53:
45:
2270:
2109:
2910:
2343:
1731:
145:
2368:(1873) . "Appendix No. 21. On the Theory of Errors of Observation".
2098:
Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery
2460:
The VLDB Journal the International Journal on Very Large Data Bases
1817:
1765:(instead of a probability density function). If no outliers occur,
422:
247:
77:
58:
33:
2064:. Studies in Computational Intelligence Vol. 5. Springer. p.
188:
In general, if the nature of the population distribution is known
2053:
654:
176:
means that roughly 1 in 22 observations will differ by twice the
122:
1685:
is more desirable to correct the erroneous value, if possible.
200:) of a given distribution, the number of outliers will follow a
19:
This article is about the statistical term. For other uses, see
16:
Observation far apart from others in statistics and data science
2554:
130:
110:
Outliers, being the most extreme observations, may include the
100:
2659:
2292:
E178: Standard Practice for Dealing With Outlying Observations
363:
Other methods flag observations based on measures such as the
280:
E178: Standard Practice for Dealing With Outlying Observations
95:
than what is deemed reasonable. This can be due to incidental
2824:"Probabilistic set-membership approach for robust regression"
2505:
Efficient algorithms for mining outliers from large data sets
2811:
Cauchy Distribution. From MathWorld--A Wolfram Web Resource
2745:
Data Analysis: A Statistical Primer for Psychology Students
277:
192:, it is possible to test if the number of outliers deviate
2413:– Appendix 21, according to the editorial note on page 515
1722:, this should be clearly stated on any subsequent report.
1814:-relaxed intersection could be suspected to be outliers.
2332:
Proceedings of the American Academy of Arts and Sciences
2056:; Chen, Guoqing; Kerre, Etienne (2005). Wets, G. (ed.).
962:> Rejection Region, the data point is an outlier. If
2637:
Smith, M.R.; Martinez, T.; Giraud-Carrier, C. (2014). "
2875:
IEE Proceedings - Vision, Image, and Signal Processing
2535:
2394:
Writings of Charles S. Peirce: A Chronological Edition
2307:"Criterion for the Rejection of Doubtful Observations"
966:≤ Rejection Region, the data point is not an outlier.
889:
838:
1641:
1610:
1568:
1404:
1375:
1338:
1046:
984:
888:
837:
709:
611:
585:
561:
434:
400:
373:
335:
315:
2770:"Simplified estimation from censored normal samples"
2585:
2502:
2252:
2060:
Intelligent Data Mining: Techniques and Applications
1795:-relaxed intersection corresponds to the set of all
1750:
considers that the uncertainty corresponding to the
951:{\displaystyle \scriptstyle \delta =|(X-mean(X))/s|}
685:to label observations as outliers or non-outliers.
681:(LOF). Some approaches may use the distance to the
2919:
1692:The two common approaches to exclude outliers are
1647:
1623:
1596:
1551:
1387:
1361:
1317:
1013:
950:
862:
819:
623:
597:
567:
544:
413:
386:
341:
321:
208:, which can generally be well-approximated by the
2082:
1718:If a data point (or points) is excluded from the
2945:
2902:
2330:(May 1877 – May 1878). "On Peirce's criterion".
1604:is the hypothesis induced by learning algorithm
645:In various domains such as, but not limited to,
1741:
1328:Practically, this formulation is unfeasible as
164:Relative probabilities in a normal distribution
2707:
2457:
2161:
2088:
1974:(2nd ed.). New York: MacMillan. pp.
2703:
2701:
2639:An Instance Level Analysis of Data Complexity
2543:LOF: Identifying Density-based Local Outliers
2503:Ramaswamy, S.; Rastogi, R.; Shim, K. (2000).
2052:
688:
537:
437:
2653:
2651:
2649:
2647:
2225:
1433:
1421:
1072:
1060:
863:{\displaystyle \scriptstyle {t_{\alpha /2}}}
236:
2664:Journal of Experimental Psychology: General
2660:"Outliers may not be automatically removed"
2326:
2248:
2246:
2181:
1754:th measurement of an unknown random vector
1725:
309:It is proposed to determine in a series of
2831:Journal of Statistical Theory and Practice
2747:. Transaction Publishers. pp. 24–25.
2698:
2785:
2644:
2471:
2195:
1769:should belong to the intersection of all
2243:
2089:Zimek, Arthur; Filzmoser, Peter (2018).
1816:
1658:
1332:is potentially infinite and calculating
695:Studentized residual § Distribution
159:
155:
27:
2168:Robust Regression and Outlier Detection
1960:
870:is the critical value from the Student
634:
2946:
2920:Balakrishnan, N.; Childs, A. (2001) ,
2872:
2821:
2387:
2378:(goes to Report p. 200, PDF's p. 215).
2364:
2010:
1930:
1734:". For instance, when sampling from a
2903:
2774:The Annals of Mathematical Statistics
2767:
2657:
2422:
2171:(3rd ed.), John Wiley & Sons
1845:
297:
129:, but an oven is at 175 °C, the
2742:
2641:". Machine Learning, 95(2): 225-256.
2259:Statistical Analysis and Data Mining
1825:-relaxed intersection of 6 sets for
1791:. As illustrated by the figure, the
103:that generated an assumed family of
2592:Data Mining and Knowledge Discovery
2226:Barnett, Vic; Lewis, Toby (1994) ,
13:
2206:10.1023/B:AIRE.0000045502.10941.a9
631:indicates data that is "far out".
14:
2975:
2896:
1665:estimation of covariance matrices
358:
2540:; Ng, R. T.; Sander, J. (2000).
1883:Data transformation (statistics)
1799:which belong to all sets except
1597:{\displaystyle g_{j}(t,\alpha )}
1025:is the assigned class label and
2954:Statistical charts and diagrams
2866:
2857:
2815:
2803:
2794:
2761:
2736:
2631:
2618:
2579:
2529:
2496:
2451:
2416:
2381:
2358:
2320:
2296:
2285:
2219:
2175:
2155:
68:, or that the population has a
2843:10.1080/15598608.2010.10411978
2184:Artificial Intelligence Review
2146:
2046:
2035:
2016:
2004:
1995:
1954:
1945:10.1080/00401706.1969.10490657
1924:
1591:
1579:
1546:
1543:
1531:
1511:
1504:
1495:
1487:
1463:
1455:
1436:
1418:
1356:
1349:
1342:
1305:
1298:
1291:
1285:
1267:
1235:
1228:
1221:
1215:
1197:
1188:
1181:
1174:
1148:
1141:
1134:
1128:
1125:
1107:
1095:
1075:
1057:
1008:
1001:
994:
943:
931:
928:
922:
901:
897:
555:for some nonnegative constant
532:
506:
484:
458:
1:
1918:
2315:Errata to the original paper
2228:Outliers in Statistical Data
1971:Introduction to Econometrics
1742:Set-membership uncertainties
1679:
1670:
605:indicates an "outlier", and
7:
2927:Encyclopedia of Mathematics
1861:
1810:that do not intersect the
246:. Others are model-based.
38:Michelson–Morley experiment
10:
2980:
2768:Dixon, W. J. (June 1960).
1868:Anomaly (natural sciences)
1388:{\displaystyle L\subset H}
692:
689:Modified Thompson Tau test
638:
579:proposed this test, where
301:
264:Grubbs's test for outliers
18:
2658:Karch, Julian D. (2023).
2604:10.1007/s10618-012-0300-z
2586:Schubert, E.; Zimek, A.;
2426:Exploratory Data Analysis
2253:Zimek, A.; Schubert, E.;
237:Definitions and detection
223:
105:probability distributions
70:heavy-tailed distribution
2940:described by NIST manual
2743:Wike, Edward L. (2006).
1852:hierarchical Bayes model
1758:is represented by a set
1726:Non-normal distributions
1631:trained on training set
1014:{\displaystyle 1-p(y|x)}
421:are the lower and upper
244:normal probability plots
21:Outlier (disambiguation)
2887:10.1049/ip-vis:19941330
2787:10.1214/aoms/1177705900
2389:Peirce, Charles Sanders
2366:Peirce, Charles Sanders
2022:Ripley, Brian D. 2004.
1898:Random sample consensus
1893:Influential observation
1748:set membership approach
1648:{\displaystyle \alpha }
878:-2 degrees of freedom,
545:{\displaystyle {\big }}
2624:Thompson .R. (1985). "
2423:Tukey, John W (1977).
1842:
1649:
1625:
1598:
1553:
1500:
1389:
1363:
1362:{\displaystyle p(h|t)}
1319:
1015:
952:
864:
821:
625:
599:
569:
546:
415:
388:
356:
343:
323:
165:
41:
2710:Psychological Methods
2563:10.1145/335191.335388
2513:10.1145/342009.335437
2482:10.1007/s007780050006
2230:(3 ed.), Wiley,
1820:
1659:Working with outliers
1650:
1635:with hyperparameters
1626:
1624:{\displaystyle g_{j}}
1599:
1554:
1470:
1390:
1364:
1320:
1016:
953:
865:
822:
626:
600:
598:{\displaystyle k=1.5}
570:
547:
416:
414:{\displaystyle Q_{3}}
389:
387:{\displaystyle Q_{1}}
354:2:558 by Chauvenet.)
352:A Manual of Astronomy
344:
324:
307:
259:Chauvenet's criterion
202:binomial distribution
163:
156:Occurrence and causes
138:than the rest of the
31:
2964:Statistical outliers
2311:Astronomical Journal
2165:; Leroy, A. (1996),
1908:Studentized residual
1888:Extreme value theory
1789:relaxed intersection
1639:
1608:
1566:
1402:
1373:
1336:
1044:
982:
886:
835:
707:
679:Local Outlier Factor
635:In anomaly detection
609:
583:
559:
432:
398:
371:
333:
313:
284:Mahalanobis distance
210:Poisson distribution
183:Poisson distribution
170:normally distributed
2822:Jaulin, L. (2010).
2809:Weisstein, Eric W.
2557:. pp. 93–104.
1736:Cauchy distribution
810:
683:k-nearest neighbors
624:{\displaystyle k=3}
365:interquartile range
82:normal distribution
2722:10.1037/met0000014
2676:10.1037/xge0001357
2429:. Addison-Wesley.
2029:2012-10-21 at the
1878:Anscombe's quartet
1846:Alternative models
1843:
1696:(or trimming) and
1645:
1621:
1594:
1549:
1385:
1359:
1315:
1313:
1263:
1170:
1094:
1011:
948:
947:
874:distribution with
860:
859:
817:
788:
621:
595:
565:
542:
411:
384:
367:. For example, if
339:
319:
304:Peirce's criterion
298:Peirce's criterion
178:standard deviation
166:
42:
2959:Robust statistics
2436:978-0-201-07616-5
2408:978-0-253-37201-7
2313:II 45 (1852) and
2271:10.1002/sam.11161
2237:978-0-471-93094-5
2110:10.1002/widm.1280
2075:978-3-540-26256-5
2024:Robust statistics
1985:978-0-02-374545-4
1903:Robust regression
1873:Novelty detection
1468:
1254:
1161:
1085:
815:
812:
771:
713:
675:anomaly detection
651:signal processing
641:Anomaly detection
568:{\displaystyle k}
342:{\displaystyle n}
322:{\displaystyle m}
66:measurement error
36:of data from the
2971:
2934:
2916:
2915:
2891:
2890:
2870:
2864:
2861:
2855:
2854:
2828:
2819:
2813:
2807:
2801:
2798:
2792:
2791:
2789:
2765:
2759:
2758:
2740:
2734:
2733:
2705:
2696:
2695:
2670:(6): 1735–1753.
2655:
2642:
2635:
2629:
2622:
2616:
2615:
2583:
2577:
2576:
2548:
2536:Breunig, M. M.;
2533:
2527:
2526:
2500:
2494:
2493:
2475:
2455:
2449:
2448:
2420:
2414:
2412:
2385:
2379:
2373:
2362:
2356:
2355:
2344:10.2307/25138498
2328:Peirce, Benjamin
2324:
2318:
2300:
2294:
2289:
2283:
2282:
2250:
2241:
2240:
2223:
2217:
2216:
2199:
2179:
2173:
2172:
2159:
2153:
2150:
2144:
2143:
2141:
2140:
2134:
2128:. Archived from
2095:
2086:
2080:
2079:
2063:
2050:
2044:
2039:
2033:
2020:
2014:
2008:
2002:
1999:
1993:
1992:
1958:
1952:
1951:
1928:
1654:
1652:
1651:
1646:
1634:
1630:
1628:
1627:
1622:
1620:
1619:
1603:
1601:
1600:
1595:
1578:
1577:
1558:
1556:
1555:
1550:
1530:
1529:
1514:
1499:
1498:
1490:
1484:
1469:
1467:
1466:
1458:
1449:
1417:
1416:
1394:
1392:
1391:
1386:
1368:
1366:
1365:
1360:
1352:
1331:
1324:
1322:
1321:
1316:
1314:
1301:
1262:
1241:
1231:
1184:
1169:
1154:
1144:
1093:
1036:
1032:
1028:
1024:
1020:
1018:
1017:
1012:
1004:
957:
955:
954:
949:
946:
938:
900:
873:
869:
867:
866:
861:
858:
857:
856:
852:
826:
824:
823:
818:
816:
814:
813:
811:
809:
804:
800:
774:
772:
767:
764:
763:
762:
758:
742:
741:
740:
736:
721:
719:
714:
712:Rejection Region
711:
630:
628:
627:
622:
604:
602:
601:
596:
574:
572:
571:
566:
551:
549:
548:
543:
541:
540:
531:
530:
518:
517:
499:
498:
483:
482:
470:
469:
451:
450:
441:
440:
420:
418:
417:
412:
410:
409:
393:
391:
390:
385:
383:
382:
348:
346:
345:
340:
328:
326:
325:
320:
174:three sigma rule
150:central tendency
99:or flaws in the
97:systematic error
2979:
2978:
2974:
2973:
2972:
2970:
2969:
2968:
2944:
2943:
2899:
2894:
2871:
2867:
2862:
2858:
2826:
2820:
2816:
2808:
2804:
2799:
2795:
2766:
2762:
2755:
2741:
2737:
2706:
2699:
2656:
2645:
2636:
2632:
2623:
2619:
2588:Kriegel, H. -P.
2584:
2580:
2573:
2546:
2534:
2530:
2523:
2501:
2497:
2456:
2452:
2437:
2421:
2417:
2409:
2386:
2382:
2363:
2359:
2325:
2321:
2303:Benjamin Peirce
2301:
2297:
2290:
2286:
2251:
2244:
2238:
2224:
2220:
2197:10.1.1.109.1943
2180:
2176:
2160:
2156:
2151:
2147:
2138:
2136:
2132:
2093:
2087:
2083:
2076:
2051:
2047:
2040:
2036:
2031:Wayback Machine
2021:
2017:
2009:
2005:
2000:
1996:
1986:
1959:
1955:
1929:
1925:
1921:
1864:
1848:
1809:
1782:
1775:
1764:
1744:
1728:
1713:Cook's distance
1682:
1673:
1661:
1640:
1637:
1636:
1632:
1615:
1611:
1609:
1606:
1605:
1573:
1569:
1567:
1564:
1563:
1525:
1521:
1510:
1494:
1486:
1485:
1474:
1462:
1454:
1453:
1448:
1412:
1408:
1403:
1400:
1399:
1374:
1371:
1370:
1348:
1337:
1334:
1333:
1329:
1312:
1311:
1297:
1258:
1239:
1238:
1227:
1180:
1165:
1152:
1151:
1140:
1089:
1078:
1047:
1045:
1042:
1041:
1034:
1030:
1026:
1022:
1000:
983:
980:
979:
942:
934:
896:
887:
884:
883:
871:
848:
844:
840:
839:
836:
833:
832:
805:
796:
792:
787:
773:
766:
765:
748:
744:
743:
732:
728:
724:
723:
722:
720:
715:
710:
708:
705:
704:
697:
691:
643:
637:
610:
607:
606:
584:
581:
580:
560:
557:
556:
536:
535:
526:
522:
513:
509:
494:
490:
478:
474:
465:
461:
446:
442:
436:
435:
433:
430:
429:
405:
401:
399:
396:
395:
378:
374:
372:
369:
368:
361:
334:
331:
330:
314:
311:
310:
306:
300:
239:
226:
204:with parameter
168:In the case of
158:
127:degrees Celsius
24:
17:
12:
11:
5:
2977:
2967:
2966:
2961:
2956:
2942:
2941:
2935:
2917:
2898:
2897:External links
2895:
2893:
2892:
2881:(4): 217–222.
2865:
2856:
2814:
2802:
2793:
2780:(2): 385–391.
2760:
2753:
2735:
2716:(3): 409–427.
2697:
2643:
2630:
2617:
2578:
2571:
2538:Kriegel, H.-P.
2528:
2521:
2495:
2473:10.1.1.43.1842
2450:
2435:
2415:
2407:
2380:
2357:
2319:
2295:
2284:
2265:(5): 363–387.
2255:Kriegel, H.-P.
2242:
2236:
2218:
2174:
2154:
2145:
2081:
2074:
2045:
2034:
2015:
2003:
1994:
1984:
1962:Maddala, G. S.
1953:
1922:
1920:
1917:
1916:
1915:
1910:
1905:
1900:
1895:
1890:
1885:
1880:
1875:
1870:
1863:
1860:
1847:
1844:
1807:
1803:of them. Sets
1780:
1773:
1762:
1743:
1740:
1727:
1724:
1681:
1678:
1672:
1669:
1660:
1657:
1644:
1618:
1614:
1593:
1590:
1587:
1584:
1581:
1576:
1572:
1560:
1559:
1548:
1545:
1542:
1539:
1536:
1533:
1528:
1524:
1520:
1517:
1513:
1509:
1506:
1503:
1497:
1493:
1489:
1483:
1480:
1477:
1473:
1465:
1461:
1457:
1452:
1447:
1444:
1441:
1438:
1435:
1432:
1429:
1426:
1423:
1420:
1415:
1411:
1407:
1384:
1381:
1378:
1358:
1355:
1351:
1347:
1344:
1341:
1326:
1325:
1310:
1307:
1304:
1300:
1296:
1293:
1290:
1287:
1284:
1281:
1278:
1275:
1272:
1269:
1266:
1261:
1257:
1253:
1250:
1247:
1244:
1242:
1240:
1237:
1234:
1230:
1226:
1223:
1220:
1217:
1214:
1211:
1208:
1205:
1202:
1199:
1196:
1193:
1190:
1187:
1183:
1179:
1176:
1173:
1168:
1164:
1160:
1157:
1155:
1153:
1150:
1147:
1143:
1139:
1136:
1133:
1130:
1127:
1124:
1121:
1118:
1115:
1112:
1109:
1106:
1103:
1100:
1097:
1092:
1088:
1084:
1081:
1079:
1077:
1074:
1071:
1068:
1065:
1062:
1059:
1056:
1053:
1050:
1049:
1010:
1007:
1003:
999:
996:
993:
990:
987:
945:
941:
937:
933:
930:
927:
924:
921:
918:
915:
912:
909:
906:
903:
899:
895:
892:
855:
851:
847:
843:
829:
828:
808:
803:
799:
795:
791:
786:
783:
780:
777:
770:
761:
757:
754:
751:
747:
739:
735:
731:
727:
718:
690:
687:
673:, the task of
639:Main article:
636:
633:
620:
617:
614:
594:
591:
588:
564:
553:
552:
539:
534:
529:
525:
521:
516:
512:
508:
505:
502:
497:
493:
489:
486:
481:
477:
473:
468:
464:
460:
457:
454:
449:
445:
439:
408:
404:
381:
377:
360:
359:Tukey's fences
357:
338:
318:
302:Main article:
299:
296:
295:
294:
291:
281:
275:
266:
261:
250:are a hybrid.
238:
235:
225:
222:
157:
154:
116:sample minimum
112:sample maximum
15:
9:
6:
4:
3:
2:
2976:
2965:
2962:
2960:
2957:
2955:
2952:
2951:
2949:
2939:
2936:
2933:
2929:
2928:
2923:
2918:
2913:
2912:
2907:
2904:Renze, John.
2901:
2900:
2888:
2884:
2880:
2876:
2869:
2860:
2852:
2848:
2844:
2840:
2836:
2832:
2825:
2818:
2812:
2806:
2797:
2788:
2783:
2779:
2775:
2771:
2764:
2756:
2754:9780202365350
2750:
2746:
2739:
2731:
2727:
2723:
2719:
2715:
2711:
2704:
2702:
2693:
2689:
2685:
2681:
2677:
2673:
2669:
2665:
2661:
2654:
2652:
2650:
2648:
2640:
2634:
2627:
2621:
2613:
2609:
2605:
2601:
2597:
2593:
2589:
2582:
2574:
2572:1-58113-217-4
2568:
2564:
2560:
2556:
2552:
2545:
2544:
2539:
2532:
2524:
2518:
2514:
2510:
2506:
2499:
2491:
2487:
2483:
2479:
2474:
2469:
2465:
2461:
2454:
2446:
2442:
2438:
2432:
2428:
2427:
2419:
2410:
2404:
2400:
2396:
2395:
2390:
2384:
2377:
2371:
2367:
2361:
2353:
2349:
2345:
2341:
2337:
2333:
2329:
2323:
2316:
2312:
2308:
2304:
2299:
2293:
2288:
2280:
2276:
2272:
2268:
2264:
2260:
2256:
2249:
2247:
2239:
2233:
2229:
2222:
2215:
2211:
2207:
2203:
2198:
2193:
2190:(2): 85–126,
2189:
2185:
2178:
2170:
2169:
2164:
2158:
2149:
2135:on 2021-11-14
2131:
2127:
2123:
2119:
2115:
2111:
2107:
2103:
2099:
2092:
2085:
2077:
2071:
2067:
2062:
2061:
2055:
2049:
2043:
2038:
2032:
2028:
2025:
2019:
2012:
2007:
1998:
1991:
1987:
1981:
1977:
1973:
1972:
1967:
1963:
1957:
1950:
1946:
1942:
1938:
1934:
1933:Technometrics
1927:
1923:
1914:
1911:
1909:
1906:
1904:
1901:
1899:
1896:
1894:
1891:
1889:
1886:
1884:
1881:
1879:
1876:
1874:
1871:
1869:
1866:
1865:
1859:
1857:
1856:mixture model
1853:
1841:= 5 (yellow).
1840:
1836:
1832:
1828:
1824:
1819:
1815:
1813:
1806:
1802:
1798:
1794:
1790:
1786:
1779:
1772:
1768:
1761:
1757:
1753:
1749:
1739:
1737:
1733:
1723:
1721:
1720:data analysis
1716:
1714:
1710:
1705:
1703:
1699:
1695:
1690:
1686:
1677:
1668:
1666:
1656:
1642:
1616:
1612:
1588:
1585:
1582:
1574:
1570:
1540:
1537:
1534:
1526:
1522:
1518:
1515:
1507:
1501:
1491:
1481:
1478:
1475:
1471:
1459:
1450:
1445:
1442:
1439:
1430:
1427:
1424:
1413:
1409:
1405:
1398:
1397:
1396:
1382:
1379:
1376:
1353:
1345:
1339:
1308:
1302:
1294:
1288:
1282:
1279:
1276:
1273:
1270:
1264:
1259:
1255:
1251:
1248:
1245:
1243:
1232:
1224:
1218:
1212:
1209:
1206:
1203:
1200:
1194:
1191:
1185:
1177:
1171:
1166:
1162:
1158:
1156:
1145:
1137:
1131:
1122:
1119:
1116:
1113:
1110:
1104:
1101:
1098:
1090:
1086:
1082:
1080:
1069:
1066:
1063:
1054:
1051:
1040:
1039:
1038:
1005:
997:
991:
988:
985:
975:
972:
967:
965:
961:
939:
935:
925:
919:
916:
913:
910:
907:
904:
893:
890:
881:
877:
853:
849:
845:
841:
806:
801:
797:
793:
789:
784:
781:
778:
775:
768:
759:
755:
752:
749:
745:
737:
733:
729:
725:
716:
703:
702:
701:
696:
686:
684:
680:
676:
672:
668:
664:
663:manufacturing
660:
656:
652:
648:
642:
632:
618:
615:
612:
592:
589:
586:
578:
562:
527:
523:
519:
514:
510:
503:
500:
495:
491:
487:
479:
475:
471:
466:
462:
455:
452:
447:
443:
428:
427:
426:
424:
406:
402:
379:
375:
366:
355:
353:
336:
316:
305:
292:
289:
285:
282:
279:
276:
274:
272:
267:
265:
262:
260:
257:
256:
255:
251:
249:
245:
234:
232:
221:
219:
215:
211:
207:
203:
199:
195:
194:significantly
191:
186:
184:
179:
175:
171:
162:
153:
151:
147:
143:
141:
137:
132:
128:
124:
119:
117:
113:
108:
106:
102:
98:
94:
89:
87:
86:mixture model
83:
79:
75:
71:
67:
62:
60:
55:
51:
47:
39:
35:
30:
26:
22:
2925:
2909:
2878:
2874:
2868:
2859:
2834:
2830:
2817:
2805:
2796:
2777:
2773:
2763:
2744:
2738:
2713:
2709:
2667:
2663:
2633:
2620:
2595:
2591:
2581:
2550:
2542:
2531:
2504:
2498:
2466:(3–4): 237.
2463:
2459:
2453:
2425:
2418:
2393:
2383:
2369:
2360:
2335:
2331:
2322:
2310:
2298:
2287:
2262:
2258:
2227:
2221:
2187:
2183:
2177:
2167:
2163:Rousseeuw, P
2157:
2148:
2137:. Retrieved
2130:the original
2104:(6): e1280.
2101:
2097:
2084:
2059:
2048:
2037:
2018:
2006:
1997:
1989:
1970:
1956:
1948:
1936:
1932:
1926:
1849:
1838:
1837:= 4 (blue),
1834:
1833:=3 (green),
1830:
1826:
1822:
1811:
1804:
1800:
1796:
1792:
1784:
1777:
1770:
1766:
1759:
1755:
1751:
1745:
1729:
1717:
1706:
1691:
1687:
1683:
1674:
1662:
1561:
1327:
976:
970:
968:
963:
959:
879:
875:
830:
698:
674:
659:econometrics
644:
554:
362:
351:
308:
270:
252:
240:
227:
217:
213:
205:
197:
189:
187:
167:
144:
120:
109:
90:
63:
49:
43:
25:
2938:Grubbs test
2837:: 155–167.
2598:: 190–237.
2338:: 348–351.
2011:Grubbs 1969
1939:(1): 1–21.
1913:Winsorizing
1698:Winsorising
671:data mining
231:King effect
93:sample mean
2948:Categories
2522:1581132174
2376:PDF Eprint
2372:: 200–224.
2139:2019-12-11
1966:"Outliers"
1919:References
1829:=2 (red),
1821:Figure 5.
1709:regression
1694:truncation
1676:outliers.
693:See also:
667:networking
647:statistics
577:John Tukey
172:data, the
146:Estimators
136:population
54:data point
46:statistics
32:Figure 1.
2932:EMS Press
2922:"Outlier"
2911:MathWorld
2906:"Outlier"
2692:258376426
2468:CiteSeerX
2192:CiteSeerX
2118:1942-4787
1732:fat tails
1680:Exclusion
1671:Retention
1643:α
1589:α
1541:α
1472:∑
1446:−
1434:⟩
1422:⟨
1380:⊂
1256:∑
1252:−
1192:−
1163:∑
1102:−
1087:∑
1073:⟩
1061:⟨
989:−
908:−
891:δ
846:α
794:α
779:−
753:−
730:α
520:−
472:−
453:−
423:quartiles
248:Box plots
212:with λ =
2851:16500768
2730:24773354
2684:37104797
2612:19036098
2490:11707259
2352:25138498
2126:53305944
2054:Ruan, Da
2027:Archived
1964:(1992).
1862:See also
1702:censored
288:leverage
269:Dixon's
242:such as
190:a priori
78:skewness
59:data set
34:Box plot
2445:3058187
2399:140–160
2374:. NOAA
2279:6724536
2214:3330313
1854:, or a
655:finance
123:average
50:outlier
2849:
2751:
2728:
2690:
2682:
2610:
2569:
2555:SIGMOD
2519:
2488:
2470:
2443:
2433:
2405:
2350:
2277:
2234:
2212:
2194:
2124:
2116:
2072:
1982:
1704:data.
1562:where
1021:where
831:where
224:Causes
140:sample
131:median
101:theory
74:robust
2847:S2CID
2827:(PDF)
2688:S2CID
2608:S2CID
2547:(PDF)
2486:S2CID
2348:JSTOR
2275:S2CID
2210:S2CID
2133:(PDF)
2122:S2CID
2094:(PDF)
958:. If
142:set.
52:is a
48:, an
2749:ISBN
2726:PMID
2680:PMID
2567:ISBN
2517:ISBN
2441:OCLC
2431:ISBN
2403:ISBN
2232:ISBN
2114:ISSN
2070:ISBN
1980:ISBN
669:and
394:and
286:and
278:ASTM
273:test
2883:doi
2879:141
2839:doi
2782:doi
2718:doi
2672:doi
2668:152
2600:doi
2559:doi
2509:doi
2478:doi
2340:doi
2267:doi
2202:doi
2106:doi
2066:318
1941:doi
1707:In
593:1.5
233:).
114:or
44:In
2950::
2930:,
2924:,
2908:.
2877:.
2845:.
2833:.
2829:.
2778:31
2776:.
2772:.
2724:.
2714:19
2712:.
2700:^
2686:.
2678:.
2666:.
2662:.
2646:^
2606:.
2596:28
2594:.
2565:.
2553:.
2549:.
2515:.
2484:.
2476:.
2462:.
2439:.
2401:.
2346:.
2336:13
2334:.
2309:,
2305:,
2273:.
2261:.
2245:^
2208:,
2200:,
2188:22
2186:,
2120:.
2112:.
2100:.
2096:.
2068:.
1988:.
1978:.
1976:89
1968:.
1947:.
1937:11
1935:.
1858:.
1746:A
1715:.
1667:.
1395::
1037::
665:,
661:,
657:,
653:,
649:,
575:.
214:pn
88:.
2914:.
2889:.
2885::
2853:.
2841::
2835:4
2790:.
2784::
2757:.
2732:.
2720::
2694:.
2674::
2614:.
2602::
2575:.
2561::
2525:.
2511::
2492:.
2480::
2464:8
2447:.
2411:.
2354:.
2342::
2317:.
2281:.
2269::
2263:5
2204::
2142:.
2108::
2102:8
2078:.
1943::
1839:q
1835:q
1831:q
1827:q
1823:q
1812:q
1808:i
1805:X
1801:q
1797:x
1793:q
1787:-
1785:q
1781:i
1778:X
1774:i
1771:X
1767:x
1763:i
1760:X
1756:x
1752:i
1633:t
1617:j
1613:g
1592:)
1586:,
1583:t
1580:(
1575:j
1571:g
1547:)
1544:)
1538:,
1535:t
1532:(
1527:j
1523:g
1519:,
1516:x
1512:|
1508:y
1505:(
1502:p
1496:|
1492:L
1488:|
1482:1
1479:=
1476:j
1464:|
1460:L
1456:|
1451:1
1443:1
1440:=
1437:)
1431:y
1428:,
1425:x
1419:(
1414:L
1410:H
1406:I
1383:H
1377:L
1357:)
1354:t
1350:|
1346:h
1343:(
1340:p
1330:H
1309:.
1306:)
1303:t
1299:|
1295:h
1292:(
1289:p
1286:)
1283:h
1280:,
1277:x
1274:,
1271:y
1268:(
1265:p
1260:H
1249:1
1246:=
1236:)
1233:t
1229:|
1225:h
1222:(
1219:p
1216:)
1213:h
1210:,
1207:x
1204:,
1201:y
1198:(
1195:p
1189:)
1186:t
1182:|
1178:h
1175:(
1172:p
1167:H
1159:=
1149:)
1146:t
1142:|
1138:h
1135:(
1132:p
1129:)
1126:)
1123:h
1120:,
1117:x
1114:,
1111:y
1108:(
1105:p
1099:1
1096:(
1091:H
1083:=
1076:)
1070:y
1067:,
1064:x
1058:(
1055:H
1052:I
1035:H
1031:t
1027:x
1023:y
1009:)
1006:x
1002:|
998:y
995:(
992:p
986:1
971:δ
964:δ
960:δ
944:|
940:s
936:/
932:)
929:)
926:X
923:(
920:n
917:a
914:e
911:m
905:X
902:(
898:|
894:=
880:n
876:n
872:t
854:2
850:/
842:t
827:;
807:2
802:2
798:/
790:t
785:+
782:2
776:n
769:n
760:)
756:1
750:n
746:(
738:2
734:/
726:t
717:=
619:3
616:=
613:k
590:=
587:k
563:k
538:]
533:)
528:1
524:Q
515:3
511:Q
507:(
504:k
501:+
496:3
492:Q
488:,
485:)
480:1
476:Q
467:3
463:Q
459:(
456:k
448:1
444:Q
438:[
407:3
403:Q
380:1
376:Q
337:n
317:m
271:Q
218:p
206:p
198:p
23:.
Text is available under the Creative Commons Attribution-ShareAlike License. Additional terms may apply.