Detecting multiple outliers in linear regression using a cluster method combined with graphical visualization (Q2271696)

From MaRDI portal
Revision as of 11:35, 2 February 2024 by Import240129110113 (talk | contribs) (Added link to MaRDI item.)
scientific article
Language Label Description Also known as
English
Detecting multiple outliers in linear regression using a cluster method combined with graphical visualization
scientific article

    Statements

    Detecting multiple outliers in linear regression using a cluster method combined with graphical visualization (English)
    0 references
    0 references
    8 August 2009
    0 references
    A new algorithm is proposed for multiple outliers detection in regression models. The idea is to apply a hierarchical clustering algorithm to the scatterplot of standardized predicted values and residuals (obtained by ordinary least squares) in order to find a single largest cluster which is considered as a set of inliers. Then the \(t\)-statistic is used to test all other observations as candidate outliers individually against the group of inliers. Results of real data sets analyses are presented.
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    hierarchical clustering
    0 references
    least squares
    0 references
    swamping
    0 references
    minimal spanning trees
    0 references