Displaying similar documents to “Algorithm 44. An abbreviated method of calculating the Mahalanobis distance or residual sum of squares in a linear regression model”

On the Equivalence between Orthogonal Regression and Linear Model with Type-II Constraints

Sandra Donevska, Eva Fišerová, Karel Hron (2011)

Acta Universitatis Palackianae Olomucensis. Facultas Rerum Naturalium. Mathematica

Similarity:

Orthogonal regression, also known as the total least squares method, regression with errors-in variables or as a calibration problem, analyzes linear relationship between variables. Comparing to the standard regression, both dependent and explanatory variables account for measurement errors. Through this paper we shortly discuss the orthogonal least squares, the least squares and the maximum likelihood methods for estimation of the orthogonal regression line. We also show that all mentioned...

An adaptive method of estimation and outlier detection in regression applicable for small to moderate sample sizes

Brenton R. Clarke (2000)

Discussiones Mathematicae Probability and Statistics

Similarity:

In small to moderate sample sizes it is important to make use of all the data when there are no outliers, for reasons of efficiency. It is equally important to guard against the possibility that there may be single or multiple outliers which can have disastrous effects on normal theory least squares estimation and inference. The purpose of this paper is to describe and illustrate the use of an adaptive regression estimation algorithm which can be used to highlight outliers, either single...

Stacked regression with restrictions

Tomasz Górecki (2005)

Discussiones Mathematicae Probability and Statistics

Similarity:

When we apply stacked regression to classification we need only discriminant indices which can be negative. In many situations, we want these indices to be positive, e.g., if we want to use them to count posterior probabilities, when we want to use stacked regression to combining classification. In such situation, we have to use leastsquares regression under the constraint βₖ ≥ 0, k = 1,2,...,K. In their earlier work [5], LeBlanc and Tibshirani used an algorithm given in [4]. However,...

A note on the strong consistency of least squares estimates

Joǎo Lita da Silva (2009)

Discussiones Mathematicae Probability and Statistics

Similarity:

The strong consistency of least squares estimates in multiples regression models with i.i.d. errors is obtained under assumptions on the design matrix and moment restrictions on the errors.

Fitting a linear regression model by combining least squares and least absolute value estimation.

Sira Allende, Carlos Bouza, Isidro Romero (1995)

Qüestiió

Similarity:

Robust estimation of the multiple regression is modeled by using a convex combination of Least Squares and Least Absolute Value criterions. A Bicriterion Parametric algorithm is developed for computing the corresponding estimates. The proposed procedure should be specially useful when outliers are expected. Its behavior is analyzed using some examples.

Kriging and masurement errors

István Fazekas, Alexander G. Kukush (2005)

Discussiones Mathematicae Probability and Statistics

Similarity:

A linear geostatistical model is considered. Properties of a universal kriging are studied when the locations of observations aremeasured with errors. Alternative prediction procedures are introduced and their least squares errors are analyzed.