Displaying similar documents to “Fitting a linear regression model by combining least squares and least absolute value estimation.”

An adaptive method of estimation and outlier detection in regression applicable for small to moderate sample sizes

Brenton R. Clarke (2000)

Discussiones Mathematicae Probability and Statistics

Similarity:

In small to moderate sample sizes it is important to make use of all the data when there are no outliers, for reasons of efficiency. It is equally important to guard against the possibility that there may be single or multiple outliers which can have disastrous effects on normal theory least squares estimation and inference. The purpose of this paper is to describe and illustrate the use of an adaptive regression estimation algorithm which can be used to highlight outliers, either single...

On the Equivalence between Orthogonal Regression and Linear Model with Type-II Constraints

Sandra Donevska, Eva Fišerová, Karel Hron (2011)

Acta Universitatis Palackianae Olomucensis. Facultas Rerum Naturalium. Mathematica

Similarity:

Orthogonal regression, also known as the total least squares method, regression with errors-in variables or as a calibration problem, analyzes linear relationship between variables. Comparing to the standard regression, both dependent and explanatory variables account for measurement errors. Through this paper we shortly discuss the orthogonal least squares, the least squares and the maximum likelihood methods for estimation of the orthogonal regression line. We also show that all mentioned...

Trimmed Estimators in Regression Framework

TomĂĄĹĄ Jurczyk (2011)

Acta Universitatis Palackianae Olomucensis. Facultas Rerum Naturalium. Mathematica

Similarity:

From the practical point of view the regression analysis and its Least Squares method is clearly one of the most used techniques of statistics. Unfortunately, if there is some problem present in the data (for example contamination), classical methods are not longer suitable. A lot of methods have been proposed to overcome these problematic situations. In this contribution we focus on special kind of methods based on trimming. There exist several approaches which use trimming off part...

Stacked regression with restrictions

Tomasz Górecki (2005)

Discussiones Mathematicae Probability and Statistics

Similarity:

When we apply stacked regression to classification we need only discriminant indices which can be negative. In many situations, we want these indices to be positive, e.g., if we want to use them to count posterior probabilities, when we want to use stacked regression to combining classification. In such situation, we have to use leastsquares regression under the constraint βₖ ≥ 0, k = 1,2,...,K. In their earlier work [5], LeBlanc and Tibshirani used an algorithm given in [4]. However,...