More about "r" and the slope edit

In this article, r^2 is called the "coefficient of determination." What is "r" called?

Also, I believe a one-dimensional regression line from an OLS fit has slope = r(sy/sx), where r is from above, and sy and sx are the standard deviations of the y and x data, respectively.

Is that true? Got a ref? If so, we should work it into the article.

dfrankow (talk) 18:23, 23 April 2020 (UTC)Reply

Regression articles discussion July 2009 edit

A discussion of content overlap of some regression-related articles has been started at Talk:Linear least squares#Merger proposal but it isn't really just a question of merging and no actual merge proposal has been made. Melcombe (talk) 11:37, 14 July 2009 (UTC)Reply


OLS edit

Don't merge them, this article is OLS. To put it simple, this article is only for OLS, so don't write much about GLS etc in this article. But we can go deeper for OLS in this article. I want to delete much material is not OLS, those material can be put in the linear regression article. For example, the section 1 should be simplified. Jackzhp (talk) 16:25, 25 March 2010 (UTC)Reply


The variance of the estimator edit

We know  , but what is  ? I don't know how to find it even for the simple case when  , but I know that the FDCR lower bound of it in this case is  . Jackzhp (talk) 16:53, 25 March 2010 (UTC)Reply

According to my calculations, the variance is equal to
 
where mii is the i-th diagonal element of the annihilator matrix M, and γ2 is the kurtosis of the distribution of the error terms. When the kurtosis is positive we can obtain an upper bound from the fact that  , and the sum of mii’s is the trace of M which is n − p:
 
 // stpasha »  00:36, 25 April 2010 (UTC)Reply

Comment from main article on assumptions edit

I'm moving this from the main article, it was included as a comment in the section on assumptions:

Can we just replace this section with the following one line?
 
Please discuss with me in the discussion page. Let's keep material only related to OLS, anything else should be deleted. If you want, please move it to linear regression article.
correlation between data points can be discussed, but not in this section. Given  , we can clear see this.
Identifiability can be discussed, but should be in the estimation section

I disagree with this proposal. Some written explanation is much more useful than a single mathematical expression. The current text is not unduly digressive. Skbkekas (talk) 22:29, 25 March 2010 (UTC)Reply

Why does "linear least-squares" redirect here? edit

This page does not make any sense to someone who is just interested in the general problem  , since this page seems extremely application specific. Did some error occur when this strange redirection happened? —Preceding unsigned comment added by Short rai (talkcontribs) 09:54, 23 May 2010 (UTC)Reply

See Ordinary least squares#Geometric approach.  // stpasha »  19:07, 23 May 2010 (UTC)Reply
But this is a major subject in virtually every subfield of applied mathematics, and you refer everyone who is not in statistics to one tiny paragraph called "geometric approach"? I think there recently was a redirect and/or merge process going on with other articles in about the topic, and there seems to have been a page called "linear least squares" earlier, but that page is impossible to find now.Short rai (talk) 23:14, 23 May 2010 (UTC)Reply
Ya, there used to be a backlink in the see also section, I have restored it now. And by the way, the OLS regression and the problem of minimization of the norm of Ax−b are exactly the same problems, only written in different notations. The difference is that statisticians use X and y for known quantities, and β for unknown. Mathematicians use those symbols in the opposite way.  // stpasha »  03:53, 24 May 2010 (UTC)Reply

Hypothesis Testing edit

Why is this section empty? Stuffisthings (talk) 13:32, 15 July 2010 (UTC)Reply

Linear Least Squares edit

FYI, there is a discussion on the usage of Linear least squares which currently redirects here, but was previously another topic, at Talk:Numerical methods for linear least squares.

76.66.198.128 (talk) 03:56, 21 October 2010 (UTC)Reply

Vertical or Euclidian Distances? edit

The article states that OLS minimizes the sum of squared distances from the points to the estimated regression line. But we are taught in standard (Euclidian) geometry that the distance between a point and a line is defined as the length of the perpindicular line segment connecting the two. This is not what OLS minimizes. Rather, it minimizes the vertical distance between the points and the line. Shouldn't the article say as much? —Preceding unsigned comment added by 76.76.220.34 (talk) 17:00, 2 November 2010 (UTC)Reply

I've changed that sentence, so that it says "vertical distances" now.  // stpasha »  20:53, 2 November 2010 (UTC)Reply
I have a problem where I do want to minimize the sum of squared Euclidean distances from a point to a set of given straight lines in the plane. Can anyone give a reference or some keywords? (The problem occurs in surveying, when many observers at known locations can see the same point at an unknown location. Each observer can measure its bearing to the target point. This gives a set of lines that ideally should intersect at the target point. But measurement errors gives an overdetermined problem if there are more than two observers.) Mikrit (talk) 15:41, 22 November 2010 (UTC)Reply
You want to look at Deming regression or Total least squares.  // stpasha »  16:47, 22 November 2010 (UTC)Reply

Example with real data edit

Are the calculations of the Akaike criterion and Schwarz criterion correct here? I know that there are many "different" forms of the AIC and SIC - but I just can't figure out how these were calculated. Certainly they seem inconsistent with the forms that are linked-to in the description that follows the calculated values.—Preceding unsigned comment added by 71.224.184.11 (talk) 05:32, 30 November 2010 (UTC)Reply

I'll try again edit

Way back in 2008 I came across this example calculation, and looked at the plot of (x,y) and noticed an odd cadence in the positioning. Some simple inspection soon showed that the original height data were in terms of inches, and whoever converted them to metric bungled the job. The conversion factor is 2.54cm to an inch and rounding to the nearest centimetre is not a good idea. This makes a visible difference and considerably changes the results of an attempt at a quadratic fit. It doesn't matter what statistical packages might be used, to whatever height of sophistication, if the input data are wrongly prepared. The saving grace is the presence of the plot of the data, but, that plot has to be looked at not just gazed at with vague approbation. In the various reorganisations, my report of this problem has been lost, and the editors ignored its content. The error remains, so I'll try again.

Height^2       Height         Const.
61.96033      -143.162      128.8128  Improper rounding of inches to whole cm.
58.5046       -131.5076     119.0205  Proper conversion, no rounding.

The original incorrectly-converted plot can be reproduced, but here is a plot of the correctly-converted heights, with a quadratic fit. Notice the now-regular spacing of the x-values, without cadence.

 
Correctly converted heights, quadratic fit.

For the incorrectly-concerted heights, the residuals are

 
Wrongly converted heights, Residuals to a quadratic fit.

Whereas for the correctly-converted height data, the residuals are much smaller. (Note the vertical scale)

 
Correctly converted heights, Residuals to a quadratic fit.

And indeed, this pattern of residuals rather suggests a higher-order fit attempt, such as with a cubic.

 
Correctly converted heights, Residuals to a cubic fit.

But assessing the merit of this escalation would be helped by the use of some more sophisticated analysis, such as might be offered by the various fancy packages, if fed correct data.

Later, it occurred to me to consider whether the weights might have been given in pounds. The results were odd in another way. Using the conversion 1KG = 2.20462234 lbs used in the USA, the weights are

115.1033 117.1095 120.1078 123.1061 126.1044 129.1247 132.123 135.1213 139.1337 142.132 146.1224 150.1348 154.1472 159.1517 164.1562
114.862  116.864  119.856  122.848  125.84   128.854  131.846 134.838  138.842  141.834 145.816  149.82   153.824  158.818  163.812

The second row being for the approximate conversion of 1KG = 2.2lbs. I am puzzled by the fractional parts. NickyMcLean (talk) 22:11, 5 September 2011 (UTC)Reply

2022 peer review about the "Sensitivity to rounding" section edit

This section is entirely unnecessary. The uncertainties quoted in the "Example with real data" section entirely encompass any variations due to the rounding issue. With rounding:  ,  ,  , which is in perfect congruence with the non-rounded central values  ,  ,  . This should also assuage any worries about the model breaking when extrapolating: the uncertainties account for it. Statistics are useless without uncertainty estimates, so I encourage the author of this sensitivity section to use them more widely in their work. I thus propose that this section be removed. — Preceding unsigned comment added by 2A02:908:C33:D160:0:0:0:9E4C (talk) 00:36, 3 May 2022 (UTC)Reply

Proposal to not call it weight edit

The term "weight" is not only wrong in this context, but also possibly confusing, since weight has a specific meaning, see also Weighted least squares. I propose to either call it "mass", which is less confusing and more correct. However, since a lot of people incorrectly refer to their mass as "weight", talking about mass might be confusing. If it is considered too confusing, I would propose a completely new example with real world data, again without the term "weight". 2001:67C:10EC:5792:8000:0:0:1179 (talk) 13:21, 18 February 2022 (UTC)Reply

Possible error in formula for standard error for coefficients edit

It seems to me that the 1/n should not be included in the formula for the standard errors for each coefficient. With 1/n, the values calculated in this example are not produced. Removing it generates the values given in the example. Would someone more knowledgeable in this subject examine this and correct the formula, if necessary? — Preceding unsigned comment added by 173.178.40.20 (talk) 17:37, 22 May 2012 (UTC)Reply

Yes, this looks wrong to me as well. — Preceding unsigned comment added by 2620:0:1009:1:BAAC:6FFF:FE7D:1EE9 (talk) 17:24, 30 November 2012 (UTC)Reply

I arrived to the same conclusion independently before checking this talk page, thus I removed the 1/n. 91.157.6.139 (talk) 12:06, 1 January 2015 (UTC)Reply

I found this error as well, and, despite the previous comment the 1/n was still there. I removed it now. — Preceding unsigned comment added by 212.213.198.88 (talk) 10:50, 20 February 2015 (UTC)Reply

On multicollinearity edit

Multicollinearity means high level of correlation between variables. OLS can handle this fine, it just needs more data to do it. However, in the article, "multicollinearity" is being used to mean perfect collinearity, ie. the data matrix does not have full column rank. This is confusing. I propose we stop using multicollinearity to mean lack of full rank, and just say "not full rank". —Preceding unsigned comment added by 24.30.13.209 (talk) 04:21, 10 March 2011 (UTC)Reply

Done. Let me know if I missed any instances of it. Duoduoduo (talk) 15:11, 10 March 2011 (UTC)Reply

Estimation edit

In the section "Simple regression model"; It is not true that  , this only holds for the true parameters and not the estimator. Rather,   is equal to the sample covariance over sample variance. Maybe use a hat over Cov and Var to signify this if you want the relation to be stressed.

Also, in my opinion a lot of time is being devoted to the use of the annihilator-matrix. It does't seem necessary to introduce the extra notation unless one wants to go into the Frisch-Waugh-Lovell theorem and this has its own separate page. — Preceding unsigned comment added by Superpronker (talkcontribs) 06:46, 1 June 2011 (UTC)Reply

I believe that a clarification on notation would help immensely. The regressor values for 1 observation is referred to as the *column* vector  . However, in the design matrix of regressor values, the values for an observation occupy a *row*. Hence, it is easy to fall into the trap of thinking of   as a row vector, leading to confusion.Craniator (talk) 05:34, 3 May 2015 (UTC)Reply

Alternative Derivations: Geometric Approach edit

In the illustration of orthogonal projection, it would be helpful to clarify that   refers to a column in the data matrix, thus clearly distinguishing it from   for the set of regressor values from one observation.Craniator (talk) 06:01, 3 May 2015 (UTC)Reply

Too technical? Should it be rewritten "one level down"? edit

I realize this article has been rated B-class, but I wonder if it is too technical for someone who does not already understand OLS. OLS is often studied in undergrad stats classes. Therefore, in the spirit of writing "one level down" (see: WP:UPFRONT), this article should ideally include an extended intro that is much more comprehensible to someone with some relatively advanced high school math training (say, through a year of high school calculus, but without linear algebra).

As currently written, almost all of the article is incomprehensible to someone who doesn't understand linear algebra. There are many ways of introducing OLS without linear algebra, so could such non-technical, intuitive approaches be put at the top of this article, leaving the more technical, formal math stuff for the bottom? I hesitate to be so bold in editing, since this is a very important article, but I think it is currently way too technical.Aroundthewayboy (talk) 04:57, 26 July 2015 (UTC)Reply

Yes, this article is basically unreadable. --The Cunctator (talk) 15:01, 12 May 2016 (UTC)Reply
I agree as well. There is also the very similar discussion in Linear_regression equally swamped by inscrutable linear algebra blather that may well be succinct and general with fancy symbols pleasing to the cognoscenti, but obstructive to those not familiar with its usage. There is yet more of much the same in Simple_linear_regression where in the "talk" I went through the derivation using calculus, that I think is rather more understandable and not just because I wrote it out. Some consolidation would be in order. NickyMcLean (talk) 12:04, 15 May 2016 (UTC)Reply
I agree too. Have some basic understanding of linear algebra and get the concept of OLS, but still find the article incomprehensible. It would probably be a good idea to have an example, like estimating a 'voter transition analysis', and not assuming the preknowledge of specialized mathematical jargon that go beyond basic analysis classes. MovGP0 (talk) 00:26, 4 December 2016 (UTC)Reply

Bias due to Ordinary Least Squares edit

There has been criticism of ordinary least squares being a biased estimator. The idea that least squares is the best comes from probability theory and the consideration of repeated measurements. Gauss showed that the sum of the square of the errors resulted in the least probable error for an estimate. The problem with trying to fit a straight line to data that is slightly unbalanced results in a biased fit since the error can be "tilted" slightly to reduce the sum of its squares. Under such circumstances the assumption that the sum of the squares is a minimum is false but unavoidable. The bias in the data is difficult to detect. --Jbergquist (talk) 23:21, 6 February 2016 (UTC)Reply

 
Ordinary Least Squares bias example

It might be helpful to include an example of the bias in an OLS fit. Here is a simple one. Notice that the fit zeros the first moment of the errors. --Jbergquist (talk) 19:31, 7 February 2016 (UTC)Reply

Dr. Kaplan's comment on this article edit

Dr. Kaplan has reviewed this Wikipedia page, and provided us with the following comments to improve its quality:


Sources:

1. Bruce Hansen's online econometrics text/lecture notes, 2016 edition: http://www.ssc.wisc.edu/~bhansen/econometrics/ 2. Wooldridge, Jeffrey M. (2010). Econometric Analysis of Cross Section and Panel Data, 2nd ed. MIT press.

Original first paragraph: <<In statistics, ordinary least squares (OLS) or linear least squares is a method for estimating the unknown parameters in a linear regression model, with the goal of minimizing the differences between the observed responses in some arbitrary dataset and the responses predicted by the linear approximation of the data (visually this is seen as the sum of the vertical distances between each data point in the set and the corresponding point on the regression line - the smaller the differences, the better the model fits the data). The resulting estimator can be expressed by a simple formula, especially in the case of a single regressor on the right-hand side.>> Comments: 1a. It's vague (at best) or misleading to say OLS estimates the unknown parameters in a linear regression model. OLS estimates the linear projection coefficient; in some cases, this corresponds to the parameters of the conditional expectation function (CEF), which is usually (but not always...) what "regression" refers to. See for example Theorem 6.2.1 in [1]. The linear projection coefficient also forms the "best linear predictor" as in Section 2.18 of [1]. 1b. The phrase "sum of the vertical distances" is wrong. The "squares" in "ordinary least squares" refers to the fact that the vertical distances are each squared, and then summed. Minimizing the sum of (absolute) vertical distances leads to the least absolute deviations (LAD) regression estimators, a.k.a. median regression.

Original second paragraph: <<The OLS estimator is consistent when the regressors are exogenous and there is no perfect multicollinearity, and optimal in the class of linear unbiased estimators when the errors are homoscedastic and serially uncorrelated. Under these conditions, the method of OLS provides minimum-variance mean-unbiased estimation when the errors have finite variances. Under the additional assumption that the errors be normally distributed, OLS is the maximum likelihood estimator. OLS is used in economics (econometrics), political science and electrical engineering (control theory and signal processing), among many areas of application. The Multi-fractional order estimator is an expanded version of OLS.>> Comments: 2a. The term "consistent" has no absolute meaning; one must say "consistent for XXX." The first sentence's use is like having a transitive verb with no direct object. I'd suggest XXX=linear projection coefficient, as in e.g. Theorem 6.2.1 in [1]. 2b. Trying to explain consistency in the intro is a tall task. The first sentence mentions "exogenous" (which is either ambiguous, or else stronger than necessary...) and multicollinearity, but it mentions nothing of sampling assumptions (iid, etc.), so it's insufficient to establish consistency. One could try to explain the result from Theorem 6.2.1 in [1], but this seems too detailed for an introduction. 2c. The term "optimal" is ambiguous. Optimal with respect to what criterion? (The next sentence answers: smallest variance, although this itself takes some defining when the variance is a matrix rather than a scalar.) 2d. The second half of the first sentence seems to be referring to the Gauss-Markov Theorem(?), but it fails to mention the critical assumption of a linear conditional expectation function. 2e. The second-to-last sentence seems to lack (grammatically) an "and" before "political science". 2f. I don't think the "Multi-fractional order estimator" (whatever that is...) should be mentioned in the introduction to an article on OLS. Perhaps something like GLS, or FGLS, or WLS, or NLLS, if you really want to mention other methods.

Original paragraph: <<Suppose the data consists of n observations { y i, x i }n i=1. Each observation includes a scalar response yi and a vector of p predictors (or regressors) xi. In a linear regression model the response variable is a linear function of the regressors: y_i = x_i ^T \beta + \varepsilon_i, \, where β is a p×1 vector of unknown parameters; εi's are unobserved scalar random variables (errors) which account for the discrepancy between the actually observed responses yi and the "predicted outcomes" xiTβ; and T denotes matrix transpose, so that xTβ is the dot product between the vectors x and β. This model can also be written in matrix notation as y = X\beta + \varepsilon, \, where y and ε are n×1 vectors, and X is an n×p matrix of regressors, which is also sometimes called the design matrix.>> Comments: 3a. It's unclear what "'predicted outcomes' $x_i'\beta$" means, esp. since it's in scare quotes(!). 3b. It's odd/unhelpful to write down a model without explaining what the $\epsilon_i$ are and/or what $\beta$ is (which is down much farther down). In contrast, see for example Section 6.1 of [1], where $\beta$ is immediately defined as the linear projection coefficient, as defined earlier (Definition 2.18.1). I understand there is some incentive to leave it ambiguous at first because there are different assumptions one could make, but logically I think it's best to start with a linear projection model because that's fundamentally what OLS estimates. Then, one can mentioned that a linear CEF/regression model implies that the (population) linear projection equals the CEF, but that's an argument about identification; one can always go further and mention that under yet more assumptions (conditional independence) the CEF derivative equals the average causal effect, etc., as in Theorem 2.30.1 in [1].

Original paragraph: <<There may be some relationship between the regressors. For instance, the third regressor may be the square of the second regressor. In this case (assuming that the first regressor is constant) we have a quadratic model in the second regressor. But this is still considered a linear model because it is linear in the βs.>> Comment: 4a. The phrase "linear model" is ambiguous. (Of course, people use it all the time.) See for example page 15 of [2]. Models can be linear in the explanatory variables, linear in parameters, both, or neither. In the case of including a squared regressor, it's still linear in parameters, but no longer linear in variables.

Original paragraph: <<There are several different frameworks in which the linear regression model can be cast in order to make the OLS technique applicable. Each of these settings produces the same formulas and same results. The only difference is the interpretation and the assumptions which have to be imposed in order for the method to give meaningful results. The choice of the applicable framework depends mostly on the nature of data in hand, and on the inference task which has to be performed.>> Comment: 5a. This isn't a bad (or at least, not wrong) paragraph, and it makes essentially my point from Comment 3b. I'd again suggest an overall structure that shows 1) under the weakest assumptions, OLS is consistent for the linear projection coefficient, 2) under additional assumptions XXX the linear projection coefficient equals [some other parameter of interest].

Original paragraph: <<One of the lines of difference in interpretation is whether to treat the regressors as random variables, or as predefined constants. In the first case (random design) the regressors xi are random and sampled together with the yi's from some population, as in an observational study. This approach allows for more natural study of the asymptotic properties of the estimators. In the other interpretation (fixed design), the regressors X are treated as known constants set by a design, and y is sampled conditionally on the values of X as in an experiment. For practical purposes, this distinction is often unimportant, since estimation and inference is carried out while conditioning on X. All results stated in this article are within the random design framework.>> 6a. The phrase "estimation and inference is carried out while conditioning on X" may be essentially true if assuming a linear CEF model, but again, that's much stronger than required. Within the more general linear projection model, it's not true.

Original paragraph: <<The exogeneity assumption is critical for the OLS theory. If it holds then the regressor variables are called exogenous. If it doesn't, then those regressors that are correlated with the error term are called endogenous,[2] and then the OLS estimates become invalid. In such case the method of instrumental variables may be used to carry out inference.>> 7a. The "exogeneity assumption" is critical for *finite-sample* OLS theory, not asymptotic OLS theory. 7b. It is still not clearly stated what $\epsilon$ is. Here, it is (implicitly) treated as a structural error term. If it were defined as a CEF error term, then the "exogeneity assumption" would hold by definition. 7c. It's a bit of an oversell to say if X is endogenous then one can use IV methods; it's *possible* in very rare cases....

I've already spent way more than 20min on this, so I must stop.


We hope Wikipedians on this talk page can take advantage of these comments and improve the quality of the article accordingly.

Dr. Kaplan has published scholarly research which seems to be relevant to this Wikipedia article:


  • Reference : David M. Kaplan & Matt Goldman, 2015. "Fractional order statistic approximation for nonparametric conditional quantile inference," Working Papers 1502, Department of Economics, University of Missouri.

ExpertIdeasBot (talk) 15:41, 19 May 2016 (UTC)Reply

Dr. Fachin's comment on this article edit

Dr. Fachin has reviewed this Wikipedia page, and provided us with the following comments to improve its quality:


Edited a couple of points myself


We hope Wikipedians on this talk page can take advantage of these comments and improve the quality of the article accordingly.

We believe Dr. Fachin has expertise on the topic of this article, since he has published relevant scholarly research:


  • Reference : Di Iorio, Francesca & Fachin, Stefano, 2012. "A note on the estimation of long-run relationships in panel equations with cross-section linkages," Economics Discussion Papers 2012-1, Kiel Institute for the World Economy.

ExpertIdeasBot (talk) 19:14, 26 July 2016 (UTC)Reply

Coefficient of determination edit

I looked for clarifications in a google-preview of Hayashi's book but could not find the derivation of the equation:

 

Is this correct? Would it be possible to expand it a bit for clarity? 129.240.43.144 (talk) 12:32, 13 November 2019 (UTC)Reply

Using the same template and standard for math symbols/equations edit

Most equations in math blocks use <math> tag or use the "Math formula" from the GUI editor. But inline equations/symbols use different templates. Some use {{mvar}}, and some use {{math}}. This causes the symbols look different in paragraphs and in math blocks. Should the templates used for math be unified? Also, the symbol standard is not unified. For example, the transpose of a matrix, some use \mathsf{T}, and some simply use T. And sometimes a vector is represented with a bold italic symbol, sometimes it is represented with a bold upright symbol. Maybe it's better to choose and follow only one standard, like the ISO standard or the IEEE standard.

IvesC (talk) 05:59, 7 April 2021 (UTC)Reply

Hi, I have been going around to some articles and have been replacing all cases of ^\mathsf{T} and ^T with ^\operatorname{T} as this is what is used in the transpose article. Any discussion, feedback, or thoughts about this would be greatly appreciated.

Magenta.lily (talk) 03:02, 2 October 2023 (UTC)Reply

What is the difference? edit

Why is this unexpected or needs to be conditioned?

'The OLS estimator is consistent for the level-one fixed effects when the regressors are exogenous and forms perfect colinearity (rank condition), consistent for the variance estimate of the residuals when regressors have finite fourth moments and—by the Gauss–Markov theorem—optimal in the class of linear unbiased estimators when the errors are homoscedastic and serially uncorrelated. Under these conditions, the method of OLS provides minimum-variance mean-unbiased estimation when the errors have finite variances.'

OLS minimizes the L2 error by construction. And (variance) = f(bias) + L2error. So with respect to all unbiased linear estimators, should'nt OLS have minimum-variance estimation? Abacons (talk) 07:54, 6 April 2023 (UTC)Reply