yego.me
💡 Stop wasting time. Read Youtube instead of watch. Download Chrome Extension

Standard deviation of residuals or root mean square deviation (RMSD) | AP Statistics | Khan Academy


4m read
·Nov 11, 2024

So we are interested in studying the relationship between the amount that folks study for a test and their score on a test, where the score is between zero and six.

What we're going to do is go look at the people who took the tests. We're going to plot for each person the amount that they studied and their score. For example, this data point is someone who studied an hour, and they got a one on the test. Then we're going to fit a regression line, and this blue regression line is the actual regression line for these four data points. Here is the equation for that regression line.

Now, there are a couple of things to keep in mind. Normally, when you're doing this type of analysis, you do it with far more than four data points. The reason why I kept this to four is because we're actually going to calculate how good a fit this regression line is by hand, and typically you would not do it by hand; we have computers for that.

The way that we're going to measure how good a fit this regression line is to the data has several names. One name is the standard deviation of the residuals; another name is the root mean square deviation, sometimes abbreviated as RMSD. Sometimes it's called root mean square error.

So what we're going to do is, for every point, we're going to calculate the residual. Then we're going to square it and add up the sum of those squared residuals. We're going to take the sum of the residuals squared, and then we're going to divide that by the number of data points we have minus two. We can talk in future videos or a more advanced statistics class about why you divide by two, but it's related to the idea that what we're calculating here is a statistic, and we're trying to estimate a true parameter as best as possible.

N minus 2 actually does the trick for us. To calculate the root mean square deviation, we would then take the square root of this. Some of you might recognize strong parallels between this and how we calculated sample standard deviation early in our statistics career, and I encourage you to think about it.

But let's actually calculate it by hand, as I mentioned earlier in this video, to see how things actually play out. To do that, I'm going to give ourselves a little table here. Let's say that is our x value in that column. Let's make this our y value. Let's make this y hat, which is going to be equal to 2.5x minus 2.

Then let's make this the residual squared, which is going to be our y value minus our y hat value; our actual minus our estimate for that given x, squared. Then we're going to sum them all up, divide by n minus 2, and take the square root.

So first, let's do this data point: that's the point 1, 1. Now, what is the estimate from our regression line? For that x value, when x is equal to 1, it's going to be 2.5 times 1 minus 2. So it's going to be 2.5 times 1 minus 2, which is equal to 0.5.

Our residual squared is going to be 1 minus 0.5, which is equal to 0.5 squared, which is going to be 0.25. All right, let's do the next data point. We have this one right over here; it is 2, 2. Now our estimate from the regression line when x equals 2 is going to be equal to 2.5 times our x value (which is 2) minus 2, which is going to be equal to 3.

So our residual squared is going to be 2 minus 3, then squared. This is negative 1 squared, which is going to be equal to 1. Then we can go to this point; that's the point 2, 3. Now, our estimate from our regression line is going to be 2.5 times our x value (which is 2) minus 2, which is going to be equal to 3.

So our residual here is going to be zero, and you can see that that point sits on the regression line. It's going to be 3 minus 3, squared, which is equal to 0. Then, last but not least, we have this point right over here: when x is 3, our y value is, this person studied 3 hours, and they got a 6 on the test. So y is equal to 6.

Our estimate from the regression line, based on that regression line, is going to be 2.5 times our x value (which is 3) minus 2, which is equal to 5.5. Our residual squared is going to be 6 minus 5.5, squared, which is 0.5 squared, which is 0.25.

Now, the next step: let me take the sum of all of these squared residuals. So this can be written as follows: the sum of the residuals squared is equal to, if I just sum all of this up, it's going to be 1.5.

If I divide that by n minus 2, that's going to be equal to, I have four data points, so I'm going to divide by 4 minus 2. I'm going to divide by 2, and then I'm going to want to take the square root of that.

This is going to get us 1.5 over 2, which is the same thing as 3/4. So it's the square root of three-fourths or the square root of 3 over 2. You could use a calculator to figure out what that is as a decimal.

But this gives us a sense of how good a fit this regression line is. The closer this is to zero, the better the fit of the regression line; the further away from zero, the worse the fit. What would be the units for the root mean square deviation?

Well, it would be in terms of whatever your units are for your y-axis. In this case, it would be the score on the test, and that's one of the other values of this calculation of taking the square root of the sum of the squares of the residuals divided by n minus 2.

So, big picture: this square root of 3 over 2 can be viewed as the approximate size of a typical or average prediction error between these points and what the regression line would have predicted. Or you could view it as the approximate size of a typical or average residual.

More Articles

View All
Dred Scott v. Sandford | The Civil War era (1844-1877) | US history | Khan Academy
Hi, this is Kim from Khan Academy. Today we’re learning more about the landmark Supreme Court case Dred Scott versus Sanford, decided in 1857. The ruling in the Dred Scott case inflamed sectional tensions over slavery, which had been growing ever more hea…
Safari Live - Day 69 | Nat Geo WILD
Well hidden at the moment, tucked away in the long grass, and you can see that one of the little cubs is suckling away. So, these are the four cubs that we saw playing with the game drive blanket not all that long ago, tugging it backwards and forwards an…
Capital by Thomas Piketty | Macroeconomics | Khan Academy
Thomas Piketty’s Capital in the 21st Century has been getting a lot of attention lately because it’s addressing an issue that matters a lot to a lot of folks: the issue of income inequality and wealth inequality. My goal here isn’t to have a view on the b…
Chimú 101 | National Geographic
(Gentle music) [Narrator] Before the Spanish conquest and before the Inca empire, one group created the most important South American civilization at the time, the Chimu. The Chimu civilization lasted almost half a millennium, from the year 1000 to aroun…
Mr. Freeman, part 40
Look closer, but don’t blink your eyes because you will lose your favorite 25th frame. There is it! Again I appear through the invisible door in the dim light of your consciousness. Let me invite you to the dance. Waltz, please. We are dividing and rolli…
Reshma Shetty Speaks at Y Combinator's Female Founders Conference 2016
[Music] Hello everyone. Um, so first off, I’d like to, uh, thank both Jessica and Susan for inviting me to be here today. It’s a real privilege and honor to speak to such a talented, amazing group of women here. Um, so, so as Cat said, my name is RMA. I’…