3
$\begingroup$

I want to start off this question by saying that I'm looking for more of a conceptual understanding of this term in a simple regression model, not a mathematical one.

In econometrics, simple linear regression has an "error term", called mu, which represents "factors other than x that effect y, or unobserved data." In a statistics class, this term is called "statistical error" or "noise" treated as epsilon, which again is random noise in the real world our model cannot capture.

In Introduction to Statistical Learning, this term is referred to as "irreducible error," and is a quantity that represents unmeasured variables that we haven't measured, or simply just unmeasurable variation in real world data. The authors bifurcate between this term and reducible error.

Here is where my confusion lies: in a theoretically, completely deterministic world, where we have an infinite amount of data, is this "irreducible error" theoretically reducible? That is to say, if in this theoretical world, our covariate, college GPA, could be mapped to our response, say "Salary," could this term technically be zero?

I understand that in the world we live in, there is always unmeasurable variation in data, and that we'll never have access to an infinite amount of data or know beforehand every single covariate, thus this error arises. But say we had 5 data points, and this was our "theoretical population," would this "irreducible error" term disappear?

For example, in Introduction to Statistical learning, they have this image: Figure 2.2

My confusion in this image arises from the fact that this data was simulated, and so we actually know the true underlying function f. But if that is the case, why is there still error? If we know f, why can't we perfectly fit this function to the data? And if this data was simulated based on f, why doesn't the simulation produce data points on the function f?

I know it may seem like a silly question, but I'm trying to understand the nuance here. Many thanks to your help in advance.

$\endgroup$

1 Answer 1

3
$\begingroup$

Yes, you know $f$. There’s still something more to the points than $f$. The $f$ just gives the conditional expected value, conditioned on the feature value(s). Then the point is drawn from a distribution that has such a mean but also has (potentially) positive variance. Thus, each point $y_i$ can be written as:

$$ y_i = f(x_i) + \varepsilon_i $$

That is, the observed point $y_i$ is equal to the expected value, given by $f(x_i)$, plus some deviation from t the expected value.

If you don’t know what the deviations are, such as if they are random variables, then you don’t know exactly what the point will be, even if you exactly know the expected value. If you do know what the deviation is, the $f(x_i)$ wasn’t really the expected value, was it? The expected value is whatever $f(x_i)$ is plus the known deviation from $f(x_i)$. In fact, beyond being the expected value, that is the certain value.

The latter case is deterministic. However, the simulation you cite has a random component, so you don’t know what the observed $y_i$ will be, even if you perfectly know its expected value given by $f$.

THE SIMULATION FROM MY COMMENTS

Run the following in R.

library(ggplot2)
# set.seed(not telling)
N <- 100
a <- -2
b <- 2
x <- seq(a, b, (b - a)/(N - 1))
Ey <- 4*x - 1
plot(x, Ey)
e <- rt(N, 1.1)
y <- Ey + e
d <- data.frame(
  x = x,
  y = y
)
ggplot(d, aes(x = x, y = y)) +
  geom_point()

I get this plot.

My plot

You did not get the same plot, even though you know the conditional expected value, the $f(x)$, is $4x - 1$? It's almost as if I denied you a piece of information.

Now set the seed as 2024.

library(ggplot2)
set.seed(2024)
N <- 100
a <- -2
b <- 2
x <- seq(a, b, (b - a)/(N - 1))
Ey <- 4*x - 1
plot(x, Ey)
e <- rt(N, 1.1)
y <- Ey + e
d <- data.frame(
  x = x,
  y = y
)
ggplot(d, aes(x = x, y = y)) +
  geom_point()
$\endgroup$
4
  • $\begingroup$ This makes sense, thank you. But I still am hung up on this idea of irreducible error. The authors say "the quantity epsilon may contain unmeasured variables that are useful in predicting Y, or it may contain unmeasurable variation." But in theory, if we had perfect information, we could get rid of irreducible error, correct? The term irreducible is being used in the context of the world we live in, which is not deterministic. I know it's a subtle point, but just want to make sure I'm understanding that correctly. $\endgroup$
    – r_squared
    Commented Aug 30 at 17:36
  • $\begingroup$ @r_squared Yes, in theory we can get rid of the error by measuring everything (at least depending on some philosophical considerations). Indeed, when we run simulations, we set a seed so others can run the exact same simulation. Without knowing the seed, you see randomness and cannot perfectly match my results. If I reveal the seed to you, however, you will get the same results. I will edit in an example of this once I am no limited to posting on mobile. (Dave needs his RStudio!) $\endgroup$
    – Dave
    Commented Aug 30 at 17:45
  • $\begingroup$ Makes perfect sense, thank you! I think I had one other clarification but I lost it at the moment...will follow-up if it comes back to me (it was super nuanced). Thanks again, Dave! $\endgroup$
    – r_squared
    Commented Aug 30 at 17:56
  • $\begingroup$ @r_squared The promised simulation has been posted. $\endgroup$
    – Dave
    Commented Sep 9 at 8:01

Your Answer

By clicking “Post Your Answer”, you agree to our terms of service and acknowledge you have read our privacy policy.

Not the answer you're looking for? Browse other questions tagged or ask your own question.