Standard error of Estimate vs Standard error of Forecast

Does anyone know why we have to use Standard Error of Forecast (Sf) rather than SEE -> Standard Error of Estimate… in a confidence interval calculation?

Pretty sure those two things are not the same… can’t remember right now.

Yeap :D. I think it must be different.

Normally we use standard error of that parameter to calculate its confidence interval like Xbar +/- sigma/(n)^1/2

As I understand SEE is the standard error of the estimate Y, while we dont use SEE

I am so confused :))

Anyone help me pls