David Bindel
@dbindel.bsky.social
Professing computing and applied math at Cornell. Numerical methods for data science, plasma physics, other stuff depending on the day. Director, Cornell Center for Applied Mathematics; Director, Simons Collaboration on Hidden Symmetries and Fusion Energy.
(A postscript: In Fall 2008, as a Courant instructor at NYU math, I taught an upper division undergrad probability course in which I gave myself free license to say things like "covariance is an inner product over a space of mean zero rvs," as a treat. At least a few students appreciated this.)
October 13, 2025 at 6:37 PM
(A postscript: In Fall 2008, as a Courant instructor at NYU math, I taught an upper division undergrad probability course in which I gave myself free license to say things like "covariance is an inner product over a space of mean zero rvs," as a treat. At least a few students appreciated this.)
Googling for this mostly brings up my own notes, which suggests to me that I'm falling prey to idiosyncracies of my own academic upbringing. Pointers to alternate terminology that might lead to more successful searches would also be welcome!
October 13, 2025 at 6:32 PM
Googling for this mostly brings up my own notes, which suggests to me that I'm falling prey to idiosyncracies of my own academic upbringing. Pointers to alternate terminology that might lead to more successful searches would also be welcome!
Now, here's the question: the concept of bias-variance decomposition is well known in statistics and machine learning. The concept of quasi-optimality is well known in approximation theory. Putting them together is not deep. Surely there must be a reference?
October 13, 2025 at 6:31 PM
Now, here's the question: the concept of bias-variance decomposition is well known in statistics and machine learning. The concept of quasi-optimality is well known in approximation theory. Putting them together is not deep. Surely there must be a reference?
One can do something similar with the noise term, and this is something I teach in my matrix computations class:
www.cs.cornell.edu/courses/cs62...
www.cs.cornell.edu/courses/cs62...
CS 6210: Matrix Computations
www.cs.cornell.edu
October 13, 2025 at 6:29 PM
One can do something similar with the noise term, and this is something I teach in my matrix computations class:
www.cs.cornell.edu/courses/cs62...
www.cs.cornell.edu/courses/cs62...
For a regularized least squares fit, we can get that 𝔼[c]-c_* is the Moore-Penrose pseudoinverse (with regularization) of a piece of r. And from there, we can show *quasi-optimality* of the squared bias: the version learned from data gives a residual within some factor of the best possible.
October 13, 2025 at 6:29 PM
For a regularized least squares fit, we can get that 𝔼[c]-c_* is the Moore-Penrose pseudoinverse (with regularization) of a piece of r. And from there, we can show *quasi-optimality* of the squared bias: the version learned from data gives a residual within some factor of the best possible.
Consider a generalized linear model s(X) = w(X) c. There is a c_* that minimizes the expected squared bias term 𝔼[r(X)²] where r(X) = f(X)-w(X) c_*; and we can show (again by orthogonality) that 𝔼[(m(X)-f(X))²] = 𝔼[(w(X) (𝔼[c]-c_*))²] + 𝔼[r(X)²].
October 13, 2025 at 6:26 PM
Consider a generalized linear model s(X) = w(X) c. There is a c_* that minimizes the expected squared bias term 𝔼[r(X)²] where r(X) = f(X)-w(X) c_*; and we can show (again by orthogonality) that 𝔼[(m(X)-f(X))²] = 𝔼[(w(X) (𝔼[c]-c_*))²] + 𝔼[r(X)²].
Now let X be random, and consider
𝔼[(f(X)-s(X))²] = Var[s(X)] + 𝔼[(m(X)-f(X))²]
where m is the mean of s with respect to the sampling variance (the expectation is just over X).
𝔼[(f(X)-s(X))²] = Var[s(X)] + 𝔼[(m(X)-f(X))²]
where m is the mean of s with respect to the sampling variance (the expectation is just over X).
October 13, 2025 at 6:23 PM
Now let X be random, and consider
𝔼[(f(X)-s(X))²] = Var[s(X)] + 𝔼[(m(X)-f(X))²]
where m is the mean of s with respect to the sampling variance (the expectation is just over X).
𝔼[(f(X)-s(X))²] = Var[s(X)] + 𝔼[(m(X)-f(X))²]
where m is the mean of s with respect to the sampling variance (the expectation is just over X).
Now suppose f(x) is a (deterministic) function and s(x) is a predictor (based on a noisy sample). Then
𝔼[(f(x)-s(x))²] = Var[s(X)]+(𝔼[s(X)]-f(x))²
This is the bias-variance decomposition (sometimes people add a term for test-time noise, but I'll drop it here). Same idea.
𝔼[(f(x)-s(x))²] = Var[s(X)]+(𝔼[s(X)]-f(x))²
This is the bias-variance decomposition (sometimes people add a term for test-time noise, but I'll drop it here). Same idea.
October 13, 2025 at 6:16 PM
Now suppose f(x) is a (deterministic) function and s(x) is a predictor (based on a noisy sample). Then
𝔼[(f(x)-s(x))²] = Var[s(X)]+(𝔼[s(X)]-f(x))²
This is the bias-variance decomposition (sometimes people add a term for test-time noise, but I'll drop it here). Same idea.
𝔼[(f(x)-s(x))²] = Var[s(X)]+(𝔼[s(X)]-f(x))²
This is the bias-variance decomposition (sometimes people add a term for test-time noise, but I'll drop it here). Same idea.
Also got to have lunch with a former student of mine yesterday. Time flies.
October 4, 2025 at 1:57 PM
Also got to have lunch with a former student of mine yesterday. Time flies.
- Asvine just released a new pen! The V800 is very pretty. I don't need more fountain pens, but can still admire.
- I get to check out a CT scanner setup tomorrow! Research reasons, nothing medical.
- The CAM students remain an excellent community. They did tie-dye shirts today.
- I get to check out a CT scanner setup tomorrow! Research reasons, nothing medical.
- The CAM students remain an excellent community. They did tie-dye shirts today.
September 28, 2025 at 12:19 AM
- Asvine just released a new pen! The V800 is very pretty. I don't need more fountain pens, but can still admire.
- I get to check out a CT scanner setup tomorrow! Research reasons, nothing medical.
- The CAM students remain an excellent community. They did tie-dye shirts today.
- I get to check out a CT scanner setup tomorrow! Research reasons, nothing medical.
- The CAM students remain an excellent community. They did tie-dye shirts today.
- I am learning about the mechanics of hair! Which is also awesome.
- Apparently, thinking these things are awesome is a thing the administration now thinks can be cured with Vitamin B. My vitamin B levels are fine, and I am happy they are wrong (even if I wish they were less in my face about it).
- Apparently, thinking these things are awesome is a thing the administration now thinks can be cured with Vitamin B. My vitamin B levels are fine, and I am happy they are wrong (even if I wish they were less in my face about it).
September 28, 2025 at 12:15 AM
- I am learning about the mechanics of hair! Which is also awesome.
- Apparently, thinking these things are awesome is a thing the administration now thinks can be cured with Vitamin B. My vitamin B levels are fine, and I am happy they are wrong (even if I wish they were less in my face about it).
- Apparently, thinking these things are awesome is a thing the administration now thinks can be cured with Vitamin B. My vitamin B levels are fine, and I am happy they are wrong (even if I wish they were less in my face about it).
- I am also writing a recommendation for a postdoc, who is awesome and should land that faculty job he wants. And if the market sucks, well, I am still happy he's awesome and working some of the time with me.
September 28, 2025 at 12:12 AM
- I am also writing a recommendation for a postdoc, who is awesome and should land that faculty job he wants. And if the market sucks, well, I am still happy he's awesome and working some of the time with me.
- I am astounded that I don't yet have a better alternative than TikZ and dvisvgm for programmatically producing the titles of math diagrams I want for notes and slides, and it's fun to Google variations of "really?" in the hopes of finding a different answer.
September 28, 2025 at 12:10 AM
- I am astounded that I don't yet have a better alternative than TikZ and dvisvgm for programmatically producing the titles of math diagrams I want for notes and slides, and it's fun to Google variations of "really?" in the hopes of finding a different answer.