Turns out it generalizes to _every distribution_ using cumulant tensors!
That's higher order variance, skewness, kurtosis, etc.
Turns out it generalizes to _every distribution_ using cumulant tensors!
That's higher order variance, skewness, kurtosis, etc.
Hopefully it can also be a way to help people become familiar with tensor diagrams.
Hopefully it can also be a way to help people become familiar with tensor diagrams.
di.ku.dk/english/news...
di.ku.dk/english/news...
(n/e)ⁿ√{2π n} ≤ n! ≤ (n/e)ⁿ(√{2π n}+1)
Inspired by the discussion on mathoverflow.net/a/458011/5429. Just had to keep hitting it with logarithmic inequalities...
(n/e)ⁿ√{2π n} ≤ n! ≤ (n/e)ⁿ(√{2π n}+1)
Inspired by the discussion on mathoverflow.net/a/458011/5429. Just had to keep hitting it with logarithmic inequalities...
Let X ~ Poisson(𝜇); Z = (X - 𝜇)/√𝜇; Y ~ Normal(0, 1).
How close is E[|X|^k] is to E[|Y|^k]?
Say we connect 𝜇 and k by 𝜇 = c k³, what is now the limit E[|X|^k]/E[|Y|^k] as k → ∞?
This was harder to solve than expected, but the answer was surprisingly pretty 🌻
Let X ~ Poisson(𝜇); Z = (X - 𝜇)/√𝜇; Y ~ Normal(0, 1).
How close is E[|X|^k] is to E[|Y|^k]?
Say we connect 𝜇 and k by 𝜇 = c k³, what is now the limit E[|X|^k]/E[|Y|^k] as k → ∞?
This was harder to solve than expected, but the answer was surprisingly pretty 🌻
AlphaZero generalized and simplified most of the tricks in chess engines like Stockfish, but one category is missing: history heuristics...
1/5
Test-Time Training (TTT) enhances AI models' abstract reasoning, achieving 53% accuracy on ARC—a 25% improvement over prior methods.
http://ekinakyurek.github.io/papers/ttt.pdf
#AI #AbstractReasoning #Innovation
AlphaZero generalized and simplified most of the tricks in chess engines like Stockfish, but one category is missing: history heuristics...
1/5
By "taking notes" as you read, ypu reduce the complexity from N^3 (N tokens at N^2 cost) to N^3/3 (1+4+9+...+N^2).
By "taking notes" as you read, ypu reduce the complexity from N^3 (N tokens at N^2 cost) to N^3/3 (1+4+9+...+N^2).