thoughts on whatever i feel like (often ml and life)
i might be wrong, opinions subject to change
tanaybiradar.com
Tested it out on an IRL Github frontend issue. Seems to work pretty well (but it's really expensive)
Tested it out on an IRL Github frontend issue. Seems to work pretty well (but it's really expensive)
It only seems to work well for a few types of clips (ex. news, landscape shots). Physics/interactions between objects are notably low-quality.
It only seems to work well for a few types of clips (ex. news, landscape shots). Physics/interactions between objects are notably low-quality.
life imitates art.
life imitates art.
today @akshitab.bsky.social @natolambert.bsky.social and I are giving our #neurips2024 tutorial on language model development.
everything from data, training, adaptation. published or not, no secrets 🫡
tues, 12/10, 9:30am PT ☕️
neurips.cc/virtual/2024...
today @akshitab.bsky.social @natolambert.bsky.social and I are giving our #neurips2024 tutorial on language model development.
everything from data, training, adaptation. published or not, no secrets 🫡
tues, 12/10, 9:30am PT ☕️
neurips.cc/virtual/2024...
the weird part: your function can be differentiable when viewing it as R2->R2, but not when viewing it as C->C.
the weird part: your function can be differentiable when viewing it as R2->R2, but not when viewing it as C->C.
"Just because we’re using the same words doesn’t mean we mean the same things.... But let's try anyway."
nathanieldrew.substack.com/p/how-to-hav...
"Just because we’re using the same words doesn’t mean we mean the same things.... But let's try anyway."
nathanieldrew.substack.com/p/how-to-hav...
now that LLMs have been in our hands for ~2 years, i feel like it takes a lot to impress me. i'm kind of surprised it took multiple prompts to get this right.
now that LLMs have been in our hands for ~2 years, i feel like it takes a lot to impress me. i'm kind of surprised it took multiple prompts to get this right.
1. View a matrix as a sum of rank-1 matrices
2. Read orthonormal bases for the 4 fundamental subspaces
3. Get a low-rank approximation, etc.
Sadly higher order tensors are much harder to factorize for NN weight compression...
1. View a matrix as a sum of rank-1 matrices
2. Read orthonormal bases for the 4 fundamental subspaces
3. Get a low-rank approximation, etc.
Sadly higher order tensors are much harder to factorize for NN weight compression...
Regardless, I wonder if neural networks and digital computing (as opposed to analog) are the wrong approach to intelligence.
Regardless, I wonder if neural networks and digital computing (as opposed to analog) are the wrong approach to intelligence.
Suppose you run a program on 2 machines with different ISAs. Outputs obtained on one machine may be *impossible* on the other. There are no guarantees about the order of memory reads/writes.
Suppose you run a program on 2 machines with different ISAs. Outputs obtained on one machine may be *impossible* on the other. There are no guarantees about the order of memory reads/writes.
Decided to start working on my own RSS feed reader that curates your content - no complicated algorithms!
github.com/TanayB11/pap...
Decided to start working on my own RSS feed reader that curates your content - no complicated algorithms!
github.com/TanayB11/pap...
in theory you should get more fulfillment from this, but i feel like you get more fulfillment from changing *someones* world instead of changing *the* world.
stevenscrawls.com/care-doesnt-...
in theory you should get more fulfillment from this, but i feel like you get more fulfillment from changing *someones* world instead of changing *the* world.
stevenscrawls.com/care-doesnt-...
"I am always sincere, but never serious." - Alan Watts
"I am always sincere, but never serious." - Alan Watts