Visiting Scholar SHI Lab @ Georgia Tech
Opinions expressed are those of my cat
So if it's easy to distinguish, I think it's best to. Avoids confusion later on.
So if it's easy to distinguish, I think it's best to. Avoids confusion later on.
Something doesn't smell right and it's not my shit
Something doesn't smell right and it's not my shit
But my larger point is that this is extremely difficult to conclude. I just think "PhD level reasoning" is a much stronger claim so needs stronger evidence.
But my larger point is that this is extremely difficult to conclude. I just think "PhD level reasoning" is a much stronger claim so needs stronger evidence.
en.wikipedia.org/wiki/ELIZA
Per the death question, well we trained it on those. So if we programmed similar responses into Eliza does that make Eliza more alive or change your answer about her? Do we default to conscious?
en.wikipedia.org/wiki/ELIZA
Per the death question, well we trained it on those. So if we programmed similar responses into Eliza does that make Eliza more alive or change your answer about her? Do we default to conscious?
So let's modify our memorization question slightly: how do you differentiate reasoning from doing a similarity search on a lookup table?
Are those different things? Is the failure in figure 1 because a reasoning failure or a search failure? How do you know?
So let's modify our memorization question slightly: how do you differentiate reasoning from doing a similarity search on a lookup table?
Are those different things? Is the failure in figure 1 because a reasoning failure or a search failure? How do you know?
We train these machines very differently and so you can't evaluate them the same way.
So go to Figure 1 and tell me if those are in distribution or not. They are all (ax, by, cz) problems
We train these machines very differently and so you can't evaluate them the same way.
So go to Figure 1 and tell me if those are in distribution or not. They are all (ax, by, cz) problems
A problem here is what is considered OOD? Take this old example, what do you consider to be OOD? The number of digits? Some factorization? Why?
bsky.app/profile/swal...
Not getting 100% on 3 digit times a 3 digit but getting a 6 digit x 4 digit should make us question everything.
Something fundamental is wrong.
A problem here is what is considered OOD? Take this old example, what do you consider to be OOD? The number of digits? Some factorization? Why?
bsky.app/profile/swal...
I want to differentiate reasoning from memorizing. We can agree here, right?
If they fail a problem that uses identical reasoning to problems that they succeed at and such problems are the same as those in the training, can you conclude the
I want to differentiate reasoning from memorizing. We can agree here, right?
If they fail a problem that uses identical reasoning to problems that they succeed at and such problems are the same as those in the training, can you conclude the
But maybe you can help me. How do we know my calculator isn't conscious? What makes it uniquely unconscious? That it doesn't talk? Doesn't pursue its own goals? How do you differentiate
But maybe you can help me. How do we know my calculator isn't conscious? What makes it uniquely unconscious? That it doesn't talk? Doesn't pursue its own goals? How do you differentiate
If we built a really sophisticated animatronic duck do you think you could easily differentiate it from a real duck?
If we built a really sophisticated animatronic duck do you think you could easily differentiate it from a real duck?
You can't apply the duck test here.
Just because it looks like a duck, swims like a duck, and quacks like a duck does not mean it isn't an advanced animatronic duck. In a normal setting we should conclude that it is very
You can't apply the duck test here.
Just because it looks like a duck, swims like a duck, and quacks like a duck does not mean it isn't an advanced animatronic duck. In a normal setting we should conclude that it is very
I'd say this work demonstrates that you cannot conclude that their outputs are reliable representations of their processing.
alignment.anthropic.com/2025/sublimi...
I'd say this work demonstrates that you cannot conclude that their outputs are reliable representations of their processing.
alignment.anthropic.com/2025/sublimi...
That Grok example above is illustrative of this.
That Grok example above is illustrative of this.
That's exactly how the bias in claims works.
That's exactly how the bias in claims works.
Does this prove our claims? No. Does it give them evidence? Yes. In the same way you are using the links you provided (ethos)
Does this prove our claims? No. Does it give them evidence? Yes. In the same way you are using the links you provided (ethos)