Huge thanks to @veredshwartz.bsky.social, Peter West, Giuseppe Carenini
Paper: huggingface.co/papers/2509....
Code will be released soon!
Huge thanks to @veredshwartz.bsky.social, Peter West, Giuseppe Carenini
Paper: huggingface.co/papers/2509....
Code will be released soon!
👉 Social reasoning in LLMs cannot be achieved through optimizing their performance on general reasoning benchmarks alone!
👉It requires explicit modeling of mental states to enable safe, fair, and effective interactions with humans.
👉 Social reasoning in LLMs cannot be achieved through optimizing their performance on general reasoning benchmarks alone!
👉It requires explicit modeling of mental states to enable safe, fair, and effective interactions with humans.
🔹 ToMA prioritizes intentions > emotions (other dimensions remain similar)
🔹 Uses +5.6% more 1st-order belief than bases, even when both are prompted equally for 0th/1st order states.
🔹 ToMA prioritizes intentions > emotions (other dimensions remain similar)
🔹 Uses +5.6% more 1st-order belief than bases, even when both are prompted equally for 0th/1st order states.
ToMA outperforms the base under all settings. Its reasoning is more strategic (e.g., compromise, accommodation). Even in failures, ToMA shows more active engagement (e.g., failed persuasion).
ToMA outperforms the base under all settings. Its reasoning is more strategic (e.g., compromise, accommodation). Even in failures, ToMA shows more active engagement (e.g., failed persuasion).