Asher Zheng
banner
asher-zheng.bsky.social
Asher Zheng
@asher-zheng.bsky.social
PhD @ UT Linguistics
Semantics/Pragmatics/NLP
https://asherz720.github.io/
Prev.@UoEdinburgh @Hanyang
By analyzing model reasoning, we find extra reasoning introduces overcomplication (img left), misunderstanding, and internal inconsistency (img right). This shows the current LLMs still lack sophisticated pragmatic understanding in many ways.
June 3, 2025 at 11:56 AM
We evaluate a range of LLMs in terms of how good they are at perceiving strategic language. We show models struggle with our metrics while showing an overall good understanding of Gricean principles. Model size tends to have a positive effect, while reasoning does not help.
June 3, 2025 at 11:56 AM
(2) BaT and PaT are valid terms that reflect strategic gains/losses, which can to some extent predict conversational outcomes. In addition, our metrics are more objective. When conditioned on cases where the outcome is made based on logical arguments, the predictive power rises.
June 3, 2025 at 11:56 AM
We also introduce CHARM, an annotated dataset of real legal cross-examination dialogues. By applying our framework, we show (1) (non-)cooperative discourse are distinct over the identified properties (img left), and BaT and PaT show such a distributional distinction (img right).
June 3, 2025 at 11:56 AM
Based on the components above, we introduce three metrics—Benefit at Turn (BaT), Penalty at Turn (PaT), and Normalized Relative Benefit at Turn (NRBaT)—to measure the strategic gains, losses, and cumulative benefits at a turn.
June 3, 2025 at 11:56 AM
Language is often strategic, but LLMs tend to play nice. How strategic are they really? Probing into that is key for future safety alignment.

👉Introducing CoBRA🐍, a framework that assesses strategic language.

Work with my amazing advisors @jessyjli.bsky.social and @David I. Beaver!
June 3, 2025 at 11:56 AM