Hal Daumé III
banner
haldaume3.bsky.social
Hal Daumé III
@haldaume3.bsky.social
Human-centered AI #HCAI, NLP & ML. Director TRAILS (Trustworthy AI in Law & Society) and AIM (AI Interdisciplinary Institute at Maryland). Formerly Microsoft Research NYC. Fun: 🧗🧑‍🍳🧘⛷️🏕️. he/him.
my teenagehood destroyed: pbrush now has AI support.
June 23, 2025 at 9:36 AM
There is a new version of the Research Plan for NIST's AI Safety Consortium (AISIC) in response to EOs. I did a diff.

Out: safety, responsibility, sociotechnical, fairness, working w fed agencies, authenticating content, watermarking, RN of CBRN, autonomous replication, ctrl of physical systems
>
March 4, 2025 at 12:29 PM
at TRAILCon today, enjoying TRAILS branded Trail Mix
February 3, 2025 at 3:03 PM
Please join us for:
AI at Work: Building and Evaluating Trust

Presented by our Trustworthy AI in Law & Society (TRIALS) institute.

Feb 3-4
Washington DC

Open to all!

Details and registration at: trails.gwu.edu/trailscon-2025
Sponsorship details at: trails.gwu.edu/media/556
January 16, 2025 at 3:22 PM
the previous is "what is intelligence" which then transitioned to "what is AI":

here, i kind of like kaplan+haenlein and also google. tesler is of course always correct and may really be the right answer.
December 18, 2024 at 3:22 PM
fwiw, I did a poll with undergrads (this was in the second day of a gened genai course) with seven definitions (all with citations inline).

i kind of like gardner and legg+hutter fwiw.
December 18, 2024 at 3:18 PM
for the linguists out there
December 18, 2024 at 8:19 AM
Premier-TACO is a Few-Shot Policy Learner: Pretraining Multitask Representation via Temporal Action-Driven Contrastive Loss

We show that a carefully constructed temporal contrastive loss leads to effective, multitask RL pretraining.

by Ruijie Zheng +al ICML’24
hal3.name/docs/daume24...
(eos)
December 16, 2024 at 10:51 AM
HateCOT: An Explanation-Enhanced Dataset for Generalizable Offensive Speech Detection via LLMs

There are lots of hate datasets with different nuances. We show how to pretrain on a COT-enhanced dataset to get great performance on data du jour.

by Huy Nghiem +al EMNLP’24
hal3.name/docs/daume24...
>
December 13, 2024 at 10:20 AM
starting to feel like slack is as bad as email.
December 12, 2024 at 1:22 PM
Do great minds think alike? Investigating Human-AI Complementarity in QA

We use item response theory to compare the capabilities of 155 people vs 70 chatbots at answering questions, teasing apart complementarities; implications for design.

by Maharshi Gor +al EMNLP’24
hal3.name/docs/daume24...
>
December 12, 2024 at 10:41 AM
Understanding the Impacts of Language Technologies’ Performance Disparities on AAL Speakers

We find AAL speakers expend significant invisible labor in order to achieve parity of outputs in LT systems; fairness measures don't capture this.

by Jay Cunningham +al ACL’24
hal3.name/docs/daume24...
>
December 11, 2024 at 8:53 AM
DrM: Mastering Visual Reinforcement Learning through Dormant Ratio Minimization

We show in RL that explicitly minimizing the dormant ratio (fraction of non-active neurons) improves exploration, rewards, etc.

by Guowei Xu, Ruijie Zheng, Yongyuan Liang +al ICLR’24
hal3.name/docs/daume24...
>
December 10, 2024 at 9:21 AM
The Impact of Explanations on Fairness in Human-AI Decision-Making: Protected vs Proxy Features

Despite hopes that explanations improve fairness, we see that when biases are hidden behind proxy features, explanations may not help.

Navita Goyal, Connor Baumler +al IUI’24
hal3.name/docs/daume23...
>
December 9, 2024 at 11:41 AM
PRISE: Learning Temporal Action Abstractions as a Sequence Compression Problem

We show that LLM-style byte-pair encoding can be used to compress action sequences to give “meta-actions” of different lengths, improving RL performance.

by Ruijie Zheng +al ICML ‘24
hal3.name/docs/daume24...
>
December 6, 2024 at 9:06 AM
An RCT on Anonymizing Reviewers to Each Other in Peer Review Discussions

We observe that anon discussions between revs led to slightly more discussion, less influence of seniority, no diff in politeness, & is slightly preferred by revs.

by Charvi Rastogi et al PLOS1’24
hal3.name/docs/daume24...
>
December 5, 2024 at 9:03 AM
December 4, 2024 at 3:16 PM
"You Gotta be a Doctor, Lin": An Investigation of Name-Based Bias of LLMs in Employment Recommendations

Chatbots routinely prefer candidates with White, female-sounding names over others, even when candidates have identical qualifications.

by Huy Nghiem et al EMNLP’24

hal3.name/docs/daume24...
>
December 4, 2024 at 9:02 AM
Large Language Models Help Humans Verify Truthfulness—Except When They Are Convincingly Wrong

Should one use chatbots or web search to fact check? Chatbots help more on avg, but people uncritically accept their suggestions much more often.

by Chenglei Si +al NAACL’24

hal3.name/docs/daume24...
>
December 3, 2024 at 9:31 AM
Successfully Guiding Humans with Imperfect Instructions by Highlighting Potential Errors and Suggesting Corrections

When generating instructions for people, we can help them by highlighting potential confabs, AND by suggesting alternatives.

by Lingjun Zhao EMNLP’24

hal3.name/docs/daume24...
>
December 2, 2024 at 9:37 AM
france: no we don’t do halloween because it’s too american and commercialized

also france: you’ve heard of black friday? let me introduce you to
December 1, 2024 at 7:52 AM
ASL STEM Wiki: Dataset and Benchmark for Interpreting STEM Articles

We develop a continuous signing dataset for ASL on a STEM subset of Wikipedia; challenges suggest problems related to fingerspelling detection, sign linking, & translation.

by Kayo Yin et al EMNLP’24
hal3.name/docs/daume24...
>
November 27, 2024 at 9:00 AM
Join fellow members of the AI Safety Institute Consortium (AISIC) on Dec 3 for the 1st annual plenary, hosted by TRAILS on behalf of NIST at UMD. (Reception on Dec 2 evening.)

This event is only open to individuals who work for a member of AISIC.

📅Register by Dec 2
👉Learn more: go.umd.edu/1u0n
November 27, 2024 at 8:58 AM
Toxicity Detection is NOT all you Need: Measuring the Gaps to Supporting Volunteer Content Moderators

Much work aims to help moderators by automation; much focusing on toxicity. Turns out that’s a small part of what volunteer mods want.

by Yang Trista Cao +al EMNLP’24

hal3.name/docs/daume24...
>
November 26, 2024 at 8:45 AM
How do Authors' Perceptions of their Papers Compare with Co-authors' Perceptions and Peer-review Decisions?

Authors overestimate the chances their papers'll be accepted; co-auths disagree about paper value about as much as auth/revs.

by Charvi Rastogi +al PLOS One'24

hal3.name/docs/daume24...

>
November 25, 2024 at 9:34 AM