Dilek Hakkani-Tur
@dilekh.bsky.social
800 followers 310 following 4 posts
UIUC, NLP, @convai-uiuc.bsky.social
Posts Media Videos Starter Packs
Reposted by Dilek Hakkani-Tur
sigdial.bsky.social
The regular submission deadline for SIGdial 2025 has now passed...

But we are still welcoming submissions through ACL Rolling Review 🎉

ARR Commitment Deadline: June 6th

Acceptance notifications will be on June 20th, and then SIGdial will be held in Avignon, France: August 25th - 27th
Reposted by Dilek Hakkani-Tur
sagnikmukherjee.bsky.social
🚨 Paper Alert: “RL Finetunes Small Subnetworks in Large Language Models”

From DeepSeek V3 Base to DeepSeek R1 Zero, a whopping 86% of parameters were NOT updated during RL training 😮😮
And this isn’t a one-off. The pattern holds across RL algorithms and models.
🧵A Deep Dive
Reposted by Dilek Hakkani-Tur
tejassrinivasan.bsky.social
LLMs are all around us, but how can we foster reliable and accountable interactions with them??

To discuss these problems, we will host the first ORIGen workshop at @colmweb.org! Submissions welcome from NLP, HCI, CogSci, and anything human-centered, due June 20 :)

origen-workshop.github.io
ORIGen 2025
Workshop on Optimal Reliance and Accountability in Interactions with Generative LMs
origen-workshop.github.io
Reposted by Dilek Hakkani-Tur
beyzabozdag.bsky.social
Thrilled to announce our new survey that explores the exciting possibilities and troubling risks of computational persuasion in the era of LLMs 🤖💬
📄Arxiv: arxiv.org/pdf/2505.07775
💻 GitHub: github.com/beyzabozdag/...
Reposted by Dilek Hakkani-Tur
sagnikmukherjee.bsky.social
🚀Our ICML 2025 paper introduces "Premise-Augmented Reasoning Chains" - a structured approach to induce explicit dependencies in reasoning chains.

By revealing the dependencies within chains, we significantly improve how LLM reasoning can be verified.

🧵[1/n]
Reposted by Dilek Hakkani-Tur
sivareddyg.bsky.social
Incredibly proud of my students @adadtur.bsky.social and Gaurav Kamath for winning a SAC award at #NAACL2025 for their work on assessing how LLMs model constituent shifts.
Reposted by Dilek Hakkani-Tur
xhluca.bsky.social
Agents like OpenAI Operator can solve complex computer tasks, but what happens when users use them to cause harm, e.g. spread misinformation?

To find out, we introduce SafeArena (safearena.github.io), a benchmark to assess the capabilities of web agents to complete harmful web tasks. A thread 👇
dilekh.bsky.social
While persuasive models are promising for social good, they can also be misused towards harmful behavior. Recent work by @beyzabozdag.bsky.social and @shuhaib.bsky.social aims to assess LLM persuasiveness and susceptibility towards persuasion.
beyzabozdag.bsky.social
[1/6] Can LLMs out-persuade each other? 🤖🧠💬

Introducing Persuade Me If You Can (PMIYC)—a new framework to evaluate (1) how persuasive LLMs are and (2) how easily they can be persuaded! 🚀

📄Arxiv: arxiv.org/abs/2503.01829
🌐Project Page: beyzabozdag.github.io/PMIYC/
dilekh.bsky.social
Data selection for instruction fine-tuning of LLMs doesn't need to be computationally costly. Great work by @wonderingishika.bsky.social! @convai-uiuc.bsky.social
wonderingishika.bsky.social
🚀Very excited about my new paper!

NN-CIFT slashes data valuation costs by 99% using tiny neural nets (205k params, just 0.0027% of 8B LLMs) while maintaining top-tier performance!
dilekh.bsky.social
Instruction data can also be synthesized using feedback based on reference examples. Please check our recent work for more information. Thanks to @shuhaib.bsky.social, Xiusi Chen, and Heng Ji!
shuhaib.bsky.social
💡 Introducing Reference-Level Feedback: A new paradigm for using feedback to improve synthetic data!
🌐 shuhaibm.github.io/refed/
🧵 [1/n]
dilekh.bsky.social
AI over-reliance is an important issue for conversational agents. Our work supported mainly by the DARPA FACT program proposes introducing positive friction to encourage users to think critically when making decisions. Great team-work, all!
@convai-uiuc.bsky.social @gokhantur.bsky.social
merterm.bsky.social
‼️ Ever wish LLMs would just... slow down for a second?

In our latest work, "Better Slow than Sorry: Introducing Positive Friction for Reliable Dialogue Systems", we delve into how strategic delays can enhance dialogue systems.

Paper Website: merterm.github.io/positive-fri...