Data management and NLP/LLMs for information quality.
https://www.eurecom.fr/~papotti/
- Query planning as constrained optimization over quality constraints and cost objective
- Gradient-based optimization to jointly choose operators and allocate error budgets across pipelines
- KV-cache–based operators to turn discrete physical choices into a runtime-quality continuum
- Query planning as constrained optimization over quality constraints and cost objective
- Gradient-based optimization to jointly choose operators and allocate error budgets across pipelines
- KV-cache–based operators to turn discrete physical choices into a runtime-quality continuum
This is the first outcome of our collaboration with Technische Universität Darmstadt within the @agencerecherche.bsky.social / @dfg.de ANR/DFG #Magiq project - more to come!
This is the first outcome of our collaboration with Technische Universität Darmstadt within the @agencerecherche.bsky.social / @dfg.de ANR/DFG #Magiq project - more to come!
bsky.app/profile/papo...
RAG uses embedding scores to pick Top-K, then treat all retrieved chunks as equal.
Parallel Context-of-Experts Decoding (PCED) uses retrieval scores to move evidence aggregation from attention to decoding.
🚀 180× faster time-to-first-token!
bsky.app/profile/papo...
Amazing work from Giulio Corallo in his industrial PhD at SAP!
Amazing work from Giulio Corallo in his industrial PhD at SAP!
🚀 Systems win: ~180× faster time-to-first-token vs long-context prefill using continuous batching and Paged Attention.
🚀 Systems win: ~180× faster time-to-first-token vs long-context prefill using continuous batching and Paged Attention.
● Keeps each document as its own 𝐞𝐱𝐩𝐞𝐫𝐭 with independent KV cache
● Runs experts in 𝐩𝐚𝐫𝐚𝐥𝐥𝐞𝐥 to get logits
● Selects next token with a 𝐫𝐞𝐭𝐫𝐢𝐞𝐯𝐚𝐥-𝐚𝐰𝐚𝐫𝐞 𝐜𝐨𝐧𝐭𝐫𝐚𝐬𝐭𝐢𝐯𝐞 𝐝𝐞𝐜𝐨𝐝𝐢𝐧𝐠 rule integrating scores as a prior
● Keeps each document as its own 𝐞𝐱𝐩𝐞𝐫𝐭 with independent KV cache
● Runs experts in 𝐩𝐚𝐫𝐚𝐥𝐥𝐞𝐥 to get logits
● Selects next token with a 𝐫𝐞𝐭𝐫𝐢𝐞𝐯𝐚𝐥-𝐚𝐰𝐚𝐫𝐞 𝐜𝐨𝐧𝐭𝐫𝐚𝐬𝐭𝐢𝐯𝐞 𝐝𝐞𝐜𝐨𝐝𝐢𝐧𝐠 rule integrating scores as a prior
#LLM #Factuality #Benchmark #RelationalFactQA #NLP #AI
#LLM #Factuality #Benchmark #RelationalFactQA #NLP #AI
Wider or longer output tables = tougher for all LLMs! 🧨
From Llama 3 and Qwen to GPT-4, no LLM goes above 25% accuracy on our stricter measure.
Wider or longer output tables = tougher for all LLMs! 🧨
From Llama 3 and Qwen to GPT-4, no LLM goes above 25% accuracy on our stricter measure.
@tanmoy-chak.bsky.social for leading this effort!
@tanmoy-chak.bsky.social for leading this effort!
@iaugenstein.bsky.social
@preslavnakov.bsky.social
@igurevych.bsky.social
@emilioferrara.bsky.social
@fil.bsky.social
@giovannizagni.bsky.social
@dcorney.com
@mbakker.bsky.social
@computermacgyver.bsky.social
@irenelarraz.bsky.social
@gretawarren.bsky.social
Excited to hear your thoughts!
#Misinformation #FactChecking #SocialMedia #Epistemology #HCI #DigitalTruth #CommunityNotes
arxiv.org/pdf/2505.20067
Excited to hear your thoughts!
#Misinformation #FactChecking #SocialMedia #Epistemology #HCI #DigitalTruth #CommunityNotes
arxiv.org/pdf/2505.20067
– Can crowds overcome bias?
– What counts as evidence?
– Who holds epistemic authority?
Our interdisciplinary analysis combines perspectives from HCI, media studies, & digital governance.
– Can crowds overcome bias?
– What counts as evidence?
– Who holds epistemic authority?
Our interdisciplinary analysis combines perspectives from HCI, media studies, & digital governance.
We argue this isn’t just a technical shift — it’s an epistemological transformation. Who gets to define what's true when everyone is the fact-checker?
We argue this isn’t just a technical shift — it’s an epistemological transformation. Who gets to define what's true when everyone is the fact-checker?