Hovy, D., Berg-Kirkpatrick, T., Vaswani, A., & Hovy E. (2013). Learning Whom to Trust With MACE. In: Proceedings of NAACL-HLT. ACL.
aclanthology.org/N13-1132.pdf
And for even more details:
aclanthology.org/Q18-1040.pdf
N/N
Hovy, D., Berg-Kirkpatrick, T., Vaswani, A., & Hovy E. (2013). Learning Whom to Trust With MACE. In: Proceedings of NAACL-HLT. ACL.
aclanthology.org/N13-1132.pdf
And for even more details:
aclanthology.org/Q18-1040.pdf
N/N
Last week, I played around with Cursor – and got it all done in ~1 hour. 🤯
If you work with any response data that needs aggregation, give it a try—and let me know what you think!
4/N
Last week, I played around with Cursor – and got it all done in ~1 hour. 🤯
If you work with any response data that needs aggregation, give it a try—and let me know what you think!
4/N
1. Annotator reliability (who’s consistent?)
2. Item difficulty (which examples spark disagreement?)
3. The most likely aggregate label (the latent “best guess”)
That “side project” ended up powering hundreds of annotation projects over the years.
3/N
1. Annotator reliability (who’s consistent?)
2. Item difficulty (which examples spark disagreement?)
3. The most likely aggregate label (the latent “best guess”)
That “side project” ended up powering hundreds of annotation projects over the years.
3/N
That summer, Taylor Berg-Kirkpatrick, Ashish Vaswani, and I built MACE (Multi-Annotator Competence Estimation).
2/N
That summer, Taylor Berg-Kirkpatrick, Ashish Vaswani, and I built MACE (Multi-Annotator Competence Estimation).
2/N
In my PhD, I had a side project to fix an annoying problem: when you ask 5 people to label the same thing, you often get different answers. But in ML (and lots of other analyses), you still need a single aggregated answer. Using the majority vote is easy–but often wrong.
1/N
In my PhD, I had a side project to fix an annoying problem: when you ask 5 people to label the same thing, you often get different answers. But in ML (and lots of other analyses), you still need a single aggregated answer. Using the majority vote is easy–but often wrong.
1/N
Join the MilaNLP team and contribute to our upcoming research projects (SALMON & TOLD)
🔗 Details + how to apply: milanlproc.github.io/open_positio...
⏰ Deadline: Jan 31, 2026
@pranav-nlp.bsky.social presented "You Cannot Sound Like GPT": Signs of language discrimination and resistance in computer science publishing.
Paper: arxiv.org/abs/2505.08127
#NLProc
@pranav-nlp.bsky.social presented "You Cannot Sound Like GPT": Signs of language discrimination and resistance in computer science publishing.
Paper: arxiv.org/abs/2505.08127
#NLProc
#NLProc
#NLProc
Join the MilaNLP team and contribute to our upcoming research projects (SALMON & TOLD)
🔗 Details + how to apply: milanlproc.github.io/open_positio...
⏰ Deadline: Jan 31, 2026
Join the MilaNLP team and contribute to our upcoming research projects (SALMON & TOLD)
🔗 Details + how to apply: milanlproc.github.io/open_positio...
⏰ Deadline: Jan 31, 2026
milanlproc.github.io/open_positio...
milanlproc.github.io/open_positio...
#NLP #multimodality #speech
#NLP #multimodality #speech
Today Henning Hoffmann presented the paper "Music for All: Representational Bias and Cross-Cultural Adaptability of Music Generation Models"
Paper: arxiv.org/pdf/2502.07328
#NLProc
Today Henning Hoffmann presented the paper "Music for All: Representational Bias and Cross-Cultural Adaptability of Music Generation Models"
Paper: arxiv.org/pdf/2502.07328
#NLProc
Paper: arxiv.org/pdf/2511.15304
#NLProc
#LLMs #jailbreaking
Paper: arxiv.org/pdf/2511.15304
#NLProc
#LLMs #jailbreaking
Join the MilaNLP team and contribute to our upcoming research projects.
🔗 More details: milanlproc.github.io/open_positio...
⏰ Deadline: Jan 31, 2026
Join the MilaNLP team and contribute to our upcoming research projects.
🔗 More details: milanlproc.github.io/open_positio...
⏰ Deadline: Jan 31, 2026
🗣️ @penzo-nicolo.bsky.social on multi-party conversations
🌍 @patriciachiril.bsky.social on NLP for socially grounded research
#NLProc
🗣️ @penzo-nicolo.bsky.social on multi-party conversations
🌍 @patriciachiril.bsky.social on NLP for socially grounded research
#NLProc
Lots to think about how we evaluate fairness in language models!
#NLProc #fairness #LLMs
Lots to think about how we evaluate fairness in language models!
#NLProc #fairness #LLMs
#NLProc
#NLProc