🎙️ #speech #AI #NLP #recsys
🐶 Apollo’s human | 🛶 Rowing to 1M meters
🌐 https://mariateleki.github.io/
🤔 Traditional F1 scores hide why disfluency removal models succeed or fail.
Our survey shows they struggle with:
⚠️ Long-term coherence
⚠️ Controllability
📚 Paper: mariateleki.github.io/pdf/A_Survey...
#StoryGeneration #GenerativeAI #NLP
Our survey shows they struggle with:
⚠️ Long-term coherence
⚠️ Controllability
📚 Paper: mariateleki.github.io/pdf/A_Survey...
#StoryGeneration #GenerativeAI #NLP
In our #INTERSPEECH2024 paper, we looked at how Google ASR vs WhisperX handle messy, real-world podcasts (82k+ episodes!):
🎙️ WhisperX → better with “uh/um”
📝 Google ASR → better with self-corrections
In our #INTERSPEECH2024 paper, we looked at how Google ASR vs WhisperX handle messy, real-world podcasts (82k+ episodes!):
🎙️ WhisperX → better with “uh/um”
📝 Google ASR → better with self-corrections
#AcademicMentoring #PhDLife
#AcademicMentoring #PhDLife
That’s why I build evaluation frameworks for speech & conversational AI — so we can stress-test systems against real-world variability.
#AIResearch #Evaluation #SpeechProcessing
That’s why I build evaluation frameworks for speech & conversational AI — so we can stress-test systems against real-world variability.
#AIResearch #Evaluation #SpeechProcessing
Because speech ≠ text.
🎙️ People pause, restart, self-correct
🌎 Background noise & accents vary
💬 Context shifts across domains
Because speech ≠ text.
🎙️ People pause, restart, self-correct
🌎 Background noise & accents vary
💬 Context shifts across domains
“I want a horror -- comedy -- movie”? 🎥
That slip-of-the-tongue can confuse recommender systems.
Our INTERSPEECH 2025 paper shows some LLMs handle it better than others.
📄 mariateleki.github.io/pdf/HorrorCo...
#INTERSPEECH2025 #ConversationalAI #RecSys
“I want a horror -- comedy -- movie”? 🎥
That slip-of-the-tongue can confuse recommender systems.
Our INTERSPEECH 2025 paper shows some LLMs handle it better than others.
📄 mariateleki.github.io/pdf/HorrorCo...
#INTERSPEECH2025 #ConversationalAI #RecSys
But can AI tell a good story?
Our Survey on LLMs for Story Generation (EMNLP Findings 2025) explores:
✨ Coherence
🎛️ Controllability
🎨 Creativity
⚖️ Authenticity
📄 mariateleki.github.io/pdf/A_Survey...
#StoryGeneration #GenerativeAI
But can AI tell a good story?
Our Survey on LLMs for Story Generation (EMNLP Findings 2025) explores:
✨ Coherence
🎛️ Controllability
🎨 Creativity
⚖️ Authenticity
📄 mariateleki.github.io/pdf/A_Survey...
#StoryGeneration #GenerativeAI
We restart, repeat, and slip.
For AI, those little disfluencies can cause big problems.
That’s why my research builds methods to make spoken language systems more robust.
#SpeechProcessing #ConversationalAI #NLP #AI
We restart, repeat, and slip.
For AI, those little disfluencies can cause big problems.
That’s why my research builds methods to make spoken language systems more robust.
#SpeechProcessing #ConversationalAI #NLP #AI
📚 Covers: controllability, coherence, and creativity
🧩 Discusses: evaluation challenges
🌍 Highlights: hybrid symbolic–neural approaches
💻 Includes: an open resource list (PRs welcome!)
📚 Covers: controllability, coherence, and creativity
🧩 Discusses: evaluation challenges
🌍 Highlights: hybrid symbolic–neural approaches
💻 Includes: an open resource list (PRs welcome!)
🤔 Traditional F1 scores hide why disfluency removal models succeed or fail.
🤔 Traditional F1 scores hide why disfluency removal models succeed or fail.
In our INTERSPEECH 2025 paper, we introduced Syn-WSSE, a psycholinguistically grounded framework for simulating whole-word substitution errors in conversational recommenders (e.g., “I want a horror—comedy movie”).
In our INTERSPEECH 2025 paper, we introduced Syn-WSSE, a psycholinguistically grounded framework for simulating whole-word substitution errors in conversational recommenders (e.g., “I want a horror—comedy movie”).
In our #INTERSPEECH2024 paper, we compared Google ASR vs WhisperX on 82k+ podcasts 🎙️
🌱 WhisperX → better with accurately transcribing “uh/um”
🌱 Google ASR → better with accurately transcribing edited nodes
🌱 Which to use? Depends on your data.
In our #INTERSPEECH2024 paper, we compared Google ASR vs WhisperX on 82k+ podcasts 🎙️
🌱 WhisperX → better with accurately transcribing “uh/um”
🌱 Google ASR → better with accurately transcribing edited nodes
🌱 Which to use? Depends on your data.
In our INTERSPEECH 2025 paper, we studied whole-word substitution errors. 🧵
In our INTERSPEECH 2025 paper, we studied whole-word substitution errors. 🧵
But bias also hides in how we speak. Our ICWSM 2025 paper showed that men’s discourse markers (“going,” “well”) are treated as more “stable” in LLM embeddings than women’s (“like,” “really”). 🧵
But bias also hides in how we speak. Our ICWSM 2025 paper showed that men’s discourse markers (“going,” “well”) are treated as more “stable” in LLM embeddings than women’s (“like,” “really”). 🧵
We’re also releasing a community resource on GitHub — please feel free to send a pull request as new systems come out.
Paper: tinyurl.com/3jmdkx72
Github: tinyurl.com/2hmrkvrt
#NLP #LLMs #EMNLP2025
We’re also releasing a community resource on GitHub — please feel free to send a pull request as new systems come out.
Paper: tinyurl.com/3jmdkx72
Github: tinyurl.com/2hmrkvrt
#NLP #LLMs #EMNLP2025