Marvin Lavechin
@marvinlavechin.bsky.social
40 followers 65 following 16 posts
Machine learning, speech processing, language acquisition and cognition. Soon @cnrs.fr @univ-amu.fr; currently postdoc at MIT, Cambridge, US.
Posts Media Videos Starter Packs
Pinned
marvinlavechin.bsky.social
Glad to share this new study comparing the performance and biases of the LENA and ACLEW algorithms in analyzing language environments in Down, Fragile X, Angelman syndromes, and populations at elevated likelihood for autism 👶📄
osf.io/preprints/ps...
🧵1/12
OSF
osf.io
Reposted by Marvin Lavechin
thueber.bsky.social
🚀 I’m excited to announce the launch of our new research chair DevAI&Speech (2025–2029), funded by the Grenoble AI Institute MIAI Cluster IA!

The project explores how human developmental processes can inspire more grounded and socially aware conversational AI (1/6).
Reposted by Marvin Lavechin
carorowland.bsky.social
Children are incredible language learning machines. But how do they do it? Our latest paper, just published in TICS, synthesizes decades of evidence to propose four components that must be built into any theory of how children learn language. 1/
www.cell.com/trends/cogni... @mpi-nl.bsky.social
Constructing language: a framework for explaining acquisition
Explaining how children build a language system is a central goal of research in language acquisition, with broad implications for language evolution, adult language processing, and artificial intelli...
www.cell.com
Reposted by Marvin Lavechin
nsaphra.bsky.social
Next we jump from analyzing text models to predictive speech models! Phoneticists have claimed for decades that humans rely more on contextual cues when processing vowels compared to consonants. Turns out so do speech models!
Reposted by Marvin Lavechin
infantstudies.bsky.social
Don't miss our next ICIS webinar! June 19 2025.
Join leading researchers for a deep dive into cutting-edge work in infancy research.
infantstudies.org/icis-online-...
#InfantResearch #InfantStudies
Text: ICIS Webinars - Thursday, June 19 2025 - Beyond Who’s Speaking When: Machine Learning Tools to Extract Rich Multi-Dimensional Features from Home Audio Recordings
marvinlavechin.bsky.social
A great opportunity to learn how speech technology can advance research on how children learn language (and vice versa)

You know, beyond surveillance and chatbots 🙊
maureendeseyssel.bsky.social
Now that @interspeech.bsky.social registration is open, time for some shameless promo!

Sign-up and join our Interspeech tutorial: Speech Technology Meets Early Language Acquisition: How Interdisciplinary Efforts Benefit Both Fields. 🗣️👶

www.interspeech2025.org/tutorials

⬇️ (1/2)
https://www.interspeech2025.org/tutorials
Your cookies are disabled, please enable them.
www.interspeech2025.org
marvinlavechin.bsky.social
👀
maureendeseyssel.bsky.social
New preprint out! 👇

We adapt the ABX task, commonly used in speech models, to investigate how multilingual text models represent form (language) vs content (meaning).

📄 arxiv.org/pdf/2505.17747

🙌 With Jie Chi, Skyler Seto, @maartjeterhoeve.bsky.social, Masha Fedzechkina & Natalie Schluter
arxiv.org
Reposted by Marvin Lavechin
cardiffbabylab.bsky.social
Interested in collecting and processing naturalistic audio data using new AI tools?

If so, join us for our free, two-day workshop; 'Long Form Audio Recordings: A to Z', generously supported by Cardiff University Doctoral Academy.
marvinlavechin.bsky.social
@mpoli.fr check this out if you haven't read it yet! Really cool work!
Reposted by Marvin Lavechin
rtommccoy.bsky.social
🤖🧠 Paper out in Nature Communications! 🧠🤖

Bayesian models can learn rapidly. Neural networks can handle messy, naturalistic data. How can we combine these strengths?

Our answer: Use meta-learning to distill Bayesian priors into a neural network!

www.nature.com/articles/s41...

1/n
A schematic of our method. On the left are shown Bayesian inference (visualized using Bayes’ rule and a portrait of the Reverend Bayes) and neural networks (visualized as a weight matrix). Then, an arrow labeled “meta-learning” combines Bayesian inference and neural networks into a “prior-trained neural network”, described as a neural network that has the priors of a Bayesian model – visualized as the same portrait of Reverend Bayes but made out of numbers. Finally, an arrow labeled “learning” goes from the prior-trained neural network to two examples of what it can learn: formal languages (visualized with a finite-state automaton) and aspects of English syntax (visualized with a parse tree for the sentence “colorless green ideas sleep furiously”).
Reposted by Marvin Lavechin
mcxfrank.bsky.social
Super excited to submit a big sabbatical project this year: "Continuous developmental changes in word
recognition support language learning across early
childhood": osf.io/preprints/ps...
title of paper (in text) plus author list Time course of word recognition for kids at different ages.
Reposted by Marvin Lavechin
marvinlavechin.bsky.social
Truly amazing collab. with:
Lisa Hamrick
Bridgette Kelleher @drbkelleher.bsky.social
Amanda Seidl

12/12
marvinlavechin.bsky.social
🔬 2) For researchers studying language development across neurodevelopmental conditions: these tools seem to perform reliably across diverse populations.

Cross-population differences found by the algorithms likely reflect real differences, not artifacts of the algo.

11/12
marvinlavechin.bsky.social
⁉️ What does all of this mean?

1) We now understand better each algorithm's strengths and weaknesses.

LENA excels at precise speaker classification with fewer false alarms (good for acoustic analyses), while ACLEW better captures the full range of language interactions.

10/12
marvinlavechin.bsky.social
🧩 Performance variations were driven more by participants' speech patterns than by diagnostic groups, with the amount of other children's vocalizations, female and male adult speech predicting the performance of both algorithms.

9/12
marvinlavechin.bsky.social
⚖️ Despite being trained exclusively on typically-developing children, both algorithms maintained consistent performance across all diagnostic groups (low-risk, Angelman, fragile X, Down, siblings of children with ASD)!

8/12
marvinlavechin.bsky.social
⚖️ What about automatic counts: CTC, AWC, CVC?
Both algorithms capture a large portion of variance in human counts (Pearson's r from .78 to .92)

ACLEW seems better on CTC (.92 vs .83)
LENA is slightly better on AWC (.82 vs .78)
ACLEW is slightly better on CVC (.88 vs .83)

7/12
marvinlavechin.bsky.social
⚖️ ACLEW, by comparison, correctly classified 69 hours out of those same 100 hours, missing only about 15 hours, confused the speaker category for 15 hours, but generated 99 hours of false alarms.

ACLEW makes a lot of mistakes but retrieved most of the speech.

6/12
marvinlavechin.bsky.social
⚖️ We found very different segmentation strategies.
For every 100 hours of speech, LENA correctly classified 45 hours, missed 41 hours, generated 27 hours of false alarms, and confused the speaker category for 14 hours.

LENA makes few mistakes but misses a lot of speech.

5/12
marvinlavechin.bsky.social
📝 In this study, we annotated 25 hours of audio from 50 children across various neurodevelopmental profiles.

We ask two questions:
1) How do LENA and ACLEW compare across their key performance metrics?
2) Do these algorithms have lower performance for certain populations?

4/12
marvinlavechin.bsky.social
🤖 Both algorithms start by segmenting the audio into speaker categories: key child, other children, male and female adult. From this segmentation step, they extract key metrics: Conversational Turn Count (CTC), Adult Word Count (AWC), and Child Vocalization Count (CVC).

3/12
marvinlavechin.bsky.social
🎙️Wearables are revolutionizing early language research! These devices + AI/speech tech help analyze vast amounts of audio from children's daily lives. Two main algorithms are used to analyze child-centered daylong recordings: the proprietary LENA and open-source ACLEW.

2/12
marvinlavechin.bsky.social
Glad to share this new study comparing the performance and biases of the LENA and ACLEW algorithms in analyzing language environments in Down, Fragile X, Angelman syndromes, and populations at elevated likelihood for autism 👶📄
osf.io/preprints/ps...
🧵1/12
OSF
osf.io