Kwanghee Choi
@juice500ml.bsky.social
๐ค 124
๐ฅ 134
๐ 20
Master's student
@ltiatcmu.bsky.social
, working on speech AI at
@shinjiw.bsky.social
reposted by
Kwanghee Choi
Marianne de Heer Kloots
about 2 months ago
Had such a great time presenting our tutorial on Interpretability Techniques for Speech Models at
#Interspeech2025
! ๐ For anyone looking for an introduction to the topic, we've now uploaded all materials to the website:
interpretingdl.github.io/speech-inter...
add a skeleton here at some point
2
39
15
Can we make discrete speech units lightweight๐ชถ and streamable๐? Excited to share our new
#Interspeech2025
paper: On-device Streaming Discrete Speech Units
arxiv.org/abs/2506.01845
(1/n)
about 2 months ago
2
1
1
www.nature.com/articles/350...
Ted Chiang. Catching crumbs from the table. Nature 405, 517 (2000). My favorite sci-fi short, which surprisingly well-summarizes what I actually do nowadays. I bet self-supervised speech models contain undiscovered theories on phonetics and phonology.
loading . . .
Catching crumbs from the table - Nature
In the face of metahuman science, humans have become metascientists.
https://www.nature.com/articles/35014679
4 months ago
0
3
0
reposted by
Kwanghee Choi
Daniel Csillag @ ICML2025
5 months ago
It's good to finally have a good reference for this stuff! Kudos to the authors.
arxiv.org/abs/2501.18374
loading . . .
Proofs for Folklore Theorems on the Radon-Nikodym Derivative
In this paper, rigorous statements and formal proofs are presented for both foundational and advanced folklore theorems on the Radon-Nikodym derivative. The cases of conditional and marginal probabili...
https://arxiv.org/abs/2501.18374
0
4
3
Can self-supervised models ๐ค understand allophony ๐ฃ? Excited to share my new
#NAACL2025
paper: Leveraging Allophony in Self-Supervised Speech Models for Atypical Pronunciation Assessment
arxiv.org/abs/2502.07029
(1/n)
5 months ago
2
15
10
reposted by
Kwanghee Choi
7 months ago
New
#NAACL2025
demo, Excited to introduce ESPnet-SDS, a new open-source toolkit for building unified web interfaces for both cascaded & end-to-end spoken dialogue system, providing real-time evaluation, and more! ๐:
arxiv.org/abs/2503.08533
Live Demo:
huggingface.co/spaces/Siddh...
1
7
5
reposted by
Kwanghee Choi
Dave Levitan
8 months ago
More from inside NIH: Per a source with knowledge, for all internal research (of which there is like $10 billion worth or so), ALL purchasing shut down as of yesterday. That means gloves, reagents, anything involved with lab work, which means a lot of that work will stop.
88
2637
1512
reposted by
Kwanghee Choi
Language Technologies Institute | CMU
9 months ago
Are you a pre-doctoral student interested in language technologies, especially focusing on safe, fair and inclusive AI? Our Summer 2025 Language Technology for All Internship could be a great fit. See the link below for more info, and to apply:
lti.cs.cmu.edu/news-and-eve...
loading . . .
CMU LTI Language Technology for All Internship 2025 - Language Technologies Institute - School of Computer Science - Carnegie Mellon University
The LTI is currently seeking applicants for the summer 2025 Language Technology for All Internship
https://lti.cs.cmu.edu/news-and-events/news/2025-01-06-call-for-interns.html
2
16
13
reposted by
Kwanghee Choi
Badr M. Abdullah, PhD
10 months ago
๐ฃ
#SpeechTech
&
#SpeechScience
people We are organizing a special session at
#Interspeech2025
on: Interpretability in Audio & Speech Technology Check out the special session website:
sites.google.com/view/intersp...
Paper submission deadline ๐ 12 February 2025
1
16
10
reposted by
Kwanghee Choi
Shinji Watanabe
10 months ago
We are excited to announce the launch of ML SUPERB 2.0 (
multilingual.superbbenchmark.org
) as part of the Interspeech 2024 official challenge! We hope this upgraded version of ML SUPERB advances universal access to speech processing worldwide. Please join it!
#Interspeech2025
1
20
10
you reached the end!!
feeds!
log in