Visiting student @ CLSP Johns Hopkins University
GitHub: https://github.com/domklement
LinkedIN: https://www.linkedin.com/in/dominik-klement/
We’re open to feedback, discussions, and collaborations. Let’s work together to shape the future of ASR and diarization technology!
[14/14]
We’re open to feedback, discussions, and collaborations. Let’s work together to shape the future of ASR and diarization technology!
[14/14]
Thanks to Alon Vinnikov, Amir Ivry, Eyal Krupka (Microsoft) for organizing the CHiME-8 NOTSOFAR-1 Challenge, and to the CHiME-8 Steering Committee for their dedication to advancing speech recognition research!
[13/14]
Thanks to Alon Vinnikov, Amir Ivry, Eyal Krupka (Microsoft) for organizing the CHiME-8 NOTSOFAR-1 Challenge, and to the CHiME-8 Steering Committee for their dedication to advancing speech recognition research!
[13/14]
[12/14]
[12/14]
[11/14]
[11/14]
[10/14]
[10/14]
We’re making our research accessible by open-sourcing training and inference codebases, and providing interactive demos:
🔗DiariZen Source Code github.com/BUTSpeechFIT...
[9/14]
We’re making our research accessible by open-sourcing training and inference codebases, and providing interactive demos:
🔗DiariZen Source Code github.com/BUTSpeechFIT...
[9/14]
4. Leveraging Self-Supervised Learning for Speaker Diarization - Accepted to ICASSP 2025. This paper introduces DiariZen - our state-of-the-art diarization model and toolkit.
arxiv.org/abs/2409.09408
[8/14]
4. Leveraging Self-Supervised Learning for Speaker Diarization - Accepted to ICASSP 2025. This paper introduces DiariZen - our state-of-the-art diarization model and toolkit.
arxiv.org/abs/2409.09408
[8/14]
3. BUT/JHU System Description for CHiME-8 NOTSOFAR-1 Challenge - The work earned the 🏆Jury Prize for being one of the most practical, efficient, and novel systems. Our robust diarization-ASR integration is capable of tackling overlapped speech.
isca-archive.org/chime_2024/p...
[7/14]
3. BUT/JHU System Description for CHiME-8 NOTSOFAR-1 Challenge - The work earned the 🏆Jury Prize for being one of the most practical, efficient, and novel systems. Our robust diarization-ASR integration is capable of tackling overlapped speech.
isca-archive.org/chime_2024/p...
[7/14]
2. Target Speaker ASR with Whisper arxiv.org/abs/2409.09543 - Accepted to ICASSP 2025. This work enhances the Whisper ASR model for target-speaker recognition, demonstrating its applicability in complex acoustic scenarios.
[6/14]
2. Target Speaker ASR with Whisper arxiv.org/abs/2409.09543 - Accepted to ICASSP 2025. This work enhances the Whisper ASR model for target-speaker recognition, demonstrating its applicability in complex acoustic scenarios.
[6/14]
[5/14]
[5/14]
1. DiCoW: Diarization-Conditioned Whisper for Target Speaker Automatic Speech Recognition - Submitted to CSL. Our diarization-conditioned approach that eliminates the need for speaker enrollment or source separation.
arxiv.org/abs/2501.00114
[4/14]
1. DiCoW: Diarization-Conditioned Whisper for Target Speaker Automatic Speech Recognition - Submitted to CSL. Our diarization-conditioned approach that eliminates the need for speaker enrollment or source separation.
arxiv.org/abs/2501.00114
[4/14]
[3/14]
[3/14]
- Simplifying Multi-Speaker ASR: Our models directly use diarization outputs as conditioning signals, bypassing the need for enrollment data or complex source separation techniques.
[2/14]
- Simplifying Multi-Speaker ASR: Our models directly use diarization outputs as conditioning signals, bypassing the need for enrollment data or complex source separation techniques.
[2/14]