Improvement of Speaker Diarization pipeline performance for Indian languages #1856
Unanswered
mukherjeesougata
asked this question in
Q&A
Replies: 2 comments
-
@hbredin can you kindly please respond to this query? |
Beta Was this translation helpful? Give feedback.
0 replies
-
@mogwai can you kindly please reply to this query? |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I am using the Emilia-Pipe pipeline to process in-the-wild raw audio containing Indian languages and extract clean speech audio with segregated speakers. This pipeline utilizes the PyAnnote Speaker Diarization 3.1 pretrained model from Hugging Face, accessed via Hugging Face tokens. However, in many cases, the pipeline fails to segregate speakers accurately. I have a few queries regarding this issue:
Looking forward to your guidance. Thank you!
Beta Was this translation helpful? Give feedback.
All reactions