[CONTRIBUTION] Speech Dataset Generator #3604
davidmartinrius
started this conversation in
Show and tell
Replies: 1 comment
-
|
Recently, I have also implemented automatic downloads from sources such as YouTube, Librivox, and TED Talks. Additionally, speakers are now stored in a chroma vector database, and the system automatically detects them. There is no need to manually assign a name to each speaker anymore; the system handles it for you. It convert the voices to embeddings, and then each audio is compared using cosine similarity to determine if a speaker is the same as another. This allows you to handle extensive data sets, and the labeling process will be automated. I updated the first message in this thread with the latest features. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi everyone!
I have just published this project on GitHub: https://github.com/davidmartinrius/speech-dataset-generator/
Now you can create datasets automatically with any audio or lists of audios.
I hope you can find it useful.
Here are the key functionalities of the project:
Dataset Generation: Creation of multilingual datasets with Mean Opinion Score (MOS).
Silence Removal: It includes a feature to remove silences from audio files, enhancing the overall quality.
Sound Quality Improvement: It improves the quality of the audio when needed.
Audio Segmentation: It can segment audio files within specified second ranges.
Transcription: The project transcribes the segmented audio, providing a textual representation.
Gender Identification: It identifies the gender of each speaker in the audio.
Pyannote Embeddings: Utilizes pyannote embeddings for speaker detection across multiple audio files.
Automatic Speaker Naming: Automatically assigns names to speakers detected in multiple audios.
Multiple Speaker Detection: Capable of detecting multiple speakers within each audio file.
Store speaker embeddings: The speakers are detected and stored in a Chroma database, so you do not need to assign a speaker name.
Syllabic and words-per-minute metrics
Multiple input sources: You can either use your own files or download content by pasting URLs from sources such as YouTube, LibriVox and TED Talks.
Feel free to explore the project at https://github.com/davidmartinrius/speech-dataset-generator
Beta Was this translation helpful? Give feedback.
All reactions