Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

These voice can not split correctly #1524

Closed
lucasjinreal opened this issue Nov 2, 2023 · 4 comments
Closed

These voice can not split correctly #1524

lucasjinreal opened this issue Nov 2, 2023 · 4 comments

Comments

@lucasjinreal
Copy link

asr_res_240-243_1_audio.zip

the output time and label

start=0.0s stop=2.4s speaker_SPEAKER_00
start=0.4s stop=1.4s speaker_SPEAKER_01

but the speaker are clearly 2 speaker first and later, how to precisely get the splitter time in the middle?

Copy link

github-actions bot commented Nov 2, 2023

Thank you for your issue.You might want to check the FAQ if you haven't done so already.

Feel free to close this issue if you found an answer in the FAQ.

If your issue is a feature request, please read this first and update your request accordingly, if needed.

If your issue is a bug report, please provide a minimum reproducible example as a link to a self-contained Google Colab notebook containing everthing needed to reproduce the bug:

  • installation
  • data preparation
  • model download
  • etc.

Providing an MRE will increase your chance of getting an answer from the community (either maintainers or other power users).

Companies relying on pyannote.audio in production may contact me via email regarding:

  • paid scientific consulting around speaker diarization and speech processing in general;
  • custom models and tailored features (via the local tech transfer office).

This is an automated reply, generated by FAQtory

@hbredin
Copy link
Member

hbredin commented Nov 2, 2023

Without providing details about the code you tried, it is kind of difficult to tell.
Here are my 2 cents applying the pretrained pyannote/segmentation-3.0 model: it looks like it does manage to do the job...

AUDIO = "asr_res_240-243_1_audio.mp3"

from pyannote.audio import Audio
io = Audio(mono="downmix", sample_rate=16000)
waveform, sample_rate = io(AUDIO)
audio = {"waveform": waveform, "sample_rate": sample_rate}

from pyannote.audio import Inference
inference = Inference("pyannote/segmentation-3.0", window="whole")
prediction = inference(audio)

from matplotlib import pyplot as plt
plt.plot(prediction)
plt.legend(['speaker#1', 'speaker#2', 'speaker#3'])

output

@lucasjinreal
Copy link
Author

@hbredin hi, the audio acutially only have 2 people, the first period is person1, and rest is a man voice.

the cliff of speaker3 seems detected the later man voice, but how can i tell, (i actually just need split 2 person), this cliff is exactly what I want?

Copy link

stale bot commented May 2, 2024

This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions.

@stale stale bot added the wontfix label May 2, 2024
@stale stale bot closed this as completed Jun 1, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants