Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to use AMI dataset to evaluate the DER performance? #12

Open
BLack-yzf opened this issue Jun 4, 2020 · 5 comments
Open

How to use AMI dataset to evaluate the DER performance? #12

BLack-yzf opened this issue Jun 4, 2020 · 5 comments

Comments

@BLack-yzf
Copy link

Hi, I have seen some authors use AMI corpus to make evaluation on diarization task. But there is no more details about how to evaluate specifically. Like how to choose the dev and test part of the AMI,and how to make the corresponding data preparation.
Is there any guidance about using AMI corpus to evaluate the task?
Thanks.
@wq2012

@hbredin
Copy link
Contributor

hbredin commented Jun 4, 2020

AMI website provides an official train/dev/test split.

For data preparation, I did most of the work in a pyannote.audio tutorial already.

@BLack-yzf
Copy link
Author

Thanks a lot!

@BLack-yzf
Copy link
Author

Hi,hbredin.
I see your repo has provided the files "MixHeadset.train.rttm" and "MixHeadset.train.uem" for Mix-headset. Is there rttm and uem files for {headset-0,headset-1,headset-2,headset-3}.
Thanks. @hbredin

@hbredin
Copy link
Contributor

hbredin commented Jun 8, 2020

I don't know. I am not the creator of the AMI corpus.
Check the official AMI website.

@BLack-yzf
Copy link
Author

Hi hbredin,
Your work on AMI data preparation has helped me a lot. Recently, I have met some problems.I hope you can give me some advice.
I have download the Mix-headset data and the corresponding rttm files you have provided.
I prepare to use the kaldi/ami recipe to make the data preparation. But I can't get the "segments" files. So, I try to use the provided rttm files to produce file "segments". When performing the "md-eval.pl" to make rttms, it causes the error that " Speaker IS1008d.Mix-Headset 1 108.664 -1.032 ". I have found that some segments overlapped. So, I think the initial process of producing file "segments" from provided rttm files is wrong. Can you provide the right "segment" files for Mix-headset data(train, dev and test)?
Hope your reply, thanks.
@hbredin

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants