Magdalena Fuentes

Results 68 comments of Magdalena Fuentes

Hey @aahlenst ! Thanks for taking the initiative on this, we're happy to help. Also, thanks @faroit for pointing this out, agree we should discuss on best ways of both...

@PRamoneda and @genisplaja also commented that tests in big datasets are slow (with `test_full_dataset`). They mentioned the idea that we could parallelize them to save some time. @rabitt thoughts?

We discussed offline with @lostanlen that we should write down somewhere the version policy when moving the indexes to Zenodo: - what happens in the future with a new release...

@lostanlen pointed out a fair question, how to include index in zenodo when someone is contributing with a new loader? Are we going to do it ourselves manually? Is it...

Give me a couple of days to check if it's worth starting from a backup version

Closed via #335. Because of size limits of GitHub LFS we decided to move big indexes to zenodo.

I like this idea! What if we make `cleanup` dependent on that boolean, and if we're not checking `checksums` by default when we download we delete the compressed files?

We'll embed the `mir-datasets.yaml` file on the mirdata repo, and add a column that indicates if the dataset has a `mirdata` loader or not. Also as part of the checklist...

This didn't go to far at the end. Thoughts on this moving forward? @rabitt @nkundiushuti

Missing entries now are: Beatport EDM key, cante100, Mridangam Stroke, Saraga. Small inconsistencies: GiantSteps tempo has audio available though in the list it says it doesn't.