ark icon indicating copy to clipboard operation
ark copied to clipboard

Determine minimum viable test data set

Open TimothyStiles opened this issue 1 year ago • 3 comments

Our minimum viable test set should be small enough to allow local development but complete (and probably large enough) to do some actually meaningful work. Thought is to perhaps keep just a large collection of b. sub strains or some model organism and their affiliated genes/proteins? rhea, chembl, and reactome.org should all be small enough that we can just keep their relevant data in the test set.

Right now the scraper is not sophisticated enough to just take b. Sub genomes. It simply just downloads all of genbank / uniprots data dumps.

TimothyStiles avatar Apr 04 '23 19:04 TimothyStiles

We should have something in the range of 100-300MB. It will otherwise clog up everyone's LFS quotas when they fork it. Additionally, if we can avoid LFS, we should that way you won't have any issues if someone forgets to check-in LFS objects.

rkrishnasanka avatar Apr 04 '23 20:04 rkrishnasanka

git lfs is nightmare <- Krishna

minimum viable dataset <- most interesting reaction Isaac isaac 👇 protein -> metabolic pathway -> map against genome

TimothyStiles avatar Apr 04 '23 20:04 TimothyStiles

checking out @rkrishnasanka's minimal viable dataset on his fork.

TimothyStiles avatar Apr 04 '23 20:04 TimothyStiles