Dag7
Dag7
Using the description field, unless we're interested in preserving the compos too (like what pouet.org does). In the second case, I would find out first a way to represent compos...
It's full of random stuff, without metadata, without guarantees of provenance. In case we import it in bulk we run the risk of putting everything there randomly, while the idea...
since form response could be saved as a csv it should be easy. This would close #231 too.
Indeed a cool idea, but are we sure that we really want to do this? I mean, of course we are, but we can't do this automatically. Every link is...
> > > Issues labeled with "sources" aren't meant to be able to be scraper per-se. They are just a source we should add to the repository. I can create...
I really don't understand here what we need to scrape / download. Everything is in Japanese, which is not readable to me, and from the message that have been quoted...
### Updates: I manually checked every file; sadly there are two big, giant, homebrew packs containing a total number of 1185 files . After tweaking a little bit the `dupe_scraper.py`,...
> @dag7dev everything that you mentioned is reproducible somewhere? or was it done manually? Everything has been done manually. If needed, as long as I will have some free time,...
>I don't it's a good idea to flood the database with those two huge compilations if we don't have enough information for the manifests and to correctly spot duplicates. Why...
> For the ROM compilations, we should do it the other way round: look for ROM/game names we don't have yet in our database (e.g. names/slugs/filenames that have no similarities...