multiMiR
multiMiR copied to clipboard
Handle massive queries
I am trying to perform massive analysis of miRNA targets and I have found your tool very usefull. However, since I am working in a computation cluster, my cluster have limited time to execute at login (This is the only node wit access to internet) and your tool download the queries and format all data to multiMiR format for each database in the same job, I have a lot of troubles to make big queries or complete all the databases. I have tryed to split big queries to many small chunks, but the attempt was not successful. Can you recommend me any solution for my problem? I suggest you to divide get_multimir function in diferent modes: i) query all databases and save the raw data to a temporal folder ii) give to get_multimir the temporal folder and the function format this data to multiMiR layout instead of download the data. In this way, I can download the without the CPU workload and then process all the data in my cluster without interruption issues. Also, it would be a great feature, an user can make a generic query and download it. Then, make more specific queries onto the downloaded data tuning the parameters without stressing the databases APIs queried by your package so the user would identify faster its ideal setting.