dbscan-python
dbscan-python copied to clipboard
Can the memory footprint be reduced?
With a 14178107*8 vector, a 108GB memory machine is quickly used up. Is there a way to reduce the memory footprint?
The train output:
Input: 14178107 points, dimension 8
scheduler = Parlay-HomeGrown
num-threads = 16
num-cell = 12333095
compute-grid = 5.06638
Hi, I couldn't think of a straight forward way to optimize the memory usage at the moment. I think the high memory usage in your case may be related to a relatively small eps value that you are using. Alternatively, you may also want to try machines with larger memory such as AWS EC2.