Fork me on GitHub
#data-science
<
2021-11-26
>
skuro13:11:14

hey hey. I'm using tablecloth for some data analysis and given the sheer size of my dataset I see that keeping everything in memory at once is fairly demanding (currently "just" 16G but it's bound to be more memory hungry in the future)

skuro13:11:40

is there any way I can limit or otherwise optimize memory usage?

otfrom14:11:20

you can lazily process the datasets if that works. Converting to the right types helps as well

otfrom14:11:43

some of the lazy and map/reduce (on one machine) style processing is in tech.ml.dataset