I repeatedly hear this question. I don't have any hands-on these scenarios. As I can learn and get many approaches/ideas to do the same, so would like to understand
1) What would be the best approach ?
2) What would be the efficient way to do this ?
According to me, I would approach to break down the huge file size to smaller files (I mean, Batches
).
Let's say, I have 2 files with data to be manipulated (Each file sorted
and un-sorted
order).
Definitely, reading such huge file results to memory error (won't be able to load file depending upon RAM).
1) How can it be achieved through Python ?
2) Best time saving and efficient method ?
3) Can Python-Pandas achieve this ? If yes, how ?
Very curious to hear from you.Please help me.