in this blog https://blogs.aws.amazon.com/bigdata/post/Tx2ANLN1PGELDJU/Best-Practices-for-Micro-Batch-Loading-on-Amazon-Redshift
I fail to see what is the main point of the load stream approach. Where does the saving come from? I understand about turning off Analyze, Sort, Vacuum, using fixed compression and all that. But why the separate tracks or empty temporary tables? You can do all this without the multiple tracks. Is there some meaning in the phrase "an optimal load path for COPY to follow" What exactly is optimal? the S3 prefix or something else? whatever is optimal, Why would it not be optimal if you load directory into the final schema? Are we assuming that reading s3 files is the bottleneck?
Could someone provide some insight?