As the question states, is there a way to estimate the amount of time it will take to write a spark dataframe to a file such as a parquet? I don't like waiting indefinitely knowing that I have the best instance to handle the task and it's already been over an hour.
So please, if anyone knows a way to optimize this and/or get a good estimate of how long it will take please post your answer below.