I have a dataframe with 120 millions of records and when i try to write it, takes 45 minutes. The DF is partitioned by a field "id_date" which is in this format: yyyymmdd. This DF is a delta table in databricks. I try with autooptimize, compact, etc from delta tables properties but i don´t get any advance. Sorry for my bad english. Anyone can help me? Thanks and regards
I am trying to write a df with many records (more than 100 million) and it takes me a long time. I need to optimize this process.