I've got a code that read data from many excel files in one folder. Here is the line responsible for reading data:
d[file] = pd.read_excel(filenames[file], nrows=ilosc_wierszy, usecols=range(kol_odp+1)).fillna("")
Some files are 1 601 KB big when the others are just 21 KB. I don't know why is that since the original file should be the same. Nevertheless on every file there is the same data. So how does it happen that python reads bigger files like 30 seconds and smaller in less than a second? In the line above I specified that I need just nrows and usecols so I thought that python should read only that and go to the next file. Why does it happen that it take so long and is there a way to make this faster?