I have a map-reduce job whose input is a big data set (let's say of size 100GB). What this map-reduce job does is splitting the big data into chunks and writing separate files, one per each data chunk. That is, the output of the job is multiple files, each of size 64MB.
The output of this map-reduce job is used as the input for another map-reduce job. As the new input consists of multiple files, each of size 64MB, does each mapper in the second map-reduce job reads only one file or it might reads more than 1 file?