I was downloading twitter data using flume into hdfs. Though i have data more than 2 GB, my chunk file splits are less than 64MB. i.e. 1st file with 300KB,2nd File - 566 KB. Why is that happening?
Asked
Active
Viewed 53 times
1 Answers
0
That is because of your flume configuration.
read this, you will have to set hdfs.rollInterval or hdfs.rollSize

vishnu viswanath
- 3,794
- 2
- 36
- 47