2

I have a simple mapreduce job , where in for some of keys , number of values are millions in number . As a result of which reducer is not able to finish . I have gone through this link Hadoop handling data skew in reducer but not able to follow if there is any best practice available for such kind of scenarios. Can anyone please suggest the best way to handle such cases in mapreduce job ?

Community
  • 1
  • 1
KBR
  • 464
  • 1
  • 7
  • 24

0 Answers0