I want to use MapReduce to filter a huge dataset for rare entities satisfying some criteria. I could speed this up a lot by terminating reducers once they violate the criteria, since they will be computing on entities that I'm not interested in.
To make up an example, say I have a corpus with billions of articles, and I want to return only articles with fewer than 100 words. The vast majority of articles have >100,000 words, so I can skip most of the work by terminating the reducers once it reaches the stopping criteria (word_count >100).