My program use a DistributedCache to cache files
JobConf conf = new JobConf(new Configuration(), ItemMining.class);
DistributedCache.addCacheFile(new URI("output1/FList.txt"), conf);
DistributedCache.addCacheFile(new URI("output1/GList.txt"), conf);
I get the files in
configure(){
..
localFiles = DistributedCache.getLocalCacheFiles(job);
FileSystem fs = FileSystem.get(job);
FSDataInputStream inF = fs.open(localFiles[0]);
..
}
The whole program can be run and get the right result on Eclipse. But when I run it in Hadoop cluster, I find that this part is not be called! Why does this happen? Do I need to set something in configuration?