I\'m running hadoop job on many input files. But if one of the files is corrupted the whole job is fails.
How can I make the job to ignore the corrupted file? maybe writ
There is also another possible way. You could use mapred.max.map.failures.percent configuration option. Of course this way of solving this problem could also hide some other problems occurring during map phase.
mapred.max.map.failures.percent