how to prevent hadoop job to fail on corrupted input file

后端 未结 3 1842
轮回少年
轮回少年 2021-02-09 02:30

I\'m running hadoop job on many input files. But if one of the files is corrupted the whole job is fails.

How can I make the job to ignore the corrupted file? maybe writ

3条回答
  •  独厮守ぢ
    2021-02-09 03:00

    There is also another possible way. You could use mapred.max.map.failures.percent configuration option. Of course this way of solving this problem could also hide some other problems occurring during map phase.

提交回复
热议问题