I am very new to hadoop , so please excuse the dumb questions.
I have the following knowledge Best usecase of Hadoop is large files thus helping in efficiency while
Flume writes to HDFS by means of HDFS sink. When Flume starts and begins to receive events, the sink opens new file and writes events into it. At some point previously opened file should be closed, and until then data in the current block being written is not visible to other redaers.
As described in the documentation, Flume HDFS sink has several file closing strategies:
rollInterval
option)rollSize
option)rollCount
option)idleTimeout
option)So, to your questions:
a) Flume writes events to currently opened file until it is closed (and new file opened).
b) Append is allowed in HDFS, but Flume does not use it. After file is closed, Flume does not append to it any data.
c) To hide currently opened file from mapreduce application use inUsePrefix
option - all files with name that starts with .
is not visible to MR jobs.