How hadoop io.file.buffer.size works?
Hi all,
I want to know how the parameter io.file.buffer.size works in hadoop?
My understanding is it's related with file read/write.
Whatever I have found, is that -------
In LineReader.java,it's used as the default buffer size for each line.
In BlockReader.newBlockReader(), it's used as the internal buffer size of the
BufferedInputStream. Also, in compression related classes, it's used as
default buffer size.
I want to know how this buffer size affects the hadoop file system and it's performance?
And it's internal working?
Help needed..
Thanks
I want to know how the parameter io.file.buffer.size works in hadoop?
My understanding is it's related with file read/write.
Whatever I have found, is that -------
In LineReader.java,it's used as the default buffer size for each line.
In BlockReader.newBlockReader(), it's used as the internal buffer size of the
BufferedInputStream. Also, in compression related classes, it's used as
default buffer size.
I want to know how this buffer size affects the hadoop file system and it's performance?
And it's internal working?
Help needed..
Thanks
Comments
sam E. likes this
2 comments
Yes, it has significantly effect on performance, too low or too high cause much performance issue
it normally good have 32KB,64KB,128KB