[prev in list] [next in list] [prev in thread] [next in thread]
List: hadoop-user
Subject: Re: splitting of big files?
From: Doug Cutting <cutting () apache ! org>
Date: 2008-05-29 17:20:18
Message-ID: 483EE5D2.7090008 () apache ! org
[Download RAW message or body]
Erik Paulson wrote:
> When reading from HDFS, how big are the network read requests, and what
> controls that? Or, more concretely, if I store files using 64Meg blocks
> in HDFS and run the simple word count example, and I get the default of
> one FileSplit/Map task per 64 meg block, how many bytes into the second 64meg
> block will a mapper read before it first passes a buffer up to the record
> reader to see if it has found an end-of-line?
This is controlled by io.file.buffer.size, which is 4k by default.
Doug
[prev in list] [next in list] [prev in thread] [next in thread]
Configure |
About |
News |
Add a list |
Sponsored by KoreLogic