"hadoop fsck <filename> -blocks" is something that I think of quickly.

http://hadoop.apache.org/common/docs/current/commands_manual.html#fsck has more 
details

Kai

Am 28.02.2012 um 02:30 schrieb Mohit Anchlia:

> How do I verify the block size of a given file? Is there a command?
> 
> On Mon, Feb 27, 2012 at 7:59 AM, Joey Echeverria <j...@cloudera.com> wrote:
> 
>> dfs.block.size can be set per job.
>> 
>> mapred.tasktracker.map.tasks.maximum is per tasktracker.
>> 
>> -Joey
>> 
>> On Mon, Feb 27, 2012 at 10:19 AM, Mohit Anchlia <mohitanch...@gmail.com>
>> wrote:
>>> Can someone please suggest if parameters like dfs.block.size,
>>> mapred.tasktracker.map.tasks.maximum are only cluster wide settings or
>> can
>>> these be set per client job configuration?
>>> 
>>> On Sat, Feb 25, 2012 at 5:43 PM, Mohit Anchlia <mohitanch...@gmail.com
>>> wrote:
>>> 
>>>> If I want to change the block size then can I use Configuration in
>>>> mapreduce job and set it when writing to the sequence file or does it
>> need
>>>> to be cluster wide setting in .xml files?
>>>> 
>>>> Also, is there a way to check the block of a given file?
>>>> 
>> 
>> 
>> 
>> --
>> Joseph Echeverria
>> Cloudera, Inc.
>> 443.305.9434
>> 

-- 
Kai Voigt
k...@123.org




Reply via email to