Home | About | Sematext search-lucene.com search-hadoop.com
 Search Hadoop and all its subprojects:

Switch to Threaded View
HDFS, mail # user - Re: namenode memory test


Copy link to this message
-
Re: namenode memory test
sudhakara st 2013-04-27, 18:59
Every file, directory and block in HDFS is represented as an object in the
namenode’s memory, Namenode consume about average of 150 bytes per each
block(object).
On Wed, Apr 24, 2013 at 12:30 PM, Mahesh Balija
<[EMAIL PROTECTED]>wrote:

> Can you manually go into the directory configured for hadoop.tmp.dir under
> core-site.xml and do an ls -l to find the disk usage details, it will have
> fsimage, edits, fstime, VERSION.
> or the basic commands like,
> hadoop fs -du
> hadoop fsck
>
>
>
> On Wed, Apr 24, 2013 at 7:56 AM, 自己 <[EMAIL PROTECTED]> wrote:
>
>> Hi, I would like to know  how much memory our data take on the name-node
>> per block, file and directory.
>> For example, the metadata size of a file.
>> When I store some files in HDFS,how can I get the memory size take on
>> the name-node?
>> Is there some tools or commands to test the memory size take on the
>> name-node?
>>
>> I'm looking forward to your reply! Thanks!
>>
>>
>>
>
--

Regards,
.....  Sudhakara.st