Is there a good way to emulate or determine the blocks that would exist on
HDFS for a given file. if I have a 135mb file and my block size is 128,
does it stand to say I would have 2 blocks, block A is byte 0-134217728 and
Block b is 134217729-141557761 deterministically? I am attempting to
calculate checksum in a manner that matches the method defined in
DFSClient.getChecksum(Path p).
NEW: Monitor These Apps!
elasticsearch, apache solr, apache hbase, hadoop, redis, casssandra, amazon cloudwatch, mysql, memcached, apache kafka, apache zookeeper, apache storm, ubuntu, centOS, red hat, debian, puppet labs, java, senseiDB