Home | About | Sematext search-lucene.com search-hadoop.com
NEW: Monitor These Apps!
elasticsearch, apache solr, apache hbase, hadoop, redis, casssandra, amazon cloudwatch, mysql, memcached, apache kafka, apache zookeeper, apache storm, ubuntu, centOS, red hat, debian, puppet labs, java, senseiDB
 Search Hadoop and all its subprojects:

Switch to Plain View
Hadoop >> mail # user >> What happens when you do a ctrl-c on a big dfs -rmr


+
bzheng 2009-03-12, 02:48
Copy link to this message
-
Re: What happens when you do a ctrl-c on a big dfs -rmr

When you issue -rmr with directory, namenode get a directory name and starts deleting files recursively. It adds the blocks belonging to files to invalidate list. NameNode then deletes those blocks lazily. So, yes it will issue command to datanodes to delete those blocks, just give it some time. You do not need to reformat HDFS.
Lohit

----- Original Message ----
From: bzheng <[EMAIL PROTECTED]>
To: [EMAIL PROTECTED]
Sent: Wednesday, March 11, 2009 7:48:41 PM
Subject: What happens when you do a ctrl-c on a big dfs -rmr
I did a ctrl-c immediately after issuing a hadoop dfs -rmr command.  The rmr
target is no longer visible from the dfs -ls command.  The number of files
deleted is huge and I don't think it can possibly delete them all between
the time the command is issued and ctrl-c.  Does this mean it leaves behind
unreachable files on the slave nodes and making them dead weights?  We can
always reformat hdfs to be sure.  But is there a way to check?  Thanks.
--
View this message in context: http://www.nabble.com/What-happens-when-you-do-a-ctrl-c-on-a-big-dfs--rmr-tp22468909p22468909.html
Sent from the Hadoop core-user mailing list archive at Nabble.com.
+
何 永强 2009-03-12, 05:54
NEW: Monitor These Apps!
elasticsearch, apache solr, apache hbase, hadoop, redis, casssandra, amazon cloudwatch, mysql, memcached, apache kafka, apache zookeeper, apache storm, ubuntu, centOS, red hat, debian, puppet labs, java, senseiDB