Home | About | Sematext search-lucene.com search-hadoop.com
NEW: Monitor These Apps!
elasticsearch, apache solr, apache hbase, hadoop, redis, casssandra, amazon cloudwatch, mysql, memcached, apache kafka, apache zookeeper, apache storm, ubuntu, centOS, red hat, debian, puppet labs, java, senseiDB
 Search Hadoop and all its subprojects:

Switch to Threaded View
Hadoop >> mail # user >> How to access contents of a Map Reduce job's working directory


Copy link to this message
-
Re: How to access contents of a Map Reduce job's working directory
Try to use FileSystem.copyToLocal API to copy the files from the setup
directory.

On Tue, Aug 2, 2011 at 5:54 AM, Shrish Bajpai <[EMAIL PROTECTED]>wrote:

> I have just started to explore Hadoop but I am stuck in a situation now.
>
> I want to run a MapReduce job in hadoop which needs to create a "setup"
> folder in working directory. During the execution the job will generate
> some additional text files within this "setup" folder. The problem is I
> dont know how to access or move this setup folder content to my local file
> system as at end of the job, the job directory will be cleaned up.
>
> It would be great if you can help.
>
> Regards
>
> Shrish
>
>
--
Join me at http://hadoopworkshop.eventbrite.com/
NEW: Monitor These Apps!
elasticsearch, apache solr, apache hbase, hadoop, redis, casssandra, amazon cloudwatch, mysql, memcached, apache kafka, apache zookeeper, apache storm, ubuntu, centOS, red hat, debian, puppet labs, java, senseiDB