Home | About | Sematext search-lucene.com search-hadoop.com
 Search Hadoop and all its subprojects:

Switch to Threaded View
Pig >> mail # user >> Copying files to Amazon S3 using Pig is slow

Copy link to this message
Re: Copying files to Amazon S3 using Pig is slow
Definitely go down the s3distcp route. I use it to copy large number of
smaller files from s3 into fewer larger ones in HDFS and it's been working
great. This also helps out with the Pig jobs running faster vs. having Pig
try to load XXXX files from s3.


On Fri, Jun 8, 2012 at 5:40 AM, James Newhaven <[EMAIL PROTECTED]>wrote:

> I want to copy 26,000 HDFS files generated by a pig script to Amazon S3.
> I am using the copyToLocal command, but I noticed the copy throughput is
> only one file per second - so it is going to take about 7 hours to copy all
> the files.
> The command I am using is: copyToLocal /tmp/files/ s3://my-bucket/
> Does anyone have any ideas how I could speed this up?
> Thanks,
> James