Home | About | Sematext search-lucene.com search-hadoop.com
 Search Hadoop and all its subprojects:

Switch to Threaded View
Hadoop, mail # user - How to package multiple jars for a Hadoop job


Copy link to this message
-
Re: How to package multiple jars for a Hadoop job
Mark Kerzner 2011-02-18, 22:26
Thank you,
Mark

On Fri, Feb 18, 2011 at 4:23 PM, Eric Sammer <[EMAIL PROTECTED]> wrote:

> Mark:
>
> You have a few options. You can:
>
> 1. Package dependent jars in a lib/ directory of the jar file.
> 2. Use something like Maven's assembly plugin to build a self contained
> jar.
>
> Either way, I'd strongly recommend using something like Maven to build your
> artifacts so they're reproducible and in line with commonly used tools. Hand
> packaging files tends to be error prone. This is less of a Hadoop-ism and
> more of a general Java development issue, though.
>
>
> On Fri, Feb 18, 2011 at 5:18 PM, Mark Kerzner <[EMAIL PROTECTED]>wrote:
>
>> Hi,
>>
>> I have a script that I use to re-package all the jars (which are output in
>> a
>> dist directory by NetBeans) - and it structures everything correctly into
>> a
>> single jar for running a MapReduce job. Here it is below, but I am not
>> sure
>> if it is the best practice. Besides, it hard-codes my paths. I am sure
>> that
>> there is a better way.
>>
>> #!/bin/sh
>> # to be run from the project directory
>> cd ../dist
>> jar -xf MR.jar
>> jar -cmf META-INF/MANIFEST.MF  /home/mark/MR.jar *
>> cd ../bin
>> echo "Repackaged for Hadoop"
>>
>> Thank you,
>> Mark
>>
>
>
>
> --
> Eric Sammer
> twitter: esammer
> data: www.cloudera.com
>