-Re: Unsatisfied link error - how to load native library without copying it in /lib/native folder
That should be:
and to include jars in for your mrjobs, you would do:
2) -libjars path_to_my1.jar,path_to_my2.jar
On Thu, Oct 25, 2012 at 6:10 PM, Dipesh Khakhkhar
<[EMAIL PROTECTED]> wrote:
> I am a new hadoop user and have few very basic questions (they might sound
> very stupid to many people so please bear with me).
> I am running a MR task and my launcher program needs to load a library using
> System.loadLibrary(somelibrary). This works fine if I put this library in
> lib/native/Linux-amd64-64. I tried the following -
> 1. provided -files=/path_to_directory_containging_my_library
> 2. provided the following in mapred-site.xml (didn't try it in core-site.xml
> or hdfs-site.xml)
> I'm using hadoop 1.0.3 and this is a single node cluster for testing
> I have a production environment where I'm running 4 data nodes and currently
> I'm copying this file in lib/native/Linux-amd64-64 folder in each node's
> hadoop installation.
> A related question regarding providing jars required for running the whole
> M/R application - currently I have edited hadoop-classpath variable in
> hadoop-env.sh. For cluster if I provide -libjars option will that work
> without editing classpath? I require this jar's classes before launching M/R
> Also how can I provide my application jar ( i.e. bin/hadoop jar myjar
> com.x.x.ProgramName ) in the data nodes? Currently I'm copying it in the
> lib directory of hadoop installation.
> Thanks in advance for answering my queries.
Apache MRUnit - Unit testing MapReduce - http://incubator.apache.org/mrunit/