Home | About | Sematext search-lucene.com search-hadoop.com
NEW: Monitor These Apps!
elasticsearch, apache solr, apache hbase, hadoop, redis, casssandra, amazon cloudwatch, mysql, memcached, apache kafka, apache zookeeper, apache storm, ubuntu, centOS, red hat, debian, puppet labs, java, senseiDB
 Search Hadoop and all its subprojects:

Switch to Threaded View
HDFS >> mail # dev >> Compiling hadoop native libraries


Copy link to this message
-
Re: Compiling hadoop native libraries
Run the following command:

$ ant -Dcompile.native=true package

Arun

On Jul 31, 2011, at 9:20 PM, Aastha Mehta wrote:

> Hi Arun,
>
> Thanks for the prompt reply. I am not sure, I understood you correctly.
> Compile/binary/tar of what? The native files? The lib/native/Linux-amd64-64/
> contains following files:
> libhadoop.a
> libhadoop.la
> libhadoop.so
> libhadoop.so.1
> libhadoop.so.1.0.0
>
> This directory is present in the package itself. So, should I make a tar of
> it and then provide it? I tried the following, but it failed:
> ant -Dcompile.native=true
> $HADOOP_HOME/lib/native/Linux-amd64-64/libhadoop.so
>
> The error I got is - "Target  lib/native/Linux-amd64-64/libhadoop.so does
> not exist in the project Hadoop".
>
> Thanks,
> Aastha.
>
> On 1 August 2011 09:44, Arun Murthy <[EMAIL PROTECTED]> wrote:
>
>> <target> could be compile or binary or tar.
>>
>> Arun
>>
>> Sent from my iPhone
>>
>> On Jul 31, 2011, at 9:05 PM, Aastha Mehta <[EMAIL PROTECTED]> wrote:
>>
>>> Hello,
>>>
>>> I am trying to run fuse_dfs_wrapper.sh from
>>> hadoop-0.20.2/src/contrib/fuse_dfs/src on a 64-bit machine. I get the
>>> following error:
>>> ./fuse_dfs: error while loading shared libraries: libhdfs.so.0: cannot
>> open
>>> shared object file: No such file or directory
>>>
>>> I searched on the net and found a response to a similar query here -
>>>
>> https://groups.google.com/a/cloudera.org/group/cdh-user/browse_thread/thread/3db7efc10cff8bbc?pli=1
>>>
>>> My hadoop package contains the native files in
>>> hadoop-0.20.2/lib/native/Linux-amd64-64/
>>>
>>> I followed to this link -
>>> http://hadoop.apache.org/common/docs/current/native_libraries.html to
>>> understand the steps to build hadoop native libraries.
>>>
>>> I have a small query regarding the building step. On the above link, it
>> is
>>> mentioned -
>>>
>>> "Once you installed the prerequisite packages use the standard hadoop
>>> build.xml file and pass along the compile.native flag (set to true) to
>> build
>>> the native hadoop library:
>>>
>>> $ ant -Dcompile.native=true <target>
>>>
>>> You should see the newly-built library in:
>>>
>>> $ build/native/<platform>/lib
>>>
>>> where <platform> is a combination of the system-properties: ${os.name
>>> }-${os.arch}-${sun.arch.data.model} (for example, Linux-i386-32)."
>>>
>>>
>>> Could someone please tell what exactly is <target> in the first step.
>>>
>>>
>>> Thanks and regards,
>>>
>>> Aastha.
>>>
>>>
>>>
>>>
>>>
>>> --
>>> Aastha Mehta
>>> B.E. (Hons.) Computer Science
>>> BITS Pilani
>>> E-mail: [EMAIL PROTECTED]
>>
>
>
>
> --
> Aastha Mehta
> B.E. (Hons.) Computer Science
> BITS Pilani
> E-mail: [EMAIL PROTECTED]
NEW: Monitor These Apps!
elasticsearch, apache solr, apache hbase, hadoop, redis, casssandra, amazon cloudwatch, mysql, memcached, apache kafka, apache zookeeper, apache storm, ubuntu, centOS, red hat, debian, puppet labs, java, senseiDB