One way to deploy Hadoop in a more formal environment is to do it
via the Bigtop distribution. Bigtop provides packages and puppet deployment
code for most Linux distributions. We try to make the experience of
deploying Hadoop as seamless as possible, since our goal can
be summarized as "trying to be Ubuntu of Hadoop-based bigdata
management platforms". More info is available over here:
We're also an Apache incubating project. That means -- if
we don't quite scratch your itch yet -- join the fun and help
us make a 100% Apache Hadoop-based bigdata management
platform a reality.
On Wed, Feb 29, 2012 at 6:56 PM, Merto Mertek <[EMAIL PROTECTED]> wrote:
> I would be glad to hear what is your development cycle and how you deploy
> new features to production cluster... How do you deploy them to the
> production cluster? With bash scripts and rsync, ant, maven or any other
> automation tool? I would be thankfull if you could point me to any resource
> describing best practices in developing, deploying and automatization of
> java project in unix/linux environment..
> On 13 February 2012 11:26, Merto Mertek <[EMAIL PROTECTED]> wrote:
>> I am interested in some general tips on how to develop and deploy new
>> versions of hadoop. I've been trying to compile a new version of hadoop
>> and place the new jar to the cluster in the lib folder, however it was not
>> picked despite the classpath was explicitly set to the lib folder. I am
>> interested in the following questions:
>> a) How to deploy a new version? Just copy the new compiled jar file to all
>> lib folders on all nodes?
>> b) Should I make just a new compile or a new release ('ant' vs 'ant tar')?
>> c) How do you develop and deploy hadoop locally and how remotely? For
>> deploying builds are you using your own sh scripts or are you using any
>> tools like ant/maven?
>> d) What is the purpose of the folder $HADOOP_HOME/share/hadoop?
>> Any other tips are welcomed..
>> Thank you