I've used puppet for the job. Adobe posted a set of puppet scripts a while
back that would deploy a tar.gz. They were my starting point (here:
configs and values are pretty old so make sure you look at the docs for
whatever version you are deploying. Even if you deploy through rpm/deb I
would highly recommend some sort of config control; be it puppet, chef, or
cf engine. The time taken to learn and set that up will pay itself back in
the long term. There are a lot of different moving pieces in a hadoop MR /
HDFS / HBase cluster and they will take some tuning.
On Wed, May 30, 2012 at 2:08 PM, Peter Naudus <[EMAIL PROTECTED]> wrote:
> Hello All,
> Is there a "community standard" / "best" way to deploy HBase to a cluster?
> We're in the process of setting up a ~15 node cluster and I'm curious how
> you all go about your deployments. Do you package the code into an RPM,
> place it into a central YUM repository, and then drive the install via
> Puppet? Or do you rsync the code and use shell scripts? Or all the above?
> Thanks so much for your input!
> ~ Peter