Home | About | Sematext search-lucene.com search-hadoop.com
NEW: Monitor These Apps!
elasticsearch, apache solr, apache hbase, hadoop, redis, casssandra, amazon cloudwatch, mysql, memcached, apache kafka, apache zookeeper, apache storm, ubuntu, centOS, red hat, debian, puppet labs, java, senseiDB
 Search Hadoop and all its subprojects:

Switch to Threaded View
HBase >> mail # user >> HBase Security Configuration


Copy link to this message
-
Re: HBase Security Configuration
Hey Konrad,

Make sure your HBase's classpath also has the Hadoop conf dir on it
(specifically hdfs-site.xml and core-site.xml). It it already does
have that, make sure they are populated with the right HDFS cluster
values (core-site needs two properties that toggle security ON, and
hdfs-site needs the HDFS server principals configured inside it -
basically just copy these core-site and hdfs-site files from your
secured HDFS cluster config over to the HBase machines/classpath).

On Tue, Apr 17, 2012 at 5:38 PM, Konrad Tendera <[EMAIL PROTECTED]> wrote:
> Hello,
> I'm trying to configure secure HBase using following instruction: https://ccp.cloudera.com/display/CDHDOC/HBase+Security+Configuration. Our cluster uses Kerberos and everything in Hadoop work fine. But when I start HBase following exception is thrown
>
> FATAL org.apache.hadoop.hbase.master.HMaster: Unhandled exception. Starting shutdown.
> org.apache.hadoop.security.AccessControlException: Authentication is required
>        at org.apache.hadoop.ipc.Client.call(Client.java:1028)
>        at org.apache.hadoop.ipc.WritableRpcEngine$Invoker.invoke(WritableRpcEngine.java:198)
>        at $Proxy9.getProtocolVersion(Unknown Source)
>        at org.apache.hadoop.ipc.WritableRpcEngine.getProxy(WritableRpcEngine.java:235)
>        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:275)
>        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:249)
>        at org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:161)
>        at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:278)
>        at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:245)
>        at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:109)
>        at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1792)
>        at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:76)
>        at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:1826)
>        at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1808)
>        at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:265)
>        at org.apache.hadoop.fs.Path.getFileSystem(Path.java:189)
>        at org.apache.hadoop.hbase.util.FSUtils.getRootDir(FSUtils.java:471)
>        at org.apache.hadoop.hbase.master.MasterFileSystem.<init>(MasterFileSystem.java:94)
>        at org.apache.hadoop.hbase.master.HMaster.finishInitialization(HMaster.java:448)
>        at org.apache.hadoop.hbase.master.HMaster.run(HMaster.java:326)
>        at java.lang.Thread.run(Thread.java:662)
>
> I can't find any info about it. I'm using Hbase 0.92 with Hadoop 0.22
>
> --
> Konrad Tendera

--
Harsh J
NEW: Monitor These Apps!
elasticsearch, apache solr, apache hbase, hadoop, redis, casssandra, amazon cloudwatch, mysql, memcached, apache kafka, apache zookeeper, apache storm, ubuntu, centOS, red hat, debian, puppet labs, java, senseiDB