Home | About | Sematext search-lucene.com search-hadoop.com
 Search Hadoop and all its subprojects:

Switch to Threaded View
Pig, mail # user - HBaseStorage in pig 0.8


Copy link to this message
-
Re: HBaseStorage in pig 0.8
Corbin Hoenes 2010-12-15, 17:38
PIG-1769 has been created sorry I lost track of this :(

On Nov 22, 2010, at 2:30 PM, Dmitriy Ryaboy wrote:

> Hm, good point. Can you create a JIRA for this?
>
> On Mon, Nov 22, 2010 at 1:16 PM, Corbin Hoenes <[EMAIL PROTECTED]> wrote:
>
>> One comment on the HBaseStorage store func.  In our load statement we
>> are allowed to prefix the table name with "hbase://" but when we call
>> store it throws an exception unless we remove hbase:// from the table
>> name:
>>
>> this works:
>> store raw into 'piggytest2' USING
>> org.apache.pig.backend.hadoop.hbase.HBaseStorage('content2:field1
>> anchor2:field1a anchor2:field2a');
>>
>> this won't
>> store raw into 'hbase://piggytest2'
>>
>> Exception:
>> Caused by: java.lang.IllegalArgumentException:
>> java.net.URISyntaxException: Relative path in absolute URI:
>> hbase://piggytest2_logs
>>
>> Would be nice to be able to prefix the store with hbase:// as well.
>>
>>
>> On Mon, Nov 22, 2010 at 12:10 PM, Dmitriy Ryaboy <[EMAIL PROTECTED]>
>> wrote:
>>>
>>> Why is it connecting to localhost?
>>> Sounds like you don't have the appropriate config files on the path.
>>> Hm, maybe  we should serialize those in the constructor so that you don't
>>> have to have them on the JT classpath (I have them on the JT classpath so
>>> this never came up). Can you confirm that this is the problem?
>>>
>>> D
>>>
>>> On Fri, Nov 19, 2010 at 10:33 PM, Corbin Hoenes <[EMAIL PROTECTED]> wrote:
>>>
>>>> Hey Jeff,
>>>>
>>>> It wasn't starting a job but I got a bit further by registering the
>> pig8
>>>> jar in my pig script.  It seemed to have a bunch of dependencies on
>> google
>>>> common collections; zookeeper etc... built into that jar.
>>>>
>>>> Now I am seeing this in the web ui logs:
>>>> 2010-11-19 23:19:44,200 INFO org.apache.zookeeper.ClientCnxn:
>> Attempting
>>>> connection to server localhost/127.0.0.1:2181
>>>> 2010-11-19 23:19:44,201 WARN org.apache.zookeeper.ClientCnxn: Exception
>>>> closing session 0x0 to sun.nio.ch.SelectionKeyImpl@65efb4be
>>>> java.net.ConnectException: Connection refused
>>>>       at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
>>>>       at
>>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:574)
>>>>       at
>>>> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:885)
>>>> 2010-11-19 23:19:44,201 WARN org.apache.zookeeper.ClientCnxn: Ignoring
>>>> exception during shutdown input
>>>> java.nio.channels.ClosedChannelException
>>>>       at
>>>> sun.nio.ch.SocketChannelImpl.shutdownInput(SocketChannelImpl.java:638)
>>>>       at
>> sun.nio.ch.SocketAdaptor.shutdownInput(SocketAdaptor.java:360)
>>>>       at
>>>> org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:951)
>>>>       at
>>>> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:922)
>>>> 2010-11-19 23:19:44,201 WARN org.apache.zookeeper.ClientCnxn: Ignoring
>>>> exception during shutdown output
>>>> java.nio.channels.ClosedChannelException
>>>>       at
>>>> sun.nio.ch.SocketChannelImpl.shutdownOutput(SocketChannelImpl.java:649)
>>>>       at
>> sun.nio.ch.SocketAdaptor.shutdownOutput(SocketAdaptor.java:368)
>>>>       at
>>>> org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:956)
>>>>       at
>>>> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:922)
>>>> 2010-11-19 23:19:44,303 WARN
>>>> org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper: Failed to create
>> /hbase
>>>> -- check quorum servers, currently=localhost:2181
>>>> org.apache.zookeeper.KeeperException$ConnectionLossException:
>>>> KeeperErrorCode = ConnectionLoss for /hbase
>>>> Looks like it doesn't know where my hbase/conf/hbase-site.xml file is?
>> Not
>>>> sure how would this get passed to the HBaseStorage class?
>>>>
>>>> On Nov 19, 2010, at 5:09 PM, Jeff Zhang wrote:
>>>>
>>>>> Does the mapreduce job start ? Could you check the logs on hadoop
>> side ?
>>>>>
>>>>>
>>>>> On Sat, Nov 20, 2010 at 7:56 AM, Corbin Hoenes <[EMAIL PROTECTED]>