Hmm. The data in my tables is not important. So, I dropped the table and 
recreated it. This doesn't seem to have resolved the issue, though.

Is there perhaps a Pig query I can run that would use a built in HBase table, 
like that .META. table, and see if it works? I don't know if that'd help or 
not, though.

Ryan

On Mar 22, 2012, at 10:02 PM, Norbert Burger wrote:

> Actually on second glance, this seems like an issue not with the HBase
> config, but with the server:port info inside your .META. table.  Have you
> tried LOADing from a different table besides "events"?  From the HBase
> shell, you can use the following command to extract server hostnames for
> each of your regions:
> 
> scan '.META.',{COLUMNS => 'info:server'}
> 
> Dropping and recreating the table should resolve the issue; if you have
> real data already, then there are tools in hbase/bin to repair regions.
> 
> Norbert
> 
> On Thu, Mar 22, 2012 at 10:44 PM, Ryan Cole <[email protected]> wrote:
> 
>> I was thinking that maybe it was because I did not have HBase config path
>> on PIG_CLASSPATH, so I added it. This did not help, though.
>> 
>> Ryan
>> 
>> On Thu, Mar 22, 2012 at 9:07 PM, Ryan Cole <[email protected]> wrote:
>> 
>>> Norbert,
>>> 
>>> I have confirmed that this is indeed an issue connecting to HBase. I
>> tried
>>> just running a Pig script that did not use HBaseStorage, and it works.
>> Here
>>> is my hbase-site.xml config file, as well as my query that I'm running:
>>> 
>>> https://gist.github.com/2166187
>>> 
>>> Also, for ease of reference, here is the query:
>>> 
>>> raw = LOAD 'hbase://events' USING
>>> org.apache.pig.backend.hadoop.hbase.HBaseStorage('event:*', '-loadKey
>>> true') AS (id:bytearray, events_map:map[]);
>>> 
>>> Maybe I need to change the rootdir in the config to be my hostname, and
>>> not localhost?
>>> 
>>> Ryan
>>> 
>>> 
>>> On Thu, Mar 22, 2012 at 8:58 PM, Norbert Burger <
>> [email protected]>wrote:
>>> 
>>>> You're encountering problems connecting to HBase (presumably your Pig
>>>> script uses HBaseStorage).  How does your hbase/conf/hbase-site.xml
>> look?
>>>> 
>>>> Norbert
>>>> 
>>>> On Thu, Mar 22, 2012 at 9:16 PM, Ryan Cole <[email protected]> wrote:
>>>> 
>>>>> Hello,
>>>>> 
>>>>> I'm new to these lists. I'm trying to get Pig working, for my first
>>>> time. I
>>>>> have setup Hadoop and HBase (on HDFS) using the psuedo-distributed
>>>> setup,
>>>>> all on one machine. I am able to run MapReduce jobs, using the
>>>> example.jar
>>>>> file included with the Hadoop release.
>>>>> 
>>>>> Whenever I try to run even the simplest query examples, using Pig, I
>> get
>>>>> the following error:
>>>>> 
>>>>> `ERROR 2017: Internal error creating job configuration.`
>>>>> 
>>>>> and, the log file has the following more specific error:
>>>>> 
>>>>> `Caused by: java.lang.IllegalArgumentException: Not a host:port pair:
>>>>> �^@^@^@^P8948@ryan-serverlocalhost,44544,1332443083936`
>>>>> 
>>>>> It looks like Pig goes through the entire compile process, of turning
>>>> the
>>>>> Pig Latin into the MapReduce code, but fails to send it off to
>> Hadoop's
>>>>> MapReduce. That's just my uneducated analysis of what I see happening,
>>>>> though. I have pasted the Grunt console output, and the log file
>>>> contents
>>>>> here: https://gist.github.com/2163762.
>>>>> 
>>>>> This is Pig v.0.9.2, HBase v. 0.92.1, Hadoop v1.0.1.
>>>>> 
>>>>> Does anyone have any idea why this may be happening? It looks obvious
>>>> that
>>>>> I have something configured improperly, but I looked for any host:post
>>>>> settings that would stand out and didn't find anything obvious.
>>>>> 
>>>>> Thanks,
>>>>> Ryan
>>>>> 
>>>> 
>>> 
>>> 
>> 

Reply via email to