Ah, that may be cause the core-site.xml has the property
io.serializations fully defined for Gora as well? You can do that as
an alternative fix, supply a core-site.xml across tasktrackers that
also carry the serialization class Gora requires. I failed to think of
that as a solution.

On Sat, Jul 28, 2012 at 6:04 AM, Sriram Ramachandrasekaran
<sri.ram...@gmail.com> wrote:
> okay. But this issue didn't present itself when run in standalone mode. :)
>
> On 28 Jul 2012 06:02, "Harsh J" <ha...@cloudera.com> wrote:
>>
>> I find it easier to run jobs via MRUnit (http://mrunit.apache.org,
>> TDD) first, or via LocalJobRunner, for debug purposes.
>>
>> On Sat, Jul 28, 2012 at 5:53 AM, Sriram Ramachandrasekaran
>> <sri.ram...@gmail.com> wrote:
>> > hello harsh,
>> > thanks for your investigations. while we were debugging, I saw the exact
>> > thing. As you pointed out, we suspected it to be a problem. So, we set
>> > the
>> > job conf object directly on Gora's query object.
>> > It goes something like this,
>> > query.setConf..(job.getConfig..())
>> >
>> > And, then I saw that it was not getting into creating a new object at
>> > getOrCreate().
>> >
>> > OTOH, i've not tried the job.xml thing. I should give it a try n I shall
>> > keep the loop posted.
>> >
>> > I would also like to hear about standard practices for debugging
>> > distributed
>> > MR tasks.
>> >
>> > -----
>> > reply from a hh device. Pl excuse typos n lack of formatting.
>> >
>> > On 28 Jul 2012 03:30, "Harsh J" <ha...@cloudera.com> wrote:
>> >>
>> >> Hi Sriram,
>> >>
>> >> I suspect the following in Gora to somehow be causing this issue:
>> >>
>> >> IOUtils source:
>> >>
>> >>
>> >> http://svn.apache.org/viewvc/gora/trunk/gora-core/src/main/java/org/apache/gora/util/IOUtils.java?view=markup
>> >> QueryBase source:
>> >>
>> >>
>> >> http://svn.apache.org/viewvc/gora/trunk/gora-core/src/main/java/org/apache/gora/query/impl/QueryBase.java?view=markup
>> >>
>> >> Notice that IOUtils.deserialize(…) calls expect a proper Configuration
>> >> object. If not passed (i.e., if null), they call the following.
>> >>
>> >> 68        private static Configuration getOrCreateConf(Configuration
>> >> conf)
>> >> {
>> >> 69          if(conf == null) {
>> >> 70            if(IOUtils.conf == null) {
>> >> 71              IOUtils.conf = new Configuration();
>> >> 72            }
>> >> 73          }
>> >> 74          return conf != null ? conf : IOUtils.conf;
>> >> 75        }
>> >>
>> >> Now QueryBase, has in its readFields method, some
>> >> IOUtils.deserialize(…) calls, that seem to pass a null for the
>> >> configuration object. The IOUtils.deserialize(…) method hence calls
>> >> this above method, and initializes a whole new Configuration object,
>> >> as the passed conf object is null.
>> >>
>> >> If it does that, it would not be loading the "job.xml" file contents,
>> >> which is the job's config file (thats something the map task's config
>> >> set alone loads, and not a file thats loaded by default). So hence,
>> >> custom serializers will disappear the moment it begins using this new
>> >> Configuration object.
>> >>
>> >> This is what you'll want to investigate and fix or notify the Gora
>> >> devs about (why QueryBase#readFields uses a null object, and if it can
>> >> reuse some set conf object). As a cheap hack fix, maybe doing the
>> >> following will make it work in an MR environment?
>> >>
>> >> IOUtils.conf = new Configuration();
>> >> IOUtils.conf.addResource("job.xml");
>> >>
>> >> I haven't tried the above, but let us know how we can be of further
>> >> assistance. An ideal fix would be to only use the MapTask's provided
>> >> Configuration object everywhere, somehow, and never re-create one.
>> >>
>> >> P.s. If you want a thread ref link to share with other devs over Gora,
>> >> here it is: http://search-hadoop.com/m/BXZA4dTUFC
>> >>
>> >> On Fri, Jul 27, 2012 at 1:24 PM, Sriram Ramachandrasekaran
>> >> <sri.ram...@gmail.com> wrote:
>> >> > Hello,
>> >> > I have an MR job that talks to HBase. I use Gora to talk to HBase.
>> >> > Gora
>> >> > also
>> >> > provides couple of classes which can be extended to write Mappers and
>> >> > Reducers, if the mappers need input from an HBase store and Reducers
>> >> > need to
>> >> > write it out to an HBase store. This is the reason why I use Gora.
>> >> >
>> >> > Now, when I run my MR job, I get an exception as below.
>> >> > (https://issues.apache.org/jira/browse/HADOOP-3093)
>> >> > java.lang.RuntimeException: java.io.IOException:
>> >> > java.lang.NullPointerException
>> >> > at
>> >> >
>> >> >
>> >> > org.apache.gora.mapreduce.GoraInputFormat.setConf(GoraInputFormat.java:115)
>> >> > at
>> >> >
>> >> > org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:62)
>> >> > at
>> >> >
>> >> >
>> >> > org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:117)
>> >> > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:723)
>> >> > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:370)
>> >> > at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
>> >> > at java.security.AccessController.doPrivileged(Native Method)
>> >> > at javax.security.auth.Subject.doAs(Subject.java:415)
>> >> > at
>> >> >
>> >> >
>> >> > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1093)
>> >> > at org.apache.hadoop.mapred.Child.main(Child.java:249)
>> >> > Caused by: java.io.IOException: java.lang.NullPointerException
>> >> > at org.apache.gora.util.IOUtils.loadFromConf(IOUtils.java:483)
>> >> > at
>> >> >
>> >> >
>> >> > org.apache.gora.mapreduce.GoraInputFormat.getQuery(GoraInputFormat.java:125)
>> >> > at
>> >> >
>> >> >
>> >> > org.apache.gora.mapreduce.GoraInputFormat.setConf(GoraInputFormat.java:112)
>> >> > ... 9 more
>> >> > Caused by: java.lang.NullPointerException
>> >> > at
>> >> >
>> >> >
>> >> > org.apache.hadoop.io.serializer.SerializationFactory.getDeserializer(SerializationFactory.java:77)
>> >> > at org.apache.gora.util.IOUtils.deserialize(IOUtils.java:205)
>> >> > at
>> >> > org.apache.gora.query.impl.QueryBase.readFields(QueryBase.java:234)
>> >> > at
>> >> >
>> >> >
>> >> > org.apache.hadoop.io.serializer.WritableSerialization$WritableDeserializer.deserialize(WritableSerialization.java:67)
>> >> > at
>> >> >
>> >> >
>> >> > org.apache.hadoop.io.serializer.WritableSerialization$WritableDeserializer.deserialize(WritableSerialization.java:40)
>> >> > at
>> >> >
>> >> >
>> >> > org.apache.hadoop.io.DefaultStringifier.fromString(DefaultStringifier.java:75)
>> >> > at
>> >> >
>> >> > org.apache.hadoop.io.DefaultStringifier.load(DefaultStringifier.java:133)
>> >> > at org.apache.gora.util.IOUtils.loadFromConf(IOUtils.java:480)
>> >> > ... 11 more
>> >> >
>> >> > I tried the following things to work through this issue.
>> >> > 0. The stack trace indicates that, when setting up a new Mapper, it
>> >> > is
>> >> > unable to deserialize something. (I could not get to understand where
>> >> > it
>> >> > fails).
>> >> > 1. I looked around the forums and realized that serialization options
>> >> > are
>> >> > not getting passed, so, I tried setting up, io.serializations config
>> >> > on
>> >> > the
>> >> > job.
>> >> >    1.1. I am not setting up the "io.serializations" myself, I use
>> >> > GoraMapReduceUtils.setIOSerializations() to do it. I verified that,
>> >> > the
>> >> > confs are getting proper serializers.
>> >> > 2. I verified in the job xml to see if these confs have got through,
>> >> > they
>> >> > were. But, it failed again.
>> >> > 3. I tried starting the hadoop job runner with debug options turned
>> >> > on
>> >> > and
>> >> > in suspend mode, -XDebug suspend=y and I also set the VM options for
>> >> > mapred
>> >> > child tasks, via the mapred.child.java.opts to see if I can debug the
>> >> > VM
>> >> > that gets spawned newly. Although I get a message on my stdout
>> >> > saying,
>> >> > opening port X and waiting, when I try to attach a remote debugger on
>> >> > that
>> >> > port, it does not work.
>> >> >
>> >> > I understand that, when SerializationFactory tries to deSerialize
>> >> > 'something', it does not find an appropriate unmarshaller and so it
>> >> > fails.
>> >> > But, I would like to know a way to find that 'something' and I would
>> >> > like to
>> >> > get some idea on how (pseudo) distributed MR jobs should be generally
>> >> > debugged. I tried searching, did not find anything useful.
>> >> >
>> >> > Any help/pointers would be greatly useful.
>> >> >
>> >> > Thanks!
>> >> >
>> >> > --
>> >> > It's just about how deep your longing is!
>> >> >
>> >>
>> >>
>> >>
>> >> --
>> >> Harsh J
>>
>>
>>
>> --
>> Harsh J



-- 
Harsh J

Reply via email to