Prasanth,

Any luck?


On Tue, Dec 24, 2013 at 4:31 PM, Bryan Jeffrey <bryan.jeff...@gmail.com>wrote:

> Prasanth,
>
> I am also traveling this week.  Your assistance would be appreciated, but
> not at the expense of your holiday!
>
> Bryan
> On Dec 24, 2013 2:23 PM, "Prasanth Jayachandran" <
> pjayachand...@hortonworks.com> wrote:
>
>> Bryan
>>
>> I have a similar setup. I will try to reproduce this issue and get back
>> to you asap.
>> Since i am traveling expect some delay.
>>
>> Thanks
>> Prasanth
>>
>> Sent from my iPhone
>>
>> On Dec 24, 2013, at 11:39 AM, Bryan Jeffrey <bryan.jeff...@gmail.com>
>> wrote:
>>
>> Hello.
>>
>> I posted this a few weeks ago, but was unable to get a response that
>> solved the issue.  I have made no headway in the mean time.  I was hoping
>> that if I re-summarized the issue that someone would have some advice
>> regarding this problem.
>> Running the following version of Hadoop: hadoop-2.2.0
>> Running the following version of Hive: hive-0.12.0
>>
>> I have a simple test system setup with (2) datanodes/node manager and (1)
>> namenode/resource manager.  Hive is running on the namenode, and contacting
>> a MySQL database for metastore.
>>
>> I have created a small table 'from_text' as follows:
>>
>> [server:10001] hive> describe from_text;
>> foo                     int                     None
>> bar                     int                     None
>> boo                     string                  None
>>
>>
>> [server:10001] hive> select * from from_text;
>> 1       2       Hello
>> 2       3       World
>>
>> I go to insert the data into my Orc table, 'orc_table':
>>
>> [server:10001] hive> describe orc_test;
>> foo                     int                     from deserializer
>> bar                     int                     from deserializer
>> boo                     string                  from deserializer
>>
>>
>> The job runs, but fails to complete with the following errors (see
>> below).  This seems to be the exact example covered in the example here:
>>
>>
>> http://hortonworks.com/blog/orcfile-in-hdp-2-better-compression-better-performance/
>>
>> The error output is below.  I have tried several things to solve the
>> issue, including re-installing Hive 0.12.0 from binary install.
>>
>> Help?
>>
>> [server:10001] hive> insert into table orc_test select * from from_text;
>> [Hive Error]: Query returned non-zero code: 2, cause: FAILED: Execution
>> Error, return code 2 from org.apache.hadoop.hive.ql.exec.mr.MapRedTask
>>
>>
>> Diagnostic Messages for this Task:
>> Error: java.lang.RuntimeException: Hive Runtime Error while closing
>> operators
>>         at
>> org.apache.hadoop.hive.ql.exec.mr.ExecMapper.close(ExecMapper.java:240)
>>         at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:61)
>>         at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:429)
>>         at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341)
>>         at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:162)
>>         at java.security.AccessController.doPrivileged(Native Method)
>>         at javax.security.auth.Subject.doAs(Subject.java:396)
>>         at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491)
>>         at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:157)
>> Caused by: java.lang.UnsupportedOperationException: This is supposed to
>> be overridden by subclasses.
>>         at
>> com.google.protobuf.GeneratedMessage.getUnknownFields(GeneratedMessage.java:180)
>>         at org.apache.hadoop.hive.ql.io.orc
>> .OrcProto$ColumnStatistics.getSerializedSize(OrcProto.java:3046)
>>         at
>> com.google.protobuf.CodedOutputStream.computeMessageSizeNoTag(CodedOutputStream.java:749)
>>         at
>> com.google.protobuf.CodedOutputStream.computeMessageSize(CodedOutputStream.java:530)
>>         at org.apache.hadoop.hive.ql.io.orc
>> .OrcProto$RowIndexEntry.getSerializedSize(OrcProto.java:4129)
>>         at
>> com.google.protobuf.CodedOutputStream.computeMessageSizeNoTag(CodedOutputStream.java:749)
>>         at
>> com.google.protobuf.CodedOutputStream.computeMessageSize(CodedOutputStream.java:530)
>>         at org.apache.hadoop.hive.ql.io.orc
>> .OrcProto$RowIndex.getSerializedSize(OrcProto.java:4641)
>>         at
>> com.google.protobuf.AbstractMessageLite.writeTo(AbstractMessageLite.java:75)
>>         at org.apache.hadoop.hive.ql.io.orc
>> .WriterImpl$TreeWriter.writeStripe(WriterImpl.java:548)
>>         at org.apache.hadoop.hive.ql.io.orc
>> .WriterImpl$StructTreeWriter.writeStripe(WriterImpl.java:1328)
>>         at org.apache.hadoop.hive.ql.io.orc
>> .WriterImpl.flushStripe(WriterImpl.java:1699)
>>         at org.apache.hadoop.hive.ql.io.orc
>> .WriterImpl.close(WriterImpl.java:1868)
>>         at org.apache.hadoop.hive.ql.io.orc
>> .OrcOutputFormat$OrcRecordWriter.close(OrcOutputFormat.java:95)
>>         at
>> org.apache.hadoop.hive.ql.exec.FileSinkOperator$FSPaths.closeWriters(FileSinkOperator.java:181)
>>         at
>> org.apache.hadoop.hive.ql.exec.FileSinkOperator.closeOp(FileSinkOperator.java:866)
>>         at
>> org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:596)
>>         at
>> org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:613)
>>         at
>> org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:613)
>>         at
>> org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:613)
>>         at
>> org.apache.hadoop.hive.ql.exec.mr.ExecMapper.close(ExecMapper.java:207)
>>         ... 8 more
>>
>>
>> On Tue, Dec 17, 2013 at 11:56 AM, Bryan Jeffrey 
>> <bryan.jeff...@gmail.com>wrote:
>>
>>> Prasanth,
>>>
>>> I downloaded the binary Hive version from the URL you specified.  I
>>> untarred the Hive tar, copied in configuration files, and started Hive.  I
>>> continue to see the same error:
>>>
>>> [server:10001] hive> describe orc_test;
>>> foo                     int                     from deserializer
>>> bar                     int                     from deserializer
>>> boo                     string                  from deserializer
>>>
>>>
>>> [server:10001] hive> describe from_text;
>>> foo                     int                     None
>>> bar                     int                     None
>>> boo                     string                  None
>>>
>>> [server:10001] hive> select * from from_text;
>>> 1       2       Hello
>>> 2       3       World
>>>
>>> [server:10001] hive> insert into table orc_test select * from from_text;
>>> [Hive Error]: Query returned non-zero code: 2, cause: FAILED: Execution
>>> Error, return code 2 from org.apache.hadoop.hive.ql.exec.mr.MapRedTask
>>>
>>> From the Hive Log:
>>> Diagnostic Messages for this Task:
>>> Error: java.lang.RuntimeException: Hive Runtime Error while closing
>>> operators
>>>         at
>>> org.apache.hadoop.hive.ql.exec.mr.ExecMapper.close(ExecMapper.java:240)
>>>         at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:61)
>>>         at
>>> org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:429)
>>>         at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341)
>>>         at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:162)
>>>         at java.security.AccessController.doPrivileged(Native Method)
>>>         at javax.security.auth.Subject.doAs(Subject.java:396)
>>>         at
>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491)
>>>         at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:157)
>>> Caused by: java.lang.UnsupportedOperationException: This is supposed to
>>> be overridden by subclasses.
>>>         at
>>> com.google.protobuf.GeneratedMessage.getUnknownFields(GeneratedMessage.java:180)
>>>         at
>>> org.apache.hadoop.hive.ql.io.orc.OrcProto$ColumnStatistics.getSerializedSize(OrcProto.java:3046)
>>>          at
>>> com.google.protobuf.CodedOutputStream.computeMessageSizeNoTag(CodedOutputStream.java:749)
>>>         at
>>> com.google.protobuf.CodedOutputStream.computeMessageSize(CodedOutputStream.java:530)
>>>         at
>>> org.apache.hadoop.hive.ql.io.orc.OrcProto$RowIndexEntry.getSerializedSize(OrcProto.java:4129)
>>>         at
>>> com.google.protobuf.CodedOutputStream.computeMessageSizeNoTag(CodedOutputStream.java:749)
>>>         at
>>> com.google.protobuf.CodedOutputStream.computeMessageSize(CodedOutputStream.java:530)
>>>         at
>>> org.apache.hadoop.hive.ql.io.orc.OrcProto$RowIndex.getSerializedSize(OrcProto.java:4641)
>>>         at
>>> com.google.protobuf.AbstractMessageLite.writeTo(AbstractMessageLite.java:75)
>>>         at
>>> org.apache.hadoop.hive.ql.io.orc.WriterImpl$TreeWriter.writeStripe(WriterImpl.java:548)
>>>         at
>>> org.apache.hadoop.hive.ql.io.orc.WriterImpl$StructTreeWriter.writeStripe(WriterImpl.java:1328)
>>>         at
>>> org.apache.hadoop.hive.ql.io.orc.WriterImpl.flushStripe(WriterImpl.java:1699)
>>>         at
>>> org.apache.hadoop.hive.ql.io.orc.WriterImpl.close(WriterImpl.java:1868)
>>>         at
>>> org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat$OrcRecordWriter.close(OrcOutputFormat.java:95)
>>>         at
>>> org.apache.hadoop.hive.ql.exec.FileSinkOperator$FSPaths.closeWriters(FileSinkOperator.java:181)
>>>         at
>>> org.apache.hadoop.hive.ql.exec.FileSinkOperator.closeOp(FileSinkOperator.java:866)
>>>         at
>>> org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:596)
>>>         at
>>> org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:613)
>>>         at
>>> org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:613)
>>>         at
>>> org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:613)
>>>         at
>>> org.apache.hadoop.hive.ql.exec.mr.ExecMapper.close(ExecMapper.java:207)
>>>         ... 8 more
>>>
>>>
>>>
>>>
>>>
>>> On Tue, Dec 17, 2013 at 2:31 AM, Prasanth Jayachandran <
>>> pjayachand...@hortonworks.com> wrote:
>>>
>>>> Bryan
>>>>
>>>> In either cases (source download or binary download) you do not need to
>>>> compile orc protobuf component. The java source from .proto files should be
>>>> already available when you download hive 0.12 release. I would recommend
>>>> re-downloading hive 0.12 binary release from
>>>> http://mirror.symnds.com/software/Apache/hive/hive-0.12.0/ and running
>>>> hive directly. After extracting the 
>>>> hive-0.12.0-bin.tar.gz<http://mirror.symnds.com/software/Apache/hive/hive-0.12.0/hive-0.12.0-bin.tar.gz>
>>>>  set
>>>> HIVE_HOME to the extracted directory and run hive. Let me know if you face
>>>> any issues.
>>>>
>>>> Thanks
>>>> Prasanth Jayachandran
>>>>
>>>> On Dec 16, 2013, at 5:19 PM, Bryan Jeffrey <bryan.jeff...@gmail.com>
>>>> wrote:
>>>>
>>>> Prasanth,
>>>>
>>>> I simply compiled the protobuf library, and then compiled the orc
>>>> protobuf component.  I did not recompile either Hive or custom UDFs/etc.
>>>>
>>>> Is a protobuf recompile the solution for this issue, or a dead end?
>>>>  Has this been seen before?  I looked for more feedback, but most of the
>>>> Orc issues were associated with Hive 0.11.0.
>>>>
>>>> I will try recompiling the 2.4 protobuf version shortly!
>>>>
>>>> Bryan
>>>>
>>>>
>>>> On Mon, Dec 16, 2013 at 8:02 PM, Prasanth Jayachandran <
>>>> pjayachand...@hortonworks.com> wrote:
>>>>
>>>>> Also what are you doing with steps 2 through 5? Compiling hive or your
>>>>> custom code?
>>>>>
>>>>> Thanks
>>>>> Prasanth Jayachandran
>>>>>
>>>>> On Dec 16, 2013, at 4:55 PM, Bryan Jeffrey <bryan.jeff...@gmail.com>
>>>>> wrote:
>>>>>
>>>>> Prasanth,
>>>>>
>>>>> I am running Hive 0.12.0 downloaded from the Apache Hive site.  I did
>>>>> not compile it.  I downloaded protobuf 2.5.0 earlier today from the Google
>>>>> Code site.  I compiled it via the following steps:
>>>>> (1) ./configure && make (to compile the C code)
>>>>> (2) protoc --java_out=src/main/java -I../src 
>>>>> ../src/google/protobuf/descriptor.proto
>>>>> ../src/google/protobuf/orc.proto
>>>>> (3) Compiled the org/apache/... directory via javac
>>>>> (4) Created jar via jar -cf protobuf-java-2.4.1.jar org
>>>>> (5) Copied my protobuf-java-2.4.1.jar over the one in hive-0.12.0/lib
>>>>> (6) Restarted hive
>>>>>
>>>>> Same results before/after protobuf modification.
>>>>>
>>>>> Bryan
>>>>>
>>>>>
>>>>> On Mon, Dec 16, 2013 at 7:34 PM, Prasanth Jayachandran <
>>>>> pjayachand...@hortonworks.com> wrote:
>>>>>
>>>>>> What version of protobuf are you using? Are you compiling hive from
>>>>>> source?
>>>>>>
>>>>>>  Thanks
>>>>>> Prasanth Jayachandran
>>>>>>
>>>>>> On Dec 16, 2013, at 4:30 PM, Bryan Jeffrey <bryan.jeff...@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>   Hello.
>>>>>>
>>>>>> Running the following version of Hadoop: hadoop-2.2.0
>>>>>> Running the following version of Hive: hive-0.12.0
>>>>>>
>>>>>> I have a simple test system setup with (2) datanodes/node manager and
>>>>>> (1) namenode/resource manager.  Hive is running on the namenode, and
>>>>>> contacting a MySQL database for metastore.
>>>>>>
>>>>>> I have created a small table 'from_text' as follows:
>>>>>>
>>>>>> [server:10001] hive> describe from_text;
>>>>>> foo                     int                     None
>>>>>> bar                     int                     None
>>>>>> boo                     string                  None
>>>>>>
>>>>>>
>>>>>> [server:10001] hive> select * from from_text;
>>>>>> 1       2       Hello
>>>>>> 2       3       World
>>>>>>
>>>>>> I go to insert the data into my Orc table, 'orc_table':
>>>>>>
>>>>>> [server:10001] hive> describe orc_test;
>>>>>> foo                     int                     from deserializer
>>>>>> bar                     int                     from deserializer
>>>>>> boo                     string                  from deserializer
>>>>>>
>>>>>>
>>>>>> The job runs, but fails to complete with the following errors (see
>>>>>> below).  This seems to be the exact example covered in the example here:
>>>>>>
>>>>>>
>>>>>> http://hortonworks.com/blog/orcfile-in-hdp-2-better-compression-better-performance/
>>>>>>
>>>>>> I took a few minutes to recompile the protbuf library as several
>>>>>> other problems mentioned that Hive 0.12 did not have the protobuf library
>>>>>> updated. That did not remedy the problem.  Any ideas?
>>>>>>
>>>>>>
>>>>>> [server:10001] hive> insert into table orc_test select * from
>>>>>> from_text;
>>>>>> [Hive Error]: Query returned non-zero code: 2, cause: FAILED:
>>>>>> Execution Error, return code 2 from
>>>>>> org.apache.hadoop.hive.ql.exec.mr.MapRedTask
>>>>>>
>>>>>>
>>>>>> Diagnostic Messages for this Task:
>>>>>> Error: java.lang.RuntimeException: Hive Runtime Error while closing
>>>>>> operators
>>>>>>         at
>>>>>> org.apache.hadoop.hive.ql.exec.mr.ExecMapper.close(ExecMapper.java:240)
>>>>>>         at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:61)
>>>>>>         at
>>>>>> org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:429)
>>>>>>         at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341)
>>>>>>         at
>>>>>> org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:162)
>>>>>>         at java.security.AccessController.doPrivileged(Native Method)
>>>>>>         at javax.security.auth.Subject.doAs(Subject.java:396)
>>>>>>         at
>>>>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491)
>>>>>>         at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:157)
>>>>>> Caused by: java.lang.UnsupportedOperationException: This is supposed
>>>>>> to be overridden by subclasses.
>>>>>>         at
>>>>>> com.google.protobuf.GeneratedMessage.getUnknownFields(GeneratedMessage.java:180)
>>>>>>         at
>>>>>> org.apache.hadoop.hive.ql.io.orc.OrcProto$ColumnStatistics.getSerializedSize(OrcProto.java:3046)
>>>>>>          at
>>>>>> com.google.protobuf.CodedOutputStream.computeMessageSizeNoTag(CodedOutputStream.java:749)
>>>>>>         at
>>>>>> com.google.protobuf.CodedOutputStream.computeMessageSize(CodedOutputStream.java:530)
>>>>>>         at
>>>>>> org.apache.hadoop.hive.ql.io.orc.OrcProto$RowIndexEntry.getSerializedSize(OrcProto.java:4129)
>>>>>>         at
>>>>>> com.google.protobuf.CodedOutputStream.computeMessageSizeNoTag(CodedOutputStream.java:749)
>>>>>>         at
>>>>>> com.google.protobuf.CodedOutputStream.computeMessageSize(CodedOutputStream.java:530)
>>>>>>         at
>>>>>> org.apache.hadoop.hive.ql.io.orc.OrcProto$RowIndex.getSerializedSize(OrcProto.java:4641)
>>>>>>         at
>>>>>> com.google.protobuf.AbstractMessageLite.writeTo(AbstractMessageLite.java:75)
>>>>>>         at
>>>>>> org.apache.hadoop.hive.ql.io.orc.WriterImpl$TreeWriter.writeStripe(WriterImpl.java:548)
>>>>>>         at
>>>>>> org.apache.hadoop.hive.ql.io.orc.WriterImpl$StructTreeWriter.writeStripe(WriterImpl.java:1328)
>>>>>>         at
>>>>>> org.apache.hadoop.hive.ql.io.orc.WriterImpl.flushStripe(WriterImpl.java:1699)
>>>>>>         at
>>>>>> org.apache.hadoop.hive.ql.io.orc.WriterImpl.close(WriterImpl.java:1868)
>>>>>>         at
>>>>>> org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat$OrcRecordWriter.close(OrcOutputFormat.java:95)
>>>>>>         at
>>>>>> org.apache.hadoop.hive.ql.exec.FileSinkOperator$FSPaths.closeWriters(FileSinkOperator.java:181)
>>>>>>         at
>>>>>> org.apache.hadoop.hive.ql.exec.FileSinkOperator.closeOp(FileSinkOperator.java:866)
>>>>>>         at
>>>>>> org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:596)
>>>>>>         at
>>>>>> org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:613)
>>>>>>         at
>>>>>> org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:613)
>>>>>>         at
>>>>>> org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:613)
>>>>>>         at
>>>>>> org.apache.hadoop.hive.ql.exec.mr.ExecMapper.close(ExecMapper.java:207)
>>>>>>         ... 8 more
>>>>>>
>>>>>>
>>>>>>
>>>>>> CONFIDENTIALITY NOTICE
>>>>>> NOTICE: This message is intended for the use of the individual or
>>>>>> entity to which it is addressed and may contain information that is
>>>>>> confidential, privileged and exempt from disclosure under applicable law.
>>>>>> If the reader of this message is not the intended recipient, you are 
>>>>>> hereby
>>>>>> notified that any printing, copying, dissemination, distribution,
>>>>>> disclosure or forwarding of this communication is strictly prohibited. If
>>>>>> you have received this communication in error, please contact the sender
>>>>>> immediately and delete it from your system. Thank You.
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> CONFIDENTIALITY NOTICE
>>>>> NOTICE: This message is intended for the use of the individual or
>>>>> entity to which it is addressed and may contain information that is
>>>>> confidential, privileged and exempt from disclosure under applicable law.
>>>>> If the reader of this message is not the intended recipient, you are 
>>>>> hereby
>>>>> notified that any printing, copying, dissemination, distribution,
>>>>> disclosure or forwarding of this communication is strictly prohibited. If
>>>>> you have received this communication in error, please contact the sender
>>>>> immediately and delete it from your system. Thank You.
>>>>>
>>>>
>>>>
>>>>
>>>> CONFIDENTIALITY NOTICE
>>>> NOTICE: This message is intended for the use of the individual or
>>>> entity to which it is addressed and may contain information that is
>>>> confidential, privileged and exempt from disclosure under applicable law.
>>>> If the reader of this message is not the intended recipient, you are hereby
>>>> notified that any printing, copying, dissemination, distribution,
>>>> disclosure or forwarding of this communication is strictly prohibited. If
>>>> you have received this communication in error, please contact the sender
>>>> immediately and delete it from your system. Thank You.
>>>>
>>>
>>>
>>
>> CONFIDENTIALITY NOTICE
>> NOTICE: This message is intended for the use of the individual or entity
>> to which it is addressed and may contain information that is confidential,
>> privileged and exempt from disclosure under applicable law. If the reader
>> of this message is not the intended recipient, you are hereby notified that
>> any printing, copying, dissemination, distribution, disclosure or
>> forwarding of this communication is strictly prohibited. If you have
>> received this communication in error, please contact the sender immediately
>> and delete it from your system. Thank You.
>
>

Reply via email to