Probably irc is better suited for this
type of question, you have an issue with conversion of types:

" Caused by: java.lang.RuntimeException: Hive internal error: conversion
of string to array<string>not supported yet."


Selecting any of the hours that do not work brings out the same error, you
can inspect the table data by looking at files in:
/wmf/data/wmf/wdqs_extract/year=2017/month=8/day=23/hour=9/
and try to find the records that have faulty data see if you can exclude
them from your query.

You can get data with: hdfs dfs -cat
/wmf/data/wmf/wdqs_extract/year=2017/month=8/day=23/hour=8/ and look at it
with the parquet-tools command line. As I mentioned earlier probably
pinging us on irc would be best.

On Sat, Sep 9, 2017 at 7:42 AM, Adrian Bielefeldt <
[email protected]> wrote:

> Hello everyone,
>
> I'm having trouble with the entries in wmf.wdqs_extract on 23.08.2017,
> specifically the hours 9-17. The call
>
> hive -e "insert overwrite local directory 'temp' row format delimited
> fields terminated by '\t' select uri_query, uri_path, user_agent, ts,
> agent_type, hour, http_status from wmf.wdqs_extract where
> uri_query<>\"\" and year='2017' and month='8' and day='23' and hour='8'"
>
> works fine, as do all hours in 0-8 and 18-23, but if I change hour to
> anything between 9 and 17 it fails with the following message:
>
> Error: java.lang.RuntimeException: Error in configuring
> object
>
>         at
> org.apache.hadoop.util.ReflectionUtils.setJobConf(
> ReflectionUtils.java:109)
>
>         at
> org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:75)
>
>         at
> org.apache.hadoop.util.ReflectionUtils.newInstance(
> ReflectionUtils.java:133)
>
>         at
> org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:449)
>
>         at
> org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
>
>         at
> org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164)
>
>         at java.security.AccessController.doPrivileged(Native
> Method)
>
>         at
> javax.security.auth.Subject.doAs(Subject.java:421)
>
>         at
> org.apache.hadoop.security.UserGroupInformation.doAs(
> UserGroupInformation.java:1796)
>
>         at
> org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
>
> Caused by: java.lang.reflect.InvocationTargetException
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:
> 57)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(
> DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:606)
>         at
> org.apache.hadoop.util.ReflectionUtils.setJobConf(
> ReflectionUtils.java:106)
>         ... 9 more
> Caused by: java.lang.RuntimeException: Error in configuring object
>         at
> org.apache.hadoop.util.ReflectionUtils.setJobConf(
> ReflectionUtils.java:109)
>         at
> org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:75)
>         at
> org.apache.hadoop.util.ReflectionUtils.newInstance(
> ReflectionUtils.java:133)
>         at org.apache.hadoop.mapred.MapRunner.configure(MapRunner.java:38)
>         ... 14 more
> Caused by: java.lang.reflect.InvocationTargetException
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:
> 57)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(
> DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:606)
>         at
> org.apache.hadoop.util.ReflectionUtils.setJobConf(
> ReflectionUtils.java:106)
>         ... 17 more
> Caused by: java.lang.RuntimeException: Map operator initialization failed
>         at
> org.apache.hadoop.hive.ql.exec.mr.ExecMapper.configure(
> ExecMapper.java:147)
>         ... 22 more
> Caused by: java.lang.RuntimeException: Hive internal error: conversion
> of string to array<string>not supported yet.
>         at
> org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorConverters$
> ListConverter.<init>(ObjectInspectorConverters.java:313)
>         at
> org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorConverters.
> getConverter(ObjectInspectorConverters.java:158)
>         at
> org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorConverters$
> StructConverter.<init>(ObjectInspectorConverters.java:374)
>         at
> org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorConverters.
> getConverter(ObjectInspectorConverters.java:155)
>         at
> org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorConverters$
> StructConverter.<init>(ObjectInspectorConverters.java:374)
>         at
> org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorConverters.
> getConverter(ObjectInspectorConverters.java:155)
>         at
> org.apache.hadoop.hive.ql.exec.MapOperator.initObjectInspector(
> MapOperator.java:199)
>         at
> org.apache.hadoop.hive.ql.exec.MapOperator.setChildren(
> MapOperator.java:355)
>         at
> org.apache.hadoop.hive.ql.exec.mr.ExecMapper.configure(
> ExecMapper.java:116)
>         ... 22 more
>
> Can anybode help me out with this?
>
> Greetings,
>
> Adrian
>
>
> _______________________________________________
> Analytics mailing list
> [email protected]
> https://lists.wikimedia.org/mailman/listinfo/analytics
>
_______________________________________________
Analytics mailing list
[email protected]
https://lists.wikimedia.org/mailman/listinfo/analytics

Reply via email to