Wikimedia-analytics will be the channel
On Sun, Sep 10, 2017 at 5:14 AM Adrian Bielefeldt < [email protected]> wrote: > Hi Nuria, > > thanks for the reply. Am I correct in assuming you mean the channel > irc://irc.freenode.net/wikidata, or is there a specific analytics channel? > > Greetings, > > Adrian > > On 09/10/2017 03:29 AM, Nuria Ruiz wrote: > > Probably irc is better suited for this > type of question, you have an issue with conversion of types: > > " Caused by: java.lang.RuntimeException: Hive internal error: conversion > of string to array<string>not supported yet." > > > Selecting any of the hours that do not work brings out the same error, you > can inspect the table data by looking at files in: > /wmf/data/wmf/wdqs_extract/year=2017/month=8/day=23/hour=9/ > and try to find the records that have faulty data see if you can exclude > them from your query. > > You can get data with: hdfs dfs -cat > /wmf/data/wmf/wdqs_extract/year=2017/month=8/day=23/hour=8/ and look at it > with the parquet-tools command line. As I mentioned earlier probably > pinging us on irc would be best. > > On Sat, Sep 9, 2017 at 7:42 AM, Adrian Bielefeldt < > [email protected]> wrote: > >> Hello everyone, >> >> I'm having trouble with the entries in wmf.wdqs_extract on 23.08.2017, >> specifically the hours 9-17. The call >> >> hive -e "insert overwrite local directory 'temp' row format delimited >> fields terminated by '\t' select uri_query, uri_path, user_agent, ts, >> agent_type, hour, http_status from wmf.wdqs_extract where >> uri_query<>\"\" and year='2017' and month='8' and day='23' and hour='8'" >> >> works fine, as do all hours in 0-8 and 18-23, but if I change hour to >> anything between 9 and 17 it fails with the following message: >> >> Error: java.lang.RuntimeException: Error in configuring >> object >> >> at >> >> org.apache.hadoop.util.ReflectionUtils.setJobConf(ReflectionUtils.java:109) >> >> at >> org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:75) >> >> at >> >> org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:133) >> >> at >> org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:449) >> >> at >> org.apache.hadoop.mapred.MapTask.run(MapTask.java:343) >> >> at >> org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164) >> >> at java.security.AccessController.doPrivileged(Native >> Method) >> >> at >> javax.security.auth.Subject.doAs(Subject.java:421) >> >> at >> >> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1796) >> >> at >> org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158) >> >> Caused by: java.lang.reflect.InvocationTargetException >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> at >> >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >> at >> >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >> at java.lang.reflect.Method.invoke(Method.java:606) >> at >> >> org.apache.hadoop.util.ReflectionUtils.setJobConf(ReflectionUtils.java:106) >> ... 9 more >> Caused by: java.lang.RuntimeException: Error in configuring object >> at >> >> org.apache.hadoop.util.ReflectionUtils.setJobConf(ReflectionUtils.java:109) >> at >> org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:75) >> at >> >> org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:133) >> at org.apache.hadoop.mapred.MapRunner.configure(MapRunner.java:38) >> ... 14 more >> Caused by: java.lang.reflect.InvocationTargetException >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> at >> >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >> at >> >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >> at java.lang.reflect.Method.invoke(Method.java:606) >> at >> >> org.apache.hadoop.util.ReflectionUtils.setJobConf(ReflectionUtils.java:106) >> ... 17 more >> Caused by: java.lang.RuntimeException: Map operator initialization failed >> at >> >> org.apache.hadoop.hive.ql.exec.mr.ExecMapper.configure(ExecMapper.java:147) >> ... 22 more >> Caused by: java.lang.RuntimeException: Hive internal error: conversion >> of string to array<string>not supported yet. >> at >> >> org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorConverters$ListConverter.<init>(ObjectInspectorConverters.java:313) >> at >> >> org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorConverters.getConverter(ObjectInspectorConverters.java:158) >> at >> >> org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorConverters$StructConverter.<init>(ObjectInspectorConverters.java:374) >> at >> >> org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorConverters.getConverter(ObjectInspectorConverters.java:155) >> at >> >> org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorConverters$StructConverter.<init>(ObjectInspectorConverters.java:374) >> at >> >> org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorConverters.getConverter(ObjectInspectorConverters.java:155) >> at >> >> org.apache.hadoop.hive.ql.exec.MapOperator.initObjectInspector(MapOperator.java:199) >> at >> >> org.apache.hadoop.hive.ql.exec.MapOperator.setChildren(MapOperator.java:355) >> at >> >> org.apache.hadoop.hive.ql.exec.mr.ExecMapper.configure(ExecMapper.java:116) >> ... 22 more >> >> Can anybode help me out with this? >> >> Greetings, >> >> Adrian >> >> >> _______________________________________________ >> Analytics mailing list >> [email protected] >> https://lists.wikimedia.org/mailman/listinfo/analytics >> > > > > _______________________________________________ > Analytics mailing > [email protected]https://lists.wikimedia.org/mailman/listinfo/analytics > > > _______________________________________________ > Analytics mailing list > [email protected] > https://lists.wikimedia.org/mailman/listinfo/analytics >
_______________________________________________ Analytics mailing list [email protected] https://lists.wikimedia.org/mailman/listinfo/analytics
