wangxiaobaidu11 edited a comment on pull request #10920:
URL: https://github.com/apache/druid/pull/10920#issuecomment-984461681
hi @JulianJaffePinterest, I hope you can get out of your sad mood soon and
take good care of yourself! Last month I tested your code. I met a problem
described below:
when run
“dataset.write().format("druid").mode(SaveMode.Overwrite).options(map).save();“
`21/12/02 15:19:40 ERROR Executor: Exception in task 0.0 in stage 0.0 (TID 0)
noc.com.fasterxml.jackson.databind.exc.InvalidTypeIdException: Could not
resolve type id 'thetaSketch' as a subtype of
`org.apache.druid.query.aggregation.AggregatorFactory`: known type ids =
[cardinality, count, doubleAny, doubleFirst, doubleLast, doubleMax, doubleMean,
doubleMin, doubleSum, filtered, floatAny, floatFirst, floatLast, floatMax,
floatMin, floatSum, grouping, histogram, hyperUnique, javascript, longAny,
longFirst, longLast, longMax, longMin, longSum, stringAny, stringFirst,
stringFirstFold, stringLast, stringLastFold]
at [Source: (String)"[
{ "type": "count", "name": "count" },
{ "type": "longSum", "name": "sum_metric1", "fieldName": "sum_metric1" },
{ "type": "longSum", "name": "sum_metric2", "fieldName": "sum_metric2" },
{ "type": "doubleSum", "name": "sum_metric3", "fieldName": "sum_metric3" },
{ "type": "floatSum", "name": "sum_metric4", "fieldName": "sum_metric4" },
{ "type": "thetaSketch", "name": "uniq_id1", "fieldName": "uniq_id1",
"isInputThetaSketch": true }
]"; line: 7, column: 13] (through reference chain: java.lang.Object[][5])
at
noc.com.fasterxml.jackson.databind.exc.InvalidTypeIdException.from(InvalidTypeIdException.java:43)
at
noc.com.fasterxml.jackson.databind.DeserializationContext.invalidTypeIdException(DeserializationContext.java:1761)
at
noc.com.fasterxml.jackson.databind.DeserializationContext.handleUnknownTypeId(DeserializationContext.java:1268)
at
noc.com.fasterxml.jackson.databind.jsontype.impl.TypeDeserializerBase._handleUnknownTypeId(TypeDeserializerBase.java:290)
at
noc.com.fasterxml.jackson.databind.jsontype.impl.TypeDeserializerBase._findDeserializer(TypeDeserializerBase.java:162)
at
noc.com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer._deserializeTypedForId(AsPropertyTypeDeserializer.java:113)
at
noc.com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer.deserializeTypedFromObject(AsPropertyTypeDeserializer.java:97)
at
noc.com.fasterxml.jackson.databind.deser.AbstractDeserializer.deserializeWithType(AbstractDeserializer.java:254)
at
noc.com.fasterxml.jackson.databind.deser.std.ObjectArrayDeserializer.deserialize(ObjectArrayDeserializer.java:197)
at
noc.com.fasterxml.jackson.databind.deser.std.ObjectArrayDeserializer.deserialize(ObjectArrayDeserializer.java:21)
at
noc.com.fasterxml.jackson.databind.ObjectMapper._readMapAndClose(ObjectMapper.java:4218)
at
noc.com.fasterxml.jackson.databind.ObjectMapper.readValue(ObjectMapper.java:3214)
at
noc.com.fasterxml.jackson.databind.ObjectMapper.readValue(ObjectMapper.java:3197)
at
org.apache.druid.spark.v2.writer.DruidDataWriterFactory$.createDataSchemaFromConfiguration(DruidDataWriterFactory.scala:99)
at
org.apache.druid.spark.v2.writer.DruidDataWriterFactory.createDataWriter(DruidDataWriterFactory.scala:70)
at
org.apache.spark.sql.execution.datasources.v2.DataWritingSparkTask$.run(WriteToDataSourceV2Exec.scala:113)
at
org.apache.spark.sql.execution.datasources.v2.WriteToDataSourceV2Exec.$anonfun$doExecute$2(WriteToDataSourceV2Exec.scala:67)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
at org.apache.spark.scheduler.Task.run(Task.scala:121)
at
org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:411)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1360)
at
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:414)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
21/12/02 15:19:40 WARN TaskSetManager: Lost task 0.0 in stage 0.0 (TID 0,
localhost, executor driver):
noc.com.fasterxml.jackson.databind.exc.InvalidTypeIdException: Could not
resolve type id 'thetaSketch' as a subtype of
`org.apache.druid.query.aggregation.AggregatorFactory`: known type ids =
[cardinality, count, doubleAny, doubleFirst, doubleLast, doubleMax, doubleMean,
doubleMin, doubleSum, filtered, floatAny, floatFirst, floatLast, floatMax,
floatMin, floatSum, grouping, histogram, hyperUnique, javascript, longAny,
longFirst, longLast, longMax, longMin, longSum, stringAny, stringFirst,
stringFirstFold, stringLast, stringLastFold]
at [Source: (String)"[
{ "type": "count", "name": "count" },
{ "type": "longSum", "name": "sum_metric1", "fieldName": "sum_metric1" },
{ "type": "longSum", "name": "sum_metric2", "fieldName": "sum_metric2" },
{ "type": "doubleSum", "name": "sum_metric3", "fieldName": "sum_metric3" },
{ "type": "floatSum", "name": "sum_metric4", "fieldName": "sum_metric4" },
{ "type": "thetaSketch", "name": "uniq_id1", "fieldName": "uniq_id1",
"isInputThetaSketch": true }
]"; line: 7, column: 13] (through reference chain: java.lang.Object[][5])
at
noc.com.fasterxml.jackson.databind.exc.InvalidTypeIdException.from(InvalidTypeIdException.java:43)
at
noc.com.fasterxml.jackson.databind.DeserializationContext.invalidTypeIdException(DeserializationContext.java:1761)
at
noc.com.fasterxml.jackson.databind.DeserializationContext.handleUnknownTypeId(DeserializationContext.java:1268)
at
noc.com.fasterxml.jackson.databind.jsontype.impl.TypeDeserializerBase._handleUnknownTypeId(TypeDeserializerBase.java:290)
at
noc.com.fasterxml.jackson.databind.jsontype.impl.TypeDeserializerBase._findDeserializer(TypeDeserializerBase.java:162)
at
noc.com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer._deserializeTypedForId(AsPropertyTypeDeserializer.java:113)
at
noc.com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer.deserializeTypedFromObject(AsPropertyTypeDeserializer.java:97)
at
noc.com.fasterxml.jackson.databind.deser.AbstractDeserializer.deserializeWithType(AbstractDeserializer.java:254)
at
noc.com.fasterxml.jackson.databind.deser.std.ObjectArrayDeserializer.deserialize(ObjectArrayDeserializer.java:197)
at
noc.com.fasterxml.jackson.databind.deser.std.ObjectArrayDeserializer.deserialize(ObjectArrayDeserializer.java:21)
at
noc.com.fasterxml.jackson.databind.ObjectMapper._readMapAndClose(ObjectMapper.java:4218)
at
noc.com.fasterxml.jackson.databind.ObjectMapper.readValue(ObjectMapper.java:3214)
at
noc.com.fasterxml.jackson.databind.ObjectMapper.readValue(ObjectMapper.java:3197)
at
org.apache.druid.spark.v2.writer.DruidDataWriterFactory$.createDataSchemaFromConfiguration(DruidDataWriterFactory.scala:99)
at
org.apache.druid.spark.v2.writer.DruidDataWriterFactory.createDataWriter(DruidDataWriterFactory.scala:70)
at
org.apache.spark.sql.execution.datasources.v2.DataWritingSparkTask$.run(WriteToDataSourceV2Exec.scala:113)
at
org.apache.spark.sql.execution.datasources.v2.WriteToDataSourceV2Exec.$anonfun$doExecute$2(WriteToDataSourceV2Exec.scala:67)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
at org.apache.spark.scheduler.Task.run(Task.scala:121)
at
org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:411)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1360)
at
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:414)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
`
> Hey @jihoonson, I had some unexpected and unfortunate personal/familial
crises to deal with these past few months. While they're not entirely in the
rear-view mirror, I should have more time again to push this to the finish
line. I've opened #11823 with the next chunk of code (the reading half of the
connector). Please let me know if you think the PR is still too big; I couldn't
find a good place to split it that wouldn't require a reviewer to know the rest
of the code anyway.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]