samarthjain opened a new issue #7482: HadoopIndexer job with input as the datasource and configured segments table doesn't work URL: https://github.com/apache/incubator-druid/issues/7482 ### Affected Version 0.14, 0.13, 0.12 The Druid version where the problem was encountered. 0.12 ### Description I was trying out the hadoop based reingestion job http://druid.io/docs/latest/ingestion/update-existing-data.html which uses the datasource itself as the input. When I ran the job, it failed because it was trying to read segment metadata from druid_segments table and not from the table, customprefix_segments, I specified in the metadataUpdateSpec. "metadataUpdateSpec": { "connectURI": "jdbc:mysql...", "password": "XXXXXXX", "segmentTable": "customprefix_segments", "type": "mysql", "user": "XXXXXXXX" }, Looking at the code, I see that the segmentTable specified in the spec is actually passed in as pending_segments table (3rd param is for pending_segments and 4th param is for segments table) https://github.com/apache/incubator-druid/blob/master/indexing-hadoop/src/main/java/org/apache/druid/indexer/updater/MetadataStorageUpdaterJobSpec.java#L92 This code has been around forever though, so would have to be careful before simply switching the order of param values.
---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: [email protected] With regards, Apache Git Services --------------------------------------------------------------------- To unsubscribe, e-mail: [email protected] For additional commands, e-mail: [email protected]
