samarthjain opened a new issue #7482: HadoopIndexer job with input as the 
datasource and configured segments table doesn't work
URL: https://github.com/apache/incubator-druid/issues/7482
 
 
   ### Affected Version
   0.14, 0.13, 0.12
   The Druid version where the problem was encountered.
   0.12
   ### Description
   
   I was trying out the hadoop based reingestion job 
http://druid.io/docs/latest/ingestion/update-existing-data.html  which uses the 
datasource itself as the input. 
   
   When I ran the job, it failed because it was trying to read segment metadata 
from druid_segments table and not from the table, customprefix_segments, I 
specified in the metadataUpdateSpec.
   
   "metadataUpdateSpec": {
         "connectURI": "jdbc:mysql...",
         "password": "XXXXXXX",
         "segmentTable": "customprefix_segments",
         "type": "mysql",
         "user": "XXXXXXXX"
   },
   
   Looking at the code, I see that the segmentTable specified in the spec is 
actually passed in as pending_segments table (3rd param is for pending_segments 
and 4th param is for segments table)
   
https://github.com/apache/incubator-druid/blob/master/indexing-hadoop/src/main/java/org/apache/druid/indexer/updater/MetadataStorageUpdaterJobSpec.java#L92
   
   This code has been around forever though, so would have to be careful before 
simply switching the order of param values.
   

----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
[email protected]


With regards,
Apache Git Services

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to