hehuiyuan commented on issue #24270: [SPARK-27343][KAFKA][SS]Avoid hardcoded 
for spark-sql-kafka-0-10
URL: https://github.com/apache/spark/pull/24270#issuecomment-490957492
 
 
   > 
   > 
   > @gaborgsomogyi do you have any other comments besides the one above and 
the two at [#24270 
(comment)](https://github.com/apache/spark/pull/24270#discussion_r274827760) ?
   > 
   > What's the action for the one above, prefix the keys with "spark.sql."? 
The Kafka configs start with "spark.kafka." _except_ for 
"spark.sql.kafkaConsumerCache.capacity" which sort of looks like an error. That 
config isn't documented anywhere, but this was noticed and brought up at 
https://issues.apache.org/jira/browse/SPARK-25466 and #22138 proposes to keep 
the spark.sql. prefix I think.
   > 
   > I don't know if we want to change it here, but seems like we want to move 
away from spark.sql. prefixes here if anything?
   
   
   
   > What's the action for the one above, prefix the keys with "spark.sql."? 
The Kafka configs start with "spark.kafka." _except_ for 
"spark.sql.kafkaConsumerCache.capacity" which sort of looks like an error.
   
   In spark streaming kafka, this parameter 
spark.streaming.kafka.consumer.cache.maxCapacity is used.
   I think that might want to distinguish it from this parameter.
   
   

----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to