cdmikechen commented on a change in pull request #4451:
URL: https://github.com/apache/hudi/pull/4451#discussion_r780733010
##########
File path:
hudi-kafka-connect/src/main/java/org/apache/hudi/connect/utils/KafkaConnectUtils.java
##########
@@ -65,6 +70,52 @@
private static final Logger LOG =
LogManager.getLogger(KafkaConnectUtils.class);
private static final String HOODIE_CONF_PREFIX = "hoodie.";
+ public static final String HADOOP_CONF_DIR = "HADOOP_CONF_DIR";
+ public static final String HADOOP_HOME = "HADOOP_HOME";
+ private static final List<Path> DEFAULT_HADOOP_CONF_FILES;
+
+ static {
+ DEFAULT_HADOOP_CONF_FILES = new ArrayList<>();
+ try {
+ String hadoopConfigPath = System.getenv(HADOOP_CONF_DIR);
+ String hadoopHomePath = System.getenv(HADOOP_HOME);
+ DEFAULT_HADOOP_CONF_FILES.addAll(getHadoopConfigFiles(hadoopConfigPath,
hadoopHomePath));
+ if (!DEFAULT_HADOOP_CONF_FILES.isEmpty()) {
+ LOG.info(String.format("Found Hadoop default config files %s",
DEFAULT_HADOOP_CONF_FILES));
+ }
Review comment:
@codope
My idea was: because the hadoop environment is set by default, users need to
know that kafka-connect has obtained the correct information. Thus, the users
can judge whether the environment that they set is wrong, or manually declare
the hadoop configuration path when registering the task.
Because the default log level is info, it may be easier for users to output
information with info.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]