If you have already updated the log4j.properties, and it still could not work, then I suggest to log in the Yarn NodeManager machine and check the log4j.properties in the container workdir is correct. Also you could have a look at the java dynamic options are correctly set.
I think it should work if the log4j.properties and java dynamic options are set correctly. BTW, could you share the new yarn logs? Best, Yang Diwakar Jha <diwakar.n...@gmail.com> 于2020年11月2日周一 下午4:32写道: > > > Hi Yang, > > Thank you so much for taking a look at the log files. I changed my > log4j.properties. Below is the actual file that I got from EMR 6.1.0 > distribution of flink 1.11. I observed that it is different from Flink 1.11 > that i downloaded so i changed it. Still I didn't see any logs. > > *Actual* > log4j.rootLogger=INFO,file > > # Log all infos in the given file > log4j.appender.file=org.apache.log4j.FileAppender > log4j.appender.file.file=${log.file} > log4j.appender.file.append=false > log4j.appender.file.layout=org.apache.log4j.PatternLayout > log4j.appender.file.layout.ConversionPattern=%d{yyyy-MM-dd HH:mm:ss,SSS} > %-5p %-60c %x - %m%n > > # suppress the irrelevant (wrong) warnings from the netty channel handler > log4j.logger.org.jboss.netty.channel.DefaultChannelPipeline=ERROR,file > > > *modified : *commented the above and added new logging from actual flink > application log4.properties file > > #log4j.rootLogger=INFO,file > > # Log all infos in the given file > #log4j.appender.file=org.apache.log4j.FileAppender > #log4j.appender.file.file=${log.file} > #log4j.appender.file.append=false > #log4j.appender.file.layout=org.apache.log4j.PatternLayout > #log4j.appender.file.layout.ConversionPattern=%d{yyyy-MM-dd HH:mm:ss,SSS} > %-5p %-60c %x - %m%n > > # suppress the irrelevant (wrong) warnings from the netty channel handler > #log4j.logger.org.jboss.netty.channel.DefaultChannelPipeline=ERROR,file > > # This affects logging for both user code and Flink > rootLogger.level = INFO > rootLogger.appenderRef.file.ref = MainAppender > > # Uncomment this if you want to _only_ change Flink's logging > #logger.flink.name = org.apache.flink > #logger.flink.level = INFO > > # The following lines keep the log level of common libraries/connectors on > # log level INFO. The root logger does not override this. You have to > manually > # change the log levels here. > logger.akka.name = akka > logger.akka.level = INFO > logger.kafka.name= org.apache.kafka > logger.kafka.level = INFO > logger.hadoop.name = org.apache.hadoop > logger.hadoop.level = INFO > logger.zookeeper.name = org.apache.zookeeper > logger.zookeeper.level = INFO > > # Log all infos in the given file > appender.main.name = MainAppender > appender.main.type = File > appender.main.append = false > appender.main.fileName = ${sys:log.file} > appender.main.layout.type = PatternLayout > appender.main.layout.pattern = %d{yyyy-MM-dd HH:mm:ss,SSS} %-5p %-60c %x - > %m%n > > # Suppress the irrelevant (wrong) warnings from the Netty channel handler > logger.netty.name = > org.apache.flink.shaded.akka.org.jboss.netty.channel.DefaultChannelPipeline > logger.netty.level = OFF > > ********************************** > I also think its related to the log4j setting but I'm not able to figure > it out. > Please let me know if you want any other log files or configuration. > > Thanks. > > On Sun, Nov 1, 2020 at 10:06 PM Yang Wang <danrtsey...@gmail.com> wrote: > >> Hi Diwakar Jha, >> >> From the logs you have provided, everything seems working as expected. >> The JobManager and TaskManager >> java processes have been started with correct dynamic options, especially >> for the logging. >> >> Could you share the content of $FLINK_HOME/conf/log4j.properties? I think >> there's something wrong with the >> log4j config file. For example, it is a log4j1 format. But we are using >> log4j2 in Flink 1.11. >> >> >> Best, >> Yang >> >> Diwakar Jha <diwakar.n...@gmail.com> 于2020年11月2日周一 上午1:57写道: >> >>> Hi >>> I'm running Flink 1.11 on EMR 6.1.0. I can see my job is running fine >>> but i'm not seeing any taskmanager/jobmanager logs. >>> I see the below error in stdout. >>> 18:29:19.834 [flink-akka.actor.default-dispatcher-28] ERROR >>> org.apache.flink.runtime.rest.handler.taskmanager. >>> TaskManagerLogFileHandler - Failed to transfer file from TaskExecutor >>> container_1604033334508_0001_01_000004. >>> java.util.concurrent.CompletionException: org.apache.flink.util. >>> FlinkException: The file LOG does not exist on the TaskExecutor. >>> >>> I'm stuck at this step for a couple of days now and not able to migrate >>> to Flink 1.11. I would appreciate it if anyone can help me. >>> i have the following setup : >>> a) i'm deploying flink using yarn. I have attached yarn application id >>> logs. >>> c) stsd setup >>> >>> metrics.reporters: stsd >>> metrics.reporter.stsd.factory.class: >>> org.apache.flink.metrics.statsd.StatsDReporterFactory >>> metrics.reporter.stsd.host: localhost >>> metrics.reporter.stsd.port: 8125 >>> >>>