[
https://issues.apache.org/jira/browse/AIRFLOW-6515?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Kaxil Naik resolved AIRFLOW-6515.
---------------------------------
Fix Version/s: 1.10.11
Resolution: Fixed
> log level of INFO/WARN when ERROR happened
> ------------------------------------------
>
> Key: AIRFLOW-6515
> URL: https://issues.apache.org/jira/browse/AIRFLOW-6515
> Project: Apache Airflow
> Issue Type: Bug
> Components: logging
> Affects Versions: 1.10.7
> Reporter: t oo
> Assignee: Will Hudgins
> Priority: Major
> Fix For: 1.10.11
>
>
> log level should be error on some (but there are false positives):
> grep -iE
> 'log\.(info|warn).*(error|exceptio|fail|unab|couldn|lost|gone|missing|not
> fou|abort|exit|could not)' -R *
> airflow/sensors/base_sensor_operator.py: self.log.info("Success
> criteria met. Exiting.")
> airflow/logging_config.py: log.warning('Unable to load the config,
> contains a configuration error.')
> airflow/operators/check_operator.py: self.log.warning("The
> following %s tests out of %s failed:", j, n)
> airflow/operators/sql_to_gcs.py: self.log.warning('Using default
> schema due to missing name or type. Please '
> airflow/operators/bash_operator.py: self.log.info('Command exited
> with return code %s', self.sub_process.returncode)
> airflow/serialization/serialized_objects.py: LOG.warning('Failed
> to stringify.', exc_info=True)
> airflow/providers/amazon/aws/operators/batch.py:
> self.log.info("AWS Batch Job has failed")
> airflow/providers/amazon/aws/hooks/s3.py:
> self.log.info(e.response["Error"]["Message"])
> airflow/providers/amazon/aws/hooks/s3.py:
> self.log.info(e.response["Error"]["Message"])
> airflow/utils/dag_processing.py: self.log.info("Exiting gracefully
> upon receiving signal %s", signum)
> airflow/utils/dag_processing.py: self.log.info("Exiting dag
> parsing loop as all files "
> airflow/utils/dag_processing.py: self.log.info("Failing jobs
> without heartbeat after %s", limit_dttm)
> airflow/utils/dag_processing.py: self.log.info("Waiting up to %s
> seconds for processes to exit...", timeout)
> airflow/utils/helpers.py: log.info("Process %s (%s) terminated with
> exit code %s", p, p.pid, p.returncode)
> airflow/models/dagrun.py: self.log.info('Marking run %s failed',
> self)
> airflow/models/dagrun.py: self.log.info('Deadlock; marking run %s
> failed', self)
> airflow/models/dagrun.py: self.log.warning("Failed to get
> task '{}' for dag '{}'. "
> airflow/gcp/sensors/gcs.py: self.log.warning("FAILURE: Inactivity
> Period passed, not enough objects found in %s", path)
> airflow/gcp/operators/spanner.py: self.log.info("The Cloud Spanner
> database was missing: "
> airflow/gcp/hooks/kubernetes_engine.py: self.log.info('Assuming
> Success: %s', error.message)
> airflow/gcp/hooks/kubernetes_engine.py: self.log.info('Assuming
> Success: %s', error.message)
> airflow/gcp/hooks/cloud_memorystore.py: self.log.info("Failovering
> Instance: %s", name)
> airflow/gcp/hooks/cloud_memorystore.py: self.log.info("Instance
> failovered: %s", name)
> airflow/gcp/hooks/bigquery.py: self.log.info(error_msg)
> airflow/gcp/hooks/bigtable.py: self.log.info("The instance '%s'
> does not exist in project '%s'. Exiting", instance_id,
> airflow/contrib/sensors/bash_sensor.py: self.log.info("Command
> exited with return code %s", sp.returncode)
> airflow/contrib/sensors/ftp_sensor.py: self.log.info('Ftp
> error encountered: %s', str(e))
> airflow/contrib/operators/azure_container_instances_operator.py:
> self.log.info("Container had exit code: %s", exit_code)
> airflow/contrib/operators/azure_container_instances_operator.py:
> self.log.info("Container exited with detail_status %s", detail_status)
> airflow/contrib/operators/azure_container_instances_operator.py:
> self.log.info("Azure provision failure")
> airflow/contrib/operators/winrm_operator.py: self.log.info("Hook
> not found, creating...")
> airflow/contrib/operators/docker_swarm_operator.py:
> self.log.info('Service status before exiting: %s', status)
> airflow/contrib/auth/backends/ldap_auth.py: log.warning("Unable to
> find group for %s %s", search_base, search_filter)
> airflow/contrib/auth/backends/ldap_auth.py: log.warning("""Missing
> attribute "%s" when looked-up in Ldap database.
> airflow/contrib/auth/backends/ldap_auth.py: log.warning("Parsing error
> when retrieving the user's group(s)."
> airflow/contrib/utils/sendgrid.py: log.warning('Failed to send out
> email with subject %s, status code: %s',
> airflow/contrib/example_dags/example_kubernetes_operator.py:
> log.warning("Could not import KubernetesPodOperator: " + str(e))
> airflow/contrib/hooks/gdrive_hook.py: self.log.info("Not found
> %s directory", current_folder)
> airflow/contrib/hooks/sqoop_hook.py: self.log.info("Command exited
> with return code %s", self.sp.returncode)
> airflow/contrib/hooks/salesforce_hook.py: log.warning("Could not
> convert field to timestamps: %s", column.name)
> airflow/contrib/hooks/qubole_check_hook.py: log.info("Qubole
> command not found")
> airflow/contrib/hooks/spark_submit_hook.py:
> self.log.info("Exception when attempting to kill Spark on K8s:")
> airflow/www/node_modules/eslint/lib/util/npm-util.js: log.info("Could
> not read package.json file. Please check that the file contains valid JSON.");
> airflow/www/node_modules/eslint/lib/config/config-initializer.js:
> log.info("Local ESLint installation not found.");
> grep: airflow/www/static/docs: No such file or directory
> airflow/www/api/experimental/endpoints.py: _log.info(error_message)
> airflow/www/api/experimental/endpoints.py: _log.info(error_message)
> airflow/www/api/experimental/endpoints.py: _log.info(error_message)
> airflow/executors/kubernetes_executor.py: self.log.info('Event: %s
> Failed', pod_id)
> airflow/executors/kubernetes_executor.py: self.log.warning("could
> not get try_number as an int: %s", labels.get('try_number', '1'))
> airflow/executors/kubernetes_executor.py:
> self.log.warning('ApiException when attempting to run task, re-queueing. '
> airflow/executors/base_executor.py: self.log.info("could not queue
> task %s", simple_task_instance.key)
> airflow/executors/debug_executor.py: self.log.info("Setting %s
> to %s", ti.key, State.UPSTREAM_FAILED)
> airflow/executors/debug_executor.py: self.log.info("Setting %s to
> %s", ti.key, State.UPSTREAM_FAILED)
> airflow/hooks/hive_hooks.py: self.log.info("Could not connect
> to %s:%s", conn.host, conn.port)
> airflow/stats.py: log.warning("Could not configure StatsClient: %s, using
> DummyStatsLogger instead.", e)
> airflow/task/task_runner/cgroup_task_runner.py:
> self.log.warning("Task failed with return code of 137. This may indicate "
> airflow/jobs/local_task_job.py: self.log.info("Task exited
> with return code %s", return_code)
> airflow/jobs/scheduler_job.py: self.log.info("Exiting gracefully upon
> receiving signal %s", signum)
> airflow/jobs/scheduler_job.py:
> self.log.warning("TaskInstance %s went missing from the database", ti)
> airflow/jobs/scheduler_job.py: self.log.info("Exited execute loop")
> airflow/jobs/scheduler_job.py: self.log.info("Exiting
> scheduler loop as all files"
> airflow/jobs/backfill_job.py: self.log.info("Backfill done. Exiting.")
> airflow/kubernetes/pod_launcher.py: self.log.info('Event with job
> id %s Failed', job_id)
> grep: logs/scheduler/latest: No such file or directory
> tests/sensors/test_timeout_sensor.py: self.log.info("Success criteria
> met. Exiting.")
> tests/gcp/operators/test_cloud_sql_system_helper.py:
> self.log.info("File doesn't exits. Creating dir...")
> tests/gcp/operators/test_cloud_sql_system_helper.py:
> self.log.info("Error while creating dir.")
> tests/gcp/utils/gcp_authenticator.py: self.log.info('Airflow DB
> Session error:' + str(ex))
> tests/gcp/utils/gcp_authenticator.py: self.log.info('Airflow DB
> Session error:' + str(ex))
> tests/gcp/utils/gcp_authenticator.py: self.log.info("The {} is
> missing".format(key_path))
> tests/contrib/utils/logging_command_executor.py:
> self.log.warning("Error when executing %s", " ".join(cmd))
> tests/contrib/utils/logging_command_executor.py:
> self.log.info("Error when executing '{}'".format(" ".join(cmd)))
--
This message was sent by Atlassian Jira
(v8.3.4#803005)