[
https://issues.apache.org/jira/browse/AIRFLOW-3279?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Paul Velthuis updated AIRFLOW-3279:
-----------------------------------
Description:
The documentation of how to install logging to a Google Cloud bucket is unclear.
I am now following the tutorial on the airflow page:
[https://airflow.apache.org/howto/write-logs.html]
Here I find it unclear what part of the 'logger' I have to adjust in the
`{{airflow/config_templates/airflow_local_settings.py}}`.
The adjustment states:
# Update the airflow.task and airflow.tas_runner blocks to be 'gcs.task'
instead of 'file.task'. 'loggers':
Unknown macro: \{ 'airflow.task'}
However what I find in the template is:
|'loggers': \{\| \|'airflow.processor': { \\\\| \\\\|'handlers': ['processor'],
\\\\| \\\\|'level': LOG_LEVEL, \\\\| \\\\|'propagate': False, \\\\| \\\\|},|
|'airflow.task': {
\|
\|'handlers': ['task'],
\|
\|'level': LOG_LEVEL,
\|
\|'propagate': False,
\|
\|},|
|'flask_appbuilder': {
\|
\|'handler': ['console'],
\|
\|'level': FAB_LOG_LEVEL,
\|
\|'propagate': True,
\|
\|}|
},
Since for me it is very important to do it right at the first time I hope some
clarity can be provided in what has to be adjusted in the logger. Is it only
the 'airflow.task' or more?
Furthermore, at step 6 it is a little unclear what remote_log_conn_id means. I
would propose to add a little more information to make this more clear.
When training to deploy the new log_config.py and following the instructions I
receive the following error:
python3.6/site-packages/airflow/logging_config.py", line 60, in
configure_logging
'Unable to load custom logging from {}'.format(logging_class_path)
ImportError: Unable to load custom logging from log_config.LOGGING_CONFIG
If I add the `logging_class_path` instead of `logging_config_class` this error
resolves.
I now get the error:
Task log handler gcs.task does not support read logs. 'NoneType' object has no
attribute 'read'
was:
The documentation of how to install logging to a Google Cloud bucket is unclear.
I am now following the tutorial on the airflow page:
[https://airflow.apache.org/howto/write-logs.html]
Here I find it unclear what part of the 'logger' I have to adjust in the
`{{airflow/config_templates/airflow_local_settings.py}}`.
The adjustment states:
# Update the airflow.task and airflow.tas_runner blocks to be 'gcs.task'
instead of 'file.task'. 'loggers':
Unknown macro: \{ 'airflow.task'}
However what I find in the template is:
|'loggers': \{\| \|'airflow.processor': { \\\| \\\|'handlers': ['processor'],
\\\| \\\|'level': LOG_LEVEL, \\\| \\\|'propagate': False, \\\| \\\|},|
|'airflow.task': \{ \\| \\|'handlers': ['task'], \\| \\|'level': LOG_LEVEL, \\|
\\|'propagate': False, \\| \\|},|
|'flask_appbuilder': \{ \\| \\|'handler': ['console'], \\| \\|'level':
FAB_LOG_LEVEL, \\| \\|'propagate': True, \\| \\|}|
},
Since for me it is very important to do it right at the first time I hope some
clarity can be provided in what has to be adjusted in the logger. Is it only
the 'airflow.task' or more?
Furthermore, at step 6 it is a little unclear what remote_log_conn_id means. I
would propose to add a little more information to make this more clear.
When training to deploy the new log_config.py and following the instructions I
receive the following error:
python3.6/site-packages/airflow/logging_config.py", line 60, in
configure_logging
'Unable to load custom logging from {}'.format(logging_class_path)
ImportError: Unable to load custom logging from log_config.LOGGING_CONFIG
If I add the `logging_class_path` instead of `logging_config_class` this error
resolves.
> Documentation for Google Logging unclear
> ----------------------------------------
>
> Key: AIRFLOW-3279
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3279
> Project: Apache Airflow
> Issue Type: Bug
> Components: configuration, Documentation, logging
> Reporter: Paul Velthuis
> Assignee: Fokko Driesprong
> Priority: Major
>
> The documentation of how to install logging to a Google Cloud bucket is
> unclear.
> I am now following the tutorial on the airflow page:
> [https://airflow.apache.org/howto/write-logs.html]
> Here I find it unclear what part of the 'logger' I have to adjust in the
> `{{airflow/config_templates/airflow_local_settings.py}}`.
>
> The adjustment states:
>
> # Update the airflow.task and airflow.tas_runner blocks to be 'gcs.task'
> instead of 'file.task'. 'loggers':
> Unknown macro: \{ 'airflow.task'}
>
> However what I find in the template is:
> |'loggers': \{\| \|'airflow.processor': { \\\\| \\\\|'handlers':
> ['processor'], \\\\| \\\\|'level': LOG_LEVEL, \\\\| \\\\|'propagate': False,
> \\\\| \\\\|},|
> |'airflow.task': {
> \|
> \|'handlers': ['task'],
> \|
> \|'level': LOG_LEVEL,
> \|
> \|'propagate': False,
> \|
> \|},|
> |'flask_appbuilder': {
> \|
> \|'handler': ['console'],
> \|
> \|'level': FAB_LOG_LEVEL,
> \|
> \|'propagate': True,
> \|
> \|}|
> },
>
> Since for me it is very important to do it right at the first time I hope
> some clarity can be provided in what has to be adjusted in the logger. Is it
> only the 'airflow.task' or more?
> Furthermore, at step 6 it is a little unclear what remote_log_conn_id means.
> I would propose to add a little more information to make this more clear.
>
> When training to deploy the new log_config.py and following the instructions
> I receive the following error:
> python3.6/site-packages/airflow/logging_config.py", line 60, in
> configure_logging
> 'Unable to load custom logging from {}'.format(logging_class_path)
> ImportError: Unable to load custom logging from log_config.LOGGING_CONFIG
> If I add the `logging_class_path` instead of `logging_config_class` this
> error resolves.
>
> I now get the error:
> Task log handler gcs.task does not support read logs. 'NoneType' object has
> no attribute 'read'
>
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)