[jira] [Commented] (AIRFLOW-3571) GoogleCloudStorageToBigQueryOperator succeeds to uploading CSV file from GCS to BiqQuery but a task is failed

2018-12-30 Thread jack (JIRA)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-3571?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16730970#comment-16730970
 ] 

jack commented on AIRFLOW-3571:
---

[~yohei] yep. You can create a ticket for all operators (example: 
https://issues.apache.org/jira/browse/AIRFLOW-3278 )

> GoogleCloudStorageToBigQueryOperator succeeds to uploading CSV file from GCS 
> to BiqQuery but a task is failed
> -
>
> Key: AIRFLOW-3571
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3571
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: contrib
>Affects Versions: 1.10.0
>Reporter: Yohei Onishi
>Assignee: Yohei Onishi
>Priority: Major
>
> I am using the following service in asia-northeast1-c zone. * GCS: 
> asia-northeast1-c
>  * BigQuery dataset and table: asia-northeast1-c
>  * Composer: asia-northeast1-c
> My task created by GoogleCloudStorageToBigQueryOperator succeeded to 
> uploading CSV file from a GCS bucket to a BigQuery table but the task was 
> failed due to the following error.
>   
> {code:java}
> [2018-12-26 21:35:47,464] {base_task_runner.py:107} INFO - Job 146: Subtask 
> bq_load_data_into_dest_table_from_gcs [2018-12-26 21:35:47,464] 
> {discovery.py:871} INFO - URL being requested: GET 
> https://www.googleapis.com/bigquery/v2/projects/my-project/jobs/job_abc123?alt=json
> [2018-12-26 21:35:47,931] {models.py:1736} ERROR - ('BigQuery job status 
> check failed. Final error was: %s', 404)
> Traceback (most recent call last)
>   File "/usr/local/lib/airflow/airflow/contrib/hooks/bigquery_hook.py", line 
> 981, in run_with_configuratio
> jobId=self.running_job_id).execute(
>   File "/usr/local/lib/python3.6/site-packages/googleapiclient/_helpers.py", 
> line 130, in positional_wrappe
> return wrapped(*args, **kwargs
>   File "/usr/local/lib/python3.6/site-packages/googleapiclient/http.py", line 
> 851, in execut
> raise HttpError(resp, content, uri=self.uri
> googleapiclient.errors.HttpError:  https://www.googleapis.com/bigquery/v2/projects/my-project/jobs/job_abc123?alt=json
>  returned "Not found: Job my-project:job_abc123"
> During handling of the above exception, another exception occurred
> Traceback (most recent call last)
>   File "/usr/local/lib/airflow/airflow/models.py", line 1633, in _run_raw_tas
> result = task_copy.execute(context=context
>   File "/usr/local/lib/airflow/airflow/contrib/operators/gcs_to_bq.py", line 
> 237, in execut
> time_partitioning=self.time_partitioning
>   File "/usr/local/lib/airflow/airflow/contrib/hooks/bigquery_hook.py", line 
> 951, in run_loa
> return self.run_with_configuration(configuration
>   File "/usr/local/lib/airflow/airflow/contrib/hooks/bigquery_hook.py", line 
> 1003, in run_with_configuratio
> err.resp.status
> Exception: ('BigQuery job status check failed. Final error was: %s', 404
> {code}
> The task failed to find a job {color:#ff}fmy-project:job_abc123{color} 
> but the correct job id is{color:#ff} 
> my-project:asia-northeast1:job_abc123{color}. (Note: this is just an example, 
> not actual id.)
>  I suppose the operator does not treat zone properly.
>   
> {code:java}
> $ bq show -j my-project:asia-northeast1:job_abc123
> Job my-project:asia-northeast1:job_abc123
> Job Type State Start Time Duration User Email Bytes Processed Bytes Billed 
> Billing Tier Labels
> -- - - -- 
> -- 
> - -- -- 
> load SUCCESS 27 Dec 05:35:47 0:00:01 my-service-account-id-email
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (AIRFLOW-3571) GoogleCloudStorageToBigQueryOperator succeeds to uploading CSV file from GCS to BiqQuery but a task is failed

2018-12-30 Thread Yohei Onishi (JIRA)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-3571?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16730967#comment-16730967
 ] 

Yohei Onishi commented on AIRFLOW-3571:
---

[https://github.com/apache/incubator-airflow/pull/4324]

4324 is merged so I can work on my issue.

> GoogleCloudStorageToBigQueryOperator succeeds to uploading CSV file from GCS 
> to BiqQuery but a task is failed
> -
>
> Key: AIRFLOW-3571
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3571
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: contrib
>Affects Versions: 1.10.0
>Reporter: Yohei Onishi
>Assignee: Yohei Onishi
>Priority: Major
>
> I am using the following service in asia-northeast1-c zone. * GCS: 
> asia-northeast1-c
>  * BigQuery dataset and table: asia-northeast1-c
>  * Composer: asia-northeast1-c
> My task created by GoogleCloudStorageToBigQueryOperator succeeded to 
> uploading CSV file from a GCS bucket to a BigQuery table but the task was 
> failed due to the following error.
>   
> {code:java}
> [2018-12-26 21:35:47,464] {base_task_runner.py:107} INFO - Job 146: Subtask 
> bq_load_data_into_dest_table_from_gcs [2018-12-26 21:35:47,464] 
> {discovery.py:871} INFO - URL being requested: GET 
> https://www.googleapis.com/bigquery/v2/projects/my-project/jobs/job_abc123?alt=json
> [2018-12-26 21:35:47,931] {models.py:1736} ERROR - ('BigQuery job status 
> check failed. Final error was: %s', 404)
> Traceback (most recent call last)
>   File "/usr/local/lib/airflow/airflow/contrib/hooks/bigquery_hook.py", line 
> 981, in run_with_configuratio
> jobId=self.running_job_id).execute(
>   File "/usr/local/lib/python3.6/site-packages/googleapiclient/_helpers.py", 
> line 130, in positional_wrappe
> return wrapped(*args, **kwargs
>   File "/usr/local/lib/python3.6/site-packages/googleapiclient/http.py", line 
> 851, in execut
> raise HttpError(resp, content, uri=self.uri
> googleapiclient.errors.HttpError:  https://www.googleapis.com/bigquery/v2/projects/my-project/jobs/job_abc123?alt=json
>  returned "Not found: Job my-project:job_abc123"
> During handling of the above exception, another exception occurred
> Traceback (most recent call last)
>   File "/usr/local/lib/airflow/airflow/models.py", line 1633, in _run_raw_tas
> result = task_copy.execute(context=context
>   File "/usr/local/lib/airflow/airflow/contrib/operators/gcs_to_bq.py", line 
> 237, in execut
> time_partitioning=self.time_partitioning
>   File "/usr/local/lib/airflow/airflow/contrib/hooks/bigquery_hook.py", line 
> 951, in run_loa
> return self.run_with_configuration(configuration
>   File "/usr/local/lib/airflow/airflow/contrib/hooks/bigquery_hook.py", line 
> 1003, in run_with_configuratio
> err.resp.status
> Exception: ('BigQuery job status check failed. Final error was: %s', 404
> {code}
> The task failed to find a job {color:#ff}fmy-project:job_abc123{color} 
> but the correct job id is{color:#ff} 
> my-project:asia-northeast1:job_abc123{color}. (Note: this is just an example, 
> not actual id.)
>  I suppose the operator does not treat zone properly.
>   
> {code:java}
> $ bq show -j my-project:asia-northeast1:job_abc123
> Job my-project:asia-northeast1:job_abc123
> Job Type State Start Time Duration User Email Bytes Processed Bytes Billed 
> Billing Tier Labels
> -- - - -- 
> -- 
> - -- -- 
> load SUCCESS 27 Dec 05:35:47 0:00:01 my-service-account-id-email
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (AIRFLOW-3571) GoogleCloudStorageToBigQueryOperator succeeds to uploading CSV file from GCS to BiqQuery but a task is failed

2018-12-29 Thread Yohei Onishi (JIRA)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-3571?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16730906#comment-16730906
 ] 

Yohei Onishi commented on AIRFLOW-3571:
---

OK will do

> GoogleCloudStorageToBigQueryOperator succeeds to uploading CSV file from GCS 
> to BiqQuery but a task is failed
> -
>
> Key: AIRFLOW-3571
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3571
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: contrib
>Affects Versions: 1.10.0
>Reporter: Yohei Onishi
>Assignee: Yohei Onishi
>Priority: Major
>
> I am using the following service in asia-northeast1-c zone. * GCS: 
> asia-northeast1-c
>  * BigQuery dataset and table: asia-northeast1-c
>  * Composer: asia-northeast1-c
> My task created by GoogleCloudStorageToBigQueryOperator succeeded to 
> uploading CSV file from a GCS bucket to a BigQuery table but the task was 
> failed due to the following error.
>   
> {code:java}
> [2018-12-26 21:35:47,464] {base_task_runner.py:107} INFO - Job 146: Subtask 
> bq_load_data_into_dest_table_from_gcs [2018-12-26 21:35:47,464] 
> {discovery.py:871} INFO - URL being requested: GET 
> https://www.googleapis.com/bigquery/v2/projects/my-project/jobs/job_abc123?alt=json
> [2018-12-26 21:35:47,931] {models.py:1736} ERROR - ('BigQuery job status 
> check failed. Final error was: %s', 404)
> Traceback (most recent call last)
>   File "/usr/local/lib/airflow/airflow/contrib/hooks/bigquery_hook.py", line 
> 981, in run_with_configuratio
> jobId=self.running_job_id).execute(
>   File "/usr/local/lib/python3.6/site-packages/googleapiclient/_helpers.py", 
> line 130, in positional_wrappe
> return wrapped(*args, **kwargs
>   File "/usr/local/lib/python3.6/site-packages/googleapiclient/http.py", line 
> 851, in execut
> raise HttpError(resp, content, uri=self.uri
> googleapiclient.errors.HttpError:  https://www.googleapis.com/bigquery/v2/projects/my-project/jobs/job_abc123?alt=json
>  returned "Not found: Job my-project:job_abc123"
> During handling of the above exception, another exception occurred
> Traceback (most recent call last)
>   File "/usr/local/lib/airflow/airflow/models.py", line 1633, in _run_raw_tas
> result = task_copy.execute(context=context
>   File "/usr/local/lib/airflow/airflow/contrib/operators/gcs_to_bq.py", line 
> 237, in execut
> time_partitioning=self.time_partitioning
>   File "/usr/local/lib/airflow/airflow/contrib/hooks/bigquery_hook.py", line 
> 951, in run_loa
> return self.run_with_configuration(configuration
>   File "/usr/local/lib/airflow/airflow/contrib/hooks/bigquery_hook.py", line 
> 1003, in run_with_configuratio
> err.resp.status
> Exception: ('BigQuery job status check failed. Final error was: %s', 404
> {code}
> The task failed to find a job {color:#ff}fmy-project:job_abc123{color} 
> but the correct job id is{color:#ff} 
> my-project:asia-northeast1:job_abc123{color}. (Note: this is just an example, 
> not actual id.)
>  I suppose the operator does not treat zone properly.
>   
> {code:java}
> $ bq show -j my-project:asia-northeast1:job_abc123
> Job my-project:asia-northeast1:job_abc123
> Job Type State Start Time Duration User Email Bytes Processed Bytes Billed 
> Billing Tier Labels
> -- - - -- 
> -- 
> - -- -- 
> load SUCCESS 27 Dec 05:35:47 0:00:01 my-service-account-id-email
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (AIRFLOW-3571) GoogleCloudStorageToBigQueryOperator succeeds to uploading CSV file from GCS to BiqQuery but a task is failed

2018-12-27 Thread Yohei Onishi (JIRA)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-3571?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16730064#comment-16730064
 ] 

Yohei Onishi commented on AIRFLOW-3571:
---

Sorry I misunderstood the PR. This just fixes 
[bigquery_hook.py|https://github.com/apache/incubator-airflow/pull/4324/files#diff-ee06f8fcbc476ea65446a30160c2a2b2]
 and [http://bigquery_operator.py|http://bigquery_operator.py./] only. Not 
GoogleCloudStorageToBigQueryOperator

 [https://github.com/apache/incubator-airflow/pull/4324]

 

> GoogleCloudStorageToBigQueryOperator succeeds to uploading CSV file from GCS 
> to BiqQuery but a task is failed
> -
>
> Key: AIRFLOW-3571
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3571
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: contrib
>Affects Versions: 1.10.0
>Reporter: Yohei Onishi
>Assignee: Yohei Onishi
>Priority: Major
>
> I am using the following service in asia-northeast1-c zone. * GCS: 
> asia-northeast1-c
>  * BigQuery dataset and table: asia-northeast1-c
>  * Composer: asia-northeast1-c
> My task created by GoogleCloudStorageToBigQueryOperator succeeded to 
> uploading CSV file from a GCS bucket to a BigQuery table but the task was 
> failed due to the following error.
>   
> {code:java}
> [2018-12-26 21:35:47,464] {base_task_runner.py:107} INFO - Job 146: Subtask 
> bq_load_data_into_dest_table_from_gcs [2018-12-26 21:35:47,464] 
> {discovery.py:871} INFO - URL being requested: GET 
> https://www.googleapis.com/bigquery/v2/projects/my-project/jobs/job_abc123?alt=json
> [2018-12-26 21:35:47,931] {models.py:1736} ERROR - ('BigQuery job status 
> check failed. Final error was: %s', 404)
> Traceback (most recent call last)
>   File "/usr/local/lib/airflow/airflow/contrib/hooks/bigquery_hook.py", line 
> 981, in run_with_configuratio
> jobId=self.running_job_id).execute(
>   File "/usr/local/lib/python3.6/site-packages/googleapiclient/_helpers.py", 
> line 130, in positional_wrappe
> return wrapped(*args, **kwargs
>   File "/usr/local/lib/python3.6/site-packages/googleapiclient/http.py", line 
> 851, in execut
> raise HttpError(resp, content, uri=self.uri
> googleapiclient.errors.HttpError:  https://www.googleapis.com/bigquery/v2/projects/my-project/jobs/job_abc123?alt=json
>  returned "Not found: Job my-project:job_abc123"
> During handling of the above exception, another exception occurred
> Traceback (most recent call last)
>   File "/usr/local/lib/airflow/airflow/models.py", line 1633, in _run_raw_tas
> result = task_copy.execute(context=context
>   File "/usr/local/lib/airflow/airflow/contrib/operators/gcs_to_bq.py", line 
> 237, in execut
> time_partitioning=self.time_partitioning
>   File "/usr/local/lib/airflow/airflow/contrib/hooks/bigquery_hook.py", line 
> 951, in run_loa
> return self.run_with_configuration(configuration
>   File "/usr/local/lib/airflow/airflow/contrib/hooks/bigquery_hook.py", line 
> 1003, in run_with_configuratio
> err.resp.status
> Exception: ('BigQuery job status check failed. Final error was: %s', 404
> {code}
> The task failed to find a job {color:#ff}fmy-project:job_abc123{color} 
> but the correct job id is{color:#ff} 
> my-project:asia-northeast1:job_abc123{color}. (Note: this is just an example, 
> not actual id.)
>  I suppose the operator does not treat zone properly.
>   
> {code:java}
> $ bq show -j my-project:asia-northeast1:job_abc123
> Job my-project:asia-northeast1:job_abc123
> Job Type State Start Time Duration User Email Bytes Processed Bytes Billed 
> Billing Tier Labels
> -- - - -- 
> -- 
> - -- -- 
> load SUCCESS 27 Dec 05:35:47 0:00:01 my-service-account-id-email
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (AIRFLOW-3571) GoogleCloudStorageToBigQueryOperator succeeds to uploading CSV file from GCS to BiqQuery but a task is failed

2018-12-27 Thread Yohei Onishi (JIRA)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-3571?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16729468#comment-16729468
 ] 

Yohei Onishi commented on AIRFLOW-3571:
---

This PR is also fixing the operator as well so I will close once the PR is 
merged.

> GoogleCloudStorageToBigQueryOperator succeeds to uploading CSV file from GCS 
> to BiqQuery but a task is failed
> -
>
> Key: AIRFLOW-3571
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3571
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: contrib
>Affects Versions: 1.10.0
>Reporter: Yohei Onishi
>Assignee: Yohei Onishi
>Priority: Major
>
> I am using the following service in asia-northeast1-c zone. * GCS: 
> asia-northeast1-c
>  * BigQuery dataset and table: asia-northeast1-c
>  * Composer: asia-northeast1-c
> My task created by GoogleCloudStorageToBigQueryOperator succeeded to 
> uploading CSV file from a GCS bucket to a BigQuery table but the task was 
> failed due to the following error.
>   
> {code:java}
> [2018-12-26 21:35:47,464] {base_task_runner.py:107} INFO - Job 146: Subtask 
> bq_load_data_into_dest_table_from_gcs [2018-12-26 21:35:47,464] 
> {discovery.py:871} INFO - URL being requested: GET 
> https://www.googleapis.com/bigquery/v2/projects/my-project/jobs/job_abc123?alt=json
> [2018-12-26 21:35:47,931] {models.py:1736} ERROR - ('BigQuery job status 
> check failed. Final error was: %s', 404)
> Traceback (most recent call last)
>   File "/usr/local/lib/airflow/airflow/contrib/hooks/bigquery_hook.py", line 
> 981, in run_with_configuratio
> jobId=self.running_job_id).execute(
>   File "/usr/local/lib/python3.6/site-packages/googleapiclient/_helpers.py", 
> line 130, in positional_wrappe
> return wrapped(*args, **kwargs
>   File "/usr/local/lib/python3.6/site-packages/googleapiclient/http.py", line 
> 851, in execut
> raise HttpError(resp, content, uri=self.uri
> googleapiclient.errors.HttpError:  https://www.googleapis.com/bigquery/v2/projects/my-project/jobs/job_abc123?alt=json
>  returned "Not found: Job my-project:job_abc123"
> During handling of the above exception, another exception occurred
> Traceback (most recent call last)
>   File "/usr/local/lib/airflow/airflow/models.py", line 1633, in _run_raw_tas
> result = task_copy.execute(context=context
>   File "/usr/local/lib/airflow/airflow/contrib/operators/gcs_to_bq.py", line 
> 237, in execut
> time_partitioning=self.time_partitioning
>   File "/usr/local/lib/airflow/airflow/contrib/hooks/bigquery_hook.py", line 
> 951, in run_loa
> return self.run_with_configuration(configuration
>   File "/usr/local/lib/airflow/airflow/contrib/hooks/bigquery_hook.py", line 
> 1003, in run_with_configuratio
> err.resp.status
> Exception: ('BigQuery job status check failed. Final error was: %s', 404
> {code}
> The task failed to find a job {color:#ff}fmy-project:job_abc123{color} 
> but the correct job id is{color:#ff} 
> my-project:asia-northeast1:job_abc123{color}. (Note: this is just an example, 
> not actual id.)
>  I suppose the operator does not treat zone properly.
>   
> {code:java}
> $ bq show -j my-project:asia-northeast1:job_abc123
> Job my-project:asia-northeast1:job_abc123
> Job Type State Start Time Duration User Email Bytes Processed Bytes Billed 
> Billing Tier Labels
> -- - - -- 
> -- 
> - -- -- 
> load SUCCESS 27 Dec 05:35:47 0:00:01 my-service-account-id-email
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (AIRFLOW-3571) GoogleCloudStorageToBigQueryOperator succeeds to uploading CSV file from GCS to BiqQuery but a task is failed

2018-12-26 Thread Yohei Onishi (JIRA)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-3571?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16729399#comment-16729399
 ] 

Yohei Onishi commented on AIRFLOW-3571:
---

Thanks, will do.

> GoogleCloudStorageToBigQueryOperator succeeds to uploading CSV file from GCS 
> to BiqQuery but a task is failed
> -
>
> Key: AIRFLOW-3571
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3571
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: contrib
>Affects Versions: 1.10.0
>Reporter: Yohei Onishi
>Assignee: Yohei Onishi
>Priority: Major
>
> I am using the following service in asia-northeast1-c zone. * GCS: 
> asia-northeast1-c
>  * BigQuery dataset and table: asia-northeast1-c
>  * Composer: asia-northeast1-c
> My task created by GoogleCloudStorageToBigQueryOperator succeeded to 
> uploading CSV file from a GCS bucket to a BigQuery table but the task was 
> failed due to the following error.
>   
> {code:java}
> [2018-12-26 21:35:47,464] {base_task_runner.py:107} INFO - Job 146: Subtask 
> bq_load_data_into_dest_table_from_gcs [2018-12-26 21:35:47,464] 
> {discovery.py:871} INFO - URL being requested: GET 
> https://www.googleapis.com/bigquery/v2/projects/my-project/jobs/job_abc123?alt=json
> [2018-12-26 21:35:47,931] {models.py:1736} ERROR - ('BigQuery job status 
> check failed. Final error was: %s', 404)
> Traceback (most recent call last)
>   File "/usr/local/lib/airflow/airflow/contrib/hooks/bigquery_hook.py", line 
> 981, in run_with_configuratio
> jobId=self.running_job_id).execute(
>   File "/usr/local/lib/python3.6/site-packages/googleapiclient/_helpers.py", 
> line 130, in positional_wrappe
> return wrapped(*args, **kwargs
>   File "/usr/local/lib/python3.6/site-packages/googleapiclient/http.py", line 
> 851, in execut
> raise HttpError(resp, content, uri=self.uri
> googleapiclient.errors.HttpError:  https://www.googleapis.com/bigquery/v2/projects/my-project/jobs/job_abc123?alt=json
>  returned "Not found: Job my-project:job_abc123"
> During handling of the above exception, another exception occurred
> Traceback (most recent call last)
>   File "/usr/local/lib/airflow/airflow/models.py", line 1633, in _run_raw_tas
> result = task_copy.execute(context=context
>   File "/usr/local/lib/airflow/airflow/contrib/operators/gcs_to_bq.py", line 
> 237, in execut
> time_partitioning=self.time_partitioning
>   File "/usr/local/lib/airflow/airflow/contrib/hooks/bigquery_hook.py", line 
> 951, in run_loa
> return self.run_with_configuration(configuration
>   File "/usr/local/lib/airflow/airflow/contrib/hooks/bigquery_hook.py", line 
> 1003, in run_with_configuratio
> err.resp.status
> Exception: ('BigQuery job status check failed. Final error was: %s', 404
> {code}
> The task failed to find a job {color:#ff}fmy-project:job_abc123{color} 
> but the correct job id is{color:#ff} 
> my-project:asia-northeast1:job_abc123{color}. (Note: this is just an example, 
> not actual id.)
>  I suppose the operator does not treat zone properly.
>   
> {code:java}
> $ bq show -j my-project:asia-northeast1:job_abc123
> Job my-project:asia-northeast1:job_abc123
> Job Type State Start Time Duration User Email Bytes Processed Bytes Billed 
> Billing Tier Labels
> -- - - -- 
> -- 
> - -- -- 
> load SUCCESS 27 Dec 05:35:47 0:00:01 my-service-account-id-email
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (AIRFLOW-3571) GoogleCloudStorageToBigQueryOperator succeeds to uploading CSV file from GCS to BiqQuery but a task is failed

2018-12-26 Thread jack (JIRA)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-3571?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16729398#comment-16729398
 ] 

jack commented on AIRFLOW-3571:
---

[~yohei] There is an open PR for adding support for locations to the 
BigQueryHook  https://github.com/apache/incubator-airflow/pull/4324

Once it's merged (I assume it would be soon) you are welcome to open PRs for 
extending the Operators to support it.

> GoogleCloudStorageToBigQueryOperator succeeds to uploading CSV file from GCS 
> to BiqQuery but a task is failed
> -
>
> Key: AIRFLOW-3571
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3571
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: contrib
>Affects Versions: 1.10.0
>Reporter: Yohei Onishi
>Priority: Major
>
> I am using the following service in asia-northeast1-c zone. * GCS: 
> asia-northeast1-c
>  * BigQuery dataset and table: asia-northeast1-c
>  * Composer: asia-northeast1-c
> My task created by GoogleCloudStorageToBigQueryOperator succeeded to 
> uploading CSV file from a GCS bucket to a BigQuery table but the task was 
> failed due to the following error.
>   
> {code:java}
> [2018-12-26 21:35:47,464] {base_task_runner.py:107} INFO - Job 146: Subtask 
> bq_load_data_into_dest_table_from_gcs [2018-12-26 21:35:47,464] 
> {discovery.py:871} INFO - URL being requested: GET 
> https://www.googleapis.com/bigquery/v2/projects/my-project/jobs/job_abc123?alt=json
> [2018-12-26 21:35:47,931] {models.py:1736} ERROR - ('BigQuery job status 
> check failed. Final error was: %s', 404)
> Traceback (most recent call last)
>   File "/usr/local/lib/airflow/airflow/contrib/hooks/bigquery_hook.py", line 
> 981, in run_with_configuratio
> jobId=self.running_job_id).execute(
>   File "/usr/local/lib/python3.6/site-packages/googleapiclient/_helpers.py", 
> line 130, in positional_wrappe
> return wrapped(*args, **kwargs
>   File "/usr/local/lib/python3.6/site-packages/googleapiclient/http.py", line 
> 851, in execut
> raise HttpError(resp, content, uri=self.uri
> googleapiclient.errors.HttpError:  https://www.googleapis.com/bigquery/v2/projects/my-project/jobs/job_abc123?alt=json
>  returned "Not found: Job my-project:job_abc123"
> During handling of the above exception, another exception occurred
> Traceback (most recent call last)
>   File "/usr/local/lib/airflow/airflow/models.py", line 1633, in _run_raw_tas
> result = task_copy.execute(context=context
>   File "/usr/local/lib/airflow/airflow/contrib/operators/gcs_to_bq.py", line 
> 237, in execut
> time_partitioning=self.time_partitioning
>   File "/usr/local/lib/airflow/airflow/contrib/hooks/bigquery_hook.py", line 
> 951, in run_loa
> return self.run_with_configuration(configuration
>   File "/usr/local/lib/airflow/airflow/contrib/hooks/bigquery_hook.py", line 
> 1003, in run_with_configuratio
> err.resp.status
> Exception: ('BigQuery job status check failed. Final error was: %s', 404
> {code}
> The task failed to find a job {color:#ff}fmy-project:job_abc123{color} 
> but the correct job id is{color:#ff} 
> my-project:asia-northeast1:job_abc123{color}. (Note: this is just an example, 
> not actual id.)
>  I suppose the operator does not treat zone properly.
>   
> {code:java}
> $ bq show -j my-project:asia-northeast1:job_abc123
> Job my-project:asia-northeast1:job_abc123
> Job Type State Start Time Duration User Email Bytes Processed Bytes Billed 
> Billing Tier Labels
> -- - - -- 
> -- 
> - -- -- 
> load SUCCESS 27 Dec 05:35:47 0:00:01 my-service-account-id-email
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (AIRFLOW-3571) GoogleCloudStorageToBigQueryOperator succeeds to uploading CSV file from GCS to BiqQuery but a task is failed

2018-12-26 Thread Yohei Onishi (JIRA)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-3571?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16729263#comment-16729263
 ] 

Yohei Onishi commented on AIRFLOW-3571:
---

It seems {color:#d04437}GoogleCloudStorageToBigQueryOperator{color} does not 
support regions other than US / EU.

> GoogleCloudStorageToBigQueryOperator succeeds to uploading CSV file from GCS 
> to BiqQuery but a task is failed
> -
>
> Key: AIRFLOW-3571
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3571
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: contrib
>Affects Versions: 1.10.0
>Reporter: Yohei Onishi
>Priority: Major
>
> I am using the following service in asia-northeast1-c zone. * GCS: 
> asia-northeast1-c
>  * BigQuery dataset and table: asia-northeast1-c
>  * Composer: asia-northeast1-c
> My task created by GoogleCloudStorageToBigQueryOperator succeeded to 
> uploading CSV file from a GCS bucket to a BigQuery table but the task was 
> failed due to the following error.
>   
> {code:java}
> [2018-12-26 21:35:47,464] {base_task_runner.py:107} INFO - Job 146: Subtask 
> bq_load_data_into_dest_table_from_gcs [2018-12-26 21:35:47,464] 
> {discovery.py:871} INFO - URL being requested: GET 
> https://www.googleapis.com/bigquery/v2/projects/my-project/jobs/job_abc123?alt=json
> [2018-12-26 21:35:47,931] {models.py:1736} ERROR - ('BigQuery job status 
> check failed. Final error was: %s', 404)
> Traceback (most recent call last)
>   File "/usr/local/lib/airflow/airflow/contrib/hooks/bigquery_hook.py", line 
> 981, in run_with_configuratio
> jobId=self.running_job_id).execute(
>   File "/usr/local/lib/python3.6/site-packages/googleapiclient/_helpers.py", 
> line 130, in positional_wrappe
> return wrapped(*args, **kwargs
>   File "/usr/local/lib/python3.6/site-packages/googleapiclient/http.py", line 
> 851, in execut
> raise HttpError(resp, content, uri=self.uri
> googleapiclient.errors.HttpError:  https://www.googleapis.com/bigquery/v2/projects/my-project/jobs/job_abc123?alt=json
>  returned "Not found: Job my-project:job_abc123"
> During handling of the above exception, another exception occurred
> Traceback (most recent call last)
>   File "/usr/local/lib/airflow/airflow/models.py", line 1633, in _run_raw_tas
> result = task_copy.execute(context=context
>   File "/usr/local/lib/airflow/airflow/contrib/operators/gcs_to_bq.py", line 
> 237, in execut
> time_partitioning=self.time_partitioning
>   File "/usr/local/lib/airflow/airflow/contrib/hooks/bigquery_hook.py", line 
> 951, in run_loa
> return self.run_with_configuration(configuration
>   File "/usr/local/lib/airflow/airflow/contrib/hooks/bigquery_hook.py", line 
> 1003, in run_with_configuratio
> err.resp.status
> Exception: ('BigQuery job status check failed. Final error was: %s', 404
> {code}
> The task failed to find a job {color:#ff}fmy-project:job_abc123{color} 
> but the correct job id is{color:#ff} 
> my-project:asia-northeast1:job_abc123{color}. (Note: this is just an example, 
> not actual id.)
>  I suppose the operator does not treat zone properly.
>   
> {code:java}
> $ bq show -j my-project:asia-northeast1:job_abc123
> Job my-project:asia-northeast1:job_abc123
> Job Type State Start Time Duration User Email Bytes Processed Bytes Billed 
> Billing Tier Labels
> -- - - -- 
> -- 
> - -- -- 
> load SUCCESS 27 Dec 05:35:47 0:00:01 my-service-account-id-email
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)