My DAG looks like this
default_args = {
'start_date': airflow.utils.dates.days_ago(0),
'retries': 0,
'dataflow_default_options': {
'project': 'test',
'tempLocation': 'gs://test/dataflow/pipelines/temp/',
'stagingLocation': 'gs://test/dataflow/pipelines/staging/',
'autoscalingAlgorithm': 'BASIC',
'maxNumWorkers': '1',
'region': 'asia-east1'
}
}
dag = DAG(
dag_id='gcs_avro_to_bq_dag',
default_args=default_args,
description='ETL for loading data from GCS(present in the avro format) to BQ',
schedule_interval=None,
dagrun_timeout=datetime.timedelta(minutes=30))
task = DataFlowJavaOperator(
task_id='gcs_avro_to_bq_flow_job',
jar='gs://test/dataflow/pipelines/jobs/test-1.0-SNAPSHOT.jar',
poll_sleep=1,
options={
'input': '{{ ts }}',
},
dag=dag)
My DAG is executing a jar file. The jar file has the code for running a dataflow job which writes data to BQ from GCS. The jar by itself executes successfully.
When I try to execute the airflow job, I see the following errors
[2020-05-20 17:20:41,934] {base_task_runner.py:101} INFO - Job 274: Subtask gcs_avro_to_bq_flow_job [2020-05-20 17:20:41,840] {gcp_api_base_hook.py:97} INFO - Getting connection using `google.auth.default()` since no key file is defined for hook.
[2020-05-20 17:20:41,937] {base_task_runner.py:101} INFO - Job 274: Subtask gcs_avro_to_bq_flow_job [2020-05-20 17:20:41,853] {discovery.py:272} INFO - URL being requested: GET https://www.googleapis.com/discovery/v1/apis/dataflow/v1b3/rest
[2020-05-20 17:20:44,338] {base_task_runner.py:101} INFO - Job 274: Subtask gcs_avro_to_bq_flow_job [2020-05-20 17:20:44,338] {discovery.py:873} INFO - URL being requested: GET https://dataflow.googleapis.com/v1b3/projects/test/locations/asia-east1/jobs/asia-east1?alt=json
[2020-05-20 17:20:45,285] {__init__.py:1631} ERROR - <HttpError 404 when requesting https://dataflow.googleapis.com/v1b3/projects/test/locations/asia-east1/jobs/asia-east1?alt=json returned "(7e83a8221abb0a9b): Information about job asia-east1 could not be found in our system. Please double check the id is correct. If it is please contact customer support.">
Traceback (most recent call last)
File "/usr/local/lib/airflow/airflow/models/__init__.py", line 1491, in _run_raw_tas
result = task_copy.execute(context=context
File "/usr/local/lib/airflow/airflow/contrib/operators/dataflow_operator.py", line 184, in execut
self.jar, self.job_class
File "/usr/local/lib/airflow/airflow/contrib/hooks/gcp_dataflow_hook.py", line 220, in start_java_dataflo
self._start_dataflow(variables, name, command_prefix, label_formatter
File "/usr/local/lib/airflow/airflow/contrib/hooks/gcp_api_base_hook.py", line 286, in wrappe
return func(self, *args, **kwargs
File "/usr/local/lib/airflow/airflow/contrib/hooks/gcp_dataflow_hook.py", line 200, in _start_dataflo
self.poll_sleep, job_id).wait_for_done(
File "/usr/local/lib/airflow/airflow/contrib/hooks/gcp_dataflow_hook.py", line 44, in __init_
self._job = self._get_job(
File "/usr/local/lib/airflow/airflow/contrib/hooks/gcp_dataflow_hook.py", line 63, in _get_jo
jobId=self._job_id).execute(num_retries=5
File "/opt/python3.6/lib/python3.6/site-packages/googleapiclient/_helpers.py", line 130, in positional_wrappe
return wrapped(*args, **kwargs
File "/opt/python3.6/lib/python3.6/site-packages/googleapiclient/http.py", line 851, in execut
raise HttpError(resp, content, uri=self.uri
I did some more digging and I can see the following API being called by airflow https://dataflow.googleapis.com/v1b3/projects/test/locations/asia-east1/jobs/asia-east1
As you can see the last param after jobs is asia-east
, so I feel the airflow job is trying to use the region I have provided in the default_args to search for the status of the dataflow job. Not sure if that is what is going on, but just wanted to state that observation. Am I missing something in my flows DAG? Also my java job logic looks like this
public class GcsAvroToBQ {
public interface Options extends PipelineOptions {
@Description("Input")
ValueProvider<String> getInput();
void setInput(ValueProvider<String> value);
}
/**
* Main entry point for executing the pipeline.
*
* @param args The command-line arguments to the pipeline.
*/
public static void main(String[] args) {
GcsAvroToBQ.Options options = PipelineOptionsFactory.fromArgs(args)
.withValidation()
.as(GcsAvroToBQ.Options.class);
options.getJobName();
run(options);
}
public static PipelineResult run(Options options) {
// Create the pipeline
Pipeline pipeline = Pipeline.create(options);
// My Pipeline logic to read Avro and upload to BQ
PCollection<TableRow> tableRowsForBQ; // Data to store in BQ
tableRowsForBQ.apply(
BigQueryIO.writeTableRows()
.to(bqDatasetName)
.withSchema(fieldSchemaListBuilder.schema())
.withCreateDisposition(BigQueryIO.Write.CreateDisposition.CREATE_IF_NEEDED)
.withWriteDisposition(BigQueryIO.Write.WriteDisposition.WRITE_APPEND));
return pipeline.run();
}
}
asia-east1
andasia-south1
, getting the same exception in both places. – Douglasscomposer-1.10.4-airflow-1.10.6
andcomposer-1.10.4-airflow-1.10.3
in GCP – Douglass