Skip to content

Commit cdb7fc9

Browse files
authored
add region tags for aws + azure tutorials (#8218)
teeny PR to add some region tags
1 parent 87c5251 commit cdb7fc9

File tree

2 files changed

+9
-7
lines changed

2 files changed

+9
-7
lines changed

composer/workflows/azureblobstoretogcsoperator_tutorial.py

Lines changed: 4 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -12,6 +12,7 @@
1212
# See the License for the specific language governing permissions and
1313
# limitations under the License.
1414

15+
# [START composer_dataanalyticstutorial_azure_dag]
1516
import datetime
1617

1718
from airflow import models
@@ -53,13 +54,12 @@
5354
f"{BQ_DESTINATION_DATASET_NAME}.{BQ_DESTINATION_TABLE_NAME}",
5455
f"{BQ_DESTINATION_DATASET_NAME}.{BQ_NORMALIZED_TABLE_NAME}",
5556
],
56-
5757
},
5858
"environment_config": {
5959
"execution_config": {
6060
"service_account": "{{var.value.dataproc_service_account}}"
6161
}
62-
}
62+
},
6363
}
6464

6565
yesterday = datetime.datetime.combine(
@@ -73,7 +73,7 @@
7373
# To email on failure or retry set 'email' arg to your email and enable
7474
# emailing here.
7575
"email_on_failure": False,
76-
"email_on_retry": False
76+
"email_on_retry": False,
7777
}
7878

7979
with models.DAG(
@@ -161,3 +161,4 @@
161161
)
162162

163163
azure_blob_to_gcs >> load_external_dataset >> bq_join_group >> create_batch
164+
# [END composer_dataanalyticstutorial_azure_dag]

composer/workflows/s3togcsoperator_tutorial.py

Lines changed: 5 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -12,6 +12,7 @@
1212
# See the License for the specific language governing permissions and
1313
# limitations under the License.
1414

15+
# [START composer_dataanalyticstutorial_aws_dag]
1516
import datetime
1617

1718
from airflow import models
@@ -49,13 +50,12 @@
4950
f"{BQ_DESTINATION_DATASET_NAME}.{BQ_DESTINATION_TABLE_NAME}",
5051
f"{BQ_DESTINATION_DATASET_NAME}.{BQ_NORMALIZED_TABLE_NAME}",
5152
],
52-
5353
},
5454
"environment_config": {
5555
"execution_config": {
5656
"service_account": "{{var.value.dataproc_service_account}}"
5757
}
58-
}
58+
},
5959
}
6060

6161
yesterday = datetime.datetime.combine(
@@ -69,7 +69,7 @@
6969
# To email on failure or retry set 'email' arg to your email and enable
7070
# emailing here.
7171
"email_on_failure": False,
72-
"email_on_retry": False
72+
"email_on_retry": False,
7373
}
7474

7575
with models.DAG(
@@ -106,7 +106,7 @@
106106
{"name": "Holiday", "type": "STRING"},
107107
],
108108
skip_leading_rows=1,
109-
write_disposition="WRITE_TRUNCATE"
109+
write_disposition="WRITE_TRUNCATE",
110110
)
111111

112112
with TaskGroup("join_bq_datasets") as bq_join_group:
@@ -148,3 +148,4 @@
148148
)
149149

150150
s3_to_gcs_op >> load_external_dataset >> bq_join_group >> create_batch
151+
# [END composer_dataanalyticstutorial_aws_dag]

0 commit comments

Comments
 (0)