Skip to content
This repository was archived by the owner on Sep 3, 2022. It is now read-only.

Add descriptive message on successful deployment to Composer #700

Merged
merged 4 commits into from
Aug 18, 2018
Merged
Show file tree
Hide file tree
Changes from 3 commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
5 changes: 4 additions & 1 deletion google/datalab/bigquery/commands/_bigquery.py
Original file line number Diff line number Diff line change
Expand Up @@ -939,7 +939,8 @@ def _pipeline_cell(args, cell_body):
try:
airflow = google.datalab.contrib.pipeline.airflow.Airflow(gcs_dag_bucket, gcs_dag_file_path)
airflow.deploy(name, airflow_spec)
error_message += "Pipeline successfully deployed! View Airflow dashboard for more details."
error_message += ("Airflow pipeline successfully deployed! View dashboard for more "
"details.\n")
except AttributeError:
return "Perhaps you're missing: import google.datalab.contrib.pipeline.airflow"

Expand All @@ -950,6 +951,8 @@ def _pipeline_cell(args, cell_body):
try:
composer = google.datalab.contrib.pipeline.composer.Composer(location, environment)
composer.deploy(name, airflow_spec)
error_message += ("Composer pipeline successfully deployed! View dashboard for more "
"details.\n")
except AttributeError:
return "Perhaps you're missing: import google.datalab.contrib.pipeline.composer"

Expand Down
3 changes: 2 additions & 1 deletion tests/bigquery/pipeline_tests.py
Original file line number Diff line number Diff line change
Expand Up @@ -650,7 +650,8 @@ def test_pipeline_cell_golden(self, mock_bucket_class, mock_get_table, mock_tabl

output = google.datalab.bigquery.commands._bigquery._pipeline_cell(args, cell_body)

error_message = "Pipeline successfully deployed! View Airflow dashboard for more details."
error_message = ("Airflow pipeline successfully deployed! View dashboard for more details.\n"
"Composer pipeline successfully deployed! View dashboard for more details.\n")
airflow_spec_pattern = """
import datetime
from airflow import DAG
Expand Down