Skip to content

Commit

Permalink
Apply suggestions from code review
Browse files Browse the repository at this point in the history
Co-authored-by: Tatiana Al-Chueyr <[email protected]>
  • Loading branch information
pankajkoti and tatiana authored Oct 2, 2024
1 parent 05db6a0 commit 8fc4ae2
Show file tree
Hide file tree
Showing 2 changed files with 1 addition and 8 deletions.
2 changes: 1 addition & 1 deletion .github/workflows/test.yml
Original file line number Diff line number Diff line change
Expand Up @@ -2,7 +2,7 @@ name: test

on:
push: # Run on pushes to the default branch
branches: [main,poc-dbt-compile-task]
branches: [main]
pull_request_target: # Also run on pull requests originated from forks
branches: [main]

Expand Down
7 changes: 0 additions & 7 deletions cosmos/operators/airflow_async.py
Original file line number Diff line number Diff line change
Expand Up @@ -136,7 +136,6 @@ def get_remote_sql(self) -> str:
relative_file_path = str(file_path).replace(project_dir_parent, "").lstrip("/")
remote_model_path = f"{remote_target_path_str}/{dbt_dag_task_group_identifier}/compiled/{relative_file_path}"

print("remote_model_path: ", remote_model_path)
object_storage_path = ObjectStoragePath(remote_model_path, conn_id=remote_target_path_conn_id)
with object_storage_path.open() as fp: # type: ignore
return fp.read() # type: ignore
Expand All @@ -145,12 +144,6 @@ def drop_table_sql(self) -> None:
model_name = self.extra_context["dbt_node_config"]["resource_name"] # type: ignore
sql = f"DROP TABLE IF EXISTS {self.gcp_project}.{self.dataset}.{model_name};"

print("sql: ", sql)
print("gcp_project: ", self.gcp_project)
print("dataset: ", self.dataset)
print("gcp_conn_id: ", self.gcp_conn_id)
print("self.configuration: ", self.configuration)
print("self.location: ", self.location)
hook = BigQueryHook(
gcp_conn_id=self.gcp_conn_id,
impersonation_chain=self.impersonation_chain,
Expand Down

0 comments on commit 8fc4ae2

Please sign in to comment.