Skip to content

Commit

Permalink
Reformatting code apache#2
Browse files Browse the repository at this point in the history
  • Loading branch information
akashsriv07 committed Feb 12, 2024
1 parent ca4fb15 commit fa2c8e9
Show file tree
Hide file tree
Showing 6 changed files with 21 additions and 18 deletions.
1 change: 0 additions & 1 deletion airflow/providers/google/cloud/operators/dataprocgdc.py
Original file line number Diff line number Diff line change
Expand Up @@ -31,7 +31,6 @@
from airflow.models import BaseOperator
from airflow.providers.cncf.kubernetes.hooks.kubernetes import KubernetesHook, _load_body_to_dict


if TYPE_CHECKING:
from kubernetes.client.models import CoreV1EventList

Expand Down
2 changes: 1 addition & 1 deletion tests/providers/google/cloud/operators/test_dataprocgdc.py
Original file line number Diff line number Diff line change
Expand Up @@ -25,8 +25,8 @@

from airflow.exceptions import AirflowException
from airflow.providers.google.cloud.operators.dataprocgdc import (
DataprocGdcCreateAppEnvironmentKrmOperator,
DataprocGDCSubmitSparkJobKrmOperator,
DataprocGdcCreateAppEnvironmentKrmOperator
)


Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -21,17 +21,20 @@
"""

from __future__ import annotations

from datetime import datetime

from airflow import DAG
from airflow.providers.google.cloud.operators.dataprocgdc import (
DataprocGdcCreateAppEnvironmentKrmOperator,
DataprocGDCSubmitSparkJobKrmOperator,
)
from airflow.providers.google.cloud.sensors.dataprocgdc import DataprocGDCKrmSensor
from airflow.providers.google.cloud.operators.dataprocgdc import (DataprocGDCSubmitSparkJobKrmOperator,
DataprocGdcCreateAppEnvironmentKrmOperator)

DAG_ID = "example_dag_dpgdc-krm"

with DAG(DAG_ID, schedule="@once", catchup=False, start_date=datetime(2024, 2, 5), tags=["example", "dataprocgdc"]

with DAG(
DAG_ID, schedule="@once", catchup=False, start_date=datetime(2024, 2, 5), tags=["example", "dataprocgdc"]
) as dag:
submitSparkJobOperator = DataprocGDCSubmitSparkJobKrmOperator(
task_id="example-dataprocgdc-submitspark-operator",
Expand Down Expand Up @@ -68,7 +71,6 @@
)



from tests.system.utils import get_test_run # noqa: E402

# Needed to run the example DAG with pytest (see: tests/system/README.md#run_via_pytest)
Expand Down
Empty file.
Original file line number Diff line number Diff line change
@@ -1,4 +1,5 @@
#Licensed to the Apache Software Foundation (ASF) under one
#
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
Expand All @@ -14,8 +15,8 @@
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.


#
---
apiVersion: "dataprocgdc.cloud.google.com/v1alpha1"
kind: ApplicationEnvironment
metadata:
Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,5 @@
#Licensed to the Apache Software Foundation (ASF) under one
#
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
Expand All @@ -14,20 +15,20 @@
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.


#
---
apiVersion: "dataprocgdc.cloud.google.com/v1alpha1"
kind: SparkApplication
metadata:
name: simple-spark
namespace:
spec:
properties:
"spark.hadoop.fs.s3a.endpoint": "***"
"spark.hadoop.fs.s3a.access.key": "***"
"spark.hadoop.fs.s3a.secret.key": "***"
"spark.hadoop.fs.s3a.path.style.access": "true"
"spark.hadoop.fs.s3a.connection.ssl.enabled": "false"
"spark.hadoop.fs.s3a.endpoint": "***"
"spark.hadoop.fs.s3a.access.key": "***"
"spark.hadoop.fs.s3a.secret.key": "***"
"spark.hadoop.fs.s3a.path.style.access": "true"
"spark.hadoop.fs.s3a.connection.ssl.enabled": "false"
sparkApplicationConfig:
jarFileUris:
- "file:///usr/lib/spark/examples/jars/spark-examples.jar"
Expand Down

0 comments on commit fa2c8e9

Please sign in to comment.