diff --git a/.github/boring-cyborg.yml b/.github/boring-cyborg.yml
index 1af41df176597..286d1b9083d7c 100644
--- a/.github/boring-cyborg.yml
+++ b/.github/boring-cyborg.yml
@@ -132,10 +132,7 @@ labelPRBasedOnFilePath:
- providers/standard/**
provider:databricks:
- - providers/src/airflow/providers/databricks/**/*
- - docs/apache-airflow-providers-databricks/**/*
- - providers/tests/databricks/**/*
- - providers/tests/system/databricks/**/*
+ - providers/databricks/**
provider:datadog:
- providers/datadog/**
diff --git a/dev/moving_providers/move_providers.py b/dev/moving_providers/move_providers.py
index b1d7519eb6dc2..0827f8fdddbc6 100755
--- a/dev/moving_providers/move_providers.py
+++ b/dev/moving_providers/move_providers.py
@@ -362,10 +362,13 @@ def move_provider_yaml(provider_id: str) -> tuple[list[str], list[str], list[str
dependencies = []
optional_dependencies = []
devel_dependencies = []
+ copied_logo = set()
for line in original_content:
if line.startswith(" logo: "):
logo_path = line[len(" logo: ") :]
logo_name = logo_path.split("/")[-1]
+ if logo_path in copied_logo:
+ continue
new_logo_dir = (
PROVIDERS_DIR_PATH / _get_provider_only_path(provider_id) / "docs" / "integration-logos"
)
@@ -378,6 +381,7 @@ def move_provider_yaml(provider_id: str) -> tuple[list[str], list[str], list[str
remove_empty_parent_dir=True,
)
line = f" logo: /docs/integration-logos/{logo_name}"
+ copied_logo.add(logo_path)
if line == "dependencies:" and not in_dependencies:
in_dependencies = True
continue
diff --git a/docs/.gitignore b/docs/.gitignore
index 1828bb5fc3576..47887c67a435c 100644
--- a/docs/.gitignore
+++ b/docs/.gitignore
@@ -21,6 +21,7 @@ apache-airflow-providers-cohere
apache-airflow-providers-common-compat
apache-airflow-providers-common-io
apache-airflow-providers-common-sql
+apache-airflow-providers-databricks
apache-airflow-providers-datadog
apache-airflow-providers-dbt-cloud
apache-airflow-providers-discord
diff --git a/docs/apache-airflow-providers-databricks/changelog.rst b/docs/apache-airflow-providers-databricks/changelog.rst
deleted file mode 100644
index 8e4f5f126efd4..0000000000000
--- a/docs/apache-airflow-providers-databricks/changelog.rst
+++ /dev/null
@@ -1,25 +0,0 @@
-
- .. Licensed to the Apache Software Foundation (ASF) under one
- or more contributor license agreements. See the NOTICE file
- distributed with this work for additional information
- regarding copyright ownership. The ASF licenses this file
- to you under the Apache License, Version 2.0 (the
- "License"); you may not use this file except in compliance
- with the License. You may obtain a copy of the License at
-
- .. http://www.apache.org/licenses/LICENSE-2.0
-
- .. Unless required by applicable law or agreed to in writing,
- software distributed under the License is distributed on an
- "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
- KIND, either express or implied. See the License for the
- specific language governing permissions and limitations
- under the License.
-
- .. NOTE! THIS FILE IS AUTOMATICALLY GENERATED AND WILL BE
- OVERWRITTEN WHEN PREPARING PACKAGES.
-
- .. IF YOU WANT TO MODIFY THIS FILE, YOU SHOULD MODIFY THE TEMPLATE
- `PROVIDER_CHANGELOG_TEMPLATE.rst.jinja2` IN the `dev/breeze/src/airflow_breeze/templates` DIRECTORY
-
-.. include:: ../../providers/src/airflow/providers/databricks/CHANGELOG.rst
diff --git a/providers/databricks/README.rst b/providers/databricks/README.rst
new file mode 100644
index 0000000000000..03db4020bc506
--- /dev/null
+++ b/providers/databricks/README.rst
@@ -0,0 +1,87 @@
+
+ .. Licensed to the Apache Software Foundation (ASF) under one
+ or more contributor license agreements. See the NOTICE file
+ distributed with this work for additional information
+ regarding copyright ownership. The ASF licenses this file
+ to you under the Apache License, Version 2.0 (the
+ "License"); you may not use this file except in compliance
+ with the License. You may obtain a copy of the License at
+
+ .. http://www.apache.org/licenses/LICENSE-2.0
+
+ .. Unless required by applicable law or agreed to in writing,
+ software distributed under the License is distributed on an
+ "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+ KIND, either express or implied. See the License for the
+ specific language governing permissions and limitations
+ under the License.
+
+ .. NOTE! THIS FILE IS AUTOMATICALLY GENERATED AND WILL BE OVERWRITTEN!
+
+ .. IF YOU WANT TO MODIFY TEMPLATE FOR THIS FILE, YOU SHOULD MODIFY THE TEMPLATE
+ `PROVIDER_README_TEMPLATE.rst.jinja2` IN the `dev/breeze/src/airflow_breeze/templates` DIRECTORY
+
+
+Package ``apache-airflow-providers-databricks``
+
+Release: ``7.0.0``
+
+
+`Databricks `__
+
+
+Provider package
+----------------
+
+This is a provider package for ``databricks`` provider. All classes for this provider package
+are in ``airflow.providers.databricks`` python package.
+
+You can find package information and changelog for the provider
+in the `documentation `_.
+
+Installation
+------------
+
+You can install this package on top of an existing Airflow 2 installation (see ``Requirements`` below
+for the minimum Airflow version supported) via
+``pip install apache-airflow-providers-databricks``
+
+The package supports the following python versions: 3.9,3.10,3.11,3.12
+
+Requirements
+------------
+
+======================================= ==================
+PIP package Version required
+======================================= ==================
+``apache-airflow`` ``>=2.9.0``
+``apache-airflow-providers-common-sql`` ``>=1.20.0``
+``requests`` ``>=2.27.0,<3``
+``databricks-sql-connector`` ``>=3.0.0``
+``aiohttp`` ``>=3.9.2,<4``
+``mergedeep`` ``>=1.3.4``
+``pandas`` ``>=2.1.2,<2.2``
+``pyarrow`` ``>=14.0.1``
+======================================= ==================
+
+Cross provider package dependencies
+-----------------------------------
+
+Those are dependencies that might be needed in order to use all the features of the package.
+You need to install the specified provider packages in order to use them.
+
+You can install such cross-provider dependencies when installing from PyPI. For example:
+
+.. code-block:: bash
+
+ pip install apache-airflow-providers-databricks[common.sql]
+
+
+============================================================================================================ ==============
+Dependent package Extra
+============================================================================================================ ==============
+`apache-airflow-providers-common-sql `_ ``common.sql``
+============================================================================================================ ==============
+
+The changelog for the provider package can be found in the
+`changelog `_.
diff --git a/providers/src/airflow/providers/databricks/.latest-doc-only-change.txt b/providers/databricks/docs/.latest-doc-only-change.txt
similarity index 100%
rename from providers/src/airflow/providers/databricks/.latest-doc-only-change.txt
rename to providers/databricks/docs/.latest-doc-only-change.txt
diff --git a/providers/src/airflow/providers/databricks/CHANGELOG.rst b/providers/databricks/docs/changelog.rst
similarity index 99%
rename from providers/src/airflow/providers/databricks/CHANGELOG.rst
rename to providers/databricks/docs/changelog.rst
index 227d297a86f49..8a2279b2745f6 100644
--- a/providers/src/airflow/providers/databricks/CHANGELOG.rst
+++ b/providers/databricks/docs/changelog.rst
@@ -139,7 +139,7 @@ Features
Misc
~~~~
-* ``Removed deprecated method referance airflow.www.auth.has_access when min airflow version >= 2.8.0 (#41747)``
+* ``Removed deprecated method reference airflow.www.auth.has_access when min airflow version >= 2.8.0 (#41747)``
* ``remove deprecated soft_fail from providers (#41710)``
6.9.0
@@ -451,7 +451,7 @@ Misc
Features
~~~~~~~~
-* ``Add "QUEUED" to RUN_LIFE_CYCLE_STATES following deployement of … (#33886)``
+* ``Add "QUEUED" to RUN_LIFE_CYCLE_STATES following deployment of … (#33886)``
* ``allow DatabricksSubmitRunOperator to accept a pipeline name for a pipeline_task (#32903)``
Misc
diff --git a/docs/apache-airflow-providers-databricks/commits.rst b/providers/databricks/docs/commits.rst
similarity index 100%
rename from docs/apache-airflow-providers-databricks/commits.rst
rename to providers/databricks/docs/commits.rst
diff --git a/docs/apache-airflow-providers-databricks/connections/databricks.rst b/providers/databricks/docs/connections/databricks.rst
similarity index 100%
rename from docs/apache-airflow-providers-databricks/connections/databricks.rst
rename to providers/databricks/docs/connections/databricks.rst
diff --git a/docs/apache-airflow-providers-databricks/img/databricks_workflow_task_group_airflow_graph_view.png b/providers/databricks/docs/img/databricks_workflow_task_group_airflow_graph_view.png
similarity index 100%
rename from docs/apache-airflow-providers-databricks/img/databricks_workflow_task_group_airflow_graph_view.png
rename to providers/databricks/docs/img/databricks_workflow_task_group_airflow_graph_view.png
diff --git a/docs/apache-airflow-providers-databricks/img/workflow_plugin_launch_task.png b/providers/databricks/docs/img/workflow_plugin_launch_task.png
similarity index 100%
rename from docs/apache-airflow-providers-databricks/img/workflow_plugin_launch_task.png
rename to providers/databricks/docs/img/workflow_plugin_launch_task.png
diff --git a/docs/apache-airflow-providers-databricks/img/workflow_plugin_single_task.png b/providers/databricks/docs/img/workflow_plugin_single_task.png
similarity index 100%
rename from docs/apache-airflow-providers-databricks/img/workflow_plugin_single_task.png
rename to providers/databricks/docs/img/workflow_plugin_single_task.png
diff --git a/docs/apache-airflow-providers-databricks/img/workflow_run_databricks_graph_view.png b/providers/databricks/docs/img/workflow_run_databricks_graph_view.png
similarity index 100%
rename from docs/apache-airflow-providers-databricks/img/workflow_run_databricks_graph_view.png
rename to providers/databricks/docs/img/workflow_run_databricks_graph_view.png
diff --git a/docs/apache-airflow-providers-databricks/index.rst b/providers/databricks/docs/index.rst
similarity index 100%
rename from docs/apache-airflow-providers-databricks/index.rst
rename to providers/databricks/docs/index.rst
diff --git a/docs/apache-airflow-providers-databricks/installing-providers-from-sources.rst b/providers/databricks/docs/installing-providers-from-sources.rst
similarity index 100%
rename from docs/apache-airflow-providers-databricks/installing-providers-from-sources.rst
rename to providers/databricks/docs/installing-providers-from-sources.rst
diff --git a/docs/integration-logos/databricks/Databricks.png b/providers/databricks/docs/integration-logos/Databricks.png
similarity index 100%
rename from docs/integration-logos/databricks/Databricks.png
rename to providers/databricks/docs/integration-logos/Databricks.png
diff --git a/docs/apache-airflow-providers-databricks/operators/copy_into.rst b/providers/databricks/docs/operators/copy_into.rst
similarity index 95%
rename from docs/apache-airflow-providers-databricks/operators/copy_into.rst
rename to providers/databricks/docs/operators/copy_into.rst
index 56eb20c662286..7b4ce43567a6d 100644
--- a/docs/apache-airflow-providers-databricks/operators/copy_into.rst
+++ b/providers/databricks/docs/operators/copy_into.rst
@@ -46,7 +46,7 @@ Importing CSV data
An example usage of the DatabricksCopyIntoOperator to import CSV data into a table is as follows:
-.. exampleinclude:: /../../providers/tests/system/databricks/example_databricks_sql.py
+.. exampleinclude:: /../../providers/databricks/tests/system/databricks/example_databricks_sql.py
:language: python
:start-after: [START howto_operator_databricks_copy_into]
:end-before: [END howto_operator_databricks_copy_into]
diff --git a/docs/apache-airflow-providers-databricks/operators/index.rst b/providers/databricks/docs/operators/index.rst
similarity index 100%
rename from docs/apache-airflow-providers-databricks/operators/index.rst
rename to providers/databricks/docs/operators/index.rst
diff --git a/docs/apache-airflow-providers-databricks/operators/jobs_create.rst b/providers/databricks/docs/operators/jobs_create.rst
similarity index 92%
rename from docs/apache-airflow-providers-databricks/operators/jobs_create.rst
rename to providers/databricks/docs/operators/jobs_create.rst
index 5a79c8244a966..ad4525196ef9d 100644
--- a/docs/apache-airflow-providers-databricks/operators/jobs_create.rst
+++ b/providers/databricks/docs/operators/jobs_create.rst
@@ -67,7 +67,7 @@ Specifying parameters as JSON
An example usage of the DatabricksCreateJobsOperator is as follows:
-.. exampleinclude:: /../../providers/tests/system/databricks/example_databricks.py
+.. exampleinclude:: /../../providers/databricks/tests/system/databricks/example_databricks.py
:language: python
:start-after: [START howto_operator_databricks_jobs_create_json]
:end-before: [END howto_operator_databricks_jobs_create_json]
@@ -77,7 +77,7 @@ Using named parameters
You can also use named parameters to initialize the operator and run the job.
-.. exampleinclude:: /../../providers/tests/system/databricks/example_databricks.py
+.. exampleinclude:: /../../providers/databricks/tests/system/databricks/example_databricks.py
:language: python
:start-after: [START howto_operator_databricks_jobs_create_named]
:end-before: [END howto_operator_databricks_jobs_create_named]
@@ -88,7 +88,7 @@ Pairing with DatabricksRunNowOperator
You can use the ``job_id`` that is returned by the DatabricksCreateJobsOperator in the
return_value XCom as an argument to the DatabricksRunNowOperator to run the job.
-.. exampleinclude:: /../../providers/tests/system/databricks/example_databricks.py
+.. exampleinclude:: /../../providers/databricks/tests/system/databricks/example_databricks.py
:language: python
:start-after: [START howto_operator_databricks_run_now]
:end-before: [END howto_operator_databricks_run_now]
diff --git a/docs/apache-airflow-providers-databricks/operators/notebook.rst b/providers/databricks/docs/operators/notebook.rst
similarity index 89%
rename from docs/apache-airflow-providers-databricks/operators/notebook.rst
rename to providers/databricks/docs/operators/notebook.rst
index bf7b04ca74caa..ba9e11b97ea64 100644
--- a/docs/apache-airflow-providers-databricks/operators/notebook.rst
+++ b/providers/databricks/docs/operators/notebook.rst
@@ -31,14 +31,14 @@ Examples
Running a notebook in Databricks on a new cluster
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
-.. exampleinclude:: /../../providers/tests/system/databricks/example_databricks.py
+.. exampleinclude:: /../../providers/databricks/tests/system/databricks/example_databricks.py
:language: python
:start-after: [START howto_operator_databricks_notebook_new_cluster]
:end-before: [END howto_operator_databricks_notebook_new_cluster]
Running a notebook in Databricks on an existing cluster
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
-.. exampleinclude:: /../../providers/tests/system/databricks/example_databricks.py
+.. exampleinclude:: /../../providers/databricks/tests/system/databricks/example_databricks.py
:language: python
:start-after: [START howto_operator_databricks_notebook_existing_cluster]
:end-before: [END howto_operator_databricks_notebook_existing_cluster]
diff --git a/docs/apache-airflow-providers-databricks/operators/repos_create.rst b/providers/databricks/docs/operators/repos_create.rst
similarity index 96%
rename from docs/apache-airflow-providers-databricks/operators/repos_create.rst
rename to providers/databricks/docs/operators/repos_create.rst
index 6b60eae27870b..d9608bad91c22 100644
--- a/docs/apache-airflow-providers-databricks/operators/repos_create.rst
+++ b/providers/databricks/docs/operators/repos_create.rst
@@ -63,7 +63,7 @@ Create a Databricks Repo
An example usage of the DatabricksReposCreateOperator is as follows:
-.. exampleinclude:: /../../providers/tests/system/databricks/example_databricks_repos.py
+.. exampleinclude:: /../../providers/databricks/tests/system/databricks/example_databricks_repos.py
:language: python
:start-after: [START howto_operator_databricks_repo_create]
:end-before: [END howto_operator_databricks_repo_create]
diff --git a/docs/apache-airflow-providers-databricks/operators/repos_delete.rst b/providers/databricks/docs/operators/repos_delete.rst
similarity index 95%
rename from docs/apache-airflow-providers-databricks/operators/repos_delete.rst
rename to providers/databricks/docs/operators/repos_delete.rst
index 3186dd131dff7..09849c6ae0c7c 100644
--- a/docs/apache-airflow-providers-databricks/operators/repos_delete.rst
+++ b/providers/databricks/docs/operators/repos_delete.rst
@@ -55,7 +55,7 @@ Deleting Databricks Repo by specifying path
An example usage of the DatabricksReposDeleteOperator is as follows:
-.. exampleinclude:: /../../providers/tests/system/databricks/example_databricks_repos.py
+.. exampleinclude:: /../../providers/databricks/tests/system/databricks/example_databricks_repos.py
:language: python
:start-after: [START howto_operator_databricks_repo_delete]
:end-before: [END howto_operator_databricks_repo_delete]
diff --git a/docs/apache-airflow-providers-databricks/operators/repos_update.rst b/providers/databricks/docs/operators/repos_update.rst
similarity index 96%
rename from docs/apache-airflow-providers-databricks/operators/repos_update.rst
rename to providers/databricks/docs/operators/repos_update.rst
index 6893ee0107e86..90f700ff581a6 100644
--- a/docs/apache-airflow-providers-databricks/operators/repos_update.rst
+++ b/providers/databricks/docs/operators/repos_update.rst
@@ -60,7 +60,7 @@ Updating Databricks Repo by specifying path
An example usage of the DatabricksReposUpdateOperator is as follows:
-.. exampleinclude:: /../../providers/tests/system/databricks/example_databricks_repos.py
+.. exampleinclude:: /../../providers/databricks/tests/system/databricks/example_databricks_repos.py
:language: python
:start-after: [START howto_operator_databricks_repo_update]
:end-before: [END howto_operator_databricks_repo_update]
diff --git a/docs/apache-airflow-providers-databricks/operators/run_now.rst b/providers/databricks/docs/operators/run_now.rst
similarity index 100%
rename from docs/apache-airflow-providers-databricks/operators/run_now.rst
rename to providers/databricks/docs/operators/run_now.rst
diff --git a/docs/apache-airflow-providers-databricks/operators/sql.rst b/providers/databricks/docs/operators/sql.rst
similarity index 88%
rename from docs/apache-airflow-providers-databricks/operators/sql.rst
rename to providers/databricks/docs/operators/sql.rst
index 33acacae1c6bd..336bb3cc01bc3 100644
--- a/docs/apache-airflow-providers-databricks/operators/sql.rst
+++ b/providers/databricks/docs/operators/sql.rst
@@ -49,7 +49,7 @@ Selecting data
An example usage of the DatabricksSqlOperator to select data from a table is as follows:
-.. exampleinclude:: /../../providers/tests/system/databricks/example_databricks_sql.py
+.. exampleinclude:: /../../providers/databricks/tests/system/databricks/example_databricks_sql.py
:language: python
:start-after: [START howto_operator_databricks_sql_select]
:end-before: [END howto_operator_databricks_sql_select]
@@ -59,7 +59,7 @@ Selecting data into a file
An example usage of the DatabricksSqlOperator to select data from a table and store in a file is as follows:
-.. exampleinclude:: /../../providers/tests/system/databricks/example_databricks_sql.py
+.. exampleinclude:: /../../providers/databricks/tests/system/databricks/example_databricks_sql.py
:language: python
:start-after: [START howto_operator_databricks_sql_select_file]
:end-before: [END howto_operator_databricks_sql_select_file]
@@ -69,7 +69,7 @@ Executing multiple statements
An example usage of the DatabricksSqlOperator to perform multiple SQL statements is as follows:
-.. exampleinclude:: /../../providers/tests/system/databricks/example_databricks_sql.py
+.. exampleinclude:: /../../providers/databricks/tests/system/databricks/example_databricks_sql.py
:language: python
:start-after: [START howto_operator_databricks_sql_multiple]
:end-before: [END howto_operator_databricks_sql_multiple]
@@ -80,7 +80,7 @@ Executing multiple statements from a file
An example usage of the DatabricksSqlOperator to perform statements from a file is as follows:
-.. exampleinclude:: /../../providers/tests/system/databricks/example_databricks_sql.py
+.. exampleinclude:: /../../providers/databricks/tests/system/databricks/example_databricks_sql.py
:language: python
:start-after: [START howto_operator_databricks_sql_multiple_file]
:end-before: [END howto_operator_databricks_sql_multiple_file]
@@ -107,7 +107,7 @@ Examples
--------
Configuring Databricks connection to be used with the Sensor.
-.. exampleinclude:: /../../providers/tests/system/databricks/example_databricks_sensors.py
+.. exampleinclude:: /../../providers/databricks/tests/system/databricks/example_databricks_sensors.py
:language: python
:dedent: 4
:start-after: [START howto_sensor_databricks_connection_setup]
@@ -115,7 +115,7 @@ Configuring Databricks connection to be used with the Sensor.
Poking the specific table with the SQL statement:
-.. exampleinclude:: /../../providers/tests/system/databricks/example_databricks_sensors.py
+.. exampleinclude:: /../../providers/databricks/tests/system/databricks/example_databricks_sensors.py
:language: python
:dedent: 4
:start-after: [START howto_sensor_databricks_sql]
@@ -154,7 +154,7 @@ Examples
--------
Configuring Databricks connection to be used with the Sensor.
-.. exampleinclude:: /../../providers/tests/system/databricks/example_databricks_sensors.py
+.. exampleinclude:: /../../providers/databricks/tests/system/databricks/example_databricks_sensors.py
:language: python
:dedent: 4
:start-after: [START howto_sensor_databricks_connection_setup]
@@ -162,7 +162,7 @@ Configuring Databricks connection to be used with the Sensor.
Poking the specific table for existence of data/partition:
-.. exampleinclude:: /../../providers/tests/system/databricks/example_databricks_sensors.py
+.. exampleinclude:: /../../providers/databricks/tests/system/databricks/example_databricks_sensors.py
:language: python
:dedent: 4
:start-after: [START howto_sensor_databricks_partition]
diff --git a/docs/apache-airflow-providers-databricks/operators/submit_run.rst b/providers/databricks/docs/operators/submit_run.rst
similarity index 96%
rename from docs/apache-airflow-providers-databricks/operators/submit_run.rst
rename to providers/databricks/docs/operators/submit_run.rst
index 7a8d13f646ce5..813af046127f2 100644
--- a/docs/apache-airflow-providers-databricks/operators/submit_run.rst
+++ b/providers/databricks/docs/operators/submit_run.rst
@@ -114,7 +114,7 @@ Specifying parameters as JSON
An example usage of the DatabricksSubmitRunOperator is as follows:
-.. exampleinclude:: /../../providers/tests/system/databricks/example_databricks.py
+.. exampleinclude:: /../../providers/databricks/tests/system/databricks/example_databricks.py
:language: python
:start-after: [START howto_operator_databricks_json]
:end-before: [END howto_operator_databricks_json]
@@ -124,7 +124,7 @@ Using named parameters
You can also use named parameters to initialize the operator and run the job.
-.. exampleinclude:: /../../providers/tests/system/databricks/example_databricks.py
+.. exampleinclude:: /../../providers/databricks/tests/system/databricks/example_databricks.py
:language: python
:start-after: [START howto_operator_databricks_named]
:end-before: [END howto_operator_databricks_named]
diff --git a/docs/apache-airflow-providers-databricks/operators/task.rst b/providers/databricks/docs/operators/task.rst
similarity index 88%
rename from docs/apache-airflow-providers-databricks/operators/task.rst
rename to providers/databricks/docs/operators/task.rst
index 47ceafe58ad04..a1b83b3bb932c 100644
--- a/docs/apache-airflow-providers-databricks/operators/task.rst
+++ b/providers/databricks/docs/operators/task.rst
@@ -33,21 +33,21 @@ Examples
Running a notebook in Databricks using DatabricksTaskOperator
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
-.. exampleinclude:: /../../providers/tests/system/databricks/example_databricks.py
+.. exampleinclude:: /../../providers/databricks/tests/system/databricks/example_databricks.py
:language: python
:start-after: [START howto_operator_databricks_task_notebook]
:end-before: [END howto_operator_databricks_task_notebook]
Running a SQL query in Databricks using DatabricksTaskOperator
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
-.. exampleinclude:: /../../providers/tests/system/databricks/example_databricks.py
+.. exampleinclude:: /../../providers/databricks/tests/system/databricks/example_databricks.py
:language: python
:start-after: [START howto_operator_databricks_task_sql]
:end-before: [END howto_operator_databricks_task_sql]
Running a python file in Databricks in using DatabricksTaskOperator
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
-.. exampleinclude:: /../../providers/tests/system/databricks/example_databricks.py
+.. exampleinclude:: /../../providers/databricks/tests/system/databricks/example_databricks.py
:language: python
:start-after: [START howto_operator_databricks_task_python]
:end-before: [END howto_operator_databricks_task_python]
diff --git a/docs/apache-airflow-providers-databricks/operators/workflow.rst b/providers/databricks/docs/operators/workflow.rst
similarity index 97%
rename from docs/apache-airflow-providers-databricks/operators/workflow.rst
rename to providers/databricks/docs/operators/workflow.rst
index b5c81050143a8..84f245ab20b23 100644
--- a/docs/apache-airflow-providers-databricks/operators/workflow.rst
+++ b/providers/databricks/docs/operators/workflow.rst
@@ -45,7 +45,7 @@ Examples
Example of what a DAG looks like with a DatabricksWorkflowTaskGroup
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
-.. exampleinclude:: /../../providers/tests/system/databricks/example_databricks_workflow.py
+.. exampleinclude:: /../../providers/databricks/tests/system/databricks/example_databricks_workflow.py
:language: python
:start-after: [START howto_databricks_workflow_notebook]
:end-before: [END howto_databricks_workflow_notebook]
diff --git a/docs/apache-airflow-providers-databricks/plugins/index.rst b/providers/databricks/docs/plugins/index.rst
similarity index 100%
rename from docs/apache-airflow-providers-databricks/plugins/index.rst
rename to providers/databricks/docs/plugins/index.rst
diff --git a/docs/apache-airflow-providers-databricks/plugins/workflow.rst b/providers/databricks/docs/plugins/workflow.rst
similarity index 100%
rename from docs/apache-airflow-providers-databricks/plugins/workflow.rst
rename to providers/databricks/docs/plugins/workflow.rst
diff --git a/docs/apache-airflow-providers-databricks/security.rst b/providers/databricks/docs/security.rst
similarity index 100%
rename from docs/apache-airflow-providers-databricks/security.rst
rename to providers/databricks/docs/security.rst
diff --git a/providers/src/airflow/providers/databricks/provider.yaml b/providers/databricks/provider.yaml
similarity index 84%
rename from providers/src/airflow/providers/databricks/provider.yaml
rename to providers/databricks/provider.yaml
index 0efe1474ec165..950c223a74f9d 100644
--- a/providers/src/airflow/providers/databricks/provider.yaml
+++ b/providers/databricks/provider.yaml
@@ -72,31 +72,6 @@ versions:
- 1.0.1
- 1.0.0
-dependencies:
- - apache-airflow>=2.9.0
- - apache-airflow-providers-common-sql>=1.20.0
- - requests>=2.27.0,<3
- - databricks-sql-connector>=3.0.0
- - aiohttp>=3.9.2, <4
- - mergedeep>=1.3.4
- - pandas>=2.1.2,<2.2
- - pyarrow>=14.0.1
-
-
-additional-extras:
- # pip install apache-airflow-providers-databricks[sdk]
- - name: sdk
- description: Install Databricks SDK
- dependencies:
- - databricks-sdk==0.10.0
- - name: azure-identity
- description: Install Azure Identity client library
- dependencies:
- - azure-identity>=1.3.1
-
-devel-dependencies:
- - deltalake>=0.12.0
-
integrations:
- integration-name: Databricks
external-doc-url: https://databricks.com/
@@ -106,14 +81,13 @@ integrations:
- /docs/apache-airflow-providers-databricks/operators/submit_run.rst
- /docs/apache-airflow-providers-databricks/operators/run_now.rst
- /docs/apache-airflow-providers-databricks/operators/task.rst
- logo: /integration-logos/databricks/Databricks.png
+ logo: /docs/integration-logos/Databricks.png
tags: [service]
- integration-name: Databricks SQL
external-doc-url: https://databricks.com/product/databricks-sql
how-to-guide:
- /docs/apache-airflow-providers-databricks/operators/sql.rst
- /docs/apache-airflow-providers-databricks/operators/copy_into.rst
- logo: /integration-logos/databricks/Databricks.png
tags: [service]
- integration-name: Databricks Repos
external-doc-url: https://docs.databricks.com/repos/index.html
@@ -121,13 +95,11 @@ integrations:
- /docs/apache-airflow-providers-databricks/operators/repos_create.rst
- /docs/apache-airflow-providers-databricks/operators/repos_update.rst
- /docs/apache-airflow-providers-databricks/operators/repos_delete.rst
- logo: /integration-logos/databricks/Databricks.png
tags: [service]
- integration-name: Databricks Workflow
external-doc-url: https://docs.databricks.com/en/workflows/index.html
how-to-guide:
- /docs/apache-airflow-providers-databricks/operators/workflow.rst
- logo: /integration-logos/databricks/Databricks.png
tags: [service]
operators:
diff --git a/providers/databricks/pyproject.toml b/providers/databricks/pyproject.toml
new file mode 100644
index 0000000000000..49152151e9caf
--- /dev/null
+++ b/providers/databricks/pyproject.toml
@@ -0,0 +1,105 @@
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements. See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership. The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License. You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing,
+# software distributed under the License is distributed on an
+# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+# KIND, either express or implied. See the License for the
+# specific language governing permissions and limitations
+# under the License.
+
+# NOTE! THIS FILE IS AUTOMATICALLY GENERATED AND WILL BE OVERWRITTEN!
+
+# IF YOU WANT TO MODIFY THIS FILE EXCEPT DEPENDENCIES, YOU SHOULD MODIFY THE TEMPLATE
+# `pyproject_TEMPLATE.toml.jinja2` IN the `dev/breeze/src/airflow_breeze/templates` DIRECTORY
+[build-system]
+requires = ["flit_core==3.10.1"]
+build-backend = "flit_core.buildapi"
+
+[project]
+name = "apache-airflow-providers-databricks"
+version = "7.0.0"
+description = "Provider package apache-airflow-providers-databricks for Apache Airflow"
+readme = "README.rst"
+authors = [
+ {name="Apache Software Foundation", email="dev@airflow.apache.org"},
+]
+maintainers = [
+ {name="Apache Software Foundation", email="dev@airflow.apache.org"},
+]
+keywords = [ "airflow-provider", "databricks", "airflow", "integration" ]
+classifiers = [
+ "Development Status :: 5 - Production/Stable",
+ "Environment :: Console",
+ "Environment :: Web Environment",
+ "Intended Audience :: Developers",
+ "Intended Audience :: System Administrators",
+ "Framework :: Apache Airflow",
+ "Framework :: Apache Airflow :: Provider",
+ "License :: OSI Approved :: Apache Software License",
+ "Programming Language :: Python :: 3.9",
+ "Programming Language :: Python :: 3.10",
+ "Programming Language :: Python :: 3.11",
+ "Programming Language :: Python :: 3.12",
+ "Topic :: System :: Monitoring",
+]
+requires-python = "~=3.9"
+
+# The dependencies should be modified in place in the generated file
+# Any change in the dependencies is preserved when the file is regenerated
+dependencies = [
+ "apache-airflow>=2.9.0",
+ "apache-airflow-providers-common-sql>=1.20.0",
+ "requests>=2.27.0,<3",
+ "databricks-sql-connector>=3.0.0",
+ "aiohttp>=3.9.2, <4",
+ "mergedeep>=1.3.4",
+ "pandas>=2.1.2,<2.2",
+ "pyarrow>=14.0.1",
+]
+
+# The optional dependencies should be modified in place in the generated file
+# Any change in the dependencies is preserved when the file is regenerated
+[project.optional-dependencies]
+# pip install apache-airflow-providers-databricks[sdk]
+"sdk" = [
+ "databricks-sdk==0.10.0",
+]
+"azure-identity" = [
+ "azure-identity>=1.3.1",
+]
+
+# The dependency groups should be modified in place in the generated file
+# Any change in the dependencies is preserved when the file is regenerated
+[dependency-groups]
+dev = [
+ "deltalake>=0.12.0",
+]
+
+[project.urls]
+"Documentation" = "https://airflow.apache.org/docs/apache-airflow-providers-databricks/7.0.0"
+"Changelog" = "https://airflow.apache.org/docs/apache-airflow-providers-databricks/7.0.0/changelog.html"
+"Bug Tracker" = "https://github.com/apache/airflow/issues"
+"Source Code" = "https://github.com/apache/airflow"
+"Slack Chat" = "https://s.apache.org/airflow-slack"
+"Twitter" = "https://x.com/ApacheAirflow"
+"YouTube" = "https://www.youtube.com/channel/UCSXwxpWZQ7XZ1WL3wqevChA/"
+
+[project.entry-points."apache_airflow_provider"]
+provider_info = "airflow.providers.databricks.get_provider_info:get_provider_info"
+
+[project.entry-points."airflow.plugins"]
+databricks_workflow = "airflow.providers.databricks.plugins.databricks_workflow:DatabricksWorkflowPlugin"
+
+[tool.flit.module]
+name = "airflow.providers.databricks"
+
+[tool.pytest.ini_options]
+ignore = "tests/system/"
diff --git a/providers/databricks/src/airflow/providers/databricks/LICENSE b/providers/databricks/src/airflow/providers/databricks/LICENSE
new file mode 100644
index 0000000000000..11069edd79019
--- /dev/null
+++ b/providers/databricks/src/airflow/providers/databricks/LICENSE
@@ -0,0 +1,201 @@
+ Apache License
+ Version 2.0, January 2004
+ http://www.apache.org/licenses/
+
+TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
+
+1. Definitions.
+
+ "License" shall mean the terms and conditions for use, reproduction,
+ and distribution as defined by Sections 1 through 9 of this document.
+
+ "Licensor" shall mean the copyright owner or entity authorized by
+ the copyright owner that is granting the License.
+
+ "Legal Entity" shall mean the union of the acting entity and all
+ other entities that control, are controlled by, or are under common
+ control with that entity. For the purposes of this definition,
+ "control" means (i) the power, direct or indirect, to cause the
+ direction or management of such entity, whether by contract or
+ otherwise, or (ii) ownership of fifty percent (50%) or more of the
+ outstanding shares, or (iii) beneficial ownership of such entity.
+
+ "You" (or "Your") shall mean an individual or Legal Entity
+ exercising permissions granted by this License.
+
+ "Source" form shall mean the preferred form for making modifications,
+ including but not limited to software source code, documentation
+ source, and configuration files.
+
+ "Object" form shall mean any form resulting from mechanical
+ transformation or translation of a Source form, including but
+ not limited to compiled object code, generated documentation,
+ and conversions to other media types.
+
+ "Work" shall mean the work of authorship, whether in Source or
+ Object form, made available under the License, as indicated by a
+ copyright notice that is included in or attached to the work
+ (an example is provided in the Appendix below).
+
+ "Derivative Works" shall mean any work, whether in Source or Object
+ form, that is based on (or derived from) the Work and for which the
+ editorial revisions, annotations, elaborations, or other modifications
+ represent, as a whole, an original work of authorship. For the purposes
+ of this License, Derivative Works shall not include works that remain
+ separable from, or merely link (or bind by name) to the interfaces of,
+ the Work and Derivative Works thereof.
+
+ "Contribution" shall mean any work of authorship, including
+ the original version of the Work and any modifications or additions
+ to that Work or Derivative Works thereof, that is intentionally
+ submitted to Licensor for inclusion in the Work by the copyright owner
+ or by an individual or Legal Entity authorized to submit on behalf of
+ the copyright owner. For the purposes of this definition, "submitted"
+ means any form of electronic, verbal, or written communication sent
+ to the Licensor or its representatives, including but not limited to
+ communication on electronic mailing lists, source code control systems,
+ and issue tracking systems that are managed by, or on behalf of, the
+ Licensor for the purpose of discussing and improving the Work, but
+ excluding communication that is conspicuously marked or otherwise
+ designated in writing by the copyright owner as "Not a Contribution."
+
+ "Contributor" shall mean Licensor and any individual or Legal Entity
+ on behalf of whom a Contribution has been received by Licensor and
+ subsequently incorporated within the Work.
+
+2. Grant of Copyright License. Subject to the terms and conditions of
+ this License, each Contributor hereby grants to You a perpetual,
+ worldwide, non-exclusive, no-charge, royalty-free, irrevocable
+ copyright license to reproduce, prepare Derivative Works of,
+ publicly display, publicly perform, sublicense, and distribute the
+ Work and such Derivative Works in Source or Object form.
+
+3. Grant of Patent License. Subject to the terms and conditions of
+ this License, each Contributor hereby grants to You a perpetual,
+ worldwide, non-exclusive, no-charge, royalty-free, irrevocable
+ (except as stated in this section) patent license to make, have made,
+ use, offer to sell, sell, import, and otherwise transfer the Work,
+ where such license applies only to those patent claims licensable
+ by such Contributor that are necessarily infringed by their
+ Contribution(s) alone or by combination of their Contribution(s)
+ with the Work to which such Contribution(s) was submitted. If You
+ institute patent litigation against any entity (including a
+ cross-claim or counterclaim in a lawsuit) alleging that the Work
+ or a Contribution incorporated within the Work constitutes direct
+ or contributory patent infringement, then any patent licenses
+ granted to You under this License for that Work shall terminate
+ as of the date such litigation is filed.
+
+4. Redistribution. You may reproduce and distribute copies of the
+ Work or Derivative Works thereof in any medium, with or without
+ modifications, and in Source or Object form, provided that You
+ meet the following conditions:
+
+ (a) You must give any other recipients of the Work or
+ Derivative Works a copy of this License; and
+
+ (b) You must cause any modified files to carry prominent notices
+ stating that You changed the files; and
+
+ (c) You must retain, in the Source form of any Derivative Works
+ that You distribute, all copyright, patent, trademark, and
+ attribution notices from the Source form of the Work,
+ excluding those notices that do not pertain to any part of
+ the Derivative Works; and
+
+ (d) If the Work includes a "NOTICE" text file as part of its
+ distribution, then any Derivative Works that You distribute must
+ include a readable copy of the attribution notices contained
+ within such NOTICE file, excluding those notices that do not
+ pertain to any part of the Derivative Works, in at least one
+ of the following places: within a NOTICE text file distributed
+ as part of the Derivative Works; within the Source form or
+ documentation, if provided along with the Derivative Works; or,
+ within a display generated by the Derivative Works, if and
+ wherever such third-party notices normally appear. The contents
+ of the NOTICE file are for informational purposes only and
+ do not modify the License. You may add Your own attribution
+ notices within Derivative Works that You distribute, alongside
+ or as an addendum to the NOTICE text from the Work, provided
+ that such additional attribution notices cannot be construed
+ as modifying the License.
+
+ You may add Your own copyright statement to Your modifications and
+ may provide additional or different license terms and conditions
+ for use, reproduction, or distribution of Your modifications, or
+ for any such Derivative Works as a whole, provided Your use,
+ reproduction, and distribution of the Work otherwise complies with
+ the conditions stated in this License.
+
+5. Submission of Contributions. Unless You explicitly state otherwise,
+ any Contribution intentionally submitted for inclusion in the Work
+ by You to the Licensor shall be under the terms and conditions of
+ this License, without any additional terms or conditions.
+ Notwithstanding the above, nothing herein shall supersede or modify
+ the terms of any separate license agreement you may have executed
+ with Licensor regarding such Contributions.
+
+6. Trademarks. This License does not grant permission to use the trade
+ names, trademarks, service marks, or product names of the Licensor,
+ except as required for reasonable and customary use in describing the
+ origin of the Work and reproducing the content of the NOTICE file.
+
+7. Disclaimer of Warranty. Unless required by applicable law or
+ agreed to in writing, Licensor provides the Work (and each
+ Contributor provides its Contributions) on an "AS IS" BASIS,
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
+ implied, including, without limitation, any warranties or conditions
+ of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
+ PARTICULAR PURPOSE. You are solely responsible for determining the
+ appropriateness of using or redistributing the Work and assume any
+ risks associated with Your exercise of permissions under this License.
+
+8. Limitation of Liability. In no event and under no legal theory,
+ whether in tort (including negligence), contract, or otherwise,
+ unless required by applicable law (such as deliberate and grossly
+ negligent acts) or agreed to in writing, shall any Contributor be
+ liable to You for damages, including any direct, indirect, special,
+ incidental, or consequential damages of any character arising as a
+ result of this License or out of the use or inability to use the
+ Work (including but not limited to damages for loss of goodwill,
+ work stoppage, computer failure or malfunction, or any and all
+ other commercial damages or losses), even if such Contributor
+ has been advised of the possibility of such damages.
+
+9. Accepting Warranty or Additional Liability. While redistributing
+ the Work or Derivative Works thereof, You may choose to offer,
+ and charge a fee for, acceptance of support, warranty, indemnity,
+ or other liability obligations and/or rights consistent with this
+ License. However, in accepting such obligations, You may act only
+ on Your own behalf and on Your sole responsibility, not on behalf
+ of any other Contributor, and only if You agree to indemnify,
+ defend, and hold each Contributor harmless for any liability
+ incurred by, or claims asserted against, such Contributor by reason
+ of your accepting any such warranty or additional liability.
+
+END OF TERMS AND CONDITIONS
+
+APPENDIX: How to apply the Apache License to your work.
+
+ To apply the Apache License to your work, attach the following
+ boilerplate notice, with the fields enclosed by brackets "[]"
+ replaced with your own identifying information. (Don't include
+ the brackets!) The text should be enclosed in the appropriate
+ comment syntax for the file format. We also recommend that a
+ file or class name and description of purpose be included on the
+ same "printed page" as the copyright notice for easier
+ identification within third-party archives.
+
+Copyright [yyyy] [name of copyright owner]
+
+Licensed under the Apache License, Version 2.0 (the "License");
+you may not use this file except in compliance with the License.
+You may obtain a copy of the License at
+
+ http://www.apache.org/licenses/LICENSE-2.0
+
+Unless required by applicable law or agreed to in writing, software
+distributed under the License is distributed on an "AS IS" BASIS,
+WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+See the License for the specific language governing permissions and
+limitations under the License.
diff --git a/providers/src/airflow/providers/databricks/__init__.py b/providers/databricks/src/airflow/providers/databricks/__init__.py
similarity index 100%
rename from providers/src/airflow/providers/databricks/__init__.py
rename to providers/databricks/src/airflow/providers/databricks/__init__.py
diff --git a/providers/src/airflow/providers/databricks/exceptions.py b/providers/databricks/src/airflow/providers/databricks/exceptions.py
similarity index 100%
rename from providers/src/airflow/providers/databricks/exceptions.py
rename to providers/databricks/src/airflow/providers/databricks/exceptions.py
diff --git a/providers/databricks/src/airflow/providers/databricks/get_provider_info.py b/providers/databricks/src/airflow/providers/databricks/get_provider_info.py
new file mode 100644
index 0000000000000..d18af30cdb2fd
--- /dev/null
+++ b/providers/databricks/src/airflow/providers/databricks/get_provider_info.py
@@ -0,0 +1,193 @@
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements. See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership. The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License. You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing,
+# software distributed under the License is distributed on an
+# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+# KIND, either express or implied. See the License for the
+# specific language governing permissions and limitations
+# under the License.
+
+# NOTE! THIS FILE IS AUTOMATICALLY GENERATED AND WILL BE OVERWRITTEN!
+#
+# IF YOU WANT TO MODIFY THIS FILE, YOU SHOULD MODIFY THE TEMPLATE
+# `get_provider_info_TEMPLATE.py.jinja2` IN the `dev/breeze/src/airflow_breeze/templates` DIRECTORY
+
+
+def get_provider_info():
+ return {
+ "package-name": "apache-airflow-providers-databricks",
+ "name": "Databricks",
+ "description": "`Databricks `__\n",
+ "state": "ready",
+ "source-date-epoch": 1734533222,
+ "versions": [
+ "7.0.0",
+ "6.13.0",
+ "6.12.0",
+ "6.11.0",
+ "6.10.0",
+ "6.9.0",
+ "6.8.0",
+ "6.7.0",
+ "6.6.0",
+ "6.5.0",
+ "6.4.0",
+ "6.3.0",
+ "6.2.0",
+ "6.1.0",
+ "6.0.0",
+ "5.0.1",
+ "5.0.0",
+ "4.7.0",
+ "4.6.0",
+ "4.5.0",
+ "4.4.0",
+ "4.3.3",
+ "4.3.2",
+ "4.3.1",
+ "4.3.0",
+ "4.2.0",
+ "4.1.0",
+ "4.0.1",
+ "4.0.0",
+ "3.4.0",
+ "3.3.0",
+ "3.2.0",
+ "3.1.0",
+ "3.0.0",
+ "2.7.0",
+ "2.6.0",
+ "2.5.0",
+ "2.4.0",
+ "2.3.0",
+ "2.2.0",
+ "2.1.0",
+ "2.0.2",
+ "2.0.1",
+ "2.0.0",
+ "1.0.1",
+ "1.0.0",
+ ],
+ "integrations": [
+ {
+ "integration-name": "Databricks",
+ "external-doc-url": "https://databricks.com/",
+ "how-to-guide": [
+ "/docs/apache-airflow-providers-databricks/operators/jobs_create.rst",
+ "/docs/apache-airflow-providers-databricks/operators/notebook.rst",
+ "/docs/apache-airflow-providers-databricks/operators/submit_run.rst",
+ "/docs/apache-airflow-providers-databricks/operators/run_now.rst",
+ "/docs/apache-airflow-providers-databricks/operators/task.rst",
+ ],
+ "logo": "/docs/integration-logos/Databricks.png",
+ "tags": ["service"],
+ },
+ {
+ "integration-name": "Databricks SQL",
+ "external-doc-url": "https://databricks.com/product/databricks-sql",
+ "how-to-guide": [
+ "/docs/apache-airflow-providers-databricks/operators/sql.rst",
+ "/docs/apache-airflow-providers-databricks/operators/copy_into.rst",
+ ],
+ "tags": ["service"],
+ },
+ {
+ "integration-name": "Databricks Repos",
+ "external-doc-url": "https://docs.databricks.com/repos/index.html",
+ "how-to-guide": [
+ "/docs/apache-airflow-providers-databricks/operators/repos_create.rst",
+ "/docs/apache-airflow-providers-databricks/operators/repos_update.rst",
+ "/docs/apache-airflow-providers-databricks/operators/repos_delete.rst",
+ ],
+ "tags": ["service"],
+ },
+ {
+ "integration-name": "Databricks Workflow",
+ "external-doc-url": "https://docs.databricks.com/en/workflows/index.html",
+ "how-to-guide": ["/docs/apache-airflow-providers-databricks/operators/workflow.rst"],
+ "tags": ["service"],
+ },
+ ],
+ "operators": [
+ {
+ "integration-name": "Databricks",
+ "python-modules": ["airflow.providers.databricks.operators.databricks"],
+ },
+ {
+ "integration-name": "Databricks SQL",
+ "python-modules": ["airflow.providers.databricks.operators.databricks_sql"],
+ },
+ {
+ "integration-name": "Databricks Repos",
+ "python-modules": ["airflow.providers.databricks.operators.databricks_repos"],
+ },
+ {
+ "integration-name": "Databricks Workflow",
+ "python-modules": ["airflow.providers.databricks.operators.databricks_workflow"],
+ },
+ ],
+ "hooks": [
+ {
+ "integration-name": "Databricks",
+ "python-modules": [
+ "airflow.providers.databricks.hooks.databricks",
+ "airflow.providers.databricks.hooks.databricks_base",
+ ],
+ },
+ {
+ "integration-name": "Databricks SQL",
+ "python-modules": ["airflow.providers.databricks.hooks.databricks_sql"],
+ },
+ ],
+ "triggers": [
+ {
+ "integration-name": "Databricks",
+ "python-modules": ["airflow.providers.databricks.triggers.databricks"],
+ }
+ ],
+ "sensors": [
+ {
+ "integration-name": "Databricks",
+ "python-modules": [
+ "airflow.providers.databricks.sensors.databricks_sql",
+ "airflow.providers.databricks.sensors.databricks_partition",
+ ],
+ }
+ ],
+ "connection-types": [
+ {
+ "hook-class-name": "airflow.providers.databricks.hooks.databricks.DatabricksHook",
+ "connection-type": "databricks",
+ }
+ ],
+ "plugins": [
+ {
+ "name": "databricks_workflow",
+ "plugin-class": "airflow.providers.databricks.plugins.databricks_workflow.DatabricksWorkflowPlugin",
+ }
+ ],
+ "extra-links": ["airflow.providers.databricks.operators.databricks.DatabricksJobRunLink"],
+ "dependencies": [
+ "apache-airflow>=2.9.0",
+ "apache-airflow-providers-common-sql>=1.20.0",
+ "requests>=2.27.0,<3",
+ "databricks-sql-connector>=3.0.0",
+ "aiohttp>=3.9.2, <4",
+ "mergedeep>=1.3.4",
+ "pandas>=2.1.2,<2.2",
+ "pyarrow>=14.0.1",
+ ],
+ "optional-dependencies": {
+ "sdk": ["databricks-sdk==0.10.0"],
+ "azure-identity": ["azure-identity>=1.3.1"],
+ },
+ "devel-dependencies": ["deltalake>=0.12.0"],
+ }
diff --git a/providers/src/airflow/providers/databricks/hooks/__init__.py b/providers/databricks/src/airflow/providers/databricks/hooks/__init__.py
similarity index 100%
rename from providers/src/airflow/providers/databricks/hooks/__init__.py
rename to providers/databricks/src/airflow/providers/databricks/hooks/__init__.py
diff --git a/providers/src/airflow/providers/databricks/hooks/databricks.py b/providers/databricks/src/airflow/providers/databricks/hooks/databricks.py
similarity index 100%
rename from providers/src/airflow/providers/databricks/hooks/databricks.py
rename to providers/databricks/src/airflow/providers/databricks/hooks/databricks.py
diff --git a/providers/src/airflow/providers/databricks/hooks/databricks_base.py b/providers/databricks/src/airflow/providers/databricks/hooks/databricks_base.py
similarity index 100%
rename from providers/src/airflow/providers/databricks/hooks/databricks_base.py
rename to providers/databricks/src/airflow/providers/databricks/hooks/databricks_base.py
diff --git a/providers/src/airflow/providers/databricks/hooks/databricks_sql.py b/providers/databricks/src/airflow/providers/databricks/hooks/databricks_sql.py
similarity index 99%
rename from providers/src/airflow/providers/databricks/hooks/databricks_sql.py
rename to providers/databricks/src/airflow/providers/databricks/hooks/databricks_sql.py
index b0f52798d9fc5..7dc15fdf5917d 100644
--- a/providers/src/airflow/providers/databricks/hooks/databricks_sql.py
+++ b/providers/databricks/src/airflow/providers/databricks/hooks/databricks_sql.py
@@ -31,14 +31,13 @@
overload,
)
-from databricks import sql # type: ignore[attr-defined]
-from databricks.sql.types import Row
-
from airflow.exceptions import AirflowException
from airflow.models.connection import Connection as AirflowConnection
from airflow.providers.common.sql.hooks.sql import DbApiHook, return_single_query_results
from airflow.providers.databricks.exceptions import DatabricksSqlExecutionError, DatabricksSqlExecutionTimeout
from airflow.providers.databricks.hooks.databricks_base import BaseDatabricksHook
+from databricks import sql # type: ignore[attr-defined]
+from databricks.sql.types import Row
if TYPE_CHECKING:
from databricks.sql.client import Connection
diff --git a/providers/src/airflow/providers/databricks/operators/__init__.py b/providers/databricks/src/airflow/providers/databricks/operators/__init__.py
similarity index 100%
rename from providers/src/airflow/providers/databricks/operators/__init__.py
rename to providers/databricks/src/airflow/providers/databricks/operators/__init__.py
diff --git a/providers/src/airflow/providers/databricks/operators/databricks.py b/providers/databricks/src/airflow/providers/databricks/operators/databricks.py
similarity index 99%
rename from providers/src/airflow/providers/databricks/operators/databricks.py
rename to providers/databricks/src/airflow/providers/databricks/operators/databricks.py
index 3c121c49a9e77..75035b13fc38d 100644
--- a/providers/src/airflow/providers/databricks/operators/databricks.py
+++ b/providers/databricks/src/airflow/providers/databricks/operators/databricks.py
@@ -1114,7 +1114,7 @@ def _get_run_json(self) -> dict[str, Any]:
raise ValueError("Must specify either existing_cluster_id, new_cluster or environments.")
return run_json
- def _launch_job(self, context: Context | None = None) -> int:
+ def _launch_job(self, context: Context | None = None) -> int | None:
"""Launch the job on Databricks."""
run_json = self._get_run_json()
self.databricks_run_id = self._hook.submit_run(run_json)
diff --git a/providers/src/airflow/providers/databricks/operators/databricks_repos.py b/providers/databricks/src/airflow/providers/databricks/operators/databricks_repos.py
similarity index 100%
rename from providers/src/airflow/providers/databricks/operators/databricks_repos.py
rename to providers/databricks/src/airflow/providers/databricks/operators/databricks_repos.py
diff --git a/providers/src/airflow/providers/databricks/operators/databricks_sql.py b/providers/databricks/src/airflow/providers/databricks/operators/databricks_sql.py
similarity index 99%
rename from providers/src/airflow/providers/databricks/operators/databricks_sql.py
rename to providers/databricks/src/airflow/providers/databricks/operators/databricks_sql.py
index a4cb062c9b8c3..06a4c2f445650 100644
--- a/providers/src/airflow/providers/databricks/operators/databricks_sql.py
+++ b/providers/databricks/src/airflow/providers/databricks/operators/databricks_sql.py
@@ -24,12 +24,11 @@
from collections.abc import Sequence
from typing import TYPE_CHECKING, Any, ClassVar
-from databricks.sql.utils import ParamEscaper
-
from airflow.exceptions import AirflowException
from airflow.models import BaseOperator
from airflow.providers.common.sql.operators.sql import SQLExecuteQueryOperator
from airflow.providers.databricks.hooks.databricks_sql import DatabricksSqlHook
+from databricks.sql.utils import ParamEscaper
if TYPE_CHECKING:
from airflow.utils.context import Context
diff --git a/providers/src/airflow/providers/databricks/operators/databricks_workflow.py b/providers/databricks/src/airflow/providers/databricks/operators/databricks_workflow.py
similarity index 100%
rename from providers/src/airflow/providers/databricks/operators/databricks_workflow.py
rename to providers/databricks/src/airflow/providers/databricks/operators/databricks_workflow.py
diff --git a/providers/src/airflow/providers/databricks/plugins/__init__.py b/providers/databricks/src/airflow/providers/databricks/plugins/__init__.py
similarity index 100%
rename from providers/src/airflow/providers/databricks/plugins/__init__.py
rename to providers/databricks/src/airflow/providers/databricks/plugins/__init__.py
diff --git a/providers/src/airflow/providers/databricks/plugins/databricks_workflow.py b/providers/databricks/src/airflow/providers/databricks/plugins/databricks_workflow.py
similarity index 100%
rename from providers/src/airflow/providers/databricks/plugins/databricks_workflow.py
rename to providers/databricks/src/airflow/providers/databricks/plugins/databricks_workflow.py
diff --git a/providers/src/airflow/providers/databricks/sensors/__init__.py b/providers/databricks/src/airflow/providers/databricks/sensors/__init__.py
similarity index 100%
rename from providers/src/airflow/providers/databricks/sensors/__init__.py
rename to providers/databricks/src/airflow/providers/databricks/sensors/__init__.py
diff --git a/providers/src/airflow/providers/databricks/sensors/databricks_partition.py b/providers/databricks/src/airflow/providers/databricks/sensors/databricks_partition.py
similarity index 99%
rename from providers/src/airflow/providers/databricks/sensors/databricks_partition.py
rename to providers/databricks/src/airflow/providers/databricks/sensors/databricks_partition.py
index df9f8cf3fe1d1..b3a10a9e9462e 100644
--- a/providers/src/airflow/providers/databricks/sensors/databricks_partition.py
+++ b/providers/databricks/src/airflow/providers/databricks/sensors/databricks_partition.py
@@ -25,12 +25,11 @@
from functools import cached_property
from typing import TYPE_CHECKING, Any, Callable
-from databricks.sql.utils import ParamEscaper
-
from airflow.exceptions import AirflowException
from airflow.providers.common.sql.hooks.sql import fetch_all_handler
from airflow.providers.databricks.hooks.databricks_sql import DatabricksSqlHook
from airflow.sensors.base import BaseSensorOperator
+from databricks.sql.utils import ParamEscaper
if TYPE_CHECKING:
try:
diff --git a/providers/src/airflow/providers/databricks/sensors/databricks_sql.py b/providers/databricks/src/airflow/providers/databricks/sensors/databricks_sql.py
similarity index 100%
rename from providers/src/airflow/providers/databricks/sensors/databricks_sql.py
rename to providers/databricks/src/airflow/providers/databricks/sensors/databricks_sql.py
diff --git a/providers/src/airflow/providers/databricks/triggers/__init__.py b/providers/databricks/src/airflow/providers/databricks/triggers/__init__.py
similarity index 100%
rename from providers/src/airflow/providers/databricks/triggers/__init__.py
rename to providers/databricks/src/airflow/providers/databricks/triggers/__init__.py
diff --git a/providers/src/airflow/providers/databricks/triggers/databricks.py b/providers/databricks/src/airflow/providers/databricks/triggers/databricks.py
similarity index 100%
rename from providers/src/airflow/providers/databricks/triggers/databricks.py
rename to providers/databricks/src/airflow/providers/databricks/triggers/databricks.py
diff --git a/providers/src/airflow/providers/databricks/utils/__init__.py b/providers/databricks/src/airflow/providers/databricks/utils/__init__.py
similarity index 100%
rename from providers/src/airflow/providers/databricks/utils/__init__.py
rename to providers/databricks/src/airflow/providers/databricks/utils/__init__.py
diff --git a/providers/src/airflow/providers/databricks/utils/databricks.py b/providers/databricks/src/airflow/providers/databricks/utils/databricks.py
similarity index 100%
rename from providers/src/airflow/providers/databricks/utils/databricks.py
rename to providers/databricks/src/airflow/providers/databricks/utils/databricks.py
diff --git a/providers/databricks/tests/conftest.py b/providers/databricks/tests/conftest.py
new file mode 100644
index 0000000000000..068fe6bbf5ae9
--- /dev/null
+++ b/providers/databricks/tests/conftest.py
@@ -0,0 +1,32 @@
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements. See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership. The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License. You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing,
+# software distributed under the License is distributed on an
+# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+# KIND, either express or implied. See the License for the
+# specific language governing permissions and limitations
+# under the License.
+from __future__ import annotations
+
+import pathlib
+
+import pytest
+
+pytest_plugins = "tests_common.pytest_plugin"
+
+
+@pytest.hookimpl(tryfirst=True)
+def pytest_configure(config: pytest.Config) -> None:
+ deprecations_ignore_path = pathlib.Path(__file__).parent.joinpath("deprecations_ignore.yml")
+ dep_path = [deprecations_ignore_path] if deprecations_ignore_path.exists() else []
+ config.inicfg["airflow_deprecations_ignore"] = (
+ config.inicfg.get("airflow_deprecations_ignore", []) + dep_path # type: ignore[assignment,operator]
+ )
diff --git a/providers/databricks/tests/provider_tests/__init__.py b/providers/databricks/tests/provider_tests/__init__.py
new file mode 100644
index 0000000000000..e8fd22856438c
--- /dev/null
+++ b/providers/databricks/tests/provider_tests/__init__.py
@@ -0,0 +1,17 @@
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements. See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership. The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License. You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing,
+# software distributed under the License is distributed on an
+# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+# KIND, either express or implied. See the License for the
+# specific language governing permissions and limitations
+# under the License.
+__path__ = __import__("pkgutil").extend_path(__path__, __name__) # type: ignore
diff --git a/providers/tests/databricks/__init__.py b/providers/databricks/tests/provider_tests/databricks/__init__.py
similarity index 100%
rename from providers/tests/databricks/__init__.py
rename to providers/databricks/tests/provider_tests/databricks/__init__.py
diff --git a/providers/tests/databricks/hooks/__init__.py b/providers/databricks/tests/provider_tests/databricks/hooks/__init__.py
similarity index 100%
rename from providers/tests/databricks/hooks/__init__.py
rename to providers/databricks/tests/provider_tests/databricks/hooks/__init__.py
diff --git a/providers/tests/databricks/hooks/test_databricks.py b/providers/databricks/tests/provider_tests/databricks/hooks/test_databricks.py
similarity index 100%
rename from providers/tests/databricks/hooks/test_databricks.py
rename to providers/databricks/tests/provider_tests/databricks/hooks/test_databricks.py
diff --git a/providers/tests/databricks/hooks/test_databricks_azure_workload_identity.py b/providers/databricks/tests/provider_tests/databricks/hooks/test_databricks_azure_workload_identity.py
similarity index 100%
rename from providers/tests/databricks/hooks/test_databricks_azure_workload_identity.py
rename to providers/databricks/tests/provider_tests/databricks/hooks/test_databricks_azure_workload_identity.py
diff --git a/providers/tests/databricks/hooks/test_databricks_azure_workload_identity_async.py b/providers/databricks/tests/provider_tests/databricks/hooks/test_databricks_azure_workload_identity_async.py
similarity index 100%
rename from providers/tests/databricks/hooks/test_databricks_azure_workload_identity_async.py
rename to providers/databricks/tests/provider_tests/databricks/hooks/test_databricks_azure_workload_identity_async.py
diff --git a/providers/tests/databricks/hooks/test_databricks_base.py b/providers/databricks/tests/provider_tests/databricks/hooks/test_databricks_base.py
similarity index 100%
rename from providers/tests/databricks/hooks/test_databricks_base.py
rename to providers/databricks/tests/provider_tests/databricks/hooks/test_databricks_base.py
diff --git a/providers/tests/databricks/hooks/test_databricks_sql.py b/providers/databricks/tests/provider_tests/databricks/hooks/test_databricks_sql.py
similarity index 100%
rename from providers/tests/databricks/hooks/test_databricks_sql.py
rename to providers/databricks/tests/provider_tests/databricks/hooks/test_databricks_sql.py
diff --git a/providers/tests/databricks/operators/__init__.py b/providers/databricks/tests/provider_tests/databricks/operators/__init__.py
similarity index 100%
rename from providers/tests/databricks/operators/__init__.py
rename to providers/databricks/tests/provider_tests/databricks/operators/__init__.py
diff --git a/providers/tests/databricks/operators/test_databricks.py b/providers/databricks/tests/provider_tests/databricks/operators/test_databricks.py
similarity index 99%
rename from providers/tests/databricks/operators/test_databricks.py
rename to providers/databricks/tests/provider_tests/databricks/operators/test_databricks.py
index ac5e556e48da8..5ff686ea0aadb 100644
--- a/providers/tests/databricks/operators/test_databricks.py
+++ b/providers/databricks/tests/provider_tests/databricks/operators/test_databricks.py
@@ -2208,7 +2208,7 @@ def test_both_new_and_existing_cluster_set(self):
exception_message = "Both new_cluster and existing_cluster_id are set. Only one should be set."
assert str(exc_info.value) == exception_message
- def test_both_new_and_existing_cluster_unset(self, caplog):
+ def test_both_new_and_existing_cluster_unset(self):
operator = DatabricksNotebookOperator(
task_id="test_task",
notebook_path="test_path",
diff --git a/providers/tests/databricks/operators/test_databricks_copy.py b/providers/databricks/tests/provider_tests/databricks/operators/test_databricks_copy.py
similarity index 100%
rename from providers/tests/databricks/operators/test_databricks_copy.py
rename to providers/databricks/tests/provider_tests/databricks/operators/test_databricks_copy.py
diff --git a/providers/tests/databricks/operators/test_databricks_repos.py b/providers/databricks/tests/provider_tests/databricks/operators/test_databricks_repos.py
similarity index 100%
rename from providers/tests/databricks/operators/test_databricks_repos.py
rename to providers/databricks/tests/provider_tests/databricks/operators/test_databricks_repos.py
diff --git a/providers/tests/databricks/operators/test_databricks_sql.py b/providers/databricks/tests/provider_tests/databricks/operators/test_databricks_sql.py
similarity index 100%
rename from providers/tests/databricks/operators/test_databricks_sql.py
rename to providers/databricks/tests/provider_tests/databricks/operators/test_databricks_sql.py
diff --git a/providers/tests/databricks/operators/test_databricks_workflow.py b/providers/databricks/tests/provider_tests/databricks/operators/test_databricks_workflow.py
similarity index 100%
rename from providers/tests/databricks/operators/test_databricks_workflow.py
rename to providers/databricks/tests/provider_tests/databricks/operators/test_databricks_workflow.py
diff --git a/providers/tests/databricks/plugins/__init__.py b/providers/databricks/tests/provider_tests/databricks/plugins/__init__.py
similarity index 100%
rename from providers/tests/databricks/plugins/__init__.py
rename to providers/databricks/tests/provider_tests/databricks/plugins/__init__.py
diff --git a/providers/tests/databricks/plugins/test_databricks_workflow.py b/providers/databricks/tests/provider_tests/databricks/plugins/test_databricks_workflow.py
similarity index 100%
rename from providers/tests/databricks/plugins/test_databricks_workflow.py
rename to providers/databricks/tests/provider_tests/databricks/plugins/test_databricks_workflow.py
diff --git a/providers/tests/databricks/sensors/__init__.py b/providers/databricks/tests/provider_tests/databricks/sensors/__init__.py
similarity index 100%
rename from providers/tests/databricks/sensors/__init__.py
rename to providers/databricks/tests/provider_tests/databricks/sensors/__init__.py
diff --git a/providers/tests/databricks/sensors/test_databricks_partition.py b/providers/databricks/tests/provider_tests/databricks/sensors/test_databricks_partition.py
similarity index 100%
rename from providers/tests/databricks/sensors/test_databricks_partition.py
rename to providers/databricks/tests/provider_tests/databricks/sensors/test_databricks_partition.py
diff --git a/providers/tests/databricks/sensors/test_databricks_sql.py b/providers/databricks/tests/provider_tests/databricks/sensors/test_databricks_sql.py
similarity index 100%
rename from providers/tests/databricks/sensors/test_databricks_sql.py
rename to providers/databricks/tests/provider_tests/databricks/sensors/test_databricks_sql.py
diff --git a/providers/tests/databricks/test_exceptions.py b/providers/databricks/tests/provider_tests/databricks/test_exceptions.py
similarity index 100%
rename from providers/tests/databricks/test_exceptions.py
rename to providers/databricks/tests/provider_tests/databricks/test_exceptions.py
diff --git a/providers/tests/databricks/triggers/__init__.py b/providers/databricks/tests/provider_tests/databricks/triggers/__init__.py
similarity index 100%
rename from providers/tests/databricks/triggers/__init__.py
rename to providers/databricks/tests/provider_tests/databricks/triggers/__init__.py
diff --git a/providers/tests/databricks/triggers/test_databricks.py b/providers/databricks/tests/provider_tests/databricks/triggers/test_databricks.py
similarity index 100%
rename from providers/tests/databricks/triggers/test_databricks.py
rename to providers/databricks/tests/provider_tests/databricks/triggers/test_databricks.py
diff --git a/providers/tests/databricks/utils/__init__.py b/providers/databricks/tests/provider_tests/databricks/utils/__init__.py
similarity index 100%
rename from providers/tests/databricks/utils/__init__.py
rename to providers/databricks/tests/provider_tests/databricks/utils/__init__.py
diff --git a/providers/tests/databricks/utils/test_databricks.py b/providers/databricks/tests/provider_tests/databricks/utils/test_databricks.py
similarity index 100%
rename from providers/tests/databricks/utils/test_databricks.py
rename to providers/databricks/tests/provider_tests/databricks/utils/test_databricks.py
diff --git a/providers/tests/system/databricks/__init__.py b/providers/databricks/tests/system/databricks/__init__.py
similarity index 100%
rename from providers/tests/system/databricks/__init__.py
rename to providers/databricks/tests/system/databricks/__init__.py
diff --git a/providers/tests/system/databricks/example_databricks.py b/providers/databricks/tests/system/databricks/example_databricks.py
similarity index 100%
rename from providers/tests/system/databricks/example_databricks.py
rename to providers/databricks/tests/system/databricks/example_databricks.py
diff --git a/providers/tests/system/databricks/example_databricks_repos.py b/providers/databricks/tests/system/databricks/example_databricks_repos.py
similarity index 100%
rename from providers/tests/system/databricks/example_databricks_repos.py
rename to providers/databricks/tests/system/databricks/example_databricks_repos.py
diff --git a/providers/tests/system/databricks/example_databricks_sensors.py b/providers/databricks/tests/system/databricks/example_databricks_sensors.py
similarity index 100%
rename from providers/tests/system/databricks/example_databricks_sensors.py
rename to providers/databricks/tests/system/databricks/example_databricks_sensors.py
diff --git a/providers/tests/system/databricks/example_databricks_sql.py b/providers/databricks/tests/system/databricks/example_databricks_sql.py
similarity index 100%
rename from providers/tests/system/databricks/example_databricks_sql.py
rename to providers/databricks/tests/system/databricks/example_databricks_sql.py
diff --git a/providers/tests/system/databricks/example_databricks_workflow.py b/providers/databricks/tests/system/databricks/example_databricks_workflow.py
similarity index 100%
rename from providers/tests/system/databricks/example_databricks_workflow.py
rename to providers/databricks/tests/system/databricks/example_databricks_workflow.py
diff --git a/pyproject.toml b/pyproject.toml
index 52144ab2b4e82..a81d635fe9e4e 100644
--- a/pyproject.toml
+++ b/pyproject.toml
@@ -643,6 +643,7 @@ dev = [
"apache-airflow-providers-common-io",
"apache-airflow-providers-common-sql",
"apache-airflow-providers-docker",
+ "apache-airflow-providers-databricks",
"apache-airflow-providers-datadog",
"apache-airflow-providers-dbt-cloud",
"apache-airflow-providers-discord",
@@ -717,6 +718,7 @@ apache-airflow-providers-cohere = { workspace = true }
apache-airflow-providers-common-compat = { workspace = true }
apache-airflow-providers-common-io = { workspace = true }
apache-airflow-providers-common-sql = { workspace = true }
+apache-airflow-providers-databricks = { workspace = true }
apache-airflow-providers-datadog = { workspace = true }
apache-airflow-providers-dbt-cloud = { workspace = true }
apache-airflow-providers-discord = { workspace = true }
@@ -790,6 +792,7 @@ members = [
"providers/common/compat",
"providers/common/io",
"providers/common/sql",
+ "providers/databricks",
"providers/datadog",
"providers/dbt/cloud",
"providers/discord",
diff --git a/scripts/ci/docker-compose/remove-sources.yml b/scripts/ci/docker-compose/remove-sources.yml
index 65d18df71275d..c576684920fe8 100644
--- a/scripts/ci/docker-compose/remove-sources.yml
+++ b/scripts/ci/docker-compose/remove-sources.yml
@@ -53,6 +53,7 @@ services:
- ../../../empty:/opt/airflow/providers/common/compat/src
- ../../../empty:/opt/airflow/providers/common/io/src
- ../../../empty:/opt/airflow/providers/common/sql/src
+ - ../../../empty:/opt/airflow/providers/databricks/src
- ../../../empty:/opt/airflow/providers/datadog/src
- ../../../empty:/opt/airflow/providers/dbt/cloud/src
- ../../../empty:/opt/airflow/providers/discord/src
diff --git a/scripts/ci/docker-compose/tests-sources.yml b/scripts/ci/docker-compose/tests-sources.yml
index 2941b44c37fb6..9a2e97c841696 100644
--- a/scripts/ci/docker-compose/tests-sources.yml
+++ b/scripts/ci/docker-compose/tests-sources.yml
@@ -60,6 +60,7 @@ services:
- ../../../providers/common/compat/tests:/opt/airflow/providers/common/compat/tests
- ../../../providers/common/io/tests:/opt/airflow/providers/common/io/tests
- ../../../providers/common/sql/tests:/opt/airflow/providers/common/sql/tests
+ - ../../../providers/databricks/tests:/opt/airflow/providers/databricks/tests
- ../../../providers/datadog/tests:/opt/airflow/providers/datadog/tests
- ../../../providers/dbt/cloud/tests:/opt/airflow/providers/dbt/cloud/tests
- ../../../providers/discord/tests:/opt/airflow/providers/discord/tests