diff --git a/dlt/destinations/impl/snowflake/utils.py b/dlt/destinations/impl/snowflake/utils.py index 0526d25061..c8c2a417ec 100644 --- a/dlt/destinations/impl/snowflake/utils.py +++ b/dlt/destinations/impl/snowflake/utils.py @@ -1,4 +1,4 @@ -from typing import Optional, Dict, Any +from typing import Optional from urllib.parse import urlparse, urlunparse from dlt.common.configuration.specs import ( diff --git a/tests/load/snowflake/test_snowflake_utils.py b/tests/load/snowflake/test_snowflake_utils.py index e1e0cd6fcd..fd35ad0914 100644 --- a/tests/load/snowflake/test_snowflake_utils.py +++ b/tests/load/snowflake/test_snowflake_utils.py @@ -100,7 +100,15 @@ def test_file_format_parquet_vectorized(test_table, local_file_path, local_stage (False, ";", "ASCII", True), ], ) -def test_file_format_csv(include_header, delimiter, encoding, on_error_continue, test_table, local_file_path, local_stage_path): +def test_file_format_csv( + include_header, + delimiter, + encoding, + on_error_continue, + test_table, + local_file_path, + local_stage_path, +): """Test CSV format handling in gen_copy_sql with various options.""" csv_config = CsvFormatConfiguration( include_header=include_header, @@ -412,7 +420,7 @@ def test_full_workflow_s3_with_aws_credentials(test_table, aws_credentials): staging_credentials=aws_credentials, ) - # Verify the final SQL + # Verify the final SQL assert_sql_contains( sql, f"COPY INTO {test_table}", @@ -486,4 +494,4 @@ def test_snowflake_azure_converter() -> None: assert ( azure_url == "azure://my_account.blob.core.windows.net/dlt-ci-test-bucket/path/to/file.parquet" - ) \ No newline at end of file + )