-
Notifications
You must be signed in to change notification settings - Fork 3.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Delta Lake connector doesn't remove content of MANAGED_TABLE
Glue tables created by Databricks
#13017
Comments
findinpath
added a commit
to findinpath/trino
that referenced
this issue
Jun 28, 2022
Avoid running into a test failure when dealing with a test depending on Databricks AWS S3 and AWS Glue environment. Consult for further details: trinodb#13017
findinpath
added a commit
to findinpath/trino
that referenced
this issue
Jun 29, 2022
Avoid running into a test failure when dealing with a test depending on Databricks AWS S3 and AWS Glue environment. Consult for further details: trinodb#13017
ebyhr
pushed a commit
that referenced
this issue
Jun 29, 2022
Avoid running into a test failure when dealing with a test depending on Databricks AWS S3 and AWS Glue environment. Consult for further details: #13017
findinpath
added a commit
to findinpath/trino
that referenced
this issue
Jul 1, 2022
Avoid running into a test failure when dealing with a test depending on Databricks AWS S3 and AWS Glue environment. Consult for further details: trinodb#13017
ebyhr
pushed a commit
that referenced
this issue
Jul 4, 2022
Avoid running into a test failure when dealing with a test depending on Databricks AWS S3 and AWS Glue environment. Consult for further details: #13017
Filed https://community.databricks.com/s/question/0D58Y000093AR43SAG/how-to-identify-s3-object-type-directory-or-file-created-by-databricks. I will update this issue once I got the response from Databricks community. |
Closed
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Scenario:
Databricks SQL
Trino
The operation succeeds in Trino. However the S3 content of the table is still present on AWS S3 after completing this operation.
This means that there can't be created new tables having the same name as the previously dropped table, because the directory corresponding to the table already exists.
Context:
The above mentioned code corresponds to what is happening in
trino/plugin/trino-hive/src/main/java/io/trino/plugin/hive/s3/TrinoS3FileSystem.java
Lines 735 to 745 in fe608f2
As seen in the listing above, the
ContentType
of this "object" is set toapplication/octet-stream
which is not the same asapplication/x-directory
trino/plugin/trino-hive/src/main/java/io/trino/plugin/hive/s3/TrinoS3FileSystem.java
Line 425 in fe608f2
This interesting detail breaks the logic used for deleting Delta Lake table directories in AWS S3.
The text was updated successfully, but these errors were encountered: