You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
for unknown column names or for unsupported types.
Delta Lake table properties
This is not part of the standard yet. We could probably follow the same convention as for Iceberg. There is an issue to get this into the Delta Lake protocol at delta-io/delta#2751
To control Bloom filter maximum size we use parquet_bloom_filter_max_bytes = BIGINT.
We use the same table properties as for Hive to configure writing of Bloom filters from Trino. That is, parquet_bloom_filter_columns = ARRAY['<column-name>'].
The text was updated successfully, but these errors were encountered:
Tasks
There are two components to this task: Decide which format table property to use, and decide which Trino table properties to use.
Iceberg table properties
Parquet Bloom filter write support can be configured by setting the table properties
https://github.com/apache/iceberg/blob/732fbfd516a3dfb2028fd6795f8f564f70e44742/core/src/main/java/org/apache/iceberg/TableProperties.java#L166-L171
Whenever we write to a table where the table properties are set for some columns we write Bloom filters for that column.
We should silently ignore
for unknown column names or for unsupported types.
Delta Lake table properties
This is not part of the standard yet. We could probably follow the same convention as for Iceberg. There is an issue to get this into the Delta Lake protocol at delta-io/delta#2751
Trino table properties
We mirror what was done for Hive in 5041496:
parquet_bloom_filter_max_bytes = BIGINT
.parquet_bloom_filter_columns = ARRAY['<column-name>']
.The text was updated successfully, but these errors were encountered: