Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fix snapshot usage in CAgg invalidation scanner #6677

Merged
merged 1 commit into from
Feb 21, 2024

Conversation

jnidzwetzki
Copy link
Contributor

@jnidzwetzki jnidzwetzki commented Feb 20, 2024

So far, the scanner in get_lowest_invalidated_time_for_hypertable() used a Snapshot that includes the data of all committed transactions (and not only the transactions that are started before the snapshot was created - SNAPSHOT_SELF). This could lead to a situation where we see the old and the new value of a parallel updated tuple. Since get_lowest_invalidated_time_for_hypertable() expects exactly one tuple to be found during the scan, we end up with an error. Since this is performed in our insert path (i.e., it is executed by the invalidation trigger), this error could reject inserts on the hypertable (ERROR: more than one invalidation watermark found).


Should be merged after: #6678

@jnidzwetzki jnidzwetzki self-assigned this Feb 20, 2024
@jnidzwetzki jnidzwetzki marked this pull request as ready for review February 20, 2024 10:15
parallel session updates the scanned value and commits during a scan, we end up in a
situation where we see the old and the new value. This causes ts_scanner_scan_one() to
fail. */
.snapshot = GetTransactionSnapshot(),
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Yes, we should do it. However, this is already done in this pending PR: #6672. There, we use the latest snapshot to see parallel updates.

# (3) Session 2 - Inserts new data of the year 2023, during the commits it reads the current invalidation watermark and blocks due to (2)
# (4) Session 1 - Refreshes a CAgg on this hypertable. During the update the invalidation watermark is deleted and a new one is written
# (5) Session 3 - Unblocks the scanner of session 2
# (6) Session 2 - Continues the scan. The new inserted invalidation watermark should not be seen since the transaction was started after s2 has started the scan in (3)
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I'm a little confused about this part -- if session 2 does a backfill into the range that was refreshed by session 1, what will be the final watermark value?

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

In an ideal world, we should see the most recent timestamp (i.e., if a parallel refresh runs in parallel and moves the watermark, it should be guaranteed that we see this update. Otherwise we might lose an invalidation since we believe we are > watermark). Said this, we have to avoid strong locks since this code runs on every transaction commit.

We have a rewrite of this code path on our agenda. The goal is to get rid of the complete invalidation trigger and perform this more asynchronously.

@jnidzwetzki jnidzwetzki force-pushed the cagg_fix_snapshot branch 3 times, most recently from 579a3d6 to 499b1c4 Compare February 21, 2024 11:51
So far, the scanner in get_lowest_invalidated_time_for_hypertable() used
a Snapshot that includes the data of all committed transactions (and not
only the transactions that are started before the snapshot was created -
SNAPSHOT_SELF). This could lead to a situation where we see the old and
the new value of a parallel updated tuple. Since
get_lowest_invalidated_time_for_hypertable() expects exactly one tuple
to be found during the scan, we end up with an error. Since this is
performed in our insert path (i.e., it is executed by the invalidation
trigger), this error could reject inserts on the hypertable.
@jnidzwetzki jnidzwetzki enabled auto-merge (rebase) February 21, 2024 15:52
Copy link

codecov bot commented Feb 21, 2024

Codecov Report

All modified and coverable lines are covered by tests ✅

Comparison is base (59f50f2) 80.06% compared to head (6fb552c) 81.40%.
Report is 21 commits behind head on main.

Additional details and impacted files
@@            Coverage Diff             @@
##             main    #6677      +/-   ##
==========================================
+ Coverage   80.06%   81.40%   +1.34%     
==========================================
  Files         190      191       +1     
  Lines       37181    36422     -759     
  Branches     9450     9465      +15     
==========================================
- Hits        29770    29651     -119     
- Misses       2997     2998       +1     
+ Partials     4414     3773     -641     

☔ View full report in Codecov by Sentry.
📢 Have feedback on the report? Share it here.

@jnidzwetzki jnidzwetzki merged commit 0d48019 into timescale:main Feb 21, 2024
45 checks passed
@jnidzwetzki jnidzwetzki deleted the cagg_fix_snapshot branch February 21, 2024 16:00
@timescale-automation
Copy link

Automated backport to 2.14.x not done: cherry-pick failed.

Git status

HEAD detached at origin/2.14.x
You are currently cherry-picking commit 0d4801988.
  (fix conflicts and run "git cherry-pick --continue")
  (use "git cherry-pick --skip" to skip this patch)
  (use "git cherry-pick --abort" to cancel the cherry-pick operation)

Changes to be committed:
	new file:   .unreleased/bugfix_6677
	modified:   tsl/src/continuous_aggs/common.h
	modified:   tsl/src/continuous_aggs/invalidation_threshold.c
	modified:   tsl/test/isolation/expected/cagg_concurrent_invalidation.out
	modified:   tsl/test/isolation/specs/cagg_concurrent_invalidation.spec

Unmerged paths:
  (use "git add <file>..." to mark resolution)
	both modified:   test/runner_isolation.sh
	both modified:   tsl/src/continuous_aggs/insert.c
	both modified:   tsl/test/isolation/expected/cagg_watermark_concurrent_update.out
	both modified:   tsl/test/isolation/expected/cagg_watermark_concurrent_update_1.out


Job log

@timescale-automation timescale-automation added the auto-backport-not-done Automated backport of this PR has failed non-retriably (e.g. conflicts) label Feb 21, 2024
fabriziomello added a commit to fabriziomello/timescaledb that referenced this pull request May 6, 2024
This release contains performance improvements and bug fixes since
the 2.14.2 release. We recommend that you upgrade at the next
available opportunity.

In addition, it includes these noteworthy features:
* Support `time_bucket` with `origin` and/or `offset` on Continuous Aggregate
* Compression improvements:
  - Improve expression pushdown
  - Add minmax sparse indexes when compressing columns with btree indexes
  - Make compression use the defaults functions
  - Vectorize filters in WHERE clause that contain text equality operators and LIKE expressions

**Deprecation warning**
* Starting on this release will not be possible to create Continuous Aggregate using `time_bucket_ng` anymore and it will be completely removed on the upcoming releases.
* Recommend users to [migrate their old Continuous Aggregate format to the new one](https://docs.timescale.com/use-timescale/latest/continuous-aggregates/migrate/) because it support will be completely removed in next releases prevent them to migrate.
* This is the last release supporting PostgreSQL 13.

**For on-premise users and this release only**, you will need to run [this SQL script](https://github.com/timescale/timescaledb-extras/blob/master/utils/2.15.X-fix_hypertable_foreign_keys.sql) after running `ALTER EXTENSION`. More details can be found in the pull request [timescale#6786](timescale#6797).

**Features**
* timescale#6382 Support for time_bucket with origin and offset in CAggs
* timescale#6696 Improve defaults for compression segment_by and order_by
* timescale#6705 Add sparse minmax indexes for compressed columns that have uncompressed btree indexes
* timescale#6754 Allow DROP CONSTRAINT on compressed hypertables
* timescale#6767 Add metadata table `_timestaledb_internal.bgw_job_stat_history` for tracking job execution history
* timescale#6798 Prevent usage of deprecated time_bucket_ng in CAgg definition
* timescale#6810 Add telemetry for access methods
* timescale#6811 Remove no longer relevant timescaledb.allow_install_without_preload GUC
* timescale#6837 Add migration path for CAggs using time_bucket_ng
* timescale#6865 Update the watermark when truncating a CAgg

**Bugfixes**
* timescale#6617 Fix error in show_chunks
* timescale#6621 Remove metadata when dropping chunks
* timescale#6677 Fix snapshot usage in CAgg invalidation scanner
* timescale#6698 Define meaning of 0 retries for jobs as no retries
* timescale#6717 Fix handling of compressed tables with primary or unique index in COPY path
* timescale#6726 Fix constify cagg_watermark using window function when querying a CAgg
* timescale#6729 Fix NULL start value handling in CAgg refresh
* timescale#6732 Fix CAgg migration with custom timezone / date format settings
* timescale#6752 Remove custom autovacuum setting from compressed chunks
* timescale#6770 Fix plantime chunk exclusion for OSM chunk
* timescale#6789 Fix deletes with subqueries and compression
* timescale#6796 Fix a crash involving a view on a hypertable
* timescale#6797 Fix foreign key constraint handling on compressed hypertables
* timescale#6816 Fix handling of chunks with no contraints
* timescale#6820 Fix a crash when the ts_hypertable_insert_blocker was called directly
* timescale#6849 Use non-orderby compressed metadata in compressed DML
* timescale#6867 Clean up compression settings when deleting compressed cagg
* timescale#6869 Fix compressed DML with constraints of form value OP column
* timescale#6870 Fix bool expression pushdown for queries on compressed chunks

**Thanks**
* @brasic for reporting a crash when the ts_hypertable_insert_blocker was called directly
* @bvanelli for reporting an issue with the jobs retry count
* @djzurawsk For reporting error when dropping chunks
* @Dzuzepppe for reporting an issue with DELETEs using subquery on compressed chunk working incorrectly.
* @hongquan For reporting a 'timestamp out of range' error during CAgg migrations
* @kevcenteno for reporting an issue with the show_chunks API showing incorrect output when 'created_before/created_after' was used with time-partitioned columns.
* @mahipv For starting working on the job history PR
* @rovo89 For reporting constify cagg_watermark not working using window function when querying a CAgg
fabriziomello added a commit to fabriziomello/timescaledb that referenced this pull request May 6, 2024
This release contains performance improvements and bug fixes since
the 2.14.2 release. We recommend that you upgrade at the next
available opportunity.

In addition, it includes these noteworthy features:
* Support `time_bucket` with `origin` and/or `offset` on Continuous Aggregate
* Compression improvements:
  - Improve expression pushdown
  - Add minmax sparse indexes when compressing columns with btree indexes
  - Make compression use the defaults functions
  - Vectorize filters in WHERE clause that contain text equality operators and LIKE expressions

**Deprecation warning**
* Starting on this release will not be possible to create Continuous Aggregate using `time_bucket_ng` anymore and it will be completely removed on the upcoming releases.
* Recommend users to [migrate their old Continuous Aggregate format to the new one](https://docs.timescale.com/use-timescale/latest/continuous-aggregates/migrate/) because it support will be completely removed in next releases prevent them to migrate.
* This is the last release supporting PostgreSQL 13.

**For on-premise users and this release only**, you will need to run [this SQL script](https://github.com/timescale/timescaledb-extras/blob/master/utils/2.15.X-fix_hypertable_foreign_keys.sql) after running `ALTER EXTENSION`. More details can be found in the pull request [timescale#6786](timescale#6797).

**Features**
* timescale#6382 Support for time_bucket with origin and offset in CAggs
* timescale#6696 Improve defaults for compression segment_by and order_by
* timescale#6705 Add sparse minmax indexes for compressed columns that have uncompressed btree indexes
* timescale#6754 Allow DROP CONSTRAINT on compressed hypertables
* timescale#6767 Add metadata table `_timestaledb_internal.bgw_job_stat_history` for tracking job execution history
* timescale#6798 Prevent usage of deprecated time_bucket_ng in CAgg definition
* timescale#6810 Add telemetry for access methods
* timescale#6811 Remove no longer relevant timescaledb.allow_install_without_preload GUC
* timescale#6837 Add migration path for CAggs using time_bucket_ng
* timescale#6865 Update the watermark when truncating a CAgg

**Bugfixes**
* timescale#6617 Fix error in show_chunks
* timescale#6621 Remove metadata when dropping chunks
* timescale#6677 Fix snapshot usage in CAgg invalidation scanner
* timescale#6698 Define meaning of 0 retries for jobs as no retries
* timescale#6717 Fix handling of compressed tables with primary or unique index in COPY path
* timescale#6726 Fix constify cagg_watermark using window function when querying a CAgg
* timescale#6729 Fix NULL start value handling in CAgg refresh
* timescale#6732 Fix CAgg migration with custom timezone / date format settings
* timescale#6752 Remove custom autovacuum setting from compressed chunks
* timescale#6770 Fix plantime chunk exclusion for OSM chunk
* timescale#6789 Fix deletes with subqueries and compression
* timescale#6796 Fix a crash involving a view on a hypertable
* timescale#6797 Fix foreign key constraint handling on compressed hypertables
* timescale#6816 Fix handling of chunks with no contraints
* timescale#6820 Fix a crash when the ts_hypertable_insert_blocker was called directly
* timescale#6849 Use non-orderby compressed metadata in compressed DML
* timescale#6867 Clean up compression settings when deleting compressed cagg
* timescale#6869 Fix compressed DML with constraints of form value OP column
* timescale#6870 Fix bool expression pushdown for queries on compressed chunks

**Thanks**
* @brasic for reporting a crash when the ts_hypertable_insert_blocker was called directly
* @bvanelli for reporting an issue with the jobs retry count
* @djzurawsk For reporting error when dropping chunks
* @Dzuzepppe for reporting an issue with DELETEs using subquery on compressed chunk working incorrectly.
* @hongquan For reporting a 'timestamp out of range' error during CAgg migrations
* @kevcenteno for reporting an issue with the show_chunks API showing incorrect output when 'created_before/created_after' was used with time-partitioned columns.
* @mahipv For starting working on the job history PR
* @rovo89 For reporting constify cagg_watermark not working using window function when querying a CAgg
@fabriziomello fabriziomello mentioned this pull request May 6, 2024
fabriziomello added a commit to fabriziomello/timescaledb that referenced this pull request May 6, 2024
This release contains performance improvements and bug fixes since
the 2.14.2 release. We recommend that you upgrade at the next
available opportunity.

In addition, it includes these noteworthy features:
* Support `time_bucket` with `origin` and/or `offset` on Continuous Aggregate
* Compression improvements:
  - Improve expression pushdown
  - Add minmax sparse indexes when compressing columns with btree indexes
  - Make compression use the defaults functions
  - Vectorize filters in WHERE clause that contain text equality operators and LIKE expressions

**Deprecation warning**
* Starting on this release will not be possible to create Continuous Aggregate using `time_bucket_ng` anymore and it will be completely removed on the upcoming releases.
* Recommend users to [migrate their old Continuous Aggregate format to the new one](https://docs.timescale.com/use-timescale/latest/continuous-aggregates/migrate/) because it support will be completely removed in next releases prevent them to migrate.
* This is the last release supporting PostgreSQL 13.

**For on-premise users and this release only**, you will need to run [this SQL script](https://github.com/timescale/timescaledb-extras/blob/master/utils/2.15.X-fix_hypertable_foreign_keys.sql) after running `ALTER EXTENSION`. More details can be found in the pull request [timescale#6786](timescale#6797).

**Features**
* timescale#6382 Support for time_bucket with origin and offset in CAggs
* timescale#6696 Improve defaults for compression segment_by and order_by
* timescale#6705 Add sparse minmax indexes for compressed columns that have uncompressed btree indexes
* timescale#6754 Allow DROP CONSTRAINT on compressed hypertables
* timescale#6767 Add metadata table `_timestaledb_internal.bgw_job_stat_history` for tracking job execution history
* timescale#6798 Prevent usage of deprecated time_bucket_ng in CAgg definition
* timescale#6810 Add telemetry for access methods
* timescale#6811 Remove no longer relevant timescaledb.allow_install_without_preload GUC
* timescale#6837 Add migration path for CAggs using time_bucket_ng
* timescale#6865 Update the watermark when truncating a CAgg

**Bugfixes**
* timescale#6617 Fix error in show_chunks
* timescale#6621 Remove metadata when dropping chunks
* timescale#6677 Fix snapshot usage in CAgg invalidation scanner
* timescale#6698 Define meaning of 0 retries for jobs as no retries
* timescale#6717 Fix handling of compressed tables with primary or unique index in COPY path
* timescale#6726 Fix constify cagg_watermark using window function when querying a CAgg
* timescale#6729 Fix NULL start value handling in CAgg refresh
* timescale#6732 Fix CAgg migration with custom timezone / date format settings
* timescale#6752 Remove custom autovacuum setting from compressed chunks
* timescale#6770 Fix plantime chunk exclusion for OSM chunk
* timescale#6789 Fix deletes with subqueries and compression
* timescale#6796 Fix a crash involving a view on a hypertable
* timescale#6797 Fix foreign key constraint handling on compressed hypertables
* timescale#6816 Fix handling of chunks with no contraints
* timescale#6820 Fix a crash when the ts_hypertable_insert_blocker was called directly
* timescale#6849 Use non-orderby compressed metadata in compressed DML
* timescale#6867 Clean up compression settings when deleting compressed cagg
* timescale#6869 Fix compressed DML with constraints of form value OP column
* timescale#6870 Fix bool expression pushdown for queries on compressed chunks

**Thanks**
* @brasic for reporting a crash when the ts_hypertable_insert_blocker was called directly
* @bvanelli for reporting an issue with the jobs retry count
* @djzurawsk For reporting error when dropping chunks
* @Dzuzepppe for reporting an issue with DELETEs using subquery on compressed chunk working incorrectly.
* @hongquan For reporting a 'timestamp out of range' error during CAgg migrations
* @kevcenteno for reporting an issue with the show_chunks API showing incorrect output when 'created_before/created_after' was used with time-partitioned columns.
* @mahipv For starting working on the job history PR
* @rovo89 For reporting constify cagg_watermark not working using window function when querying a CAgg
fabriziomello added a commit that referenced this pull request May 6, 2024
This release contains performance improvements and bug fixes since
the 2.14.2 release. We recommend that you upgrade at the next
available opportunity.

In addition, it includes these noteworthy features:
* Support `time_bucket` with `origin` and/or `offset` on Continuous Aggregate
* Compression improvements:
  - Improve expression pushdown
  - Add minmax sparse indexes when compressing columns with btree indexes
  - Make compression use the defaults functions
  - Vectorize filters in WHERE clause that contain text equality operators and LIKE expressions

**Deprecation warning**
* Starting on this release will not be possible to create Continuous Aggregate using `time_bucket_ng` anymore and it will be completely removed on the upcoming releases.
* Recommend users to [migrate their old Continuous Aggregate format to the new one](https://docs.timescale.com/use-timescale/latest/continuous-aggregates/migrate/) because it support will be completely removed in next releases prevent them to migrate.
* This is the last release supporting PostgreSQL 13.

**For on-premise users and this release only**, you will need to run [this SQL script](https://github.com/timescale/timescaledb-extras/blob/master/utils/2.15.X-fix_hypertable_foreign_keys.sql) after running `ALTER EXTENSION`. More details can be found in the pull request [#6786](#6797).

**Features**
* #6382 Support for time_bucket with origin and offset in CAggs
* #6696 Improve defaults for compression segment_by and order_by
* #6705 Add sparse minmax indexes for compressed columns that have uncompressed btree indexes
* #6754 Allow DROP CONSTRAINT on compressed hypertables
* #6767 Add metadata table `_timestaledb_internal.bgw_job_stat_history` for tracking job execution history
* #6798 Prevent usage of deprecated time_bucket_ng in CAgg definition
* #6810 Add telemetry for access methods
* #6811 Remove no longer relevant timescaledb.allow_install_without_preload GUC
* #6837 Add migration path for CAggs using time_bucket_ng
* #6865 Update the watermark when truncating a CAgg

**Bugfixes**
* #6617 Fix error in show_chunks
* #6621 Remove metadata when dropping chunks
* #6677 Fix snapshot usage in CAgg invalidation scanner
* #6698 Define meaning of 0 retries for jobs as no retries
* #6717 Fix handling of compressed tables with primary or unique index in COPY path
* #6726 Fix constify cagg_watermark using window function when querying a CAgg
* #6729 Fix NULL start value handling in CAgg refresh
* #6732 Fix CAgg migration with custom timezone / date format settings
* #6752 Remove custom autovacuum setting from compressed chunks
* #6770 Fix plantime chunk exclusion for OSM chunk
* #6789 Fix deletes with subqueries and compression
* #6796 Fix a crash involving a view on a hypertable
* #6797 Fix foreign key constraint handling on compressed hypertables
* #6816 Fix handling of chunks with no contraints
* #6820 Fix a crash when the ts_hypertable_insert_blocker was called directly
* #6849 Use non-orderby compressed metadata in compressed DML
* #6867 Clean up compression settings when deleting compressed cagg
* #6869 Fix compressed DML with constraints of form value OP column
* #6870 Fix bool expression pushdown for queries on compressed chunks

**Thanks**
* @brasic for reporting a crash when the ts_hypertable_insert_blocker was called directly
* @bvanelli for reporting an issue with the jobs retry count
* @djzurawsk For reporting error when dropping chunks
* @Dzuzepppe for reporting an issue with DELETEs using subquery on compressed chunk working incorrectly.
* @hongquan For reporting a 'timestamp out of range' error during CAgg migrations
* @kevcenteno for reporting an issue with the show_chunks API showing incorrect output when 'created_before/created_after' was used with time-partitioned columns.
* @mahipv For starting working on the job history PR
* @rovo89 For reporting constify cagg_watermark not working using window function when querying a CAgg
fabriziomello added a commit to fabriziomello/timescaledb that referenced this pull request May 6, 2024
This release contains performance improvements and bug fixes since
the 2.14.2 release. We recommend that you upgrade at the next
available opportunity.

In addition, it includes these noteworthy features:
* Support `time_bucket` with `origin` and/or `offset` on Continuous Aggregate
* Compression improvements:
  - Improve expression pushdown
  - Add minmax sparse indexes when compressing columns with btree indexes
  - Make compression use the defaults functions
  - Vectorize filters in WHERE clause that contain text equality operators and LIKE expressions

**Deprecation warning**
* Starting on this release will not be possible to create Continuous Aggregate using `time_bucket_ng` anymore and it will be completely removed on the upcoming releases.
* Recommend users to [migrate their old Continuous Aggregate format to the new one](https://docs.timescale.com/use-timescale/latest/continuous-aggregates/migrate/) because it support will be completely removed in next releases prevent them to migrate.
* This is the last release supporting PostgreSQL 13.

**For on-premise users and this release only**, you will need to run [this SQL script](https://github.com/timescale/timescaledb-extras/blob/master/utils/2.15.X-fix_hypertable_foreign_keys.sql) after running `ALTER EXTENSION`. More details can be found in the pull request [timescale#6786](timescale#6797).

**Features**
* timescale#6382 Support for time_bucket with origin and offset in CAggs
* timescale#6696 Improve defaults for compression segment_by and order_by
* timescale#6705 Add sparse minmax indexes for compressed columns that have uncompressed btree indexes
* timescale#6754 Allow DROP CONSTRAINT on compressed hypertables
* timescale#6767 Add metadata table `_timestaledb_internal.bgw_job_stat_history` for tracking job execution history
* timescale#6798 Prevent usage of deprecated time_bucket_ng in CAgg definition
* timescale#6810 Add telemetry for access methods
* timescale#6811 Remove no longer relevant timescaledb.allow_install_without_preload GUC
* timescale#6837 Add migration path for CAggs using time_bucket_ng
* timescale#6865 Update the watermark when truncating a CAgg

**Bugfixes**
* timescale#6617 Fix error in show_chunks
* timescale#6621 Remove metadata when dropping chunks
* timescale#6677 Fix snapshot usage in CAgg invalidation scanner
* timescale#6698 Define meaning of 0 retries for jobs as no retries
* timescale#6717 Fix handling of compressed tables with primary or unique index in COPY path
* timescale#6726 Fix constify cagg_watermark using window function when querying a CAgg
* timescale#6729 Fix NULL start value handling in CAgg refresh
* timescale#6732 Fix CAgg migration with custom timezone / date format settings
* timescale#6752 Remove custom autovacuum setting from compressed chunks
* timescale#6770 Fix plantime chunk exclusion for OSM chunk
* timescale#6789 Fix deletes with subqueries and compression
* timescale#6796 Fix a crash involving a view on a hypertable
* timescale#6797 Fix foreign key constraint handling on compressed hypertables
* timescale#6816 Fix handling of chunks with no contraints
* timescale#6820 Fix a crash when the ts_hypertable_insert_blocker was called directly
* timescale#6849 Use non-orderby compressed metadata in compressed DML
* timescale#6867 Clean up compression settings when deleting compressed cagg
* timescale#6869 Fix compressed DML with constraints of form value OP column
* timescale#6870 Fix bool expression pushdown for queries on compressed chunks

**Thanks**
* @brasic for reporting a crash when the ts_hypertable_insert_blocker was called directly
* @bvanelli for reporting an issue with the jobs retry count
* @djzurawsk For reporting error when dropping chunks
* @Dzuzepppe for reporting an issue with DELETEs using subquery on compressed chunk working incorrectly.
* @hongquan For reporting a 'timestamp out of range' error during CAgg migrations
* @kevcenteno for reporting an issue with the show_chunks API showing incorrect output when 'created_before/created_after' was used with time-partitioned columns.
* @mahipv For starting working on the job history PR
* @rovo89 For reporting constify cagg_watermark not working using window function when querying a CAgg
fabriziomello added a commit that referenced this pull request May 7, 2024
This release contains performance improvements and bug fixes since
the 2.14.2 release. We recommend that you upgrade at the next
available opportunity.

In addition, it includes these noteworthy features:
* Support `time_bucket` with `origin` and/or `offset` on Continuous Aggregate
* Compression improvements:
  - Improve expression pushdown
  - Add minmax sparse indexes when compressing columns with btree indexes
  - Make compression use the defaults functions
  - Vectorize filters in WHERE clause that contain text equality operators and LIKE expressions

**Deprecation warning**
* Starting on this release will not be possible to create Continuous Aggregate using `time_bucket_ng` anymore and it will be completely removed on the upcoming releases.
* Recommend users to [migrate their old Continuous Aggregate format to the new one](https://docs.timescale.com/use-timescale/latest/continuous-aggregates/migrate/) because it support will be completely removed in next releases prevent them to migrate.
* This is the last release supporting PostgreSQL 13.

**For on-premise users and this release only**, you will need to run [this SQL script](https://github.com/timescale/timescaledb-extras/blob/master/utils/2.15.X-fix_hypertable_foreign_keys.sql) after running `ALTER EXTENSION`. More details can be found in the pull request [#6786](#6797).

**Features**
* #6382 Support for time_bucket with origin and offset in CAggs
* #6696 Improve defaults for compression segment_by and order_by
* #6705 Add sparse minmax indexes for compressed columns that have uncompressed btree indexes
* #6754 Allow DROP CONSTRAINT on compressed hypertables
* #6767 Add metadata table `_timestaledb_internal.bgw_job_stat_history` for tracking job execution history
* #6798 Prevent usage of deprecated time_bucket_ng in CAgg definition
* #6810 Add telemetry for access methods
* #6811 Remove no longer relevant timescaledb.allow_install_without_preload GUC
* #6837 Add migration path for CAggs using time_bucket_ng
* #6865 Update the watermark when truncating a CAgg

**Bugfixes**
* #6617 Fix error in show_chunks
* #6621 Remove metadata when dropping chunks
* #6677 Fix snapshot usage in CAgg invalidation scanner
* #6698 Define meaning of 0 retries for jobs as no retries
* #6717 Fix handling of compressed tables with primary or unique index in COPY path
* #6726 Fix constify cagg_watermark using window function when querying a CAgg
* #6729 Fix NULL start value handling in CAgg refresh
* #6732 Fix CAgg migration with custom timezone / date format settings
* #6752 Remove custom autovacuum setting from compressed chunks
* #6770 Fix plantime chunk exclusion for OSM chunk
* #6789 Fix deletes with subqueries and compression
* #6796 Fix a crash involving a view on a hypertable
* #6797 Fix foreign key constraint handling on compressed hypertables
* #6816 Fix handling of chunks with no contraints
* #6820 Fix a crash when the ts_hypertable_insert_blocker was called directly
* #6849 Use non-orderby compressed metadata in compressed DML
* #6867 Clean up compression settings when deleting compressed cagg
* #6869 Fix compressed DML with constraints of form value OP column
* #6870 Fix bool expression pushdown for queries on compressed chunks

**Thanks**
* @brasic for reporting a crash when the ts_hypertable_insert_blocker was called directly
* @bvanelli for reporting an issue with the jobs retry count
* @djzurawsk For reporting error when dropping chunks
* @Dzuzepppe for reporting an issue with DELETEs using subquery on compressed chunk working incorrectly.
* @hongquan For reporting a 'timestamp out of range' error during CAgg migrations
* @kevcenteno for reporting an issue with the show_chunks API showing incorrect output when 'created_before/created_after' was used with time-partitioned columns.
* @mahipv For starting working on the job history PR
* @rovo89 For reporting constify cagg_watermark not working using window function when querying a CAgg
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
auto-backport-not-done Automated backport of this PR has failed non-retriably (e.g. conflicts) bug
Projects
None yet
Development

Successfully merging this pull request may close these issues.

[Bug]: ERROR XX000 (internal_error) more than one invalidation watermark found
4 participants