-
Notifications
You must be signed in to change notification settings - Fork 897
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fix snapshot usage in CAgg invalidation scanner #6677
Conversation
0b3707c
to
1685b78
Compare
1685b78
to
ee41bf6
Compare
tsl/src/continuous_aggs/insert.c
Outdated
parallel session updates the scanned value and commits during a scan, we end up in a | ||
situation where we see the old and the new value. This causes ts_scanner_scan_one() to | ||
fail. */ | ||
.snapshot = GetTransactionSnapshot(), |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Wondering if we should do the same in https://github.com/timescale/timescaledb/blob/main/tsl/src/continuous_aggs/invalidation_threshold.c#L163
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yes, we should do it. However, this is already done in this pending PR: #6672. There, we use the latest snapshot to see parallel updates.
# (3) Session 2 - Inserts new data of the year 2023, during the commits it reads the current invalidation watermark and blocks due to (2) | ||
# (4) Session 1 - Refreshes a CAgg on this hypertable. During the update the invalidation watermark is deleted and a new one is written | ||
# (5) Session 3 - Unblocks the scanner of session 2 | ||
# (6) Session 2 - Continues the scan. The new inserted invalidation watermark should not be seen since the transaction was started after s2 has started the scan in (3) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I'm a little confused about this part -- if session 2 does a backfill into the range that was refreshed by session 1, what will be the final watermark value?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
In an ideal world, we should see the most recent timestamp (i.e., if a parallel refresh runs in parallel and moves the watermark, it should be guaranteed that we see this update. Otherwise we might lose an invalidation since we believe we are > watermark
). Said this, we have to avoid strong locks since this code runs on every transaction commit.
We have a rewrite of this code path on our agenda. The goal is to get rid of the complete invalidation trigger and perform this more asynchronously.
579a3d6
to
499b1c4
Compare
So far, the scanner in get_lowest_invalidated_time_for_hypertable() used a Snapshot that includes the data of all committed transactions (and not only the transactions that are started before the snapshot was created - SNAPSHOT_SELF). This could lead to a situation where we see the old and the new value of a parallel updated tuple. Since get_lowest_invalidated_time_for_hypertable() expects exactly one tuple to be found during the scan, we end up with an error. Since this is performed in our insert path (i.e., it is executed by the invalidation trigger), this error could reject inserts on the hypertable.
499b1c4
to
6fb552c
Compare
Codecov ReportAll modified and coverable lines are covered by tests ✅
Additional details and impacted files@@ Coverage Diff @@
## main #6677 +/- ##
==========================================
+ Coverage 80.06% 81.40% +1.34%
==========================================
Files 190 191 +1
Lines 37181 36422 -759
Branches 9450 9465 +15
==========================================
- Hits 29770 29651 -119
- Misses 2997 2998 +1
+ Partials 4414 3773 -641 ☔ View full report in Codecov by Sentry. |
Automated backport to 2.14.x not done: cherry-pick failed. Git status
|
This release contains performance improvements and bug fixes since the 2.14.2 release. We recommend that you upgrade at the next available opportunity. In addition, it includes these noteworthy features: * Support `time_bucket` with `origin` and/or `offset` on Continuous Aggregate * Compression improvements: - Improve expression pushdown - Add minmax sparse indexes when compressing columns with btree indexes - Make compression use the defaults functions - Vectorize filters in WHERE clause that contain text equality operators and LIKE expressions **Deprecation warning** * Starting on this release will not be possible to create Continuous Aggregate using `time_bucket_ng` anymore and it will be completely removed on the upcoming releases. * Recommend users to [migrate their old Continuous Aggregate format to the new one](https://docs.timescale.com/use-timescale/latest/continuous-aggregates/migrate/) because it support will be completely removed in next releases prevent them to migrate. * This is the last release supporting PostgreSQL 13. **For on-premise users and this release only**, you will need to run [this SQL script](https://github.com/timescale/timescaledb-extras/blob/master/utils/2.15.X-fix_hypertable_foreign_keys.sql) after running `ALTER EXTENSION`. More details can be found in the pull request [timescale#6786](timescale#6797). **Features** * timescale#6382 Support for time_bucket with origin and offset in CAggs * timescale#6696 Improve defaults for compression segment_by and order_by * timescale#6705 Add sparse minmax indexes for compressed columns that have uncompressed btree indexes * timescale#6754 Allow DROP CONSTRAINT on compressed hypertables * timescale#6767 Add metadata table `_timestaledb_internal.bgw_job_stat_history` for tracking job execution history * timescale#6798 Prevent usage of deprecated time_bucket_ng in CAgg definition * timescale#6810 Add telemetry for access methods * timescale#6811 Remove no longer relevant timescaledb.allow_install_without_preload GUC * timescale#6837 Add migration path for CAggs using time_bucket_ng * timescale#6865 Update the watermark when truncating a CAgg **Bugfixes** * timescale#6617 Fix error in show_chunks * timescale#6621 Remove metadata when dropping chunks * timescale#6677 Fix snapshot usage in CAgg invalidation scanner * timescale#6698 Define meaning of 0 retries for jobs as no retries * timescale#6717 Fix handling of compressed tables with primary or unique index in COPY path * timescale#6726 Fix constify cagg_watermark using window function when querying a CAgg * timescale#6729 Fix NULL start value handling in CAgg refresh * timescale#6732 Fix CAgg migration with custom timezone / date format settings * timescale#6752 Remove custom autovacuum setting from compressed chunks * timescale#6770 Fix plantime chunk exclusion for OSM chunk * timescale#6789 Fix deletes with subqueries and compression * timescale#6796 Fix a crash involving a view on a hypertable * timescale#6797 Fix foreign key constraint handling on compressed hypertables * timescale#6816 Fix handling of chunks with no contraints * timescale#6820 Fix a crash when the ts_hypertable_insert_blocker was called directly * timescale#6849 Use non-orderby compressed metadata in compressed DML * timescale#6867 Clean up compression settings when deleting compressed cagg * timescale#6869 Fix compressed DML with constraints of form value OP column * timescale#6870 Fix bool expression pushdown for queries on compressed chunks **Thanks** * @brasic for reporting a crash when the ts_hypertable_insert_blocker was called directly * @bvanelli for reporting an issue with the jobs retry count * @djzurawsk For reporting error when dropping chunks * @Dzuzepppe for reporting an issue with DELETEs using subquery on compressed chunk working incorrectly. * @hongquan For reporting a 'timestamp out of range' error during CAgg migrations * @kevcenteno for reporting an issue with the show_chunks API showing incorrect output when 'created_before/created_after' was used with time-partitioned columns. * @mahipv For starting working on the job history PR * @rovo89 For reporting constify cagg_watermark not working using window function when querying a CAgg
This release contains performance improvements and bug fixes since the 2.14.2 release. We recommend that you upgrade at the next available opportunity. In addition, it includes these noteworthy features: * Support `time_bucket` with `origin` and/or `offset` on Continuous Aggregate * Compression improvements: - Improve expression pushdown - Add minmax sparse indexes when compressing columns with btree indexes - Make compression use the defaults functions - Vectorize filters in WHERE clause that contain text equality operators and LIKE expressions **Deprecation warning** * Starting on this release will not be possible to create Continuous Aggregate using `time_bucket_ng` anymore and it will be completely removed on the upcoming releases. * Recommend users to [migrate their old Continuous Aggregate format to the new one](https://docs.timescale.com/use-timescale/latest/continuous-aggregates/migrate/) because it support will be completely removed in next releases prevent them to migrate. * This is the last release supporting PostgreSQL 13. **For on-premise users and this release only**, you will need to run [this SQL script](https://github.com/timescale/timescaledb-extras/blob/master/utils/2.15.X-fix_hypertable_foreign_keys.sql) after running `ALTER EXTENSION`. More details can be found in the pull request [timescale#6786](timescale#6797). **Features** * timescale#6382 Support for time_bucket with origin and offset in CAggs * timescale#6696 Improve defaults for compression segment_by and order_by * timescale#6705 Add sparse minmax indexes for compressed columns that have uncompressed btree indexes * timescale#6754 Allow DROP CONSTRAINT on compressed hypertables * timescale#6767 Add metadata table `_timestaledb_internal.bgw_job_stat_history` for tracking job execution history * timescale#6798 Prevent usage of deprecated time_bucket_ng in CAgg definition * timescale#6810 Add telemetry for access methods * timescale#6811 Remove no longer relevant timescaledb.allow_install_without_preload GUC * timescale#6837 Add migration path for CAggs using time_bucket_ng * timescale#6865 Update the watermark when truncating a CAgg **Bugfixes** * timescale#6617 Fix error in show_chunks * timescale#6621 Remove metadata when dropping chunks * timescale#6677 Fix snapshot usage in CAgg invalidation scanner * timescale#6698 Define meaning of 0 retries for jobs as no retries * timescale#6717 Fix handling of compressed tables with primary or unique index in COPY path * timescale#6726 Fix constify cagg_watermark using window function when querying a CAgg * timescale#6729 Fix NULL start value handling in CAgg refresh * timescale#6732 Fix CAgg migration with custom timezone / date format settings * timescale#6752 Remove custom autovacuum setting from compressed chunks * timescale#6770 Fix plantime chunk exclusion for OSM chunk * timescale#6789 Fix deletes with subqueries and compression * timescale#6796 Fix a crash involving a view on a hypertable * timescale#6797 Fix foreign key constraint handling on compressed hypertables * timescale#6816 Fix handling of chunks with no contraints * timescale#6820 Fix a crash when the ts_hypertable_insert_blocker was called directly * timescale#6849 Use non-orderby compressed metadata in compressed DML * timescale#6867 Clean up compression settings when deleting compressed cagg * timescale#6869 Fix compressed DML with constraints of form value OP column * timescale#6870 Fix bool expression pushdown for queries on compressed chunks **Thanks** * @brasic for reporting a crash when the ts_hypertable_insert_blocker was called directly * @bvanelli for reporting an issue with the jobs retry count * @djzurawsk For reporting error when dropping chunks * @Dzuzepppe for reporting an issue with DELETEs using subquery on compressed chunk working incorrectly. * @hongquan For reporting a 'timestamp out of range' error during CAgg migrations * @kevcenteno for reporting an issue with the show_chunks API showing incorrect output when 'created_before/created_after' was used with time-partitioned columns. * @mahipv For starting working on the job history PR * @rovo89 For reporting constify cagg_watermark not working using window function when querying a CAgg
This release contains performance improvements and bug fixes since the 2.14.2 release. We recommend that you upgrade at the next available opportunity. In addition, it includes these noteworthy features: * Support `time_bucket` with `origin` and/or `offset` on Continuous Aggregate * Compression improvements: - Improve expression pushdown - Add minmax sparse indexes when compressing columns with btree indexes - Make compression use the defaults functions - Vectorize filters in WHERE clause that contain text equality operators and LIKE expressions **Deprecation warning** * Starting on this release will not be possible to create Continuous Aggregate using `time_bucket_ng` anymore and it will be completely removed on the upcoming releases. * Recommend users to [migrate their old Continuous Aggregate format to the new one](https://docs.timescale.com/use-timescale/latest/continuous-aggregates/migrate/) because it support will be completely removed in next releases prevent them to migrate. * This is the last release supporting PostgreSQL 13. **For on-premise users and this release only**, you will need to run [this SQL script](https://github.com/timescale/timescaledb-extras/blob/master/utils/2.15.X-fix_hypertable_foreign_keys.sql) after running `ALTER EXTENSION`. More details can be found in the pull request [timescale#6786](timescale#6797). **Features** * timescale#6382 Support for time_bucket with origin and offset in CAggs * timescale#6696 Improve defaults for compression segment_by and order_by * timescale#6705 Add sparse minmax indexes for compressed columns that have uncompressed btree indexes * timescale#6754 Allow DROP CONSTRAINT on compressed hypertables * timescale#6767 Add metadata table `_timestaledb_internal.bgw_job_stat_history` for tracking job execution history * timescale#6798 Prevent usage of deprecated time_bucket_ng in CAgg definition * timescale#6810 Add telemetry for access methods * timescale#6811 Remove no longer relevant timescaledb.allow_install_without_preload GUC * timescale#6837 Add migration path for CAggs using time_bucket_ng * timescale#6865 Update the watermark when truncating a CAgg **Bugfixes** * timescale#6617 Fix error in show_chunks * timescale#6621 Remove metadata when dropping chunks * timescale#6677 Fix snapshot usage in CAgg invalidation scanner * timescale#6698 Define meaning of 0 retries for jobs as no retries * timescale#6717 Fix handling of compressed tables with primary or unique index in COPY path * timescale#6726 Fix constify cagg_watermark using window function when querying a CAgg * timescale#6729 Fix NULL start value handling in CAgg refresh * timescale#6732 Fix CAgg migration with custom timezone / date format settings * timescale#6752 Remove custom autovacuum setting from compressed chunks * timescale#6770 Fix plantime chunk exclusion for OSM chunk * timescale#6789 Fix deletes with subqueries and compression * timescale#6796 Fix a crash involving a view on a hypertable * timescale#6797 Fix foreign key constraint handling on compressed hypertables * timescale#6816 Fix handling of chunks with no contraints * timescale#6820 Fix a crash when the ts_hypertable_insert_blocker was called directly * timescale#6849 Use non-orderby compressed metadata in compressed DML * timescale#6867 Clean up compression settings when deleting compressed cagg * timescale#6869 Fix compressed DML with constraints of form value OP column * timescale#6870 Fix bool expression pushdown for queries on compressed chunks **Thanks** * @brasic for reporting a crash when the ts_hypertable_insert_blocker was called directly * @bvanelli for reporting an issue with the jobs retry count * @djzurawsk For reporting error when dropping chunks * @Dzuzepppe for reporting an issue with DELETEs using subquery on compressed chunk working incorrectly. * @hongquan For reporting a 'timestamp out of range' error during CAgg migrations * @kevcenteno for reporting an issue with the show_chunks API showing incorrect output when 'created_before/created_after' was used with time-partitioned columns. * @mahipv For starting working on the job history PR * @rovo89 For reporting constify cagg_watermark not working using window function when querying a CAgg
This release contains performance improvements and bug fixes since the 2.14.2 release. We recommend that you upgrade at the next available opportunity. In addition, it includes these noteworthy features: * Support `time_bucket` with `origin` and/or `offset` on Continuous Aggregate * Compression improvements: - Improve expression pushdown - Add minmax sparse indexes when compressing columns with btree indexes - Make compression use the defaults functions - Vectorize filters in WHERE clause that contain text equality operators and LIKE expressions **Deprecation warning** * Starting on this release will not be possible to create Continuous Aggregate using `time_bucket_ng` anymore and it will be completely removed on the upcoming releases. * Recommend users to [migrate their old Continuous Aggregate format to the new one](https://docs.timescale.com/use-timescale/latest/continuous-aggregates/migrate/) because it support will be completely removed in next releases prevent them to migrate. * This is the last release supporting PostgreSQL 13. **For on-premise users and this release only**, you will need to run [this SQL script](https://github.com/timescale/timescaledb-extras/blob/master/utils/2.15.X-fix_hypertable_foreign_keys.sql) after running `ALTER EXTENSION`. More details can be found in the pull request [#6786](#6797). **Features** * #6382 Support for time_bucket with origin and offset in CAggs * #6696 Improve defaults for compression segment_by and order_by * #6705 Add sparse minmax indexes for compressed columns that have uncompressed btree indexes * #6754 Allow DROP CONSTRAINT on compressed hypertables * #6767 Add metadata table `_timestaledb_internal.bgw_job_stat_history` for tracking job execution history * #6798 Prevent usage of deprecated time_bucket_ng in CAgg definition * #6810 Add telemetry for access methods * #6811 Remove no longer relevant timescaledb.allow_install_without_preload GUC * #6837 Add migration path for CAggs using time_bucket_ng * #6865 Update the watermark when truncating a CAgg **Bugfixes** * #6617 Fix error in show_chunks * #6621 Remove metadata when dropping chunks * #6677 Fix snapshot usage in CAgg invalidation scanner * #6698 Define meaning of 0 retries for jobs as no retries * #6717 Fix handling of compressed tables with primary or unique index in COPY path * #6726 Fix constify cagg_watermark using window function when querying a CAgg * #6729 Fix NULL start value handling in CAgg refresh * #6732 Fix CAgg migration with custom timezone / date format settings * #6752 Remove custom autovacuum setting from compressed chunks * #6770 Fix plantime chunk exclusion for OSM chunk * #6789 Fix deletes with subqueries and compression * #6796 Fix a crash involving a view on a hypertable * #6797 Fix foreign key constraint handling on compressed hypertables * #6816 Fix handling of chunks with no contraints * #6820 Fix a crash when the ts_hypertable_insert_blocker was called directly * #6849 Use non-orderby compressed metadata in compressed DML * #6867 Clean up compression settings when deleting compressed cagg * #6869 Fix compressed DML with constraints of form value OP column * #6870 Fix bool expression pushdown for queries on compressed chunks **Thanks** * @brasic for reporting a crash when the ts_hypertable_insert_blocker was called directly * @bvanelli for reporting an issue with the jobs retry count * @djzurawsk For reporting error when dropping chunks * @Dzuzepppe for reporting an issue with DELETEs using subquery on compressed chunk working incorrectly. * @hongquan For reporting a 'timestamp out of range' error during CAgg migrations * @kevcenteno for reporting an issue with the show_chunks API showing incorrect output when 'created_before/created_after' was used with time-partitioned columns. * @mahipv For starting working on the job history PR * @rovo89 For reporting constify cagg_watermark not working using window function when querying a CAgg
This release contains performance improvements and bug fixes since the 2.14.2 release. We recommend that you upgrade at the next available opportunity. In addition, it includes these noteworthy features: * Support `time_bucket` with `origin` and/or `offset` on Continuous Aggregate * Compression improvements: - Improve expression pushdown - Add minmax sparse indexes when compressing columns with btree indexes - Make compression use the defaults functions - Vectorize filters in WHERE clause that contain text equality operators and LIKE expressions **Deprecation warning** * Starting on this release will not be possible to create Continuous Aggregate using `time_bucket_ng` anymore and it will be completely removed on the upcoming releases. * Recommend users to [migrate their old Continuous Aggregate format to the new one](https://docs.timescale.com/use-timescale/latest/continuous-aggregates/migrate/) because it support will be completely removed in next releases prevent them to migrate. * This is the last release supporting PostgreSQL 13. **For on-premise users and this release only**, you will need to run [this SQL script](https://github.com/timescale/timescaledb-extras/blob/master/utils/2.15.X-fix_hypertable_foreign_keys.sql) after running `ALTER EXTENSION`. More details can be found in the pull request [timescale#6786](timescale#6797). **Features** * timescale#6382 Support for time_bucket with origin and offset in CAggs * timescale#6696 Improve defaults for compression segment_by and order_by * timescale#6705 Add sparse minmax indexes for compressed columns that have uncompressed btree indexes * timescale#6754 Allow DROP CONSTRAINT on compressed hypertables * timescale#6767 Add metadata table `_timestaledb_internal.bgw_job_stat_history` for tracking job execution history * timescale#6798 Prevent usage of deprecated time_bucket_ng in CAgg definition * timescale#6810 Add telemetry for access methods * timescale#6811 Remove no longer relevant timescaledb.allow_install_without_preload GUC * timescale#6837 Add migration path for CAggs using time_bucket_ng * timescale#6865 Update the watermark when truncating a CAgg **Bugfixes** * timescale#6617 Fix error in show_chunks * timescale#6621 Remove metadata when dropping chunks * timescale#6677 Fix snapshot usage in CAgg invalidation scanner * timescale#6698 Define meaning of 0 retries for jobs as no retries * timescale#6717 Fix handling of compressed tables with primary or unique index in COPY path * timescale#6726 Fix constify cagg_watermark using window function when querying a CAgg * timescale#6729 Fix NULL start value handling in CAgg refresh * timescale#6732 Fix CAgg migration with custom timezone / date format settings * timescale#6752 Remove custom autovacuum setting from compressed chunks * timescale#6770 Fix plantime chunk exclusion for OSM chunk * timescale#6789 Fix deletes with subqueries and compression * timescale#6796 Fix a crash involving a view on a hypertable * timescale#6797 Fix foreign key constraint handling on compressed hypertables * timescale#6816 Fix handling of chunks with no contraints * timescale#6820 Fix a crash when the ts_hypertable_insert_blocker was called directly * timescale#6849 Use non-orderby compressed metadata in compressed DML * timescale#6867 Clean up compression settings when deleting compressed cagg * timescale#6869 Fix compressed DML with constraints of form value OP column * timescale#6870 Fix bool expression pushdown for queries on compressed chunks **Thanks** * @brasic for reporting a crash when the ts_hypertable_insert_blocker was called directly * @bvanelli for reporting an issue with the jobs retry count * @djzurawsk For reporting error when dropping chunks * @Dzuzepppe for reporting an issue with DELETEs using subquery on compressed chunk working incorrectly. * @hongquan For reporting a 'timestamp out of range' error during CAgg migrations * @kevcenteno for reporting an issue with the show_chunks API showing incorrect output when 'created_before/created_after' was used with time-partitioned columns. * @mahipv For starting working on the job history PR * @rovo89 For reporting constify cagg_watermark not working using window function when querying a CAgg
This release contains performance improvements and bug fixes since the 2.14.2 release. We recommend that you upgrade at the next available opportunity. In addition, it includes these noteworthy features: * Support `time_bucket` with `origin` and/or `offset` on Continuous Aggregate * Compression improvements: - Improve expression pushdown - Add minmax sparse indexes when compressing columns with btree indexes - Make compression use the defaults functions - Vectorize filters in WHERE clause that contain text equality operators and LIKE expressions **Deprecation warning** * Starting on this release will not be possible to create Continuous Aggregate using `time_bucket_ng` anymore and it will be completely removed on the upcoming releases. * Recommend users to [migrate their old Continuous Aggregate format to the new one](https://docs.timescale.com/use-timescale/latest/continuous-aggregates/migrate/) because it support will be completely removed in next releases prevent them to migrate. * This is the last release supporting PostgreSQL 13. **For on-premise users and this release only**, you will need to run [this SQL script](https://github.com/timescale/timescaledb-extras/blob/master/utils/2.15.X-fix_hypertable_foreign_keys.sql) after running `ALTER EXTENSION`. More details can be found in the pull request [#6786](#6797). **Features** * #6382 Support for time_bucket with origin and offset in CAggs * #6696 Improve defaults for compression segment_by and order_by * #6705 Add sparse minmax indexes for compressed columns that have uncompressed btree indexes * #6754 Allow DROP CONSTRAINT on compressed hypertables * #6767 Add metadata table `_timestaledb_internal.bgw_job_stat_history` for tracking job execution history * #6798 Prevent usage of deprecated time_bucket_ng in CAgg definition * #6810 Add telemetry for access methods * #6811 Remove no longer relevant timescaledb.allow_install_without_preload GUC * #6837 Add migration path for CAggs using time_bucket_ng * #6865 Update the watermark when truncating a CAgg **Bugfixes** * #6617 Fix error in show_chunks * #6621 Remove metadata when dropping chunks * #6677 Fix snapshot usage in CAgg invalidation scanner * #6698 Define meaning of 0 retries for jobs as no retries * #6717 Fix handling of compressed tables with primary or unique index in COPY path * #6726 Fix constify cagg_watermark using window function when querying a CAgg * #6729 Fix NULL start value handling in CAgg refresh * #6732 Fix CAgg migration with custom timezone / date format settings * #6752 Remove custom autovacuum setting from compressed chunks * #6770 Fix plantime chunk exclusion for OSM chunk * #6789 Fix deletes with subqueries and compression * #6796 Fix a crash involving a view on a hypertable * #6797 Fix foreign key constraint handling on compressed hypertables * #6816 Fix handling of chunks with no contraints * #6820 Fix a crash when the ts_hypertable_insert_blocker was called directly * #6849 Use non-orderby compressed metadata in compressed DML * #6867 Clean up compression settings when deleting compressed cagg * #6869 Fix compressed DML with constraints of form value OP column * #6870 Fix bool expression pushdown for queries on compressed chunks **Thanks** * @brasic for reporting a crash when the ts_hypertable_insert_blocker was called directly * @bvanelli for reporting an issue with the jobs retry count * @djzurawsk For reporting error when dropping chunks * @Dzuzepppe for reporting an issue with DELETEs using subquery on compressed chunk working incorrectly. * @hongquan For reporting a 'timestamp out of range' error during CAgg migrations * @kevcenteno for reporting an issue with the show_chunks API showing incorrect output when 'created_before/created_after' was used with time-partitioned columns. * @mahipv For starting working on the job history PR * @rovo89 For reporting constify cagg_watermark not working using window function when querying a CAgg
So far, the scanner in
get_lowest_invalidated_time_for_hypertable()
used a Snapshot that includes the data of all committed transactions (and not only the transactions that are started before the snapshot was created -SNAPSHOT_SELF
). This could lead to a situation where we see the old and the new value of a parallel updated tuple. Sinceget_lowest_invalidated_time_for_hypertable()
expects exactly one tuple to be found during the scan, we end up with an error. Since this is performed in our insert path (i.e., it is executed by the invalidation trigger), this error could reject inserts on the hypertable (ERROR: more than one invalidation watermark found
).Should be merged after: #6678