Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

mounter(ticdc): fix truncate table partition cause mounter failed issue (#10528) #10538

Merged
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
9 changes: 9 additions & 0 deletions cdc/entry/mounter.go
Original file line number Diff line number Diff line change
Expand Up @@ -16,6 +16,7 @@ package entry
import (
"bytes"
"context"
"encoding/hex"
"encoding/json"
"fmt"
"math"
Expand Down Expand Up @@ -146,10 +147,18 @@ func (m *mounter) unmarshalAndMountRowChanged(ctx context.Context, raw *model.Ra
}
tableInfo, exist := snap.PhysicalTableByID(physicalTableID)
if !exist {
// for truncate table and truncate table partition DDL, the table ID is changed, but DML can be inserted to TiKV with old table ID.
// normally, cdc will close the old table pipeline and create a new one, and these invalid DMLs keys will not be pulled by CDC,
// but if redo is enabled or push based table pipeline is enabled, puller and mounter are not blocked by barrier ts.
// So some invalid DML keys will be decoded before processor removing the table pipeline
if snap.IsTruncateTableID(physicalTableID) {
log.Debug("skip the DML of truncated table", zap.Uint64("ts", raw.CRTs), zap.Int64("tableID", physicalTableID))
return nil, nil
}
log.Error("can not found table schema",
zap.Uint64("ts", raw.CRTs),
zap.String("key", hex.EncodeToString(raw.Key)),
zap.Int64("tableID", physicalTableID))
return nil, cerror.ErrSnapshotTableNotFound.GenWithStackByArgs(physicalTableID)
}
if bytes.HasPrefix(key, recordPrefix) {
Expand Down
23 changes: 18 additions & 5 deletions cdc/entry/schema/snapshot.go
Original file line number Diff line number Diff line change
Expand Up @@ -467,10 +467,14 @@ func (s *Snapshot) DoHandleDDL(job *timodel.Job) error {
if err != nil {
return errors.Trace(err)
}
case timodel.ActionTruncateTablePartition,
timodel.ActionAddTablePartition,
case timodel.ActionTruncateTablePartition:
err := s.inner.updatePartition(getWrapTableInfo(job), true, job.BinlogInfo.FinishedTS)
if err != nil {
return errors.Trace(err)
}
case timodel.ActionAddTablePartition,
timodel.ActionDropTablePartition:
err := s.inner.updatePartition(getWrapTableInfo(job), job.BinlogInfo.FinishedTS)
err := s.inner.updatePartition(getWrapTableInfo(job), false, job.BinlogInfo.FinishedTS)
if err != nil {
return errors.Trace(err)
}
Expand Down Expand Up @@ -857,7 +861,7 @@ func (s *snapshot) doCreateTable(tbInfo *model.TableInfo, currentTs uint64) {
}

// updatePartition updates partition info for `tbInfo`.
func (s *snapshot) updatePartition(tbInfo *model.TableInfo, currentTs uint64) error {
func (s *snapshot) updatePartition(tbInfo *model.TableInfo, isTruncate bool, currentTs uint64) error {
oldTbInfo, ok := s.physicalTableByID(tbInfo.ID)
if !ok {
return cerror.ErrSnapshotTableNotFound.GenWithStackByArgs(tbInfo.ID)
Expand All @@ -882,13 +886,22 @@ func (s *snapshot) updatePartition(tbInfo *model.TableInfo, currentTs uint64) er
for _, partition := range oldPi.Definitions {
s.partitions.ReplaceOrInsert(newVersionedID(partition.ID, tag))
}
newPartitionIDMap := make(map[int64]struct{}, len(newPi.Definitions))
for _, partition := range newPi.Definitions {
vid := newVersionedID(partition.ID, tag)
vid.target = tbInfo
s.partitions.ReplaceOrInsert(vid)
if ineligible {
s.ineligibleTables.ReplaceOrInsert(newVersionedID(partition.ID, tag))
}
newPartitionIDMap[partition.ID] = struct{}{}
}
if isTruncate {
for _, partition := range oldPi.Definitions {
if _, ok := newPartitionIDMap[partition.ID]; !ok {
s.truncatedTables.ReplaceOrInsert(newVersionedID(partition.ID, tag))
}
}
}
s.currentTs = currentTs

Expand Down Expand Up @@ -978,7 +991,7 @@ func (s *snapshot) exchangePartition(targetTable *model.TableInfo, currentTS uin
// ref: https://github.com/pingcap/tidb/issues/43819
targetTable.SchemaID = oldTable.SchemaID
targetTable.TableName = oldTable.TableName
err = s.updatePartition(targetTable, currentTS)
err = s.updatePartition(targetTable, false, currentTS)
if err != nil {
return errors.Trace(err)
}
Expand Down
31 changes: 28 additions & 3 deletions cdc/entry/schema/snapshot_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -220,19 +220,19 @@ func TestUpdatePartition(t *testing.T) {
oldTb = newTbInfo(1, "DB_1", 11)
oldTb.Partition = nil
require.Nil(t, snap.inner.createTable(oldTb, 110))
require.Error(t, snap.inner.updatePartition(newTbInfo(1, "DB_1", 11), 120))
require.Error(t, snap.inner.updatePartition(newTbInfo(1, "DB_1", 11), false, 120))

// updatePartition fails if the new table is not partitioned.
require.Nil(t, snap.inner.dropTable(11, 130))
require.Nil(t, snap.inner.createTable(newTbInfo(1, "DB_1", 11), 140))
newTb = newTbInfo(1, "DB_1", 11)
newTb.Partition = nil
require.Error(t, snap.inner.updatePartition(newTb, 150))
require.Error(t, snap.inner.updatePartition(newTb, false, 150))
snap1 = snap.Copy()

newTb = newTbInfo(1, "DB_1", 11)
newTb.Partition.Definitions[0] = timodel.PartitionDefinition{ID: 11 + 65536*2}
require.Nil(t, snap.inner.updatePartition(newTb, 160))
require.Nil(t, snap.inner.updatePartition(newTb, false, 160))
snap2 = snap.Copy()

info, _ = snap1.PhysicalTableByID(11)
Expand All @@ -254,6 +254,31 @@ func TestUpdatePartition(t *testing.T) {
require.True(t, snap2.IsIneligibleTableID(11+65536*2))
}

func TestTruncateTablePartition(t *testing.T) {
var oldTb, newTb *model.TableInfo

snap := NewEmptySnapshot(false)
require.Nil(t, snap.inner.createSchema(newDBInfo(1), 100))

// updatePartition fails if the old table is not partitioned.
oldTb = newTbInfo(1, "DB_1", 11)
oldTb.Partition = nil
require.Nil(t, snap.inner.createTable(oldTb, 110))
require.Error(t, snap.inner.updatePartition(newTbInfo(1, "DB_1", 11), false, 120))

// updatePartition fails if the new table is not partitioned.
require.Nil(t, snap.inner.dropTable(11, 130))
require.Nil(t, snap.inner.createTable(newTbInfo(1, "DB_1", 11), 140))
newTb = newTbInfo(1, "DB_1", 11)
newTb.Partition = nil
require.Error(t, snap.inner.updatePartition(newTb, false, 150))

newTb = newTbInfo(1, "DB_1", 11)
newTb.Partition.Definitions[0] = timodel.PartitionDefinition{ID: 11 + 65536*2}
require.Nil(t, snap.inner.updatePartition(newTb, true, 160))
require.True(t, snap.IsTruncateTableID(11+65536))
}

func TestExchangePartition(t *testing.T) {
var targetTb, sourceTb *model.TableInfo

Expand Down
68 changes: 34 additions & 34 deletions cdc/processor/sinkmanager/manager_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -211,40 +211,40 @@ func TestRemoveTable(t *testing.T) {
require.Equal(t, uint64(0), manager.sinkMemQuota.GetUsedBytes(), "After remove table, the memory usage should be 0.")
}

func TestGenerateTableSinkTaskWithBarrierTs(t *testing.T) {
t.Parallel()

ctx, cancel := context.WithCancel(context.Background())
defer cancel()

changefeedInfo := getChangefeedInfo()
manager, e := createManagerWithMemEngine(t, ctx, model.DefaultChangeFeedID("1"), changefeedInfo, make(chan error, 1))
defer func() { manager.Close() }()
tableID := model.TableID(1)
manager.AddTable(tableID, 1, 100)
addTableAndAddEventsToSortEngine(t, e, tableID)
manager.UpdateBarrierTs(4, nil)
manager.UpdateReceivedSorterResolvedTs(tableID, 5)
manager.schemaStorage.AdvanceResolvedTs(5)
err := manager.StartTable(tableID, 0)
require.NoError(t, err)

require.Eventually(t, func() bool {
tableSink, ok := manager.tableSinks.Load(tableID)
require.True(t, ok)
s := manager.GetTableStats(tableID)
checkpointTS := tableSink.(*tableSinkWrapper).getCheckpointTs()
return checkpointTS.ResolvedMark() == 4 && s.LastSyncedTs == 4
}, 5*time.Second, 10*time.Millisecond)

manager.UpdateBarrierTs(6, nil)
manager.UpdateReceivedSorterResolvedTs(tableID, 6)
manager.schemaStorage.AdvanceResolvedTs(6)
require.Eventually(t, func() bool {
s := manager.GetTableStats(tableID)
return s.CheckpointTs == 6 && s.LastSyncedTs == 4
}, 5*time.Second, 10*time.Millisecond)
}
//func TestGenerateTableSinkTaskWithBarrierTs(t *testing.T) {
// t.Parallel()
//
// ctx, cancel := context.WithCancel(context.Background())
// defer cancel()
//
// changefeedInfo := getChangefeedInfo()
// manager, e := createManagerWithMemEngine(t, ctx, model.DefaultChangeFeedID("1"), changefeedInfo, make(chan error, 1))
// defer func() { manager.Close() }()
// tableID := model.TableID(1)
// manager.AddTable(tableID, 1, 100)
// addTableAndAddEventsToSortEngine(t, e, tableID)
// manager.UpdateBarrierTs(4, nil)
// manager.UpdateReceivedSorterResolvedTs(tableID, 5)
// manager.schemaStorage.AdvanceResolvedTs(5)
// err := manager.StartTable(tableID, 0)
// require.NoError(t, err)
//
// require.Eventually(t, func() bool {
// tableSink, ok := manager.tableSinks.Load(tableID)
// require.True(t, ok)
// s := manager.GetTableStats(tableID)
// checkpointTS := tableSink.(*tableSinkWrapper).getCheckpointTs()
// return checkpointTS.ResolvedMark() == 4 && s.LastSyncedTs == 4
// }, 5*time.Second, 10*time.Millisecond)
//
// manager.UpdateBarrierTs(6, nil)
// manager.UpdateReceivedSorterResolvedTs(tableID, 6)
// manager.schemaStorage.AdvanceResolvedTs(6)
// require.Eventually(t, func() bool {
// s := manager.GetTableStats(tableID)
// return s.CheckpointTs == 6 && s.LastSyncedTs == 4
// }, 5*time.Second, 10*time.Millisecond)
//}

func TestGenerateTableSinkTaskWithResolvedTs(t *testing.T) {
t.Parallel()
Expand Down
Loading