add more ut #352
Matrix: Clients / UT & CodeCov
Matrix: Internal / UT & CodeCov
Matrix: Services / UT & CodeCov
Matrix: StaticAnalysis
ValidateGradleWrapper
8s
StaticAnalysisAndUnitTestsCompletionCheck
0s
Annotations
75 errors and 19 warnings
StaticAnalysis (17)
Process completed with exit code 1.
|
Internal / UT & CodeCov (8)
Process completed with exit code 1.
|
Internal / UT & CodeCov (17)
Process completed with exit code 1.
|
Internal / UT & CodeCov (11)
Process completed with exit code 1.
|
TestVeniceDelegateMode.testScatterForMultiGetWithHelixAssistedRouting:
services/venice-router/src/test/java/com/linkedin/venice/router/api/TestVeniceDelegateMode.java#L724
java.lang.AssertionError: expected [true] but found [false]
|
RouterRequestThrottlingTest.testMultiKeyThrottling[2](COMPUTE):
services/venice-router/src/test/java/com/linkedin/venice/router/throttle/RouterRequestThrottlingTest.java#L162
java.lang.AssertionError: expected [true] but found [false]
|
TestVeniceDelegateMode.testScatterForMultiGetWithHelixAssistedRouting:
services/venice-router/src/test/java/com/linkedin/venice/router/api/TestVeniceDelegateMode.java#L724
java.lang.AssertionError: expected [true] but found [false]
|
RouterRequestThrottlingTest.testMultiKeyThrottling[2](COMPUTE):
services/venice-router/src/test/java/com/linkedin/venice/router/throttle/RouterRequestThrottlingTest.java#L162
java.lang.AssertionError: expected [true] but found [false]
|
TestVeniceDelegateMode.testScatterForMultiGetWithHelixAssistedRouting:
services/venice-router/src/test/java/com/linkedin/venice/router/api/TestVeniceDelegateMode.java#L724
java.lang.AssertionError: expected [true] but found [false]
|
RouterRequestThrottlingTest.testMultiKeyThrottling[0](MULTI_GET):
services/venice-router/src/test/java/com/linkedin/venice/router/throttle/RouterRequestThrottlingTest.java#L162
java.lang.AssertionError: expected [true] but found [false]
|
RouterRequestThrottlingTest.testMultiKeyThrottling[2](COMPUTE):
services/venice-router/src/test/java/com/linkedin/venice/router/throttle/RouterRequestThrottlingTest.java#L162
java.lang.AssertionError: expected [true] but found [false]
|
TestVeniceDelegateMode.testScatterForMultiGetWithHelixAssistedRouting:
services/venice-router/src/test/java/com/linkedin/venice/router/api/TestVeniceDelegateMode.java#L724
java.lang.AssertionError: expected [true] but found [false]
|
RouterRequestThrottlingTest.testMultiKeyThrottling[0](MULTI_GET):
services/venice-router/src/test/java/com/linkedin/venice/router/throttle/RouterRequestThrottlingTest.java#L162
java.lang.AssertionError: expected [true] but found [false]
|
RouterRequestThrottlingTest.testMultiKeyThrottling[2](COMPUTE):
services/venice-router/src/test/java/com/linkedin/venice/router/throttle/RouterRequestThrottlingTest.java#L162
java.lang.AssertionError: expected [true] but found [false]
|
TestVeniceDelegateMode.testScatterForMultiGetWithHelixAssistedRouting:
services/venice-router/src/test/java/com/linkedin/venice/router/api/TestVeniceDelegateMode.java#L724
java.lang.AssertionError: expected [true] but found [false]
|
AggServerReadQuotaUsageStatsTest.testAggServerQuotaUsageStats:
services/venice-server/src/test/java/com/linkedin/venice/stats/AggServerReadQuotaUsageStatsTest.java#L53
java.lang.AssertionError: expected [16.666666666666668] but found [16.604675876726887]
|
TestVeniceDelegateMode.testScatterForMultiGetWithHelixAssistedRouting:
services/venice-router/src/test/java/com/linkedin/venice/router/api/TestVeniceDelegateMode.java#L724
java.lang.AssertionError: expected [true] but found [false]
|
AggServerReadQuotaUsageStatsTest.testAggServerQuotaUsageStats:
services/venice-server/src/test/java/com/linkedin/venice/stats/AggServerReadQuotaUsageStatsTest.java#L53
java.lang.AssertionError: expected [16.666666666666668] but found [16.604675876726887]
|
com.linkedin.venice.helix.TestControllerKMERegistrationFromMessageHeader ► testKMERegistrationThroughAdminTopicChannel:
internal/venice-test-common/src/main/java/com/linkedin/venice/utils/TestUtils.java#L245
Failed test found in:
IntegrationTests_1060-artifacts/internal/venice-test-common/build/test-results/integrationTests_1060/TEST-com.linkedin.venice.helix.TestControllerKMERegistrationFromMessageHeader.xml
internal/venice-test-common/build/test-results/integrationTests_1060/TEST-com.linkedin.venice.helix.TestControllerKMERegistrationFromMessageHeader.xml
Error:
java.lang.AssertionError: com.linkedin.venice.exceptions.VeniceNoStoreException: Store: venice_system_store_KAFKA_MESSAGE_ENVELOPE does not exist.
|
com.linkedin.venice.helixrebalance.TestRebalanceByDefaultStrategy ► testRollingUpgrade:
internal/venice-test-common/src/main/java/com/linkedin/venice/utils/TestUtils.java#L168
Failed test found in:
IntegrationTests_1060-artifacts/internal/venice-test-common/build/test-results/integrationTests_1060/TEST-com.linkedin.venice.helixrebalance.TestRebalanceByDefaultStrategy.xml
internal/venice-test-common/build/test-results/integrationTests_1060/TEST-com.linkedin.venice.helixrebalance.TestRebalanceByDefaultStrategy.xml
Error:
java.lang.AssertionError: Non-deterministic condition not met. expected [true] but found [false]
|
NativeMetadataRepositoryTest.testNativeMetadataRepositoryStats:
clients/da-vinci-client/src/test/java/com/linkedin/davinci/repository/NativeMetadataRepositoryTest.java#L123
java.lang.AssertionError: expected [1000.0] but found [0.0]
|
BatchGetAvroStoreClientUnitTest.testSimpleStreamingBatchGet:
clients/venice-client/src/test/java/com/linkedin/venice/fastclient/BatchGetAvroStoreClientUnitTest.java#L1
org.testng.internal.thread.ThreadTimeoutException: Method com.linkedin.venice.fastclient.BatchGetAvroStoreClientUnitTest.testSimpleStreamingBatchGet() didn't finish within the time-out 5000
|
NativeMetadataRepositoryTest.testNativeMetadataRepositoryStats:
clients/da-vinci-client/src/test/java/com/linkedin/davinci/repository/NativeMetadataRepositoryTest.java#L123
java.lang.AssertionError: expected [1000.0] but found [0.0]
|
BatchGetAvroStoreClientUnitTest.testSimpleStreamingBatchGet:
clients/venice-client/src/test/java/com/linkedin/venice/fastclient/BatchGetAvroStoreClientUnitTest.java#L1
org.testng.internal.thread.ThreadTimeoutException: Method com.linkedin.venice.fastclient.BatchGetAvroStoreClientUnitTest.testSimpleStreamingBatchGet() didn't finish within the time-out 5000
|
SITWithPWiseWithoutBufferAfterLeaderTest.testProcessConsumerActionsError:
clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseWithoutBufferAfterLeaderTest.java#L1
Wanted but not invoked:
leaderFollowerStoreIngestionTask.reportError(
<any string>,
1,
<Capturing argument>
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.reportError(StoreIngestionTask.java:4146)
However, there were exactly 40 interactions with this mock:
leaderFollowerStoreIngestionTask.subscribePartition(
TestTopic_797a883f18_2aa3c2a4_v1-1
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.runTest(StoreIngestionTaskTest.java:854)
leaderFollowerStoreIngestionTask.subscribePartition(
TestTopic_797a883f18_2aa3c2a4_v1-1,
true
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.subscribePartition(StoreIngestionTask.java:593)
leaderFollowerStoreIngestionTask.throwIfNotRunning();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.subscribePartition(StoreIngestionTask.java:607)
leaderFollowerStoreIngestionTask.isRunning();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.throwIfNotRunning(StoreIngestionTask.java:554)
leaderFollowerStoreIngestionTask.getIsRunning();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.isRunning(StoreIngestionTask.java:3989)
leaderFollowerStoreIngestionTask.nextSeqNum();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.subscribePartition(StoreIngestionTask.java:616)
leaderFollowerStoreIngestionTask.run();
-> at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539)
leaderFollowerStoreIngestionTask.isRunning();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1592)
leaderFollowerStoreIngestionTask.getIsRunning();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.isRunning(StoreIngestionTask.java:3989)
leaderFollowerStoreIngestionTask.updateIngestionRoleIfStoreChanged(
Mock for Store, hashCode: 100090655
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1594)
leaderFollowerStoreIngestionTask.isHybridMode();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.updateIngestionRoleIfStoreChanged(StoreIngestionTask.java:1511)
leaderFollowerStoreIngestionTask.processConsumerActions(
Mock for Store, hashCode: 100090655
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1595)
leaderFollowerStoreIngestionTask.processConsumerAction(
KafkaTaskMessage{type=SUBSCRIBE, topicPartition=TestTopic_797a883f18_2aa3c2a4_v1-1, attempts=1, sequenceNumber=1, createdTimestampInMs=1732155484357},
Mock for Store, hashCode: 100090655
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerActions(StoreIngestionTask.java:1864)
leaderFollowerStoreIngestionTask.processCommonConsumerAction(
KafkaTaskMessage{type=SUBSCRIBE, topicPartition=TestTopic_797a883f18_2aa3c2a4_v1-1, attempts=1, sequenceNumber=1, createdTimestampInMs=1732155484357}
);
-> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.processConsumerAction(LeaderFollowerStoreIngestionTask.java:556)
leaderFollowerStoreIngestionTask.reportIfCatchUpVersionTopicOffset(
PCS{replicaId=TestTopic_797a883f18_2aa3c2a4_v1-1, hybrid=false, latestProcessedLocalVersionTopicOffset=-1, latestProcessedUpstreamVersionTopicOffset=-1, latestProcessedUpstreamRTOffsetMap={}, latestIgnoredUpstreamRTOffsetMap={}, latestRTOffsetTriedToProduceToVTMap{}, offsetRecord=OffsetRecord{localVersionTopicOffset=-1, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}, errorReported=false, started=false, lagCaughtUp=false, processedRecordSizeSinceLastSync=0, leaderFollowerState=STANDBY}
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processCommonConsumerAction(StoreIngestionTask.java:2107)
leaderFollowerStoreIngestionTask.updateLeaderTopicOnFollower(
PCS{replicaId=TestTopic_797a883f18_2aa3c2a4_v1-1, hybrid=false, latestProcessedLocalVersionTopicOffset=-1, latestProcessedUpstreamVersionTopicOffset=-1, latestProcessedUpstreamRTOffsetMap={}, latestIgnoredUpstreamRTOffsetMap={}, latestRTOffsetTriedToProduceToVTMap{}, offsetRecord=OffsetRecord{localVersionTopicOffset=-1, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}, errorReported=false, started=false, lagCaughtUp=false, processedRecordSizeSinceLastSync=0, leaderFollowerState=STANDBY}
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processCommonConsumerAction(StoreIngestionTask.java:2112)
leaderFollowerStoreIngestionTask.consumerSubscribe(
TestTopic_797a883f18_2aa3c2a4_v1-1,
-1L,
"local_InMemoryKafkaBroker:41625"
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processCommonConsumerAction(StoreIngestionTask.java:2118)
leaderFollowerStoreIngestionTask.createKafkaConsumerProperties(
{"kafka.bootstrap.servers" = "local_InMemoryKafkaBroker:41625"},
"local_InMemoryKafkaBroker:41625",
false
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.consumerSubscribe(StoreIngestionTask.java:3497)
leaderFollowerStoreIngestionTask.getVersionTopic();
-> at com.linkedin.davinci.kafka.consumer.StorePartitionDataReceiver.destinationIdentifier(StorePartitionDataReceiver.java:86)
leaderFollowerStoreIngestionTask.getVersionTopic();
-> at com.linkedin.davinci.kafka.consumer.StorePartitionDataReceiver.destinationIdentifier(StorePartitionDataReceiver.java:86)
leaderFollowerStoreIngestionTask.checkLongRunningTaskState();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1596)
leaderFollowerStoreIngestionTask.getVersionTopic();
-> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.checkLongRunningTaskState(LeaderFollowerStoreIngestionTask.java:755)
leaderFollowerStoreIngestionTask.checkIngestionProgress(
Mock for Store, hashCode: 100090655
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1597)
leaderFollowerStoreIngestionTask.consumerHasAnySubscription();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.checkIngestionProgress(StoreIngestionTask.java:1464)
leaderFollowerStoreIngestionTask.maybeSendIngestionHeartbeat();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1598)
leaderFollowerStoreIngestionTask.isHybridMode();
-> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.maybeSendIngestionHeartbeat(LeaderFollowerStoreIngestionTask.java:3740)
leaderFollowerStoreIngestionTask.isRunning();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1592)
leaderFollowerStoreIngestionTask.getIsRunning();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.isRunning(StoreIngestionTask.java:3989)
leaderFollowerStoreIngestionTask.updateIngestionRoleIfStoreChanged(
Mock for Store, hashCode: 100090655
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1594)
leaderFollowerStoreIngestionTask.isHybridMode();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.updateIngestionRoleIfStoreChanged(StoreIngestionTask.java:1511)
leaderFollowerStoreIngestionTask.processConsumerActions(
Mock for Store, hashCode: 100090655
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1595)
leaderFollowerStoreIngestionTask.checkLongRunningTaskState();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1596)
leaderFollowerStoreIngestionTask.getVersionTopic();
-> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.checkLongRunningTaskState(LeaderFollowerStoreIngestionTask.java:755)
leaderFollowerStoreIngestionTask.checkIngestionProgress(
Mock for Store, hashCode: 100090655
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1597)
leaderFollowerStoreIngestionTask.consumerHasAnySubscription();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.checkIngestionProgress(StoreIngestionTask.java:1464)
leaderFollowerStoreIngestionTask.processConsumerActions(
Mock for Store, hashCode: 100090655
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testProcessConsumerActionsError$178(StoreIngestionTaskTest.java:4982)
leaderFollowerStoreIngestionTask.processConsumerActions(
Mock for Store, hashCode: 100090655
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testProcessConsumerActionsError$178(StoreIngestionTaskTest.java:4982)
leaderFollowerStoreIngestionTask.processConsumerActions(
Mock for Store, hashCode: 100090655
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testProcessConsumerActionsError$178(StoreIngestionTaskTest.java:4982)
leaderFollowerStoreIngestionTask.processConsumerActions(
Mock for Store, hashCode: 100090655
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testProcessConsumerActionsError$178(StoreIngestionTaskTest.java:4982)
leaderFollowerStoreIngestionTask.processConsumerActions(
Mock for Store, hashCode: 100090655
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testProcessConsumerActionsError$178(StoreIngestionTaskTest.java:4982)
|
SITWithTWiseWithoutBufferAfterLeaderTest.testResubscribeAfterRoleChange:
clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithTWiseWithoutBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.TooFewActualInvocations:
abstractStorageEngine.putWithReplicationMetadata(
1,
<any>,
<any>,
<any>
);
Wanted 200 times:
-> at com.linkedin.davinci.store.AbstractStorageEngine.putWithReplicationMetadata(AbstractStorageEngine.java:487)
But was 103 times:
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
|
NativeMetadataRepositoryTest.testNativeMetadataRepositoryStats:
clients/da-vinci-client/src/test/java/com/linkedin/davinci/repository/NativeMetadataRepositoryTest.java#L123
java.lang.AssertionError: expected [1000.0] but found [0.0]
|
BatchGetAvroStoreClientUnitTest.testSimpleStreamingBatchGet:
clients/venice-client/src/test/java/com/linkedin/venice/fastclient/BatchGetAvroStoreClientUnitTest.java#L65
org.testng.internal.thread.ThreadTimeoutException: Method com.linkedin.venice.fastclient.BatchGetAvroStoreClientUnitTest.testSimpleStreamingBatchGet() didn't finish within the time-out 5000
|
SITWithPWiseWithoutBufferAfterLeaderTest.testProcessConsumerActionsError:
clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseWithoutBufferAfterLeaderTest.java#L1
Wanted but not invoked:
leaderFollowerStoreIngestionTask.reportError(
<any string>,
1,
<Capturing argument>
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.reportError(StoreIngestionTask.java:4146)
However, there were exactly 40 interactions with this mock:
leaderFollowerStoreIngestionTask.subscribePartition(
TestTopic_797a883f18_2aa3c2a4_v1-1
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.runTest(StoreIngestionTaskTest.java:854)
leaderFollowerStoreIngestionTask.subscribePartition(
TestTopic_797a883f18_2aa3c2a4_v1-1,
true
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.subscribePartition(StoreIngestionTask.java:593)
leaderFollowerStoreIngestionTask.throwIfNotRunning();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.subscribePartition(StoreIngestionTask.java:607)
leaderFollowerStoreIngestionTask.isRunning();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.throwIfNotRunning(StoreIngestionTask.java:554)
leaderFollowerStoreIngestionTask.getIsRunning();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.isRunning(StoreIngestionTask.java:3989)
leaderFollowerStoreIngestionTask.nextSeqNum();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.subscribePartition(StoreIngestionTask.java:616)
leaderFollowerStoreIngestionTask.run();
-> at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539)
leaderFollowerStoreIngestionTask.isRunning();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1592)
leaderFollowerStoreIngestionTask.getIsRunning();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.isRunning(StoreIngestionTask.java:3989)
leaderFollowerStoreIngestionTask.updateIngestionRoleIfStoreChanged(
Mock for Store, hashCode: 100090655
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1594)
leaderFollowerStoreIngestionTask.isHybridMode();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.updateIngestionRoleIfStoreChanged(StoreIngestionTask.java:1511)
leaderFollowerStoreIngestionTask.processConsumerActions(
Mock for Store, hashCode: 100090655
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1595)
leaderFollowerStoreIngestionTask.processConsumerAction(
KafkaTaskMessage{type=SUBSCRIBE, topicPartition=TestTopic_797a883f18_2aa3c2a4_v1-1, attempts=1, sequenceNumber=1, createdTimestampInMs=1732155484357},
Mock for Store, hashCode: 100090655
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerActions(StoreIngestionTask.java:1864)
leaderFollowerStoreIngestionTask.processCommonConsumerAction(
KafkaTaskMessage{type=SUBSCRIBE, topicPartition=TestTopic_797a883f18_2aa3c2a4_v1-1, attempts=1, sequenceNumber=1, createdTimestampInMs=1732155484357}
);
-> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.processConsumerAction(LeaderFollowerStoreIngestionTask.java:556)
leaderFollowerStoreIngestionTask.reportIfCatchUpVersionTopicOffset(
PCS{replicaId=TestTopic_797a883f18_2aa3c2a4_v1-1, hybrid=false, latestProcessedLocalVersionTopicOffset=-1, latestProcessedUpstreamVersionTopicOffset=-1, latestProcessedUpstreamRTOffsetMap={}, latestIgnoredUpstreamRTOffsetMap={}, latestRTOffsetTriedToProduceToVTMap{}, offsetRecord=OffsetRecord{localVersionTopicOffset=-1, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}, errorReported=false, started=false, lagCaughtUp=false, processedRecordSizeSinceLastSync=0, leaderFollowerState=STANDBY}
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processCommonConsumerAction(StoreIngestionTask.java:2107)
leaderFollowerStoreIngestionTask.updateLeaderTopicOnFollower(
PCS{replicaId=TestTopic_797a883f18_2aa3c2a4_v1-1, hybrid=false, latestProcessedLocalVersionTopicOffset=-1, latestProcessedUpstreamVersionTopicOffset=-1, latestProcessedUpstreamRTOffsetMap={}, latestIgnoredUpstreamRTOffsetMap={}, latestRTOffsetTriedToProduceToVTMap{}, offsetRecord=OffsetRecord{localVersionTopicOffset=-1, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}, errorReported=false, started=false, lagCaughtUp=false, processedRecordSizeSinceLastSync=0, leaderFollowerState=STANDBY}
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processCommonConsumerAction(StoreIngestionTask.java:2112)
leaderFollowerStoreIngestionTask.consumerSubscribe(
TestTopic_797a883f18_2aa3c2a4_v1-1,
-1L,
"local_InMemoryKafkaBroker:41625"
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processCommonConsumerAction(StoreIngestionTask.java:2118)
leaderFollowerStoreIngestionTask.createKafkaConsumerProperties(
{"kafka.bootstrap.servers" = "local_InMemoryKafkaBroker:41625"},
"local_InMemoryKafkaBroker:41625",
false
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.consumerSubscribe(StoreIngestionTask.java:3497)
leaderFollowerStoreIngestionTask.getVersionTopic();
-> at com.linkedin.davinci.kafka.consumer.StorePartitionDataReceiver.destinationIdentifier(StorePartitionDataReceiver.java:86)
leaderFollowerStoreIngestionTask.getVersionTopic();
-> at com.linkedin.davinci.kafka.consumer.StorePartitionDataReceiver.destinationIdentifier(StorePartitionDataReceiver.java:86)
leaderFollowerStoreIngestionTask.checkLongRunningTaskState();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1596)
leaderFollowerStoreIngestionTask.getVersionTopic();
-> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.checkLongRunningTaskState(LeaderFollowerStoreIngestionTask.java:755)
leaderFollowerStoreIngestionTask.checkIngestionProgress(
Mock for Store, hashCode: 100090655
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1597)
leaderFollowerStoreIngestionTask.consumerHasAnySubscription();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.checkIngestionProgress(StoreIngestionTask.java:1464)
leaderFollowerStoreIngestionTask.maybeSendIngestionHeartbeat();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1598)
leaderFollowerStoreIngestionTask.isHybridMode();
-> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.maybeSendIngestionHeartbeat(LeaderFollowerStoreIngestionTask.java:3740)
leaderFollowerStoreIngestionTask.isRunning();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1592)
leaderFollowerStoreIngestionTask.getIsRunning();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.isRunning(StoreIngestionTask.java:3989)
leaderFollowerStoreIngestionTask.updateIngestionRoleIfStoreChanged(
Mock for Store, hashCode: 100090655
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1594)
leaderFollowerStoreIngestionTask.isHybridMode();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.updateIngestionRoleIfStoreChanged(StoreIngestionTask.java:1511)
leaderFollowerStoreIngestionTask.processConsumerActions(
Mock for Store, hashCode: 100090655
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1595)
leaderFollowerStoreIngestionTask.checkLongRunningTaskState();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1596)
leaderFollowerStoreIngestionTask.getVersionTopic();
-> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.checkLongRunningTaskState(LeaderFollowerStoreIngestionTask.java:755)
leaderFollowerStoreIngestionTask.checkIngestionProgress(
Mock for Store, hashCode: 100090655
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1597)
leaderFollowerStoreIngestionTask.consumerHasAnySubscription();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.checkIngestionProgress(StoreIngestionTask.java:1464)
leaderFollowerStoreIngestionTask.processConsumerActions(
Mock for Store, hashCode: 100090655
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testProcessConsumerActionsError$178(StoreIngestionTaskTest.java:4982)
leaderFollowerStoreIngestionTask.processConsumerActions(
Mock for Store, hashCode: 100090655
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testProcessConsumerActionsError$178(StoreIngestionTaskTest.java:4982)
leaderFollowerStoreIngestionTask.processConsumerActions(
Mock for Store, hashCode: 100090655
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testProcessConsumerActionsError$178(StoreIngestionTaskTest.java:4982)
leaderFollowerStoreIngestionTask.processConsumerActions(
Mock for Store, hashCode: 100090655
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testProcessConsumerActionsError$178(StoreIngestionTaskTest.java:4982)
leaderFollowerStoreIngestionTask.processConsumerActions(
Mock for Store, hashCode: 100090655
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testProcessConsumerActionsError$178(StoreIngestionTaskTest.java:4982)
|
SITWithTWiseWithoutBufferAfterLeaderTest.testResubscribeAfterRoleChange:
clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithTWiseWithoutBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.TooFewActualInvocations:
abstractStorageEngine.putWithReplicationMetadata(
1,
<any>,
<any>,
<any>
);
Wanted 200 times:
-> at com.linkedin.davinci.store.AbstractStorageEngine.putWithReplicationMetadata(AbstractStorageEngine.java:487)
But was 103 times:
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
-> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
|
NativeMetadataRepositoryTest.testNativeMetadataRepositoryStats:
clients/da-vinci-client/src/test/java/com/linkedin/davinci/repository/NativeMetadataRepositoryTest.java#L123
java.lang.AssertionError: expected [1000.0] but found [0.0]
|
BatchGetAvroStoreClientUnitTest.testSimpleStreamingBatchGet:
clients/venice-client/src/test/java/com/linkedin/venice/fastclient/BatchGetAvroStoreClientUnitTest.java#L65
org.testng.internal.thread.ThreadTimeoutException: Method com.linkedin.venice.fastclient.BatchGetAvroStoreClientUnitTest.testSimpleStreamingBatchGet() didn't finish within the time-out 5000
|
SITWithPWiseAndBufferAfterLeaderTest.testNotifier[1](AA_OFF):
clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseAndBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.ArgumentsAreDifferent: Argument(s) are different! Wanted:
storageMetadataService.put(
"TestTopic_741ebadc76_a26eff6d_v1",
1,
OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testNotifier$44(StoreIngestionTaskTest.java:1576)
Actual invocations have different arguments:
storageMetadataService.getLastOffset(
"TestTopic_741ebadc76_a26eff6d_v1",
1
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49)
storageMetadataService.getLastOffset(
"TestTopic_741ebadc76_a26eff6d_v1",
2
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49)
storageMetadataService.computeStoreVersionState(
"TestTopic_741ebadc76_a26eff6d_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$980/0x0000000800821040@6754ad51
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_741ebadc76_a26eff6d_v1",
1,
OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155537967, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.put(
"TestTopic_741ebadc76_a26eff6d_v1",
1,
OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155537967, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.computeStoreVersionState(
"TestTopic_741ebadc76_a26eff6d_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$980/0x0000000800821040@5a229b96
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_741ebadc76_a26eff6d_v1",
2,
OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155537967, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.put(
"TestTopic_741ebadc76_a26eff6d_v1",
2,
OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155537967, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.computeStoreVersionState(
"TestTopic_741ebadc76_a26eff6d_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$980/0x0000000800821040@6b221577
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_741ebadc76_a26eff6d_v1",
2,
OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155537967, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.computeStoreVersionState(
"TestTopic_741ebadc76_a26eff6d_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$980/0x0000000800821040@249aeb86
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_741ebadc76_a26eff6d_v1",
2,
OffsetRecord{localVersionTopicOffset=6, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155537967, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
|
SITWithPWiseAndBufferAfterLeaderTest.testNotifier[1](AA_OFF):
clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseAndBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.ArgumentsAreDifferent: Argument(s) are different! Wanted:
storageMetadataService.put(
"TestTopic_85002c61fc_a5691add_v1",
1,
OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testNotifier$44(StoreIngestionTaskTest.java:1576)
Actual invocations have different arguments:
storageMetadataService.getLastOffset(
"TestTopic_85002c61fc_a5691add_v1",
1
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49)
storageMetadataService.getLastOffset(
"TestTopic_85002c61fc_a5691add_v1",
2
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49)
storageMetadataService.computeStoreVersionState(
"TestTopic_85002c61fc_a5691add_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8040@3429676f
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.computeStoreVersionState(
"TestTopic_85002c61fc_a5691add_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8040@32bf0026
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_85002c61fc_a5691add_v1",
1,
OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155610483, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.put(
"TestTopic_85002c61fc_a5691add_v1",
2,
OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155610483, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.put(
"TestTopic_85002c61fc_a5691add_v1",
1,
OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155610484, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.put(
"TestTopic_85002c61fc_a5691add_v1",
2,
OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155610484, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.computeStoreVersionState(
"TestTopic_85002c61fc_a5691add_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8040@684f26c4
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.computeStoreVersionState(
"TestTopic_85002c61fc_a5691add_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8040@9ee6fd1
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
|
SITWithPWiseAndBufferAfterLeaderTest.testNotifier[1](AA_OFF):
clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseAndBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.ArgumentsAreDifferent: Argument(s) are different! Wanted:
storageMetadataService.put(
"TestTopic_8b6f1cde2d_2be9bdc6_v1",
2,
OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testNotifier$44(StoreIngestionTaskTest.java:1578)
Actual invocations have different arguments:
storageMetadataService.getLastOffset(
"TestTopic_8b6f1cde2d_2be9bdc6_v1",
1
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49)
storageMetadataService.getLastOffset(
"TestTopic_8b6f1cde2d_2be9bdc6_v1",
2
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49)
storageMetadataService.computeStoreVersionState(
"TestTopic_8b6f1cde2d_2be9bdc6_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$539/0x0000000800692840@4848ec67
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_8b6f1cde2d_2be9bdc6_v1",
1,
OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155638101, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.computeStoreVersionState(
"TestTopic_8b6f1cde2d_2be9bdc6_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$539/0x0000000800692840@1183e952
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_8b6f1cde2d_2be9bdc6_v1",
2,
OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155638101, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.put(
"TestTopic_8b6f1cde2d_2be9bdc6_v1",
1,
OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155638102, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.put(
"TestTopic_8b6f1cde2d_2be9bdc6_v1",
2,
OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155638102, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.computeStoreVersionState(
"TestTopic_8b6f1cde2d_2be9bdc6_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$539/0x0000000800692840@6fc42eff
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.computeStoreVersionState(
"TestTopic_8b6f1cde2d_2be9bdc6_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$539/0x0000000800692840@69525c03
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_8b6f1cde2d_2be9bdc6_v1",
1,
OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155638102, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
|
SITWithPWiseAndBufferAfterLeaderTest.testNotifier[1](AA_OFF):
clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseAndBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.ArgumentsAreDifferent: Argument(s) are different! Wanted:
storageMetadataService.put(
"TestTopic_91f45a3f4d_44d54fa9_v1",
2,
OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testNotifier$44(StoreIngestionTaskTest.java:1578)
Actual invocations have different arguments:
storageMetadataService.getLastOffset(
"TestTopic_91f45a3f4d_44d54fa9_v1",
1
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49)
storageMetadataService.getLastOffset(
"TestTopic_91f45a3f4d_44d54fa9_v1",
2
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49)
storageMetadataService.computeStoreVersionState(
"TestTopic_91f45a3f4d_44d54fa9_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8440@1183e952
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_91f45a3f4d_44d54fa9_v1",
1,
OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155666106, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.computeStoreVersionState(
"TestTopic_91f45a3f4d_44d54fa9_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8440@6fc42eff
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_91f45a3f4d_44d54fa9_v1",
1,
OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155666107, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.put(
"TestTopic_91f45a3f4d_44d54fa9_v1",
2,
OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155666106, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.computeStoreVersionState(
"TestTopic_91f45a3f4d_44d54fa9_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8440@69525c03
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_91f45a3f4d_44d54fa9_v1",
2,
OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155666107, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.put(
"TestTopic_91f45a3f4d_44d54fa9_v1",
1,
OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155666107, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.computeStoreVersionState(
"TestTopic_91f45a3f4d_44d54fa9_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8440@66f6ae9a
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
|
SITWithPWiseAndBufferAfterLeaderTest.testNotifier[1](AA_OFF):
clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseAndBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.ArgumentsAreDifferent: Argument(s) are different! Wanted:
storageMetadataService.put(
"TestTopic_9864f5f3d2_105ca834_v1",
2,
OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testNotifier$44(StoreIngestionTaskTest.java:1578)
Actual invocations have different arguments:
storageMetadataService.getLastOffset(
"TestTopic_9864f5f3d2_105ca834_v1",
1
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49)
storageMetadataService.getLastOffset(
"TestTopic_9864f5f3d2_105ca834_v1",
2
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49)
storageMetadataService.computeStoreVersionState(
"TestTopic_9864f5f3d2_105ca834_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$541/0x0000000800692c40@4848ec67
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_9864f5f3d2_105ca834_v1",
1,
OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155693766, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.computeStoreVersionState(
"TestTopic_9864f5f3d2_105ca834_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$541/0x0000000800692c40@1183e952
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_9864f5f3d2_105ca834_v1",
2,
OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155693766, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.put(
"TestTopic_9864f5f3d2_105ca834_v1",
1,
OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155693767, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.computeStoreVersionState(
"TestTopic_9864f5f3d2_105ca834_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$541/0x0000000800692c40@6fc42eff
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_9864f5f3d2_105ca834_v1",
2,
OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155693767, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.computeStoreVersionState(
"TestTopic_9864f5f3d2_105ca834_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$541/0x0000000800692c40@69525c03
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_9864f5f3d2_105ca834_v1",
1,
OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155693767, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
|
SITWithTWiseAndBufferAfterLeaderTest.testRecordLevelMetricForCurrentVersion[0](false):
clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithTWiseAndBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.TooFewActualInvocations:
hostLevelIngestionStats.recordTotalBytesConsumed(
<any long>
);
Wanted 2 times:
-> at com.linkedin.davinci.stats.HostLevelIngestionStats.recordTotalBytesConsumed(HostLevelIngestionStats.java:492)
But was 1 time:
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerRecord(StoreIngestionTask.java:2534)
|
NativeMetadataRepositoryTest.testNativeMetadataRepositoryStats:
clients/da-vinci-client/src/test/java/com/linkedin/davinci/repository/NativeMetadataRepositoryTest.java#L123
java.lang.AssertionError: expected [1000.0] but found [0.0]
|
BatchGetAvroStoreClientUnitTest.testSimpleStreamingBatchGet:
clients/venice-client/src/test/java/com/linkedin/venice/fastclient/BatchGetAvroStoreClientUnitTest.java#L65
org.testng.internal.thread.ThreadTimeoutException: Method com.linkedin.venice.fastclient.BatchGetAvroStoreClientUnitTest.testSimpleStreamingBatchGet() didn't finish within the time-out 5000
|
SITWithPWiseAndBufferAfterLeaderTest.testNotifier[1](AA_OFF):
clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseAndBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.ArgumentsAreDifferent: Argument(s) are different! Wanted:
storageMetadataService.put(
"TestTopic_741ebadc76_a26eff6d_v1",
1,
OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testNotifier$44(StoreIngestionTaskTest.java:1576)
Actual invocations have different arguments:
storageMetadataService.getLastOffset(
"TestTopic_741ebadc76_a26eff6d_v1",
1
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49)
storageMetadataService.getLastOffset(
"TestTopic_741ebadc76_a26eff6d_v1",
2
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49)
storageMetadataService.computeStoreVersionState(
"TestTopic_741ebadc76_a26eff6d_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$980/0x0000000800821040@6754ad51
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_741ebadc76_a26eff6d_v1",
1,
OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155537967, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.put(
"TestTopic_741ebadc76_a26eff6d_v1",
1,
OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155537967, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.computeStoreVersionState(
"TestTopic_741ebadc76_a26eff6d_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$980/0x0000000800821040@5a229b96
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_741ebadc76_a26eff6d_v1",
2,
OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155537967, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.put(
"TestTopic_741ebadc76_a26eff6d_v1",
2,
OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155537967, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.computeStoreVersionState(
"TestTopic_741ebadc76_a26eff6d_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$980/0x0000000800821040@6b221577
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_741ebadc76_a26eff6d_v1",
2,
OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155537967, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.computeStoreVersionState(
"TestTopic_741ebadc76_a26eff6d_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$980/0x0000000800821040@249aeb86
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_741ebadc76_a26eff6d_v1",
2,
OffsetRecord{localVersionTopicOffset=6, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155537967, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
|
SITWithPWiseAndBufferAfterLeaderTest.testNotifier[1](AA_OFF):
clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseAndBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.ArgumentsAreDifferent: Argument(s) are different! Wanted:
storageMetadataService.put(
"TestTopic_85002c61fc_a5691add_v1",
1,
OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testNotifier$44(StoreIngestionTaskTest.java:1576)
Actual invocations have different arguments:
storageMetadataService.getLastOffset(
"TestTopic_85002c61fc_a5691add_v1",
1
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49)
storageMetadataService.getLastOffset(
"TestTopic_85002c61fc_a5691add_v1",
2
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49)
storageMetadataService.computeStoreVersionState(
"TestTopic_85002c61fc_a5691add_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8040@3429676f
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.computeStoreVersionState(
"TestTopic_85002c61fc_a5691add_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8040@32bf0026
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_85002c61fc_a5691add_v1",
1,
OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155610483, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.put(
"TestTopic_85002c61fc_a5691add_v1",
2,
OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155610483, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.put(
"TestTopic_85002c61fc_a5691add_v1",
1,
OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155610484, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.put(
"TestTopic_85002c61fc_a5691add_v1",
2,
OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155610484, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.computeStoreVersionState(
"TestTopic_85002c61fc_a5691add_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8040@684f26c4
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.computeStoreVersionState(
"TestTopic_85002c61fc_a5691add_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8040@9ee6fd1
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
|
SITWithPWiseAndBufferAfterLeaderTest.testNotifier[1](AA_OFF):
clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseAndBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.ArgumentsAreDifferent: Argument(s) are different! Wanted:
storageMetadataService.put(
"TestTopic_8b6f1cde2d_2be9bdc6_v1",
2,
OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testNotifier$44(StoreIngestionTaskTest.java:1578)
Actual invocations have different arguments:
storageMetadataService.getLastOffset(
"TestTopic_8b6f1cde2d_2be9bdc6_v1",
1
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49)
storageMetadataService.getLastOffset(
"TestTopic_8b6f1cde2d_2be9bdc6_v1",
2
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49)
storageMetadataService.computeStoreVersionState(
"TestTopic_8b6f1cde2d_2be9bdc6_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$539/0x0000000800692840@4848ec67
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_8b6f1cde2d_2be9bdc6_v1",
1,
OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155638101, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.computeStoreVersionState(
"TestTopic_8b6f1cde2d_2be9bdc6_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$539/0x0000000800692840@1183e952
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_8b6f1cde2d_2be9bdc6_v1",
2,
OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155638101, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.put(
"TestTopic_8b6f1cde2d_2be9bdc6_v1",
1,
OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155638102, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.put(
"TestTopic_8b6f1cde2d_2be9bdc6_v1",
2,
OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155638102, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.computeStoreVersionState(
"TestTopic_8b6f1cde2d_2be9bdc6_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$539/0x0000000800692840@6fc42eff
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.computeStoreVersionState(
"TestTopic_8b6f1cde2d_2be9bdc6_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$539/0x0000000800692840@69525c03
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_8b6f1cde2d_2be9bdc6_v1",
1,
OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155638102, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
|
SITWithPWiseAndBufferAfterLeaderTest.testNotifier[1](AA_OFF):
clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseAndBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.ArgumentsAreDifferent: Argument(s) are different! Wanted:
storageMetadataService.put(
"TestTopic_91f45a3f4d_44d54fa9_v1",
2,
OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testNotifier$44(StoreIngestionTaskTest.java:1578)
Actual invocations have different arguments:
storageMetadataService.getLastOffset(
"TestTopic_91f45a3f4d_44d54fa9_v1",
1
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49)
storageMetadataService.getLastOffset(
"TestTopic_91f45a3f4d_44d54fa9_v1",
2
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49)
storageMetadataService.computeStoreVersionState(
"TestTopic_91f45a3f4d_44d54fa9_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8440@1183e952
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_91f45a3f4d_44d54fa9_v1",
1,
OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155666106, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.computeStoreVersionState(
"TestTopic_91f45a3f4d_44d54fa9_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8440@6fc42eff
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_91f45a3f4d_44d54fa9_v1",
1,
OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155666107, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.put(
"TestTopic_91f45a3f4d_44d54fa9_v1",
2,
OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155666106, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.computeStoreVersionState(
"TestTopic_91f45a3f4d_44d54fa9_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8440@69525c03
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_91f45a3f4d_44d54fa9_v1",
2,
OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155666107, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.put(
"TestTopic_91f45a3f4d_44d54fa9_v1",
1,
OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155666107, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.computeStoreVersionState(
"TestTopic_91f45a3f4d_44d54fa9_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8440@66f6ae9a
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
|
SITWithPWiseAndBufferAfterLeaderTest.testNotifier[1](AA_OFF):
clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseAndBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.ArgumentsAreDifferent: Argument(s) are different! Wanted:
storageMetadataService.put(
"TestTopic_9864f5f3d2_105ca834_v1",
2,
OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testNotifier$44(StoreIngestionTaskTest.java:1578)
Actual invocations have different arguments:
storageMetadataService.getLastOffset(
"TestTopic_9864f5f3d2_105ca834_v1",
1
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49)
storageMetadataService.getLastOffset(
"TestTopic_9864f5f3d2_105ca834_v1",
2
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49)
storageMetadataService.computeStoreVersionState(
"TestTopic_9864f5f3d2_105ca834_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$541/0x0000000800692c40@4848ec67
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_9864f5f3d2_105ca834_v1",
1,
OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155693766, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.computeStoreVersionState(
"TestTopic_9864f5f3d2_105ca834_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$541/0x0000000800692c40@1183e952
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_9864f5f3d2_105ca834_v1",
2,
OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155693766, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.put(
"TestTopic_9864f5f3d2_105ca834_v1",
1,
OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155693767, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.computeStoreVersionState(
"TestTopic_9864f5f3d2_105ca834_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$541/0x0000000800692c40@6fc42eff
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_9864f5f3d2_105ca834_v1",
2,
OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155693767, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.computeStoreVersionState(
"TestTopic_9864f5f3d2_105ca834_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$541/0x0000000800692c40@69525c03
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
storageMetadataService.put(
"TestTopic_9864f5f3d2_105ca834_v1",
1,
OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155693767, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
|
SITWithTWiseAndBufferAfterLeaderTest.testRecordLevelMetricForCurrentVersion[0](false):
clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithTWiseAndBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.TooFewActualInvocations:
hostLevelIngestionStats.recordTotalBytesConsumed(
<any long>
);
Wanted 2 times:
-> at com.linkedin.davinci.stats.HostLevelIngestionStats.recordTotalBytesConsumed(HostLevelIngestionStats.java:492)
But was 1 time:
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerRecord(StoreIngestionTask.java:2534)
|
NativeMetadataRepositoryTest.testNativeMetadataRepositoryStats:
clients/da-vinci-client/src/test/java/com/linkedin/davinci/repository/NativeMetadataRepositoryTest.java#L123
java.lang.AssertionError: expected [1000.0] but found [0.0]
|
BatchGetAvroStoreClientUnitTest.testSimpleStreamingBatchGet:
clients/venice-client/src/test/java/com/linkedin/venice/fastclient/BatchGetAvroStoreClientUnitTest.java#L65
org.testng.internal.thread.ThreadTimeoutException: Method com.linkedin.venice.fastclient.BatchGetAvroStoreClientUnitTest.testSimpleStreamingBatchGet() didn't finish within the time-out 5000
|
Clients / UT & CodeCov (11)
Process completed with exit code 1.
|
StaticAnalysisAndUnitTestsCompletionCheck
Process completed with exit code 1.
|
com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush ► testHybridStorePartitionCount:
integrations/venice-samza/src/main/java/com/linkedin/venice/samza/VeniceSystemProducer.java#L362
Failed test found in:
IntegrationTests_1050-artifacts/internal/venice-test-common/build/test-results/integrationTests_1050/TEST-com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush.xml
internal/venice-test-common/build/test-results/integrationTests_1050/TEST-com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush.xml
Error:
org.apache.samza.SamzaException: Failed to send request to Controller, error: An error occurred during controller request. controller = http://localhost:41815, route = /request_topic, params = [cluster_name=venice-cluster0, push_in_sorted_order=false, rewind_time_in_seconds_override=-1, is_write_compute_enabled=false, store_size=1, push_job_id=DC_1_store_a7b4dd6b44_9eb0a4f4, push_type=STREAM, defer_version_swap=false, hostname=localhost, repush_source_version=-1, store_name=store_a7b4dd6b44_9eb0a4f4, start_of_push=true, batch_job_heartbeat_enabled=false], timeout = 600000. Exception type: class com.linkedin.venice.exceptions.VeniceHttpException. Detailed message: Http Status 500 - Could not find an ONLINE hybrid store version for streaming writes with pushJobId: DC_1_store_a7b4dd6b44_9eb0a4f4 on store: store_a7b4dd6b44_9eb0a4f4 in cluster: venice-cluster0
|
com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush ► testHybridStorePartitionCount:
integrations/venice-samza/src/main/java/com/linkedin/venice/samza/VeniceSystemProducer.java#L362
Failed test found in:
IntegrationTests_1050-artifacts/internal/venice-test-common/build/test-results/integrationTests_1050/TEST-com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush.xml
internal/venice-test-common/build/test-results/integrationTests_1050/TEST-com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush.xml
Error:
org.apache.samza.SamzaException: Failed to send request to Controller, error: An error occurred during controller request. controller = http://localhost:41175, route = /request_topic, params = [cluster_name=venice-cluster0, push_in_sorted_order=false, rewind_time_in_seconds_override=-1, is_write_compute_enabled=false, store_size=1, push_job_id=DC_1_store_d4471ecfa0_b1345b9d, push_type=STREAM, defer_version_swap=false, hostname=localhost, repush_source_version=-1, store_name=store_d4471ecfa0_b1345b9d, start_of_push=true, batch_job_heartbeat_enabled=false], timeout = 600000. Exception type: class com.linkedin.venice.exceptions.VeniceHttpException. Detailed message: Http Status 500 - Could not find an ONLINE hybrid store version for streaming writes with pushJobId: DC_1_store_d4471ecfa0_b1345b9d on store: store_d4471ecfa0_b1345b9d in cluster: venice-cluster0
|
com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush ► testHybridStorePartitionCount:
integrations/venice-samza/src/main/java/com/linkedin/venice/samza/VeniceSystemProducer.java#L362
Failed test found in:
IntegrationTests_1050-artifacts/internal/venice-test-common/build/test-results/integrationTests_1050/TEST-com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush.xml
internal/venice-test-common/build/test-results/integrationTests_1050/TEST-com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush.xml
Error:
org.apache.samza.SamzaException: Failed to send request to Controller, error: An error occurred during controller request. controller = http://localhost:36367, route = /request_topic, params = [cluster_name=venice-cluster0, push_in_sorted_order=false, rewind_time_in_seconds_override=-1, is_write_compute_enabled=false, store_size=1, push_job_id=DC_1_store_ecd864f4cf_a4becf16, push_type=STREAM, defer_version_swap=false, hostname=localhost, repush_source_version=-1, store_name=store_ecd864f4cf_a4becf16, start_of_push=true, batch_job_heartbeat_enabled=false], timeout = 600000. Exception type: class com.linkedin.venice.exceptions.VeniceHttpException. Detailed message: Http Status 500 - Could not find an ONLINE hybrid store version for streaming writes with pushJobId: DC_1_store_ecd864f4cf_a4becf16 on store: store_ecd864f4cf_a4becf16 in cluster: venice-cluster0
|
com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush ► testHybridStorePartitionCount:
integrations/venice-samza/src/main/java/com/linkedin/venice/samza/VeniceSystemProducer.java#L362
Failed test found in:
IntegrationTests_1050-artifacts/internal/venice-test-common/build/test-results/integrationTests_1050/TEST-com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush.xml
internal/venice-test-common/build/test-results/integrationTests_1050/TEST-com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush.xml
Error:
org.apache.samza.SamzaException: Failed to send request to Controller, error: An error occurred during controller request. controller = http://localhost:33617, route = /request_topic, params = [cluster_name=venice-cluster0, push_in_sorted_order=false, rewind_time_in_seconds_override=-1, is_write_compute_enabled=false, store_size=1, push_job_id=DC_1_store_1069edc26ff_c941d0b2, push_type=STREAM, defer_version_swap=false, hostname=localhost, repush_source_version=-1, store_name=store_1069edc26ff_c941d0b2, start_of_push=true, batch_job_heartbeat_enabled=false], timeout = 600000. Exception type: class com.linkedin.venice.exceptions.VeniceHttpException. Detailed message: Http Status 500 - Could not find an ONLINE hybrid store version for streaming writes with pushJobId: DC_1_store_1069edc26ff_c941d0b2 on store: store_1069edc26ff_c941d0b2 in cluster: venice-cluster0
|
com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush ► testHybridStorePartitionCount:
integrations/venice-samza/src/main/java/com/linkedin/venice/samza/VeniceSystemProducer.java#L362
Failed test found in:
IntegrationTests_1050-artifacts/internal/venice-test-common/build/test-results/integrationTests_1050/TEST-com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush.xml
internal/venice-test-common/build/test-results/integrationTests_1050/TEST-com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush.xml
Error:
org.apache.samza.SamzaException: Failed to send request to Controller, error: An error occurred during controller request. controller = http://localhost:36147, route = /request_topic, params = [cluster_name=venice-cluster0, push_in_sorted_order=false, rewind_time_in_seconds_override=-1, is_write_compute_enabled=false, store_size=1, push_job_id=DC_1_store_12140bc2fee_6ef70689, push_type=STREAM, defer_version_swap=false, hostname=localhost, repush_source_version=-1, store_name=store_12140bc2fee_6ef70689, start_of_push=true, batch_job_heartbeat_enabled=false], timeout = 600000. Exception type: class com.linkedin.venice.exceptions.VeniceHttpException. Detailed message: Http Status 500 - Could not find an ONLINE hybrid store version for streaming writes with pushJobId: DC_1_store_12140bc2fee_6ef70689 on store: store_12140bc2fee_6ef70689 in cluster: venice-cluster0
|
com.linkedin.venice.endToEnd.TestBatchForRocksDB ► testBatchJobSnapshots[2](true):
internal/venice-test-common/src/integrationTest/java/com/linkedin/venice/endToEnd/TestBatch.java#L1428
Failed test found in:
IntegrationTests_1230-artifacts/internal/venice-test-common/build/test-results/integrationTests_1230/TEST-com.linkedin.venice.endToEnd.TestBatchForRocksDB.xml
internal/venice-test-common/build/test-results/integrationTests_1230/TEST-com.linkedin.venice.endToEnd.TestBatchForRocksDB.xml
Error:
java.util.concurrent.ExecutionException: com.linkedin.venice.client.exceptions.VeniceClientHttpException: http status: 503, com.linkedin.alpini.router.api.RouterException: store_348547b201_a1f39293_v1, partition 2 is not available to serve request of type: SINGLE_GET
|
com.linkedin.venice.endToEnd.BlobP2PTransferAmongServersTest ► testBlobP2PTransferAmongServersForBatchStore:
internal/venice-test-common/src/integrationTest/java/com/linkedin/venice/endToEnd/BlobP2PTransferAmongServersTest.java#L123
Failed test found in:
IntegrationTests_1130-artifacts/internal/venice-test-common/build/test-results/integrationTests_1130/TEST-com.linkedin.venice.endToEnd.BlobP2PTransferAmongServersTest.xml
internal/venice-test-common/build/test-results/integrationTests_1130/TEST-com.linkedin.venice.endToEnd.BlobP2PTransferAmongServersTest.xml
Error:
java.lang.AssertionError: expected [true] but found [false]
|
com.linkedin.venice.endToEnd.BlobP2PTransferAmongServersTest ► testBlobP2PTransferAmongServersForHybridStore:
integrations/venice-samza/src/main/java/com/linkedin/venice/samza/VeniceSystemProducer.java#L362
Failed test found in:
IntegrationTests_1130-artifacts/internal/venice-test-common/build/test-results/integrationTests_1130/TEST-com.linkedin.venice.endToEnd.BlobP2PTransferAmongServersTest.xml
internal/venice-test-common/build/test-results/integrationTests_1130/TEST-com.linkedin.venice.endToEnd.BlobP2PTransferAmongServersTest.xml
Error:
org.apache.samza.SamzaException: Failed to send request to Controller, error: An error occurred during controller request. controller = http://localhost:41265, route = /request_topic, params = [cluster_name=venice-cluster_50b960d02c_1e278e46, push_in_sorted_order=false, rewind_time_in_seconds_override=-1, is_write_compute_enabled=false, store_size=1, push_job_id=venice-push-id_544dd68afb_8bca5409, push_type=STREAM, defer_version_swap=false, hostname=localhost, repush_source_version=-1, store_name=test-store-hybrid, start_of_push=true, batch_job_heartbeat_enabled=false], timeout = 600000. Exception type: class com.linkedin.venice.exceptions.VeniceHttpException. Detailed message: Http Status 500 - Could not find an ONLINE hybrid store version for streaming writes with pushJobId: venice-push-id_544dd68afb_8bca5409 on store: test-store-hybrid in cluster: venice-cluster_50b960d02c_1e278e46
|
com.linkedin.venice.endToEnd.MetaSystemStoreTest ► testThinClientMetaStoreBasedRepositoryWithLargeValueSchemas:
internal/venice-test-common/src/integrationTest/java/com/linkedin/venice/endToEnd/MetaSystemStoreTest.java#L315
Failed test found in:
IntegrationTests_1500-artifacts/internal/venice-test-common/build/test-results/integrationTests_1500/TEST-com.linkedin.venice.endToEnd.MetaSystemStoreTest.xml
internal/venice-test-common/build/test-results/integrationTests_1500/TEST-com.linkedin.venice.endToEnd.MetaSystemStoreTest.xml
Error:
java.lang.AssertionError: Number of value schemas should be the same between meta system store and controller expected [2] but found [1]
|
com.linkedin.venice.endToEnd.TestHybrid ► testLeaderCanReleaseLatch:
integrations/venice-samza/src/main/java/com/linkedin/venice/samza/VeniceSystemProducer.java#L362
Failed test found in:
IntegrationTests_1400-artifacts/internal/venice-test-common/build/test-results/integrationTests_1400/TEST-com.linkedin.venice.endToEnd.TestHybrid.xml
internal/venice-test-common/build/test-results/integrationTests_1400/TEST-com.linkedin.venice.endToEnd.TestHybrid.xml
Error:
org.apache.samza.SamzaException: Failed to send request to Controller, error: An error occurred during controller request. controller = http://localhost:33903, route = /request_topic, params = [cluster_name=venice-cluster_3e14cc0c94_8858fc09, push_in_sorted_order=false, rewind_time_in_seconds_override=-1, is_write_compute_enabled=false, store_size=1, push_job_id=venice-push-id_5b76066303_506309b0, push_type=STREAM, defer_version_swap=false, hostname=localhost, repush_source_version=-1, store_name=test-store_5b14fe8ce2_681f93b6, start_of_push=true, batch_job_heartbeat_enabled=false], timeout = 600000. Exception type: class com.linkedin.venice.exceptions.VeniceHttpException. Detailed message: Http Status 500 - Could not find an ONLINE hybrid store version for streaming writes with pushJobId: venice-push-id_5b76066303_506309b0 on store: test-store_5b14fe8ce2_681f93b6 in cluster: venice-cluster_3e14cc0c94_8858fc09
|
com.linkedin.venice.controller.server.TestAdminSparkServer ► controllerClientCanGetStoreInfo:
internal/venice-test-common/src/integrationTest/java/com/linkedin/venice/controller/server/TestAdminSparkServer.java#L462
Failed test found in:
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkServer.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkServer.xml
Error:
java.lang.AssertionError: Store from new store-version should only have one version expected [1] but found [0]
|
com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers ► requestTopicIsIdempotent:
internal/venice-test-common/src/integrationTest/java/com/linkedin/venice/controller/server/TestAdminSparkServerWithMultiServers.java#L275
Failed test found in:
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers.xml
Error:
java.lang.AssertionError: An error occurred during controller request. controller = http://localhost:36255, route = /request_topic, params = [cluster_name=venice-cluster_8adec8af64_b23c05ec, push_in_sorted_order=true, rewind_time_in_seconds_override=-1, is_write_compute_enabled=false, store_size=1, push_job_id=pushId_900aa200f7_57476ca9, push_type=STREAM, defer_version_swap=false, hostname=localhost, repush_source_version=-1, store_name=StreamStore_900aa1e8f0_82f34c97, start_of_push=true, batch_job_heartbeat_enabled=false], timeout = 600000. Exception type: class com.linkedin.venice.exceptions.VeniceHttpException. Detailed message: Http Status 500 - Could not find an ONLINE hybrid store version for streaming writes with pushJobId: pushId_900aa200f7_57476ca9 on store: StreamStore_900aa1e8f0_82f34c97 in cluster: venice-cluster_8adec8af64_b23c05ec
|
com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers ► requestTopicIsIdempotent:
internal/venice-test-common/src/integrationTest/java/com/linkedin/venice/controller/server/TestAdminSparkServerWithMultiServers.java#L275
Failed test found in:
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers.xml
Error:
java.lang.AssertionError: An error occurred during controller request. controller = http://localhost:39995, route = /request_topic, params = [cluster_name=venice-cluster_a73f935bae_7da89c5a, push_in_sorted_order=true, rewind_time_in_seconds_override=-1, is_write_compute_enabled=false, store_size=1, push_job_id=pushId_abd1f4d48f_7c6792ff, push_type=STREAM, defer_version_swap=false, hostname=localhost, repush_source_version=-1, store_name=StreamStore_abd1f4bf31_2392aec2, start_of_push=true, batch_job_heartbeat_enabled=false], timeout = 600000. Exception type: class com.linkedin.venice.exceptions.VeniceHttpException. Detailed message: Http Status 500 - Could not find an ONLINE hybrid store version for streaming writes with pushJobId: pushId_abd1f4d48f_7c6792ff on store: StreamStore_abd1f4bf31_2392aec2 in cluster: venice-cluster_a73f935bae_7da89c5a
|
com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers ► requestTopicIsIdempotent:
internal/venice-test-common/src/integrationTest/java/com/linkedin/venice/controller/server/TestAdminSparkServerWithMultiServers.java#L275
Failed test found in:
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers.xml
Error:
java.lang.AssertionError: An error occurred during controller request. controller = http://localhost:45849, route = /request_topic, params = [cluster_name=venice-cluster_b2fd564a39_4b27bcc9, push_in_sorted_order=true, rewind_time_in_seconds_override=-1, is_write_compute_enabled=false, store_size=1, push_job_id=pushId_b78d2f647b_90fbcd2e, push_type=STREAM, defer_version_swap=false, hostname=localhost, repush_source_version=-1, store_name=StreamStore_b78d2f411a_6207ae1b, start_of_push=true, batch_job_heartbeat_enabled=false], timeout = 600000. Exception type: class com.linkedin.venice.exceptions.VeniceHttpException. Detailed message: Http Status 500 - Could not find an ONLINE hybrid store version for streaming writes with pushJobId: pushId_b78d2f647b_90fbcd2e on store: StreamStore_b78d2f411a_6207ae1b in cluster: venice-cluster_b2fd564a39_4b27bcc9
|
com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers ► requestTopicIsIdempotent:
internal/venice-test-common/src/integrationTest/java/com/linkedin/venice/controller/server/TestAdminSparkServerWithMultiServers.java#L275
Failed test found in:
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers.xml
Error:
java.lang.AssertionError: An error occurred during controller request. controller = http://localhost:34409, route = /request_topic, params = [cluster_name=venice-cluster_beca85400b_c645f9c3, push_in_sorted_order=true, rewind_time_in_seconds_override=-1, is_write_compute_enabled=false, store_size=1, push_job_id=pushId_c5adc61453_5d23ea01, push_type=STREAM, defer_version_swap=false, hostname=localhost, repush_source_version=-1, store_name=StreamStore_c5adc5f57d_4d794319, start_of_push=true, batch_job_heartbeat_enabled=false], timeout = 600000. Exception type: class com.linkedin.venice.exceptions.VeniceHttpException. Detailed message: Http Status 500 - Could not find an ONLINE hybrid store version for streaming writes with pushJobId: pushId_c5adc61453_5d23ea01 on store: StreamStore_c5adc5f57d_4d794319 in cluster: venice-cluster_beca85400b_c645f9c3
|
com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers ► requestTopicIsIdempotent:
internal/venice-test-common/src/integrationTest/java/com/linkedin/venice/controller/server/TestAdminSparkServerWithMultiServers.java#L275
Failed test found in:
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers.xml
Error:
java.lang.AssertionError: An error occurred during controller request. controller = http://localhost:39097, route = /request_topic, params = [cluster_name=venice-cluster_cd0f1b525a_e81b93c9, push_in_sorted_order=true, rewind_time_in_seconds_override=-1, is_write_compute_enabled=false, store_size=1, push_job_id=pushId_d1aca92d8b_696ca337, push_type=STREAM, defer_version_swap=false, hostname=localhost, repush_source_version=-1, store_name=StreamStore_d1aca90a0d_831f1407, start_of_push=true, batch_job_heartbeat_enabled=false], timeout = 600000. Exception type: class com.linkedin.venice.exceptions.VeniceHttpException. Detailed message: Http Status 500 - Could not find an ONLINE hybrid store version for streaming writes with pushJobId: pushId_d1aca92d8b_696ca337 on store: StreamStore_d1aca90a0d_831f1407 in cluster: venice-cluster_cd0f1b525a_e81b93c9
|
com.linkedin.venice.controller.server.TestAdminSparkWithMocks ► testGetRealTimeTopicUsesAdmin:
internal/venice-test-common/src/integrationTest/java/com/linkedin/venice/controller/server/TestAdminSparkWithMocks.java#L109
Failed test found in:
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
Error:
java.lang.AssertionError: unexpected error: No hybrid version found for store: store. Create a hybrid version before starting a stream push. expected [false] but found [true]
|
com.linkedin.venice.controller.server.TestAdminSparkWithMocks ► testSamzaReplicationPolicyMode[0](false, false, false):
internal/venice-test-common/src/integrationTest/java/com/linkedin/venice/controller/server/TestAdminSparkWithMocks.java#L370
Failed test found in:
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml
Error:
java.lang.AssertionError: unexpected error: No hybrid version found for store: store. Create a hybrid version before starting a stream push. expected [false] but found [true]
|
com.linkedin.venice.endToEnd.DaVinciClientTest ► testTypeChangeRecordTransformer[0](DaVinciConfig{managed=true, isolated=false, storageClass=MEMORY_BACKED_BY_DISK, cacheConfig=null, largeBatchRequestSplitThreshold=100}):
clients/venice-push-job/src/main/java/com/linkedin/venice/hadoop/DefaultInputDataInfoProvider.java#L342
Failed test found in:
IntegrationTests_1010-artifacts/internal/venice-test-common/build/test-results/integrationTests_1010/TEST-com.linkedin.venice.endToEnd.DaVinciClientTest.xml
IntegrationTests_1010-artifacts/internal/venice-test-common/build/test-results/integrationTests_1010/TEST-com.linkedin.venice.endToEnd.DaVinciClientTest.xml
IntegrationTests_1010-artifacts/internal/venice-test-common/build/test-results/integrationTests_1010/TEST-com.linkedin.venice.endToEnd.DaVinciClientTest.xml
IntegrationTests_1010-artifacts/internal/venice-test-common/build/test-results/integrationTests_1010/TEST-com.linkedin.venice.endToEnd.DaVinciClientTest.xml
IntegrationTests_1010-artifacts/internal/venice-test-common/build/test-results/integrationTests_1010/TEST-com.linkedin.venice.endToEnd.DaVinciClientTest.xml
IntegrationTests_1010-artifacts/internal/venice-test-common/build/test-results/integrationTests_1010/TEST-com.linkedin.venice.endToEnd.DaVinciClientTest.xml
internal/venice-test-common/build/test-results/integrationTests_1010/TEST-com.linkedin.venice.endToEnd.DaVinciClientTest.xml
internal/venice-test-common/build/test-results/integrationTests_1010/TEST-com.linkedin.venice.endToEnd.DaVinciClientTest.xml
internal/venice-test-common/build/test-results/integrationTests_1010/TEST-com.linkedin.venice.endToEnd.DaVinciClientTest.xml
internal/venice-test-common/build/test-results/integrationTests_1010/TEST-com.linkedin.venice.endToEnd.DaVinciClientTest.xml
internal/venice-test-common/build/test-results/integrationTests_1010/TEST-com.linkedin.venice.endToEnd.DaVinciClientTest.xml
internal/venice-test-common/build/test-results/integrationTests_1010/TEST-com.linkedin.venice.endToEnd.DaVinciClientTest.xml
Error:
com.linkedin.venice.hadoop.exceptions.VeniceInconsistentSchemaException: Inconsistent file Avro schema found. File: int2string.avro.
|
com.linkedin.venice.controller.TestParentControllerWithMultiDataCenter ► testDeferredVersionSwap:
internal/venice-test-common/src/integrationTest/java/com/linkedin/venice/utils/IntegrationTestPushUtils.java#L313
Failed test found in:
IntegrationTests_1430-artifacts/internal/venice-test-common/build/test-results/integrationTests_1430/TEST-com.linkedin.venice.controller.TestParentControllerWithMultiDataCenter.xml
internal/venice-test-common/build/test-results/integrationTests_1430/TEST-com.linkedin.venice.controller.TestParentControllerWithMultiDataCenter.xml
Error:
com.linkedin.venice.exceptions.VeniceException: Could not create store store_f17625870a_199f179a
|
com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment ► setUp:
internal/venice-test-common/src/main/java/com/linkedin/venice/utils/TestUtils.java#L245
Failed test found in:
IntegrationTests_1430-artifacts/internal/venice-test-common/build/test-results/integrationTests_1430/TEST-com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment.xml
internal/venice-test-common/build/test-results/integrationTests_1430/TEST-com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment.xml
Error:
java.lang.AssertionError: com.linkedin.venice.pubsub.api.exceptions.PubSubOpTimeoutException: Timeout while creating topic: venice_system_store_participant_store_cluster_test-cluster_134a2c9f9a1_407899d0_rt. Topic still does not exist after 30000ms.
|
com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment ► setUp:
internal/venice-test-common/src/main/java/com/linkedin/venice/utils/TestUtils.java#L245
Failed test found in:
IntegrationTests_1430-artifacts/internal/venice-test-common/build/test-results/integrationTests_1430/TEST-com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment.xml
internal/venice-test-common/build/test-results/integrationTests_1430/TEST-com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment.xml
Error:
java.lang.AssertionError: com.linkedin.venice.pubsub.api.exceptions.PubSubOpTimeoutException: Timeout while creating topic: venice_system_store_participant_store_cluster_test-cluster_154540d1b92_4b0d3c8f_rt. Topic still does not exist after 30000ms.
|
com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment ► setUp:
internal/venice-test-common/src/main/java/com/linkedin/venice/utils/TestUtils.java#L245
Failed test found in:
IntegrationTests_1430-artifacts/internal/venice-test-common/build/test-results/integrationTests_1430/TEST-com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment.xml
internal/venice-test-common/build/test-results/integrationTests_1430/TEST-com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment.xml
Error:
java.lang.AssertionError: com.linkedin.venice.pubsub.api.exceptions.PubSubOpTimeoutException: Timeout while creating topic: venice_system_store_participant_store_cluster_test-cluster_1612ea1a815_d4a9153f_rt. Topic still does not exist after 30000ms.
|
com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment ► setUp:
internal/venice-test-common/src/main/java/com/linkedin/venice/utils/TestUtils.java#L245
Failed test found in:
IntegrationTests_1430-artifacts/internal/venice-test-common/build/test-results/integrationTests_1430/TEST-com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment.xml
internal/venice-test-common/build/test-results/integrationTests_1430/TEST-com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment.xml
Error:
java.lang.AssertionError: com.linkedin.venice.pubsub.api.exceptions.PubSubOpTimeoutException: Timeout while creating topic: venice_system_store_participant_store_cluster_test-cluster_16e226e0dc6_41e11ec_rt. Topic still does not exist after 30000ms.
|
com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment ► setUp:
internal/venice-test-common/src/main/java/com/linkedin/venice/utils/TestUtils.java#L245
Failed test found in:
IntegrationTests_1430-artifacts/internal/venice-test-common/build/test-results/integrationTests_1430/TEST-com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment.xml
internal/venice-test-common/build/test-results/integrationTests_1430/TEST-com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment.xml
Error:
java.lang.AssertionError: com.linkedin.venice.pubsub.api.exceptions.PubSubOpTimeoutException: Timeout while creating topic: venice_system_store_participant_store_cluster_test-cluster_17ac2a5b473_aae542a5_rt. Topic still does not exist after 30000ms.
|
StaticAnalysis (17)
Retention days cannot be greater than the maximum allowed retention set within the repository. Using 10 instead.
|
Internal / UT & CodeCov (8)
The following actions use a deprecated Node.js version and will be forced to run on node20: mikepenz/action-junit-report@v3. For more info: https://github.blog/changelog/2024-03-07-github-actions-all-actions-will-run-on-node20-instead-of-node16-by-default/
|
Internal / UT & CodeCov (8)
Retention days cannot be greater than the maximum allowed retention set within the repository. Using 10 instead.
|
Internal / UT & CodeCov (17)
The following actions use a deprecated Node.js version and will be forced to run on node20: mikepenz/action-junit-report@v3. For more info: https://github.blog/changelog/2024-03-07-github-actions-all-actions-will-run-on-node20-instead-of-node16-by-default/
|
Internal / UT & CodeCov (17)
Retention days cannot be greater than the maximum allowed retention set within the repository. Using 10 instead.
|
Internal / UT & CodeCov (11)
The following actions use a deprecated Node.js version and will be forced to run on node20: mikepenz/action-junit-report@v3. For more info: https://github.blog/changelog/2024-03-07-github-actions-all-actions-will-run-on-node20-instead-of-node16-by-default/
|
Internal / UT & CodeCov (11)
Retention days cannot be greater than the maximum allowed retention set within the repository. Using 10 instead.
|
Services / UT & CodeCov (17)
The following actions use a deprecated Node.js version and will be forced to run on node20: mikepenz/action-junit-report@v3. For more info: https://github.blog/changelog/2024-03-07-github-actions-all-actions-will-run-on-node20-instead-of-node16-by-default/
|
Services / UT & CodeCov (17)
Retention days cannot be greater than the maximum allowed retention set within the repository. Using 10 instead.
|
Services / UT & CodeCov (8)
The following actions use a deprecated Node.js version and will be forced to run on node20: mikepenz/action-junit-report@v3. For more info: https://github.blog/changelog/2024-03-07-github-actions-all-actions-will-run-on-node20-instead-of-node16-by-default/
|
Services / UT & CodeCov (8)
Retention days cannot be greater than the maximum allowed retention set within the repository. Using 10 instead.
|
Services / UT & CodeCov (11)
The following actions use a deprecated Node.js version and will be forced to run on node20: mikepenz/action-junit-report@v3. For more info: https://github.blog/changelog/2024-03-07-github-actions-all-actions-will-run-on-node20-instead-of-node16-by-default/
|
Services / UT & CodeCov (11)
Retention days cannot be greater than the maximum allowed retention set within the repository. Using 10 instead.
|
Clients / UT & CodeCov (8)
The following actions use a deprecated Node.js version and will be forced to run on node20: mikepenz/action-junit-report@v3. For more info: https://github.blog/changelog/2024-03-07-github-actions-all-actions-will-run-on-node20-instead-of-node16-by-default/
|
Clients / UT & CodeCov (8)
Retention days cannot be greater than the maximum allowed retention set within the repository. Using 10 instead.
|
Clients / UT & CodeCov (17)
The following actions use a deprecated Node.js version and will be forced to run on node20: mikepenz/action-junit-report@v3. For more info: https://github.blog/changelog/2024-03-07-github-actions-all-actions-will-run-on-node20-instead-of-node16-by-default/
|
Clients / UT & CodeCov (17)
Retention days cannot be greater than the maximum allowed retention set within the repository. Using 10 instead.
|
Clients / UT & CodeCov (11)
The following actions use a deprecated Node.js version and will be forced to run on node20: mikepenz/action-junit-report@v3. For more info: https://github.blog/changelog/2024-03-07-github-actions-all-actions-will-run-on-node20-instead-of-node16-by-default/
|
Clients / UT & CodeCov (11)
Retention days cannot be greater than the maximum allowed retention set within the repository. Using 10 instead.
|
Artifacts
Produced during runtime
Name | Size | |
---|---|---|
StaticAnalysis
Expired
|
694 KB |
|
clients-jdk11
Expired
|
10.5 MB |
|
clients-jdk17
Expired
|
11.7 MB |
|
clients-jdk8
Expired
|
11.7 MB |
|
internal-jdk11
Expired
|
3.01 MB |
|
internal-jdk17
Expired
|
3.03 MB |
|
internal-jdk8
Expired
|
3 MB |
|
services-jdk11
Expired
|
2.42 MB |
|
services-jdk17
Expired
|
2.43 MB |
|
services-jdk8
Expired
|
2.39 MB |
|