Skip to content

add more ut

add more ut #352

Triggered via push November 21, 2024 02:12
Status Failure
Total duration 20m 58s
Artifacts 10
Matrix: Clients / UT & CodeCov
Matrix: Internal / UT & CodeCov
Matrix: Services / UT & CodeCov
Matrix: StaticAnalysis
ValidateGradleWrapper
8s
ValidateGradleWrapper
StaticAnalysisAndUnitTestsCompletionCheck
0s
StaticAnalysisAndUnitTestsCompletionCheck
Fit to window
Zoom out
Zoom in

Annotations

75 errors and 19 warnings
StaticAnalysis (17)
Process completed with exit code 1.
Internal / UT & CodeCov (8)
Process completed with exit code 1.
Internal / UT & CodeCov (17)
Process completed with exit code 1.
Internal / UT & CodeCov (11)
Process completed with exit code 1.
com.linkedin.venice.helix.TestControllerKMERegistrationFromMessageHeader ► testKMERegistrationThroughAdminTopicChannel: internal/venice-test-common/src/main/java/com/linkedin/venice/utils/TestUtils.java#L245
Failed test found in: IntegrationTests_1060-artifacts/internal/venice-test-common/build/test-results/integrationTests_1060/TEST-com.linkedin.venice.helix.TestControllerKMERegistrationFromMessageHeader.xml internal/venice-test-common/build/test-results/integrationTests_1060/TEST-com.linkedin.venice.helix.TestControllerKMERegistrationFromMessageHeader.xml Error: java.lang.AssertionError: com.linkedin.venice.exceptions.VeniceNoStoreException: Store: venice_system_store_KAFKA_MESSAGE_ENVELOPE does not exist.
com.linkedin.venice.helixrebalance.TestRebalanceByDefaultStrategy ► testRollingUpgrade: internal/venice-test-common/src/main/java/com/linkedin/venice/utils/TestUtils.java#L168
Failed test found in: IntegrationTests_1060-artifacts/internal/venice-test-common/build/test-results/integrationTests_1060/TEST-com.linkedin.venice.helixrebalance.TestRebalanceByDefaultStrategy.xml internal/venice-test-common/build/test-results/integrationTests_1060/TEST-com.linkedin.venice.helixrebalance.TestRebalanceByDefaultStrategy.xml Error: java.lang.AssertionError: Non-deterministic condition not met. expected [true] but found [false]
BatchGetAvroStoreClientUnitTest.testSimpleStreamingBatchGet: clients/venice-client/src/test/java/com/linkedin/venice/fastclient/BatchGetAvroStoreClientUnitTest.java#L1
org.testng.internal.thread.ThreadTimeoutException: Method com.linkedin.venice.fastclient.BatchGetAvroStoreClientUnitTest.testSimpleStreamingBatchGet() didn't finish within the time-out 5000
BatchGetAvroStoreClientUnitTest.testSimpleStreamingBatchGet: clients/venice-client/src/test/java/com/linkedin/venice/fastclient/BatchGetAvroStoreClientUnitTest.java#L1
org.testng.internal.thread.ThreadTimeoutException: Method com.linkedin.venice.fastclient.BatchGetAvroStoreClientUnitTest.testSimpleStreamingBatchGet() didn't finish within the time-out 5000
SITWithPWiseWithoutBufferAfterLeaderTest.testProcessConsumerActionsError: clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseWithoutBufferAfterLeaderTest.java#L1
Wanted but not invoked: leaderFollowerStoreIngestionTask.reportError( <any string>, 1, <Capturing argument> ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.reportError(StoreIngestionTask.java:4146) However, there were exactly 40 interactions with this mock: leaderFollowerStoreIngestionTask.subscribePartition( TestTopic_797a883f18_2aa3c2a4_v1-1 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.runTest(StoreIngestionTaskTest.java:854) leaderFollowerStoreIngestionTask.subscribePartition( TestTopic_797a883f18_2aa3c2a4_v1-1, true ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.subscribePartition(StoreIngestionTask.java:593) leaderFollowerStoreIngestionTask.throwIfNotRunning(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.subscribePartition(StoreIngestionTask.java:607) leaderFollowerStoreIngestionTask.isRunning(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.throwIfNotRunning(StoreIngestionTask.java:554) leaderFollowerStoreIngestionTask.getIsRunning(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.isRunning(StoreIngestionTask.java:3989) leaderFollowerStoreIngestionTask.nextSeqNum(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.subscribePartition(StoreIngestionTask.java:616) leaderFollowerStoreIngestionTask.run(); -> at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) leaderFollowerStoreIngestionTask.isRunning(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1592) leaderFollowerStoreIngestionTask.getIsRunning(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.isRunning(StoreIngestionTask.java:3989) leaderFollowerStoreIngestionTask.updateIngestionRoleIfStoreChanged( Mock for Store, hashCode: 100090655 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1594) leaderFollowerStoreIngestionTask.isHybridMode(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.updateIngestionRoleIfStoreChanged(StoreIngestionTask.java:1511) leaderFollowerStoreIngestionTask.processConsumerActions( Mock for Store, hashCode: 100090655 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1595) leaderFollowerStoreIngestionTask.processConsumerAction( KafkaTaskMessage{type=SUBSCRIBE, topicPartition=TestTopic_797a883f18_2aa3c2a4_v1-1, attempts=1, sequenceNumber=1, createdTimestampInMs=1732155484357}, Mock for Store, hashCode: 100090655 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerActions(StoreIngestionTask.java:1864) leaderFollowerStoreIngestionTask.processCommonConsumerAction( KafkaTaskMessage{type=SUBSCRIBE, topicPartition=TestTopic_797a883f18_2aa3c2a4_v1-1, attempts=1, sequenceNumber=1, createdTimestampInMs=1732155484357} ); -> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.processConsumerAction(LeaderFollowerStoreIngestionTask.java:556) leaderFollowerStoreIngestionTask.reportIfCatchUpVersionTopicOffset( PCS{replicaId=TestTopic_797a883f18_2aa3c2a4_v1-1, hybrid=false, latestProcessedLocalVersionTopicOffset=-1, latestProcessedUpstreamVersionTopicOffset=-1, latestProcessedUpstreamRTOffsetMap={}, latestIgnoredUpstreamRTOffsetMap={}, latestRTOffsetTriedToProduceToVTMap{}, offsetRecord=OffsetRecord{localVersionTopicOffset=-1, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}, errorReported=false, started=false, lagCaughtUp=false, processedRecordSizeSinceLastSync=0, leaderFollowerState=STANDBY} ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processCommonConsumerAction(StoreIngestionTask.java:2107) leaderFollowerStoreIngestionTask.updateLeaderTopicOnFollower( PCS{replicaId=TestTopic_797a883f18_2aa3c2a4_v1-1, hybrid=false, latestProcessedLocalVersionTopicOffset=-1, latestProcessedUpstreamVersionTopicOffset=-1, latestProcessedUpstreamRTOffsetMap={}, latestIgnoredUpstreamRTOffsetMap={}, latestRTOffsetTriedToProduceToVTMap{}, offsetRecord=OffsetRecord{localVersionTopicOffset=-1, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}, errorReported=false, started=false, lagCaughtUp=false, processedRecordSizeSinceLastSync=0, leaderFollowerState=STANDBY} ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processCommonConsumerAction(StoreIngestionTask.java:2112) leaderFollowerStoreIngestionTask.consumerSubscribe( TestTopic_797a883f18_2aa3c2a4_v1-1, -1L, "local_InMemoryKafkaBroker:41625" ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processCommonConsumerAction(StoreIngestionTask.java:2118) leaderFollowerStoreIngestionTask.createKafkaConsumerProperties( {"kafka.bootstrap.servers" = "local_InMemoryKafkaBroker:41625"}, "local_InMemoryKafkaBroker:41625", false ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.consumerSubscribe(StoreIngestionTask.java:3497) leaderFollowerStoreIngestionTask.getVersionTopic(); -> at com.linkedin.davinci.kafka.consumer.StorePartitionDataReceiver.destinationIdentifier(StorePartitionDataReceiver.java:86) leaderFollowerStoreIngestionTask.getVersionTopic(); -> at com.linkedin.davinci.kafka.consumer.StorePartitionDataReceiver.destinationIdentifier(StorePartitionDataReceiver.java:86) leaderFollowerStoreIngestionTask.checkLongRunningTaskState(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1596) leaderFollowerStoreIngestionTask.getVersionTopic(); -> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.checkLongRunningTaskState(LeaderFollowerStoreIngestionTask.java:755) leaderFollowerStoreIngestionTask.checkIngestionProgress( Mock for Store, hashCode: 100090655 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1597) leaderFollowerStoreIngestionTask.consumerHasAnySubscription(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.checkIngestionProgress(StoreIngestionTask.java:1464) leaderFollowerStoreIngestionTask.maybeSendIngestionHeartbeat(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1598) leaderFollowerStoreIngestionTask.isHybridMode(); -> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.maybeSendIngestionHeartbeat(LeaderFollowerStoreIngestionTask.java:3740) leaderFollowerStoreIngestionTask.isRunning(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1592) leaderFollowerStoreIngestionTask.getIsRunning(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.isRunning(StoreIngestionTask.java:3989) leaderFollowerStoreIngestionTask.updateIngestionRoleIfStoreChanged( Mock for Store, hashCode: 100090655 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1594) leaderFollowerStoreIngestionTask.isHybridMode(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.updateIngestionRoleIfStoreChanged(StoreIngestionTask.java:1511) leaderFollowerStoreIngestionTask.processConsumerActions( Mock for Store, hashCode: 100090655 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1595) leaderFollowerStoreIngestionTask.checkLongRunningTaskState(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1596) leaderFollowerStoreIngestionTask.getVersionTopic(); -> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.checkLongRunningTaskState(LeaderFollowerStoreIngestionTask.java:755) leaderFollowerStoreIngestionTask.checkIngestionProgress( Mock for Store, hashCode: 100090655 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1597) leaderFollowerStoreIngestionTask.consumerHasAnySubscription(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.checkIngestionProgress(StoreIngestionTask.java:1464) leaderFollowerStoreIngestionTask.processConsumerActions( Mock for Store, hashCode: 100090655 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testProcessConsumerActionsError$178(StoreIngestionTaskTest.java:4982) leaderFollowerStoreIngestionTask.processConsumerActions( Mock for Store, hashCode: 100090655 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testProcessConsumerActionsError$178(StoreIngestionTaskTest.java:4982) leaderFollowerStoreIngestionTask.processConsumerActions( Mock for Store, hashCode: 100090655 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testProcessConsumerActionsError$178(StoreIngestionTaskTest.java:4982) leaderFollowerStoreIngestionTask.processConsumerActions( Mock for Store, hashCode: 100090655 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testProcessConsumerActionsError$178(StoreIngestionTaskTest.java:4982) leaderFollowerStoreIngestionTask.processConsumerActions( Mock for Store, hashCode: 100090655 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testProcessConsumerActionsError$178(StoreIngestionTaskTest.java:4982)
SITWithTWiseWithoutBufferAfterLeaderTest.testResubscribeAfterRoleChange: clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithTWiseWithoutBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.TooFewActualInvocations: abstractStorageEngine.putWithReplicationMetadata( 1, <any>, <any>, <any> ); Wanted 200 times: -> at com.linkedin.davinci.store.AbstractStorageEngine.putWithReplicationMetadata(AbstractStorageEngine.java:487) But was 103 times: -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
BatchGetAvroStoreClientUnitTest.testSimpleStreamingBatchGet: clients/venice-client/src/test/java/com/linkedin/venice/fastclient/BatchGetAvroStoreClientUnitTest.java#L65
org.testng.internal.thread.ThreadTimeoutException: Method com.linkedin.venice.fastclient.BatchGetAvroStoreClientUnitTest.testSimpleStreamingBatchGet() didn't finish within the time-out 5000
SITWithPWiseWithoutBufferAfterLeaderTest.testProcessConsumerActionsError: clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseWithoutBufferAfterLeaderTest.java#L1
Wanted but not invoked: leaderFollowerStoreIngestionTask.reportError( <any string>, 1, <Capturing argument> ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.reportError(StoreIngestionTask.java:4146) However, there were exactly 40 interactions with this mock: leaderFollowerStoreIngestionTask.subscribePartition( TestTopic_797a883f18_2aa3c2a4_v1-1 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.runTest(StoreIngestionTaskTest.java:854) leaderFollowerStoreIngestionTask.subscribePartition( TestTopic_797a883f18_2aa3c2a4_v1-1, true ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.subscribePartition(StoreIngestionTask.java:593) leaderFollowerStoreIngestionTask.throwIfNotRunning(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.subscribePartition(StoreIngestionTask.java:607) leaderFollowerStoreIngestionTask.isRunning(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.throwIfNotRunning(StoreIngestionTask.java:554) leaderFollowerStoreIngestionTask.getIsRunning(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.isRunning(StoreIngestionTask.java:3989) leaderFollowerStoreIngestionTask.nextSeqNum(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.subscribePartition(StoreIngestionTask.java:616) leaderFollowerStoreIngestionTask.run(); -> at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) leaderFollowerStoreIngestionTask.isRunning(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1592) leaderFollowerStoreIngestionTask.getIsRunning(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.isRunning(StoreIngestionTask.java:3989) leaderFollowerStoreIngestionTask.updateIngestionRoleIfStoreChanged( Mock for Store, hashCode: 100090655 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1594) leaderFollowerStoreIngestionTask.isHybridMode(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.updateIngestionRoleIfStoreChanged(StoreIngestionTask.java:1511) leaderFollowerStoreIngestionTask.processConsumerActions( Mock for Store, hashCode: 100090655 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1595) leaderFollowerStoreIngestionTask.processConsumerAction( KafkaTaskMessage{type=SUBSCRIBE, topicPartition=TestTopic_797a883f18_2aa3c2a4_v1-1, attempts=1, sequenceNumber=1, createdTimestampInMs=1732155484357}, Mock for Store, hashCode: 100090655 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerActions(StoreIngestionTask.java:1864) leaderFollowerStoreIngestionTask.processCommonConsumerAction( KafkaTaskMessage{type=SUBSCRIBE, topicPartition=TestTopic_797a883f18_2aa3c2a4_v1-1, attempts=1, sequenceNumber=1, createdTimestampInMs=1732155484357} ); -> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.processConsumerAction(LeaderFollowerStoreIngestionTask.java:556) leaderFollowerStoreIngestionTask.reportIfCatchUpVersionTopicOffset( PCS{replicaId=TestTopic_797a883f18_2aa3c2a4_v1-1, hybrid=false, latestProcessedLocalVersionTopicOffset=-1, latestProcessedUpstreamVersionTopicOffset=-1, latestProcessedUpstreamRTOffsetMap={}, latestIgnoredUpstreamRTOffsetMap={}, latestRTOffsetTriedToProduceToVTMap{}, offsetRecord=OffsetRecord{localVersionTopicOffset=-1, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}, errorReported=false, started=false, lagCaughtUp=false, processedRecordSizeSinceLastSync=0, leaderFollowerState=STANDBY} ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processCommonConsumerAction(StoreIngestionTask.java:2107) leaderFollowerStoreIngestionTask.updateLeaderTopicOnFollower( PCS{replicaId=TestTopic_797a883f18_2aa3c2a4_v1-1, hybrid=false, latestProcessedLocalVersionTopicOffset=-1, latestProcessedUpstreamVersionTopicOffset=-1, latestProcessedUpstreamRTOffsetMap={}, latestIgnoredUpstreamRTOffsetMap={}, latestRTOffsetTriedToProduceToVTMap{}, offsetRecord=OffsetRecord{localVersionTopicOffset=-1, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}}, errorReported=false, started=false, lagCaughtUp=false, processedRecordSizeSinceLastSync=0, leaderFollowerState=STANDBY} ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processCommonConsumerAction(StoreIngestionTask.java:2112) leaderFollowerStoreIngestionTask.consumerSubscribe( TestTopic_797a883f18_2aa3c2a4_v1-1, -1L, "local_InMemoryKafkaBroker:41625" ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processCommonConsumerAction(StoreIngestionTask.java:2118) leaderFollowerStoreIngestionTask.createKafkaConsumerProperties( {"kafka.bootstrap.servers" = "local_InMemoryKafkaBroker:41625"}, "local_InMemoryKafkaBroker:41625", false ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.consumerSubscribe(StoreIngestionTask.java:3497) leaderFollowerStoreIngestionTask.getVersionTopic(); -> at com.linkedin.davinci.kafka.consumer.StorePartitionDataReceiver.destinationIdentifier(StorePartitionDataReceiver.java:86) leaderFollowerStoreIngestionTask.getVersionTopic(); -> at com.linkedin.davinci.kafka.consumer.StorePartitionDataReceiver.destinationIdentifier(StorePartitionDataReceiver.java:86) leaderFollowerStoreIngestionTask.checkLongRunningTaskState(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1596) leaderFollowerStoreIngestionTask.getVersionTopic(); -> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.checkLongRunningTaskState(LeaderFollowerStoreIngestionTask.java:755) leaderFollowerStoreIngestionTask.checkIngestionProgress( Mock for Store, hashCode: 100090655 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1597) leaderFollowerStoreIngestionTask.consumerHasAnySubscription(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.checkIngestionProgress(StoreIngestionTask.java:1464) leaderFollowerStoreIngestionTask.maybeSendIngestionHeartbeat(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1598) leaderFollowerStoreIngestionTask.isHybridMode(); -> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.maybeSendIngestionHeartbeat(LeaderFollowerStoreIngestionTask.java:3740) leaderFollowerStoreIngestionTask.isRunning(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1592) leaderFollowerStoreIngestionTask.getIsRunning(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.isRunning(StoreIngestionTask.java:3989) leaderFollowerStoreIngestionTask.updateIngestionRoleIfStoreChanged( Mock for Store, hashCode: 100090655 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1594) leaderFollowerStoreIngestionTask.isHybridMode(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.updateIngestionRoleIfStoreChanged(StoreIngestionTask.java:1511) leaderFollowerStoreIngestionTask.processConsumerActions( Mock for Store, hashCode: 100090655 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1595) leaderFollowerStoreIngestionTask.checkLongRunningTaskState(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1596) leaderFollowerStoreIngestionTask.getVersionTopic(); -> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.checkLongRunningTaskState(LeaderFollowerStoreIngestionTask.java:755) leaderFollowerStoreIngestionTask.checkIngestionProgress( Mock for Store, hashCode: 100090655 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1597) leaderFollowerStoreIngestionTask.consumerHasAnySubscription(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.checkIngestionProgress(StoreIngestionTask.java:1464) leaderFollowerStoreIngestionTask.processConsumerActions( Mock for Store, hashCode: 100090655 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testProcessConsumerActionsError$178(StoreIngestionTaskTest.java:4982) leaderFollowerStoreIngestionTask.processConsumerActions( Mock for Store, hashCode: 100090655 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testProcessConsumerActionsError$178(StoreIngestionTaskTest.java:4982) leaderFollowerStoreIngestionTask.processConsumerActions( Mock for Store, hashCode: 100090655 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testProcessConsumerActionsError$178(StoreIngestionTaskTest.java:4982) leaderFollowerStoreIngestionTask.processConsumerActions( Mock for Store, hashCode: 100090655 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testProcessConsumerActionsError$178(StoreIngestionTaskTest.java:4982) leaderFollowerStoreIngestionTask.processConsumerActions( Mock for Store, hashCode: 100090655 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testProcessConsumerActionsError$178(StoreIngestionTaskTest.java:4982)
SITWithTWiseWithoutBufferAfterLeaderTest.testResubscribeAfterRoleChange: clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithTWiseWithoutBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.TooFewActualInvocations: abstractStorageEngine.putWithReplicationMetadata( 1, <any>, <any>, <any> ); Wanted 200 times: -> at com.linkedin.davinci.store.AbstractStorageEngine.putWithReplicationMetadata(AbstractStorageEngine.java:487) But was 103 times: -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160) -> at com.linkedin.davinci.kafka.consumer.DeepCopyStorageEngine.putWithReplicationMetadata(DeepCopyStorageEngine.java:160)
BatchGetAvroStoreClientUnitTest.testSimpleStreamingBatchGet: clients/venice-client/src/test/java/com/linkedin/venice/fastclient/BatchGetAvroStoreClientUnitTest.java#L65
org.testng.internal.thread.ThreadTimeoutException: Method com.linkedin.venice.fastclient.BatchGetAvroStoreClientUnitTest.testSimpleStreamingBatchGet() didn't finish within the time-out 5000
SITWithPWiseAndBufferAfterLeaderTest.testNotifier[1](AA_OFF): clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseAndBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.ArgumentsAreDifferent: Argument(s) are different! Wanted: storageMetadataService.put( "TestTopic_741ebadc76_a26eff6d_v1", 1, OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testNotifier$44(StoreIngestionTaskTest.java:1576) Actual invocations have different arguments: storageMetadataService.getLastOffset( "TestTopic_741ebadc76_a26eff6d_v1", 1 ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49) storageMetadataService.getLastOffset( "TestTopic_741ebadc76_a26eff6d_v1", 2 ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49) storageMetadataService.computeStoreVersionState( "TestTopic_741ebadc76_a26eff6d_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$980/0x0000000800821040@6754ad51 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_741ebadc76_a26eff6d_v1", 1, OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155537967, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.put( "TestTopic_741ebadc76_a26eff6d_v1", 1, OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155537967, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.computeStoreVersionState( "TestTopic_741ebadc76_a26eff6d_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$980/0x0000000800821040@5a229b96 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_741ebadc76_a26eff6d_v1", 2, OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155537967, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.put( "TestTopic_741ebadc76_a26eff6d_v1", 2, OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155537967, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.computeStoreVersionState( "TestTopic_741ebadc76_a26eff6d_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$980/0x0000000800821040@6b221577 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_741ebadc76_a26eff6d_v1", 2, OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155537967, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.computeStoreVersionState( "TestTopic_741ebadc76_a26eff6d_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$980/0x0000000800821040@249aeb86 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_741ebadc76_a26eff6d_v1", 2, OffsetRecord{localVersionTopicOffset=6, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155537967, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
SITWithPWiseAndBufferAfterLeaderTest.testNotifier[1](AA_OFF): clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseAndBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.ArgumentsAreDifferent: Argument(s) are different! Wanted: storageMetadataService.put( "TestTopic_85002c61fc_a5691add_v1", 1, OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testNotifier$44(StoreIngestionTaskTest.java:1576) Actual invocations have different arguments: storageMetadataService.getLastOffset( "TestTopic_85002c61fc_a5691add_v1", 1 ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49) storageMetadataService.getLastOffset( "TestTopic_85002c61fc_a5691add_v1", 2 ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49) storageMetadataService.computeStoreVersionState( "TestTopic_85002c61fc_a5691add_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8040@3429676f ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.computeStoreVersionState( "TestTopic_85002c61fc_a5691add_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8040@32bf0026 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_85002c61fc_a5691add_v1", 1, OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155610483, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.put( "TestTopic_85002c61fc_a5691add_v1", 2, OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155610483, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.put( "TestTopic_85002c61fc_a5691add_v1", 1, OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155610484, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.put( "TestTopic_85002c61fc_a5691add_v1", 2, OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155610484, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.computeStoreVersionState( "TestTopic_85002c61fc_a5691add_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8040@684f26c4 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.computeStoreVersionState( "TestTopic_85002c61fc_a5691add_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8040@9ee6fd1 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
SITWithPWiseAndBufferAfterLeaderTest.testNotifier[1](AA_OFF): clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseAndBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.ArgumentsAreDifferent: Argument(s) are different! Wanted: storageMetadataService.put( "TestTopic_8b6f1cde2d_2be9bdc6_v1", 2, OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testNotifier$44(StoreIngestionTaskTest.java:1578) Actual invocations have different arguments: storageMetadataService.getLastOffset( "TestTopic_8b6f1cde2d_2be9bdc6_v1", 1 ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49) storageMetadataService.getLastOffset( "TestTopic_8b6f1cde2d_2be9bdc6_v1", 2 ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49) storageMetadataService.computeStoreVersionState( "TestTopic_8b6f1cde2d_2be9bdc6_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$539/0x0000000800692840@4848ec67 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_8b6f1cde2d_2be9bdc6_v1", 1, OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155638101, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.computeStoreVersionState( "TestTopic_8b6f1cde2d_2be9bdc6_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$539/0x0000000800692840@1183e952 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_8b6f1cde2d_2be9bdc6_v1", 2, OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155638101, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.put( "TestTopic_8b6f1cde2d_2be9bdc6_v1", 1, OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155638102, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.put( "TestTopic_8b6f1cde2d_2be9bdc6_v1", 2, OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155638102, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.computeStoreVersionState( "TestTopic_8b6f1cde2d_2be9bdc6_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$539/0x0000000800692840@6fc42eff ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.computeStoreVersionState( "TestTopic_8b6f1cde2d_2be9bdc6_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$539/0x0000000800692840@69525c03 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_8b6f1cde2d_2be9bdc6_v1", 1, OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155638102, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
SITWithPWiseAndBufferAfterLeaderTest.testNotifier[1](AA_OFF): clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseAndBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.ArgumentsAreDifferent: Argument(s) are different! Wanted: storageMetadataService.put( "TestTopic_91f45a3f4d_44d54fa9_v1", 2, OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testNotifier$44(StoreIngestionTaskTest.java:1578) Actual invocations have different arguments: storageMetadataService.getLastOffset( "TestTopic_91f45a3f4d_44d54fa9_v1", 1 ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49) storageMetadataService.getLastOffset( "TestTopic_91f45a3f4d_44d54fa9_v1", 2 ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49) storageMetadataService.computeStoreVersionState( "TestTopic_91f45a3f4d_44d54fa9_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8440@1183e952 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_91f45a3f4d_44d54fa9_v1", 1, OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155666106, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.computeStoreVersionState( "TestTopic_91f45a3f4d_44d54fa9_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8440@6fc42eff ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_91f45a3f4d_44d54fa9_v1", 1, OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155666107, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.put( "TestTopic_91f45a3f4d_44d54fa9_v1", 2, OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155666106, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.computeStoreVersionState( "TestTopic_91f45a3f4d_44d54fa9_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8440@69525c03 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_91f45a3f4d_44d54fa9_v1", 2, OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155666107, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.put( "TestTopic_91f45a3f4d_44d54fa9_v1", 1, OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155666107, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.computeStoreVersionState( "TestTopic_91f45a3f4d_44d54fa9_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8440@66f6ae9a ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
SITWithPWiseAndBufferAfterLeaderTest.testNotifier[1](AA_OFF): clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseAndBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.ArgumentsAreDifferent: Argument(s) are different! Wanted: storageMetadataService.put( "TestTopic_9864f5f3d2_105ca834_v1", 2, OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testNotifier$44(StoreIngestionTaskTest.java:1578) Actual invocations have different arguments: storageMetadataService.getLastOffset( "TestTopic_9864f5f3d2_105ca834_v1", 1 ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49) storageMetadataService.getLastOffset( "TestTopic_9864f5f3d2_105ca834_v1", 2 ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49) storageMetadataService.computeStoreVersionState( "TestTopic_9864f5f3d2_105ca834_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$541/0x0000000800692c40@4848ec67 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_9864f5f3d2_105ca834_v1", 1, OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155693766, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.computeStoreVersionState( "TestTopic_9864f5f3d2_105ca834_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$541/0x0000000800692c40@1183e952 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_9864f5f3d2_105ca834_v1", 2, OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155693766, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.put( "TestTopic_9864f5f3d2_105ca834_v1", 1, OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155693767, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.computeStoreVersionState( "TestTopic_9864f5f3d2_105ca834_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$541/0x0000000800692c40@6fc42eff ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_9864f5f3d2_105ca834_v1", 2, OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155693767, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.computeStoreVersionState( "TestTopic_9864f5f3d2_105ca834_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$541/0x0000000800692c40@69525c03 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_9864f5f3d2_105ca834_v1", 1, OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155693767, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
SITWithTWiseAndBufferAfterLeaderTest.testRecordLevelMetricForCurrentVersion[0](false): clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithTWiseAndBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.TooFewActualInvocations: hostLevelIngestionStats.recordTotalBytesConsumed( <any long> ); Wanted 2 times: -> at com.linkedin.davinci.stats.HostLevelIngestionStats.recordTotalBytesConsumed(HostLevelIngestionStats.java:492) But was 1 time: -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerRecord(StoreIngestionTask.java:2534)
BatchGetAvroStoreClientUnitTest.testSimpleStreamingBatchGet: clients/venice-client/src/test/java/com/linkedin/venice/fastclient/BatchGetAvroStoreClientUnitTest.java#L65
org.testng.internal.thread.ThreadTimeoutException: Method com.linkedin.venice.fastclient.BatchGetAvroStoreClientUnitTest.testSimpleStreamingBatchGet() didn't finish within the time-out 5000
SITWithPWiseAndBufferAfterLeaderTest.testNotifier[1](AA_OFF): clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseAndBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.ArgumentsAreDifferent: Argument(s) are different! Wanted: storageMetadataService.put( "TestTopic_741ebadc76_a26eff6d_v1", 1, OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testNotifier$44(StoreIngestionTaskTest.java:1576) Actual invocations have different arguments: storageMetadataService.getLastOffset( "TestTopic_741ebadc76_a26eff6d_v1", 1 ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49) storageMetadataService.getLastOffset( "TestTopic_741ebadc76_a26eff6d_v1", 2 ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49) storageMetadataService.computeStoreVersionState( "TestTopic_741ebadc76_a26eff6d_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$980/0x0000000800821040@6754ad51 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_741ebadc76_a26eff6d_v1", 1, OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155537967, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.put( "TestTopic_741ebadc76_a26eff6d_v1", 1, OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155537967, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.computeStoreVersionState( "TestTopic_741ebadc76_a26eff6d_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$980/0x0000000800821040@5a229b96 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_741ebadc76_a26eff6d_v1", 2, OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155537967, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.put( "TestTopic_741ebadc76_a26eff6d_v1", 2, OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155537967, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.computeStoreVersionState( "TestTopic_741ebadc76_a26eff6d_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$980/0x0000000800821040@6b221577 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_741ebadc76_a26eff6d_v1", 2, OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155537967, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.computeStoreVersionState( "TestTopic_741ebadc76_a26eff6d_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$980/0x0000000800821040@249aeb86 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_741ebadc76_a26eff6d_v1", 2, OffsetRecord{localVersionTopicOffset=6, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155537967, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
SITWithPWiseAndBufferAfterLeaderTest.testNotifier[1](AA_OFF): clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseAndBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.ArgumentsAreDifferent: Argument(s) are different! Wanted: storageMetadataService.put( "TestTopic_85002c61fc_a5691add_v1", 1, OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testNotifier$44(StoreIngestionTaskTest.java:1576) Actual invocations have different arguments: storageMetadataService.getLastOffset( "TestTopic_85002c61fc_a5691add_v1", 1 ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49) storageMetadataService.getLastOffset( "TestTopic_85002c61fc_a5691add_v1", 2 ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49) storageMetadataService.computeStoreVersionState( "TestTopic_85002c61fc_a5691add_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8040@3429676f ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.computeStoreVersionState( "TestTopic_85002c61fc_a5691add_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8040@32bf0026 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_85002c61fc_a5691add_v1", 1, OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155610483, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.put( "TestTopic_85002c61fc_a5691add_v1", 2, OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155610483, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.put( "TestTopic_85002c61fc_a5691add_v1", 1, OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155610484, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.put( "TestTopic_85002c61fc_a5691add_v1", 2, OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155610484, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.computeStoreVersionState( "TestTopic_85002c61fc_a5691add_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8040@684f26c4 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.computeStoreVersionState( "TestTopic_85002c61fc_a5691add_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8040@9ee6fd1 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
SITWithPWiseAndBufferAfterLeaderTest.testNotifier[1](AA_OFF): clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseAndBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.ArgumentsAreDifferent: Argument(s) are different! Wanted: storageMetadataService.put( "TestTopic_8b6f1cde2d_2be9bdc6_v1", 2, OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testNotifier$44(StoreIngestionTaskTest.java:1578) Actual invocations have different arguments: storageMetadataService.getLastOffset( "TestTopic_8b6f1cde2d_2be9bdc6_v1", 1 ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49) storageMetadataService.getLastOffset( "TestTopic_8b6f1cde2d_2be9bdc6_v1", 2 ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49) storageMetadataService.computeStoreVersionState( "TestTopic_8b6f1cde2d_2be9bdc6_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$539/0x0000000800692840@4848ec67 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_8b6f1cde2d_2be9bdc6_v1", 1, OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155638101, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.computeStoreVersionState( "TestTopic_8b6f1cde2d_2be9bdc6_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$539/0x0000000800692840@1183e952 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_8b6f1cde2d_2be9bdc6_v1", 2, OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155638101, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.put( "TestTopic_8b6f1cde2d_2be9bdc6_v1", 1, OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155638102, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.put( "TestTopic_8b6f1cde2d_2be9bdc6_v1", 2, OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155638102, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.computeStoreVersionState( "TestTopic_8b6f1cde2d_2be9bdc6_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$539/0x0000000800692840@6fc42eff ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.computeStoreVersionState( "TestTopic_8b6f1cde2d_2be9bdc6_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$539/0x0000000800692840@69525c03 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_8b6f1cde2d_2be9bdc6_v1", 1, OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155638102, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
SITWithPWiseAndBufferAfterLeaderTest.testNotifier[1](AA_OFF): clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseAndBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.ArgumentsAreDifferent: Argument(s) are different! Wanted: storageMetadataService.put( "TestTopic_91f45a3f4d_44d54fa9_v1", 2, OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testNotifier$44(StoreIngestionTaskTest.java:1578) Actual invocations have different arguments: storageMetadataService.getLastOffset( "TestTopic_91f45a3f4d_44d54fa9_v1", 1 ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49) storageMetadataService.getLastOffset( "TestTopic_91f45a3f4d_44d54fa9_v1", 2 ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49) storageMetadataService.computeStoreVersionState( "TestTopic_91f45a3f4d_44d54fa9_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8440@1183e952 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_91f45a3f4d_44d54fa9_v1", 1, OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155666106, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.computeStoreVersionState( "TestTopic_91f45a3f4d_44d54fa9_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8440@6fc42eff ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_91f45a3f4d_44d54fa9_v1", 1, OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155666107, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.put( "TestTopic_91f45a3f4d_44d54fa9_v1", 2, OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155666106, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.computeStoreVersionState( "TestTopic_91f45a3f4d_44d54fa9_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8440@69525c03 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_91f45a3f4d_44d54fa9_v1", 2, OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155666107, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.put( "TestTopic_91f45a3f4d_44d54fa9_v1", 1, OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155666107, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.computeStoreVersionState( "TestTopic_91f45a3f4d_44d54fa9_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$543/0x00000008006a8440@66f6ae9a ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33)
SITWithPWiseAndBufferAfterLeaderTest.testNotifier[1](AA_OFF): clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseAndBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.ArgumentsAreDifferent: Argument(s) are different! Wanted: storageMetadataService.put( "TestTopic_9864f5f3d2_105ca834_v1", 2, OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testNotifier$44(StoreIngestionTaskTest.java:1578) Actual invocations have different arguments: storageMetadataService.getLastOffset( "TestTopic_9864f5f3d2_105ca834_v1", 1 ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49) storageMetadataService.getLastOffset( "TestTopic_9864f5f3d2_105ca834_v1", 2 ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49) storageMetadataService.computeStoreVersionState( "TestTopic_9864f5f3d2_105ca834_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$541/0x0000000800692c40@4848ec67 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_9864f5f3d2_105ca834_v1", 1, OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155693766, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.computeStoreVersionState( "TestTopic_9864f5f3d2_105ca834_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$541/0x0000000800692c40@1183e952 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_9864f5f3d2_105ca834_v1", 2, OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155693766, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.put( "TestTopic_9864f5f3d2_105ca834_v1", 1, OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155693767, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.computeStoreVersionState( "TestTopic_9864f5f3d2_105ca834_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$541/0x0000000800692c40@6fc42eff ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_9864f5f3d2_105ca834_v1", 2, OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155693767, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.computeStoreVersionState( "TestTopic_9864f5f3d2_105ca834_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$541/0x0000000800692c40@69525c03 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:33) storageMetadataService.put( "TestTopic_9864f5f3d2_105ca834_v1", 1, OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1732155693767, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
SITWithTWiseAndBufferAfterLeaderTest.testRecordLevelMetricForCurrentVersion[0](false): clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithTWiseAndBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.TooFewActualInvocations: hostLevelIngestionStats.recordTotalBytesConsumed( <any long> ); Wanted 2 times: -> at com.linkedin.davinci.stats.HostLevelIngestionStats.recordTotalBytesConsumed(HostLevelIngestionStats.java:492) But was 1 time: -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerRecord(StoreIngestionTask.java:2534)
BatchGetAvroStoreClientUnitTest.testSimpleStreamingBatchGet: clients/venice-client/src/test/java/com/linkedin/venice/fastclient/BatchGetAvroStoreClientUnitTest.java#L65
org.testng.internal.thread.ThreadTimeoutException: Method com.linkedin.venice.fastclient.BatchGetAvroStoreClientUnitTest.testSimpleStreamingBatchGet() didn't finish within the time-out 5000
Clients / UT & CodeCov (11)
Process completed with exit code 1.
StaticAnalysisAndUnitTestsCompletionCheck
Process completed with exit code 1.
com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush ► testHybridStorePartitionCount: integrations/venice-samza/src/main/java/com/linkedin/venice/samza/VeniceSystemProducer.java#L362
Failed test found in: IntegrationTests_1050-artifacts/internal/venice-test-common/build/test-results/integrationTests_1050/TEST-com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush.xml internal/venice-test-common/build/test-results/integrationTests_1050/TEST-com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush.xml Error: org.apache.samza.SamzaException: Failed to send request to Controller, error: An error occurred during controller request. controller = http://localhost:41815, route = /request_topic, params = [cluster_name=venice-cluster0, push_in_sorted_order=false, rewind_time_in_seconds_override=-1, is_write_compute_enabled=false, store_size=1, push_job_id=DC_1_store_a7b4dd6b44_9eb0a4f4, push_type=STREAM, defer_version_swap=false, hostname=localhost, repush_source_version=-1, store_name=store_a7b4dd6b44_9eb0a4f4, start_of_push=true, batch_job_heartbeat_enabled=false], timeout = 600000. Exception type: class com.linkedin.venice.exceptions.VeniceHttpException. Detailed message: Http Status 500 - Could not find an ONLINE hybrid store version for streaming writes with pushJobId: DC_1_store_a7b4dd6b44_9eb0a4f4 on store: store_a7b4dd6b44_9eb0a4f4 in cluster: venice-cluster0
com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush ► testHybridStorePartitionCount: integrations/venice-samza/src/main/java/com/linkedin/venice/samza/VeniceSystemProducer.java#L362
Failed test found in: IntegrationTests_1050-artifacts/internal/venice-test-common/build/test-results/integrationTests_1050/TEST-com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush.xml internal/venice-test-common/build/test-results/integrationTests_1050/TEST-com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush.xml Error: org.apache.samza.SamzaException: Failed to send request to Controller, error: An error occurred during controller request. controller = http://localhost:41175, route = /request_topic, params = [cluster_name=venice-cluster0, push_in_sorted_order=false, rewind_time_in_seconds_override=-1, is_write_compute_enabled=false, store_size=1, push_job_id=DC_1_store_d4471ecfa0_b1345b9d, push_type=STREAM, defer_version_swap=false, hostname=localhost, repush_source_version=-1, store_name=store_d4471ecfa0_b1345b9d, start_of_push=true, batch_job_heartbeat_enabled=false], timeout = 600000. Exception type: class com.linkedin.venice.exceptions.VeniceHttpException. Detailed message: Http Status 500 - Could not find an ONLINE hybrid store version for streaming writes with pushJobId: DC_1_store_d4471ecfa0_b1345b9d on store: store_d4471ecfa0_b1345b9d in cluster: venice-cluster0
com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush ► testHybridStorePartitionCount: integrations/venice-samza/src/main/java/com/linkedin/venice/samza/VeniceSystemProducer.java#L362
Failed test found in: IntegrationTests_1050-artifacts/internal/venice-test-common/build/test-results/integrationTests_1050/TEST-com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush.xml internal/venice-test-common/build/test-results/integrationTests_1050/TEST-com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush.xml Error: org.apache.samza.SamzaException: Failed to send request to Controller, error: An error occurred during controller request. controller = http://localhost:36367, route = /request_topic, params = [cluster_name=venice-cluster0, push_in_sorted_order=false, rewind_time_in_seconds_override=-1, is_write_compute_enabled=false, store_size=1, push_job_id=DC_1_store_ecd864f4cf_a4becf16, push_type=STREAM, defer_version_swap=false, hostname=localhost, repush_source_version=-1, store_name=store_ecd864f4cf_a4becf16, start_of_push=true, batch_job_heartbeat_enabled=false], timeout = 600000. Exception type: class com.linkedin.venice.exceptions.VeniceHttpException. Detailed message: Http Status 500 - Could not find an ONLINE hybrid store version for streaming writes with pushJobId: DC_1_store_ecd864f4cf_a4becf16 on store: store_ecd864f4cf_a4becf16 in cluster: venice-cluster0
com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush ► testHybridStorePartitionCount: integrations/venice-samza/src/main/java/com/linkedin/venice/samza/VeniceSystemProducer.java#L362
Failed test found in: IntegrationTests_1050-artifacts/internal/venice-test-common/build/test-results/integrationTests_1050/TEST-com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush.xml internal/venice-test-common/build/test-results/integrationTests_1050/TEST-com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush.xml Error: org.apache.samza.SamzaException: Failed to send request to Controller, error: An error occurred during controller request. controller = http://localhost:33617, route = /request_topic, params = [cluster_name=venice-cluster0, push_in_sorted_order=false, rewind_time_in_seconds_override=-1, is_write_compute_enabled=false, store_size=1, push_job_id=DC_1_store_1069edc26ff_c941d0b2, push_type=STREAM, defer_version_swap=false, hostname=localhost, repush_source_version=-1, store_name=store_1069edc26ff_c941d0b2, start_of_push=true, batch_job_heartbeat_enabled=false], timeout = 600000. Exception type: class com.linkedin.venice.exceptions.VeniceHttpException. Detailed message: Http Status 500 - Could not find an ONLINE hybrid store version for streaming writes with pushJobId: DC_1_store_1069edc26ff_c941d0b2 on store: store_1069edc26ff_c941d0b2 in cluster: venice-cluster0
com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush ► testHybridStorePartitionCount: integrations/venice-samza/src/main/java/com/linkedin/venice/samza/VeniceSystemProducer.java#L362
Failed test found in: IntegrationTests_1050-artifacts/internal/venice-test-common/build/test-results/integrationTests_1050/TEST-com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush.xml internal/venice-test-common/build/test-results/integrationTests_1050/TEST-com.linkedin.venice.endToEnd.TestActiveActiveReplicationForIncPush.xml Error: org.apache.samza.SamzaException: Failed to send request to Controller, error: An error occurred during controller request. controller = http://localhost:36147, route = /request_topic, params = [cluster_name=venice-cluster0, push_in_sorted_order=false, rewind_time_in_seconds_override=-1, is_write_compute_enabled=false, store_size=1, push_job_id=DC_1_store_12140bc2fee_6ef70689, push_type=STREAM, defer_version_swap=false, hostname=localhost, repush_source_version=-1, store_name=store_12140bc2fee_6ef70689, start_of_push=true, batch_job_heartbeat_enabled=false], timeout = 600000. Exception type: class com.linkedin.venice.exceptions.VeniceHttpException. Detailed message: Http Status 500 - Could not find an ONLINE hybrid store version for streaming writes with pushJobId: DC_1_store_12140bc2fee_6ef70689 on store: store_12140bc2fee_6ef70689 in cluster: venice-cluster0
com.linkedin.venice.endToEnd.TestBatchForRocksDB ► testBatchJobSnapshots[2](true): internal/venice-test-common/src/integrationTest/java/com/linkedin/venice/endToEnd/TestBatch.java#L1428
Failed test found in: IntegrationTests_1230-artifacts/internal/venice-test-common/build/test-results/integrationTests_1230/TEST-com.linkedin.venice.endToEnd.TestBatchForRocksDB.xml internal/venice-test-common/build/test-results/integrationTests_1230/TEST-com.linkedin.venice.endToEnd.TestBatchForRocksDB.xml Error: java.util.concurrent.ExecutionException: com.linkedin.venice.client.exceptions.VeniceClientHttpException: http status: 503, com.linkedin.alpini.router.api.RouterException: store_348547b201_a1f39293_v1, partition 2 is not available to serve request of type: SINGLE_GET
com.linkedin.venice.endToEnd.BlobP2PTransferAmongServersTest ► testBlobP2PTransferAmongServersForBatchStore: internal/venice-test-common/src/integrationTest/java/com/linkedin/venice/endToEnd/BlobP2PTransferAmongServersTest.java#L123
Failed test found in: IntegrationTests_1130-artifacts/internal/venice-test-common/build/test-results/integrationTests_1130/TEST-com.linkedin.venice.endToEnd.BlobP2PTransferAmongServersTest.xml internal/venice-test-common/build/test-results/integrationTests_1130/TEST-com.linkedin.venice.endToEnd.BlobP2PTransferAmongServersTest.xml Error: java.lang.AssertionError: expected [true] but found [false]
com.linkedin.venice.endToEnd.BlobP2PTransferAmongServersTest ► testBlobP2PTransferAmongServersForHybridStore: integrations/venice-samza/src/main/java/com/linkedin/venice/samza/VeniceSystemProducer.java#L362
Failed test found in: IntegrationTests_1130-artifacts/internal/venice-test-common/build/test-results/integrationTests_1130/TEST-com.linkedin.venice.endToEnd.BlobP2PTransferAmongServersTest.xml internal/venice-test-common/build/test-results/integrationTests_1130/TEST-com.linkedin.venice.endToEnd.BlobP2PTransferAmongServersTest.xml Error: org.apache.samza.SamzaException: Failed to send request to Controller, error: An error occurred during controller request. controller = http://localhost:41265, route = /request_topic, params = [cluster_name=venice-cluster_50b960d02c_1e278e46, push_in_sorted_order=false, rewind_time_in_seconds_override=-1, is_write_compute_enabled=false, store_size=1, push_job_id=venice-push-id_544dd68afb_8bca5409, push_type=STREAM, defer_version_swap=false, hostname=localhost, repush_source_version=-1, store_name=test-store-hybrid, start_of_push=true, batch_job_heartbeat_enabled=false], timeout = 600000. Exception type: class com.linkedin.venice.exceptions.VeniceHttpException. Detailed message: Http Status 500 - Could not find an ONLINE hybrid store version for streaming writes with pushJobId: venice-push-id_544dd68afb_8bca5409 on store: test-store-hybrid in cluster: venice-cluster_50b960d02c_1e278e46
com.linkedin.venice.endToEnd.MetaSystemStoreTest ► testThinClientMetaStoreBasedRepositoryWithLargeValueSchemas: internal/venice-test-common/src/integrationTest/java/com/linkedin/venice/endToEnd/MetaSystemStoreTest.java#L315
Failed test found in: IntegrationTests_1500-artifacts/internal/venice-test-common/build/test-results/integrationTests_1500/TEST-com.linkedin.venice.endToEnd.MetaSystemStoreTest.xml internal/venice-test-common/build/test-results/integrationTests_1500/TEST-com.linkedin.venice.endToEnd.MetaSystemStoreTest.xml Error: java.lang.AssertionError: Number of value schemas should be the same between meta system store and controller expected [2] but found [1]
com.linkedin.venice.endToEnd.TestHybrid ► testLeaderCanReleaseLatch: integrations/venice-samza/src/main/java/com/linkedin/venice/samza/VeniceSystemProducer.java#L362
Failed test found in: IntegrationTests_1400-artifacts/internal/venice-test-common/build/test-results/integrationTests_1400/TEST-com.linkedin.venice.endToEnd.TestHybrid.xml internal/venice-test-common/build/test-results/integrationTests_1400/TEST-com.linkedin.venice.endToEnd.TestHybrid.xml Error: org.apache.samza.SamzaException: Failed to send request to Controller, error: An error occurred during controller request. controller = http://localhost:33903, route = /request_topic, params = [cluster_name=venice-cluster_3e14cc0c94_8858fc09, push_in_sorted_order=false, rewind_time_in_seconds_override=-1, is_write_compute_enabled=false, store_size=1, push_job_id=venice-push-id_5b76066303_506309b0, push_type=STREAM, defer_version_swap=false, hostname=localhost, repush_source_version=-1, store_name=test-store_5b14fe8ce2_681f93b6, start_of_push=true, batch_job_heartbeat_enabled=false], timeout = 600000. Exception type: class com.linkedin.venice.exceptions.VeniceHttpException. Detailed message: Http Status 500 - Could not find an ONLINE hybrid store version for streaming writes with pushJobId: venice-push-id_5b76066303_506309b0 on store: test-store_5b14fe8ce2_681f93b6 in cluster: venice-cluster_3e14cc0c94_8858fc09
com.linkedin.venice.controller.server.TestAdminSparkServer ► controllerClientCanGetStoreInfo: internal/venice-test-common/src/integrationTest/java/com/linkedin/venice/controller/server/TestAdminSparkServer.java#L462
Failed test found in: IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkServer.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkServer.xml Error: java.lang.AssertionError: Store from new store-version should only have one version expected [1] but found [0]
com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers ► requestTopicIsIdempotent: internal/venice-test-common/src/integrationTest/java/com/linkedin/venice/controller/server/TestAdminSparkServerWithMultiServers.java#L275
Failed test found in: IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers.xml Error: java.lang.AssertionError: An error occurred during controller request. controller = http://localhost:36255, route = /request_topic, params = [cluster_name=venice-cluster_8adec8af64_b23c05ec, push_in_sorted_order=true, rewind_time_in_seconds_override=-1, is_write_compute_enabled=false, store_size=1, push_job_id=pushId_900aa200f7_57476ca9, push_type=STREAM, defer_version_swap=false, hostname=localhost, repush_source_version=-1, store_name=StreamStore_900aa1e8f0_82f34c97, start_of_push=true, batch_job_heartbeat_enabled=false], timeout = 600000. Exception type: class com.linkedin.venice.exceptions.VeniceHttpException. Detailed message: Http Status 500 - Could not find an ONLINE hybrid store version for streaming writes with pushJobId: pushId_900aa200f7_57476ca9 on store: StreamStore_900aa1e8f0_82f34c97 in cluster: venice-cluster_8adec8af64_b23c05ec
com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers ► requestTopicIsIdempotent: internal/venice-test-common/src/integrationTest/java/com/linkedin/venice/controller/server/TestAdminSparkServerWithMultiServers.java#L275
Failed test found in: IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers.xml Error: java.lang.AssertionError: An error occurred during controller request. controller = http://localhost:39995, route = /request_topic, params = [cluster_name=venice-cluster_a73f935bae_7da89c5a, push_in_sorted_order=true, rewind_time_in_seconds_override=-1, is_write_compute_enabled=false, store_size=1, push_job_id=pushId_abd1f4d48f_7c6792ff, push_type=STREAM, defer_version_swap=false, hostname=localhost, repush_source_version=-1, store_name=StreamStore_abd1f4bf31_2392aec2, start_of_push=true, batch_job_heartbeat_enabled=false], timeout = 600000. Exception type: class com.linkedin.venice.exceptions.VeniceHttpException. Detailed message: Http Status 500 - Could not find an ONLINE hybrid store version for streaming writes with pushJobId: pushId_abd1f4d48f_7c6792ff on store: StreamStore_abd1f4bf31_2392aec2 in cluster: venice-cluster_a73f935bae_7da89c5a
com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers ► requestTopicIsIdempotent: internal/venice-test-common/src/integrationTest/java/com/linkedin/venice/controller/server/TestAdminSparkServerWithMultiServers.java#L275
Failed test found in: IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers.xml Error: java.lang.AssertionError: An error occurred during controller request. controller = http://localhost:45849, route = /request_topic, params = [cluster_name=venice-cluster_b2fd564a39_4b27bcc9, push_in_sorted_order=true, rewind_time_in_seconds_override=-1, is_write_compute_enabled=false, store_size=1, push_job_id=pushId_b78d2f647b_90fbcd2e, push_type=STREAM, defer_version_swap=false, hostname=localhost, repush_source_version=-1, store_name=StreamStore_b78d2f411a_6207ae1b, start_of_push=true, batch_job_heartbeat_enabled=false], timeout = 600000. Exception type: class com.linkedin.venice.exceptions.VeniceHttpException. Detailed message: Http Status 500 - Could not find an ONLINE hybrid store version for streaming writes with pushJobId: pushId_b78d2f647b_90fbcd2e on store: StreamStore_b78d2f411a_6207ae1b in cluster: venice-cluster_b2fd564a39_4b27bcc9
com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers ► requestTopicIsIdempotent: internal/venice-test-common/src/integrationTest/java/com/linkedin/venice/controller/server/TestAdminSparkServerWithMultiServers.java#L275
Failed test found in: IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers.xml Error: java.lang.AssertionError: An error occurred during controller request. controller = http://localhost:34409, route = /request_topic, params = [cluster_name=venice-cluster_beca85400b_c645f9c3, push_in_sorted_order=true, rewind_time_in_seconds_override=-1, is_write_compute_enabled=false, store_size=1, push_job_id=pushId_c5adc61453_5d23ea01, push_type=STREAM, defer_version_swap=false, hostname=localhost, repush_source_version=-1, store_name=StreamStore_c5adc5f57d_4d794319, start_of_push=true, batch_job_heartbeat_enabled=false], timeout = 600000. Exception type: class com.linkedin.venice.exceptions.VeniceHttpException. Detailed message: Http Status 500 - Could not find an ONLINE hybrid store version for streaming writes with pushJobId: pushId_c5adc61453_5d23ea01 on store: StreamStore_c5adc5f57d_4d794319 in cluster: venice-cluster_beca85400b_c645f9c3
com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers ► requestTopicIsIdempotent: internal/venice-test-common/src/integrationTest/java/com/linkedin/venice/controller/server/TestAdminSparkServerWithMultiServers.java#L275
Failed test found in: IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkServerWithMultiServers.xml Error: java.lang.AssertionError: An error occurred during controller request. controller = http://localhost:39097, route = /request_topic, params = [cluster_name=venice-cluster_cd0f1b525a_e81b93c9, push_in_sorted_order=true, rewind_time_in_seconds_override=-1, is_write_compute_enabled=false, store_size=1, push_job_id=pushId_d1aca92d8b_696ca337, push_type=STREAM, defer_version_swap=false, hostname=localhost, repush_source_version=-1, store_name=StreamStore_d1aca90a0d_831f1407, start_of_push=true, batch_job_heartbeat_enabled=false], timeout = 600000. Exception type: class com.linkedin.venice.exceptions.VeniceHttpException. Detailed message: Http Status 500 - Could not find an ONLINE hybrid store version for streaming writes with pushJobId: pushId_d1aca92d8b_696ca337 on store: StreamStore_d1aca90a0d_831f1407 in cluster: venice-cluster_cd0f1b525a_e81b93c9
com.linkedin.venice.controller.server.TestAdminSparkWithMocks ► testGetRealTimeTopicUsesAdmin: internal/venice-test-common/src/integrationTest/java/com/linkedin/venice/controller/server/TestAdminSparkWithMocks.java#L109
Failed test found in: IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml Error: java.lang.AssertionError: unexpected error: No hybrid version found for store: store. Create a hybrid version before starting a stream push. expected [false] but found [true]
com.linkedin.venice.controller.server.TestAdminSparkWithMocks ► testSamzaReplicationPolicyMode[0](false, false, false): internal/venice-test-common/src/integrationTest/java/com/linkedin/venice/controller/server/TestAdminSparkWithMocks.java#L370
Failed test found in: IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml IntegrationTests_1410-artifacts/internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml internal/venice-test-common/build/test-results/integrationTests_1410/TEST-com.linkedin.venice.controller.server.TestAdminSparkWithMocks.xml Error: java.lang.AssertionError: unexpected error: No hybrid version found for store: store. Create a hybrid version before starting a stream push. expected [false] but found [true]
com.linkedin.venice.endToEnd.DaVinciClientTest ► testTypeChangeRecordTransformer[0](DaVinciConfig{managed=true, isolated=false, storageClass=MEMORY_BACKED_BY_DISK, cacheConfig=null, largeBatchRequestSplitThreshold=100}): clients/venice-push-job/src/main/java/com/linkedin/venice/hadoop/DefaultInputDataInfoProvider.java#L342
Failed test found in: IntegrationTests_1010-artifacts/internal/venice-test-common/build/test-results/integrationTests_1010/TEST-com.linkedin.venice.endToEnd.DaVinciClientTest.xml IntegrationTests_1010-artifacts/internal/venice-test-common/build/test-results/integrationTests_1010/TEST-com.linkedin.venice.endToEnd.DaVinciClientTest.xml IntegrationTests_1010-artifacts/internal/venice-test-common/build/test-results/integrationTests_1010/TEST-com.linkedin.venice.endToEnd.DaVinciClientTest.xml IntegrationTests_1010-artifacts/internal/venice-test-common/build/test-results/integrationTests_1010/TEST-com.linkedin.venice.endToEnd.DaVinciClientTest.xml IntegrationTests_1010-artifacts/internal/venice-test-common/build/test-results/integrationTests_1010/TEST-com.linkedin.venice.endToEnd.DaVinciClientTest.xml IntegrationTests_1010-artifacts/internal/venice-test-common/build/test-results/integrationTests_1010/TEST-com.linkedin.venice.endToEnd.DaVinciClientTest.xml internal/venice-test-common/build/test-results/integrationTests_1010/TEST-com.linkedin.venice.endToEnd.DaVinciClientTest.xml internal/venice-test-common/build/test-results/integrationTests_1010/TEST-com.linkedin.venice.endToEnd.DaVinciClientTest.xml internal/venice-test-common/build/test-results/integrationTests_1010/TEST-com.linkedin.venice.endToEnd.DaVinciClientTest.xml internal/venice-test-common/build/test-results/integrationTests_1010/TEST-com.linkedin.venice.endToEnd.DaVinciClientTest.xml internal/venice-test-common/build/test-results/integrationTests_1010/TEST-com.linkedin.venice.endToEnd.DaVinciClientTest.xml internal/venice-test-common/build/test-results/integrationTests_1010/TEST-com.linkedin.venice.endToEnd.DaVinciClientTest.xml Error: com.linkedin.venice.hadoop.exceptions.VeniceInconsistentSchemaException: Inconsistent file Avro schema found. File: int2string.avro.
com.linkedin.venice.controller.TestParentControllerWithMultiDataCenter ► testDeferredVersionSwap: internal/venice-test-common/src/integrationTest/java/com/linkedin/venice/utils/IntegrationTestPushUtils.java#L313
Failed test found in: IntegrationTests_1430-artifacts/internal/venice-test-common/build/test-results/integrationTests_1430/TEST-com.linkedin.venice.controller.TestParentControllerWithMultiDataCenter.xml internal/venice-test-common/build/test-results/integrationTests_1430/TEST-com.linkedin.venice.controller.TestParentControllerWithMultiDataCenter.xml Error: com.linkedin.venice.exceptions.VeniceException: Could not create store store_f17625870a_199f179a
com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment ► setUp: internal/venice-test-common/src/main/java/com/linkedin/venice/utils/TestUtils.java#L245
Failed test found in: IntegrationTests_1430-artifacts/internal/venice-test-common/build/test-results/integrationTests_1430/TEST-com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment.xml internal/venice-test-common/build/test-results/integrationTests_1430/TEST-com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment.xml Error: java.lang.AssertionError: com.linkedin.venice.pubsub.api.exceptions.PubSubOpTimeoutException: Timeout while creating topic: venice_system_store_participant_store_cluster_test-cluster_134a2c9f9a1_407899d0_rt. Topic still does not exist after 30000ms.
com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment ► setUp: internal/venice-test-common/src/main/java/com/linkedin/venice/utils/TestUtils.java#L245
Failed test found in: IntegrationTests_1430-artifacts/internal/venice-test-common/build/test-results/integrationTests_1430/TEST-com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment.xml internal/venice-test-common/build/test-results/integrationTests_1430/TEST-com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment.xml Error: java.lang.AssertionError: com.linkedin.venice.pubsub.api.exceptions.PubSubOpTimeoutException: Timeout while creating topic: venice_system_store_participant_store_cluster_test-cluster_154540d1b92_4b0d3c8f_rt. Topic still does not exist after 30000ms.
com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment ► setUp: internal/venice-test-common/src/main/java/com/linkedin/venice/utils/TestUtils.java#L245
Failed test found in: IntegrationTests_1430-artifacts/internal/venice-test-common/build/test-results/integrationTests_1430/TEST-com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment.xml internal/venice-test-common/build/test-results/integrationTests_1430/TEST-com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment.xml Error: java.lang.AssertionError: com.linkedin.venice.pubsub.api.exceptions.PubSubOpTimeoutException: Timeout while creating topic: venice_system_store_participant_store_cluster_test-cluster_1612ea1a815_d4a9153f_rt. Topic still does not exist after 30000ms.
com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment ► setUp: internal/venice-test-common/src/main/java/com/linkedin/venice/utils/TestUtils.java#L245
Failed test found in: IntegrationTests_1430-artifacts/internal/venice-test-common/build/test-results/integrationTests_1430/TEST-com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment.xml internal/venice-test-common/build/test-results/integrationTests_1430/TEST-com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment.xml Error: java.lang.AssertionError: com.linkedin.venice.pubsub.api.exceptions.PubSubOpTimeoutException: Timeout while creating topic: venice_system_store_participant_store_cluster_test-cluster_16e226e0dc6_41e11ec_rt. Topic still does not exist after 30000ms.
com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment ► setUp: internal/venice-test-common/src/main/java/com/linkedin/venice/utils/TestUtils.java#L245
Failed test found in: IntegrationTests_1430-artifacts/internal/venice-test-common/build/test-results/integrationTests_1430/TEST-com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment.xml internal/venice-test-common/build/test-results/integrationTests_1430/TEST-com.linkedin.venice.controller.TestVeniceHelixAdminWithSharedEnvironment.xml Error: java.lang.AssertionError: com.linkedin.venice.pubsub.api.exceptions.PubSubOpTimeoutException: Timeout while creating topic: venice_system_store_participant_store_cluster_test-cluster_17ac2a5b473_aae542a5_rt. Topic still does not exist after 30000ms.
StaticAnalysis (17)
Retention days cannot be greater than the maximum allowed retention set within the repository. Using 10 instead.
Internal / UT & CodeCov (8)
The following actions use a deprecated Node.js version and will be forced to run on node20: mikepenz/action-junit-report@v3. For more info: https://github.blog/changelog/2024-03-07-github-actions-all-actions-will-run-on-node20-instead-of-node16-by-default/
Internal / UT & CodeCov (8)
Retention days cannot be greater than the maximum allowed retention set within the repository. Using 10 instead.
Internal / UT & CodeCov (17)
The following actions use a deprecated Node.js version and will be forced to run on node20: mikepenz/action-junit-report@v3. For more info: https://github.blog/changelog/2024-03-07-github-actions-all-actions-will-run-on-node20-instead-of-node16-by-default/
Internal / UT & CodeCov (17)
Retention days cannot be greater than the maximum allowed retention set within the repository. Using 10 instead.
Internal / UT & CodeCov (11)
The following actions use a deprecated Node.js version and will be forced to run on node20: mikepenz/action-junit-report@v3. For more info: https://github.blog/changelog/2024-03-07-github-actions-all-actions-will-run-on-node20-instead-of-node16-by-default/
Internal / UT & CodeCov (11)
Retention days cannot be greater than the maximum allowed retention set within the repository. Using 10 instead.
Services / UT & CodeCov (17)
The following actions use a deprecated Node.js version and will be forced to run on node20: mikepenz/action-junit-report@v3. For more info: https://github.blog/changelog/2024-03-07-github-actions-all-actions-will-run-on-node20-instead-of-node16-by-default/
Services / UT & CodeCov (17)
Retention days cannot be greater than the maximum allowed retention set within the repository. Using 10 instead.
Services / UT & CodeCov (8)
The following actions use a deprecated Node.js version and will be forced to run on node20: mikepenz/action-junit-report@v3. For more info: https://github.blog/changelog/2024-03-07-github-actions-all-actions-will-run-on-node20-instead-of-node16-by-default/
Services / UT & CodeCov (8)
Retention days cannot be greater than the maximum allowed retention set within the repository. Using 10 instead.
Services / UT & CodeCov (11)
The following actions use a deprecated Node.js version and will be forced to run on node20: mikepenz/action-junit-report@v3. For more info: https://github.blog/changelog/2024-03-07-github-actions-all-actions-will-run-on-node20-instead-of-node16-by-default/
Services / UT & CodeCov (11)
Retention days cannot be greater than the maximum allowed retention set within the repository. Using 10 instead.
Clients / UT & CodeCov (8)
The following actions use a deprecated Node.js version and will be forced to run on node20: mikepenz/action-junit-report@v3. For more info: https://github.blog/changelog/2024-03-07-github-actions-all-actions-will-run-on-node20-instead-of-node16-by-default/
Clients / UT & CodeCov (8)
Retention days cannot be greater than the maximum allowed retention set within the repository. Using 10 instead.
Clients / UT & CodeCov (17)
The following actions use a deprecated Node.js version and will be forced to run on node20: mikepenz/action-junit-report@v3. For more info: https://github.blog/changelog/2024-03-07-github-actions-all-actions-will-run-on-node20-instead-of-node16-by-default/
Clients / UT & CodeCov (17)
Retention days cannot be greater than the maximum allowed retention set within the repository. Using 10 instead.
Clients / UT & CodeCov (11)
The following actions use a deprecated Node.js version and will be forced to run on node20: mikepenz/action-junit-report@v3. For more info: https://github.blog/changelog/2024-03-07-github-actions-all-actions-will-run-on-node20-instead-of-node16-by-default/
Clients / UT & CodeCov (11)
Retention days cannot be greater than the maximum allowed retention set within the repository. Using 10 instead.

Artifacts

Produced during runtime
Name Size
StaticAnalysis Expired
694 KB
clients-jdk11 Expired
10.5 MB
clients-jdk17 Expired
11.7 MB
clients-jdk8 Expired
11.7 MB
internal-jdk11 Expired
3.01 MB
internal-jdk17 Expired
3.03 MB
internal-jdk8 Expired
3 MB
services-jdk11 Expired
2.42 MB
services-jdk17 Expired
2.43 MB
services-jdk8 Expired
2.39 MB