Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Run ipv6 lane in control plane jobs #4106

Merged
merged 1 commit into from
Feb 27, 2024

Conversation

tssurya
Copy link
Member

@tssurya tssurya commented Jan 29, 2024

- What this PR does and why is it needed
This commit fixes things here and there
in e2e's to make control planes run on
v6 as well. Note that some tests are
skipped since they cannot be run on
github runners with v6 enabled.

NOTE: One of the multicast test is moved
from e2e.go to multicast.go where it belongs.

- Special notes for reviewers
We must re-enable all the skipped tests: ci-ipv6 Add support for IPV6 e2e's to run upstream

Supercedes #3512

@tssurya
Copy link
Member Author

tssurya commented Jan 29, 2024

unrelated UT flake:

2024-01-29T18:02:58.2816353Z 
2024-01-29T18:02:58.2816665Z �[91m�[1mSummarizing 1 Failure:�[0m
2024-01-29T18:02:58.2817038Z 
2024-01-29T18:02:58.2817843Z �[91m�[1m[Fail] �[0m�[90mHybrid SDN Master Operations �[0m�[91m�[1m[It] handles a HO node is switched to a OVN node �[0m
2024-01-29T18:02:58.2819352Z �[37m/home/runner/work/ovn-kubernetes/ovn-kubernetes/go-controller/pkg/ovn/hybrid_test.go:1370�[0m
2024-01-29T18:02:58.2819970Z 
2024-01-29T18:02:58.2820219Z �[1m�[91mRan 380 of 381 Specs in 676.931 seconds�[0m
2024-01-29T18:02:58.2820963Z �[1m�[91mFAIL!�[0m -- �[32m�[1m379 Passed�[0m | �[91m�[1m1 Failed�[0m | �[33m�[1m1 Pending�[0m | �[36m�[1m0 Skipped�[0m
2024-01-29T18:02:58.2821586Z --- FAIL: TestClusterNode (677.14s)
2024-01-29T18:02:58.2821875Z FAIL
2024-01-29T18:02:58.3025125Z 	github.com/ovn-org/ovn-kubernetes/go-controller/pkg/ovn	coverage: 70.9% of statements
2024-01-29T18:02:58.4920465Z FAIL	github.com/ovn-org/ovn-kubernetes/go-controller/pkg/ovn	684.127s
2024-01-29T18:02:58.5058121Z FAIL
2024-01-29T18:02:58.5611374Z make: *** [Makefile:56: check] Error 1
2024-01-29T18:02:58.5633838Z ##[error]Process completed with exit code 2.
2024-01-29T18:02:58.5736484Z Post job cleanup.
2024-01-29T18:02:58.6612702Z [command]/usr/bin/git version
2024-01-29T18:02:58.6647095Z git version 2.43.0

@tssurya
Copy link
Member Author

tssurya commented Jan 29, 2024

flake reported at #4031
retesting

@trozet
Copy link
Contributor

trozet commented Jan 29, 2024

we also need ipv6 or dualstack external-gateway jobs @tssurya

@tssurya tssurya force-pushed the run-ipv6-lanes-cp-take-2 branch 4 times, most recently from e9aadbd to 367e847 Compare February 5, 2024 20:00
@coveralls
Copy link

coveralls commented Feb 5, 2024

Coverage Status

coverage: 51.877% (-0.08%) from 51.954%
when pulling 8cf27a9 on tssurya:run-ipv6-lanes-cp-take-2
into 5f83f42 on ovn-org:master.

@tssurya
Copy link
Member Author

tssurya commented Feb 6, 2024

ok I rebased and now all the new tests that were added:

  [FAIL] e2e egress IP validation [BeforeEach] [secondary-host-eip] Using different methods to disable a node or pod availability for egress IPv4
  /home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/egressip.go:441
  [FAIL] e2e egress IP validation [BeforeEach] [secondary-host-eip] Using different methods to disable a node or pod availability for egress IPv6 uncompressed
  /home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/egressip.go:441
  [FAIL] e2e egress IP validation [BeforeEach] [secondary-host-eip] Using different methods to disable a node or pod availability for egress IPv6 compressed
  /home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/egressip.go:441
  [FAIL] e2e egress IP validation [BeforeEach] [secondary-host-eip] Using different methods to disable a node or pod availability for egress
  /home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/egressip.go:441
  [FAIL] e2e egress IP validation [BeforeEach] [secondary-host-eip] Multiple EgressIP objects and their Egress IP hosted on the same interface

the multi-nic EIP ones are failing...ugh

@tssurya
Copy link
Member Author

tssurya commented Feb 6, 2024

hitting:

2024-02-05T20:52:56.7728279Z   �[38;5;9m[FAILED] failed to isolate IPv6 networks: failed to run "ip6tables -t filter -A FORWARD -i br-9f19c7528247 -o br-71a50b341ee7 -j DROP": exit status 4 (ip6tables v1.8.7 (nf_tables): Could not fetch rule set generation id: Permission denied (you must be root)

on all EIP tests. lmc

@tssurya
Copy link
Member Author

tssurya commented Feb 6, 2024

its related to:

func isolateIPv6Networks(networkA, networkB string) error {
	if containerRuntime != "docker" {
		panic("unsupported container runtime")
	}
	var bridgeInfNames []string
	// docker creates bridges by appending 12 chars from network ID to 'br-'
	bridgeIDLimit := 12
	for _, network := range []string{networkA, networkB} {
		// output will be wrapped in single quotes
		id, err := runCommand(containerRuntime, "inspect", network, "--format", "'{{.Id}}'")
		if err != nil {
			return err
		}
		if len(id) <= bridgeIDLimit+1 {
			return fmt.Errorf("invalid bridge ID %q", id)
		}
		bridgeInfName := fmt.Sprintf("br-%s", id[1:bridgeIDLimit+1])
		// validate bridge exists
		_, err = runCommand("ip", "link", "show", bridgeInfName)
		if err != nil {
			return fmt.Errorf("bridge %q doesnt exist: %v", bridgeInfName, err)
		}
		bridgeInfNames = append(bridgeInfNames, bridgeInfName)
	}
	if len(bridgeInfNames) != 2 {
		return fmt.Errorf("expected two bridge names but found %d", len(bridgeInfNames))
	}
	_, err := runCommand("ip6tables", "-t", "filter", "-A", "FORWARD", "-i", bridgeInfNames[0], "-o", bridgeInfNames[1], "-j", "DROP")
	if err != nil {
		return err
	}
	_, err = runCommand("ip6tables", "-t", "filter", "-A", "FORWARD", "-i", bridgeInfNames[1], "-o", bridgeInfNames[0], "-j", "DROP")
	return err
}

@tssurya
Copy link
Member Author

tssurya commented Feb 6, 2024

fixed failed to isolate IPv6 networks by adding sudo priviledges, icky but we need this for now till we can figure out a better way with github runners.

@tssurya
Copy link
Member Author

tssurya commented Feb 6, 2024

For EgressIPs the IPV6 infrastructure is not healthy. For example, I see healthchecks failing:

W0206 08:55:40.689232      16 egressip_healthcheck.go:162] Could not connect to ovn-worker ([fd00:10:244:2::2]:9107): context deadline exceeded
W0206 08:55:40.689232      16 egressip_healthcheck.go:162] Could not connect to ovn-worker ([fd00:10:244:2::2]:9107): context deadline exceeded
I0206 08:55:40.689275      16 egressip_event_handler.go:124] Node: ovn-worker has been labeled, adding it for egress assignment
W0206 08:55:40.689291      16 egressip_event_handler.go:131] Node: ovn-worker has been labeled, but node is not ready and reachable, cannot use it for egress assignment
W0206 08:55:40.689291      16 egressip_event_handler.go:131] Node: ovn-worker has been labeled, but node is not ready and reachable, cannot use it for egress assignment
I0206 08:55:40.689370      16 obj_retry.go:555] Update event received for resource *v1.Node, old object is equal to new: true
I0206 08:55:40.689467      16 obj_retry.go:555] Update event received for resource *v1.Node, old object is equal to new: true
I0206 08:55:40.689467      16 event.go:298] Event(v1.ObjectReference{Kind:"EgressIP", Namespace:"", Name:"egressip", UID:"", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Warning' reason: 'NoMatchingNodeFound' no assignable nodes for EgressIP: egressip, please tag at least one node with label: k8s.ovn.org/egress-assignable

I will skip them for now but I am going to open cards to fix these

@tssurya
Copy link
Member Author

tssurya commented Feb 6, 2024

LOL! EFW e2e's are failing with v6 for different reasons. In run https://github.com/ovn-org/ovn-kubernetes/actions/runs/7796371932/job/21261668937?pr=4106 if we take 2024-02-06T08:52:04.3321817Z �[0me2e egress firewall policy validation �[0m�[1mShould validate the egress firewall allows inbound connections�[0m

2024-02-06T08:52:08.3739810Z   �[1mSTEP:�[0m Creating the nodePort service �[38;5;243m@ 02/06/24 08:52:08.373�[0m
2024-02-06T08:52:08.3874208Z   �[1mSTEP:�[0m Waiting for the endpoints to pop up �[38;5;243m@ 02/06/24 08:52:08.387�[0m
2024-02-06T08:52:09.3880844Z   Feb  6 08:52:09.387: INFO: Waiting for amount of service:service-for-pods endpoints to be 1
2024-02-06T08:52:09.7190354Z   �[1mSTEP:�[0m Verifying connectivity from pod e2e-egress-fw-pod to external container [fc00:f853:ccd:e793::6]:1234 �[38;5;243m@ 02/06/24 08:52:09.718�[0m
2024-02-06T08:52:09.7193444Z   Feb  6 08:52:09.718: INFO: Running '/usr/local/bin/kubectl --server=https://[::1]:37173 --kubeconfig=/home/runner/ovn.conf --namespace=egress-firewall-policy-9920 exec e2e-egress-fw-pod --container=e2e-egress-fw-pod-container -- nc -vz -w 5 fc00:f853:ccd:e793::6 1234'
2024-02-06T08:52:09.8484416Z   Feb  6 08:52:09.848: INFO: stderr: "Connection to fc00:f853:ccd:e793::6 1234 port [tcp/*] succeeded!\n"
2024-02-06T08:52:09.8485981Z   Feb  6 08:52:09.848: INFO: stdout: ""
2024-02-06T08:52:09.8507235Z   �[1mSTEP:�[0m Verifying connectivity from external container fc00:f853:ccd:e793::6 to nodePort svc [fc00:f853:ccd:e793::2]:31397 �[38;5;243m@ 02/06/24 08:52:09.85�[0m
2024-02-06T08:52:09.8509326Z   Feb  6 08:52:09.850: INFO: Running command [docker exec e2e-egress-fw-external-container nc -vz -w 5 fc00:f853:ccd:e793::2 31397]
2024-02-06T08:52:10.9533836Z   Feb  6 08:52:10.953: INFO: Applying EgressFirewall configuration: [apply --namespace=egress-firewall-policy-9920 -f egress-fw.yml] 
2024-02-06T08:52:10.9536580Z   Feb  6 08:52:10.953: INFO: Running '/usr/local/bin/kubectl --server=https://[::1]:37173 --kubeconfig=/home/runner/ovn.conf --namespace=egress-firewall-policy-9920 apply --namespace=egress-firewall-policy-9920 -f egress-fw.yml'
2024-02-06T08:52:11.0768793Z   Feb  6 08:52:11.076: INFO: stderr: ""
2024-02-06T08:52:11.0769768Z   Feb  6 08:52:11.076: INFO: stdout: "egressfirewall.k8s.ovn.org/default created\n"
2024-02-06T08:52:11.0771737Z   Feb  6 08:52:11.076: INFO: Running '/usr/local/bin/kubectl --server=https://[::1]:37173 --kubeconfig=/home/runner/ovn.conf --namespace=egress-firewall-policy-9920 get egressfirewall default'
2024-02-06T08:52:11.1471062Z   Feb  6 08:52:11.146: INFO: stderr: ""
2024-02-06T08:52:11.1472422Z   Feb  6 08:52:11.146: INFO: stdout: "NAME      EGRESSFIREWALL STATUS\ndefault   \n"
2024-02-06T08:52:13.1474328Z   Feb  6 08:52:13.147: INFO: Running '/usr/local/bin/kubectl --server=https://[::1]:37173 --kubeconfig=/home/runner/ovn.conf --namespace=egress-firewall-policy-9920 get egressfirewall default'
2024-02-06T08:52:13.1995756Z   Feb  6 08:52:13.199: INFO: stderr: ""
2024-02-06T08:52:13.1996606Z   Feb  6 08:52:13.199: INFO: stdout: "NAME      EGRESSFIREWALL STATUS\ndefault   \n"
2024-02-06T08:52:15.2009384Z   Feb  6 08:52:15.200: INFO: Running '/usr/local/bin/kubectl --server=https://[::1]:37173 --kubeconfig=/home/runner/ovn.conf --namespace=egress-firewall-policy-9920 get egressfirewall default'
2024-02-06T08:52:15.2520862Z   Feb  6 08:52:15.251: INFO: stderr: ""
2024-02-06T08:52:15.2522035Z   Feb  6 08:52:15.251: INFO: stdout: "NAME      EGRESSFIREWALL STATUS\ndefault   \n"
2024-02-06T08:52:17.2533242Z   Feb  6 08:52:17.252: INFO: Running '/usr/local/bin/kubectl --server=https://[::1]:37173 --kubeconfig=/home/runner/ovn.conf --namespace=egress-firewall-policy-9920 get egressfirewall default'
2024-02-06T08:52:17.3054684Z   Feb  6 08:52:17.305: INFO: stderr: ""
2024-02-06T08:52:17.3055354Z   Feb  6 08:52:17.305: INFO: stdout: "NAME      EGRESSFIREWALL STATUS\ndefault   \n"
2024-02-06T08:52:19.3060760Z   Feb  6 08:52:19.305: INFO: Running '/usr/local/bin/kubectl --server=https://[::1]:37173 --kubeconfig=/home/runner/ovn.conf --namespace=egress-firewall-policy-9920 get egressfirewall default'
2024-02-06T08:52:19.3604400Z   Feb  6 08:52:19.360: INFO: stderr: ""
2024-02-06T08:52:19.3605051Z   Feb  6 08:52:19.360: INFO: stdout: "NAME      EGRESSFIREWALL STATUS\ndefault   \n"
2024-02-06T08:52:21.3124676Z   �[38;5;9m[FAILED]�[0m in [It] - /home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/egress_firewall.go:547 �[38;5;243m@ 02/06/24 08:52:21.312�[0m
2024-02-06T08:52:21.9984362Z   �[1mSTEP:�[0m dump namespace information after failure �[38;5;243m@ 02/06/24 08:52:21.998�[0m
2024-02-06T08:52:21.9985910Z   �[1mSTEP:�[0m Collecting events from namespace "egress-firewall-policy-9920". �[38;5;243m@ 02/06/24 08:52:21.998�[0m
2024-02-06T08:52:22.0007724Z   �[1mSTEP:�[0m Found 3 events. �[38;5;243m@ 02/06/24 08:52:22�[0m
2024-02-06T08:52:22.0010304Z   Feb  6 08:52:22.000: INFO: At 2024-02-06 08:52:07 +0000 UTC - event for e2e-egress-fw-pod: {kubelet ovn-worker2} Pulled: Container image "registry.k8s.io/e2e-test-images/agnhost:2.26" already present on machine
2024-02-06T08:52:22.0015153Z   Feb  6 08:52:22.000: INFO: At 2024-02-06 08:52:07 +0000 UTC - event for e2e-egress-fw-pod: {kubelet ovn-worker2} Created: Created container e2e-egress-fw-pod-container
2024-02-06T08:52:22.0017526Z   Feb  6 08:52:22.000: INFO: At 2024-02-06 08:52:07 +0000 UTC - event for e2e-egress-fw-pod: {kubelet ovn-worker2} Started: Started container e2e-egress-fw-pod-container
2024-02-06T08:52:22.0030447Z   Feb  6 08:52:22.002: INFO: POD                NODE         PHASE    GRACE  CONDITIONS
2024-02-06T08:52:22.0033385Z   Feb  6 08:52:22.002: INFO: e2e-egress-fw-pod  ovn-worker2  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2024-02-06 08:52:04 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2024-02-06 08:52:08 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2024-02-06 08:52:08 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2024-02-06 08:52:04 +0000 UTC  }]
2024-02-06T08:52:22.0035049Z   Feb  6 08:52:22.002: INFO: 
2024-02-06T08:52:22.0161661Z   Feb  6 08:52:22.016: INFO: 
2024-02-06T08:52:22.0162505Z   Logging node info for node ovn-control-plan

So for some reason around 52 min mark we are not able to grep for the status in EFW. Should we be increasing the timeout period because in IC this just became asynchronous? I cannot find any useful logs in status-manager code that shows the status patch happening. Clearly something went wrong...

I0206 08:22:04.425105      16 clustermanager.go:117] Starting the cluster manager
I0206 08:22:05.238369      16 status_manager.go:196] Starting StatusManager with typed managers: map[adminpolicybasedexternalroutes:0xc000193d00 egressfirewalls:0xc0004ec1c0]

leader election ===>
I0206 09:00:07.459059      17 leaderelection.go:250] attempting to acquire leader lease ovn-kubernetes/ovn-kubernetes-master...
I0206 09:00:07.459144      17 metrics.go:487] Starting metrics server to serve at address "[fc00:f853:ccd:e793::3]:9411"
I0206 09:00:07.467356      17 leaderelection.go:260] successfully acquired lease ovn-kubernetes/ovn-kubernetes-master
I0206 09:00:07.467511      17 ovnkube.go:378] Won leader election; in active mode
I0206 09:00:07.467530      17 event.go:298] Event(v1.ObjectReference{Kind:"Lease", Namespace:"ovn-kubernetes", Name:"ovn-kubernetes-master", UID:"70e62b2d-498a-41d5-b8fb-170a61fd8bcd", APIVersion:"coordination.k8s.io/v1", ResourceVersion:"8621", FieldPath:""}): type: 'Normal' reason: 'LeaderElection' ovn-control-plane became leader
I0206 09:00:07.467929      17 egressservice_cluster.go:97] Setting up event handlers for Egress Services
I0206 09:00:07.468066      17 unidled_annotation_controller.go:35] Setting up event handlers for services
I0206 09:00:07.468223      17 clustermanager.go:117] Starting the cluster manager
I0206 09:00:11.009622      17 status_manager.go:196] Starting StatusManager with typed managers: map[adminpolicybasedexternalroutes:0xc000114ac0 egressfirewalls:0xc000114e00]

Will skip EFW and open a new issue

@tssurya
Copy link
Member Author

tssurya commented Feb 6, 2024

I don't get this:

Summarizing 6 Failures:
  [FAIL] Services when a nodePort service targeting a pod with hostNetwork:true is created when tests are run towards the agnhost echo service [It] queries to the nodePort service shall work for UDP
  /home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/service.go:444
  [FAIL] Services of type NodePort [It] should listen on each host addresses
  /home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/service.go:734
  [FAIL] Service Hairpin SNAT [It] Should ensure service hairpin traffic is NOT SNATed to hairpin masquerade IP; GR LB
  /home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/service.go:907
  [FAIL] e2e EgressQoS validation Should validate correct DSCP value on EgressQoS resource changes [It] ipv6 pod before resource
  /home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/egressqos.go:321
  [FAIL] e2e EgressQoS validation Should validate correct DSCP value on EgressQoS resource changes [It] ipv6 pod after resource
  /home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/egressqos.go:321
  [FAIL] e2e EgressQoS validation Should validate correct DSCP value on pod labels changes [It] ipv6 pod
  /home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/egressqos.go:321

https://github.com/ovn-org/ovn-kubernetes/actions/runs/7799278970/job/21270727244?pr=4106

@tssurya
Copy link
Member Author

tssurya commented Feb 19, 2024

saving flakes: https://github.com/ovn-org/ovn-kubernetes/actions/runs/7933330375/job/21683028471?pr=4106
haven't seen them before.. if I do will create new cards

@tssurya
Copy link
Member Author

tssurya commented Feb 20, 2024

man, LGW is flaking and timing out and exceeding the 3 hours..
Run No1& 2: https://github.com/ovn-org/ovn-kubernetes/actions/runs/7962558670/job/21737185458?pr=4106 and https://github.com/ovn-org/ovn-kubernetes/actions/runs/7962558670/job/21763230994?pr=4106

2024-02-20T13:39:08.2310998Z   Feb 20 13:39:08.230: INFO: Unable to fetch unidling-3488/execpod-noendpointsvqxbr/agnhost-container logs: the server rejected our request for an unknown reason (get pods execpod-noendpointsvqxbr)

what is this? all the unidling stuff is failing :/

startTime: \"2024-02-20T13:29:06Z\"",
2024-02-20T13:34:07.4987786Z           fullStackTrace: "k8s.io/kubernetes/test/e2e/framework/pod.WaitTimeoutForPodRunningInNamespace({0x2b8f810, 0xc000062190}, {0x2bb78c0?, 0xc002e70340?}, {0xc0015445e8, 0x18}, {0xc000ebb610, 0xd}, 0x0?)\n\t/home/runner/go/pkg/mod/k8s.io/kubernetes@v1.28.3/test/e2e/framework/pod/wait.go:459 +0x1a4\nk8s.io/kubernetes/test/e2e/framework/pod.WaitForPodNameRunningInNamespace(...)\n\t/home/runner/go/pkg/mod/k8s.io/kubernetes@v1.28.3/test/e2e/framework/pod/wait.go:443\nk8s.io/kubernetes/test/e2e/framework/pod.CreateExecPodOrFail({0x2b8f810, 0xc000062190}, {0x2bb78c0?, 0xc002e70340}, {0xc000fbc060, 0xd}, {0x276aba9, 0x13}, 0xc0007c1f18)\n\t/home/runner/go/pkg/mod/k8s.io/kubernetes@v1.28.3/test/e2e/framework/pod/resource.go:368 +0x2a6\ngit.luolix.top/ovn-org/ovn-kubernetes/test/e2e.glob..func30.4.1()\n\t/home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/unidling.go:263 +0x345",
2024-02-20T13:34:07.4994497Z       }
2024-02-20T13:34:07.4995029Z   occurred�[0m
2024-02-20T13:34:07.4996945Z   �[38;5;9mIn �[1m[BeforeEach]�[0m�[38;5;9m at: �[1m/home/runner/go/pkg/mod/k8s.io/kubernetes@v1.28.3/test/e2e/framework/pod/resource.go:369�[0m �[38;5;243m@ 02/20/24 13:34:06.444�[0m
2024-02-20T13:34:07.4998340Z 
2024-02-20T13:34:07.4999238Z   There were �[1m�[38;5;9madditional failures�[0m detected.  To view them in detail run �[1mginkgo -vv�[0m
2024-02-20T13:34:07.5000466Z �[38;5;243m------------------------------�[0m
2024-02-20T13:34:07.5003260Z �[0mUnidling �[38;5;243mWith non annotated service �[0m�[1mShould generate a NeedPods event when adding the annotation�[0m
2024-02-20T13:34:07.5004977Z �[38;5;243m/home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/unidling.go:281�[0m
2024-02-20T13:34:07.5006368Z   �[1mSTEP:�[0m Creating a kubernetes client �[38;5;243m@ 02/20/24 13:34:07.446�[0m
2024-02-20T13:34:07.5007443Z   Feb 20 13:34:07.446: INFO: >>> kubeConfig: /home/runner/ovn.conf
2024-02-20T13:34:07.5008842Z   �[1mSTEP:�[0m Building a namespace api object, basename unidling �[38;5;243m@ 02/20/24 13:34:07.447�[0m
2024-02-20T13:34:07.5010433Z   Feb 20 13:34:07.454: INFO: Skipping waiting for service account
2024-02-20T13:34:07.5011995Z   �[1mSTEP:�[0m creating an annotated service with no endpoints and idle annotation �[38;5;243m@ 02/20/24 13:34:07.461�[0m
2024-02-20T13:34:07.5013833Z   �[1mSTEP:�[0m creating execpod-noendpoints on node ovn-control-plane �[38;5;243m@ 02/20/24 13:34:07.474�[0m
2024-02-20T13:34:07.5015093Z   Feb 20 13:34:07.474: INFO: Creating new exec pod
2024-02-20T13:34:07.5020873Z   W0220 13:34:07.482766   61806 warnings.go:70] would violate PodSecurity "restricted:latest": allowPrivilegeEscalation != false (container "agnhost-container" must set securityContext.allowPrivilegeEscalation=false), unrestricted capabilities (container "agnhost-container" must set securityContext.capabilities.drop=["ALL"]), runAsNonRoot != true (pod or container "agnhost-container" must set securityContext.runAsNonRoot=true), seccompProfile (pod or container "agnhost-container" must set securityContext.seccompProfile.type to "RuntimeDefault" or "Localhost")
2024-02-20T13:39:07.4853161Z   Feb 20 13:39:07.484: INFO: Unexpected error occurred: Timed out after 300.000s.
2024-02-20T13:39:07.4854322Z   Expected Pod to be in <v1.PodPhase>: "Running"
2024-02-20T13:39:07.4855227Z   Got instead:
2024-02-20T13:34:06.4558645Z   �[38;5;9m[FAILED]�[0m in [BeforeEach] - /home/runner/go/pkg/mod/k8s.io/kubernetes@v1.28.3/test/e2e/framework/pod/resource.go:369 �[38;5;243m@ 02/20/24 13:34:06.444�[0m
2024-02-20T13:34:07.2127107Z   �[1mSTEP:�[0m dump namespace information after failure �[38;5;243m@ 02/20/24 13:34:07.212�[0m
2024-02-20T13:34:07.2128669Z   �[1mSTEP:�[0m Collecting events from namespace "unidling-2516". �[38;5;243m@ 02/20/24 13:34:07.212�[0m
2024-02-20T13:34:07.2161079Z   �[1mSTEP:�[0m Found 2 events. �[38;5;243m@ 02/20/24 13:34:07.215�[0m
2024-02-20T13:34:07.2166004Z   Feb 20 13:34:07.215: INFO: At 2024-02-20 13:29:06 +0000 UTC - event for execpod-noendpointsjwpzw: {kubelet ovn-control-plane} FailedCreatePodSandBox: Failed to create pod sandbox: rpc error: code = Unknown desc = failed to setup network for sandbox "3757e6ea86f4c0bf78b4d52dd3448a1a98d12d870f2d52b676044a19205e9349": plugin type="ovn-k8s-cni-overlay" name="ovn-kubernetes" failed (add): failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused
2024-02-20T13:34:07.2177626Z   Feb 20 13:34:07.215: INFO: At 2024-02-20 13:29:07 +0000 UTC - event for execpod-noendpointsjwpzw: {kubelet ovn-control-plane} SandboxChanged: Pod sandbox changed, it will be killed and re-created.
2024-02-20T13:34:07.2179703Z   Feb 20 13:34:07.217: INFO: POD                       NODE               PHASE    GRACE  CONDITIONS
2024-02-20T13:34:07.2185086Z   Feb 20 13:34:07.217: INFO: execpod-noendpointsjwpzw  ovn-control-plane  Pending         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2024-02-20 13:29:06 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2024-02-20 13:29:06 +0000 UTC ContainersNotReady containers with unready status: [agnhost-container]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2024-02-20 13:29:06 +0000 UTC ContainersNotReady containers with unready status: [agnhost-container]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2024-02-20 13:29:06 +0000 UTC  }]
2024-02-20T13:34:07.2188265Z   Feb 20 13:34:07.217: INFO: 
2024-02-20T13:34:07.2271782Z   Feb 20 13:34:07.226: INFO: Unable to fetch unidling-2516/execpod-noendpointsjwpzw/agnhost-container logs: the server rejected our request for an unknown reason (get pods execpod-noendpointsjwpzw)
2024-02-20T13:34:07.2302317Z   Feb 20 13:34:07.230: INFO: 
2024-02-20T13:34:07.2303041Z   Logging node info for node ovn-control-plane

@tssurya
Copy link
Member Author

tssurya commented Feb 20, 2024

CI seems slow...

2024-02-20T11:42:07.2141633Z   Feb 20 11:42:07.214: INFO: Unexpected number of Endpoints, got 2, expected 3
2024-02-20T11:42:08.2127409Z   Feb 20 11:42:08.212: INFO: Waiting for amount of service:nodeportsvclocal endpoints to be 3
2024-02-20T11:42:08.2148223Z   Feb 20 11:42:08.214: INFO: Unexpected number of Endpoints, got 2, expected 3
2024-02-20T11:42:08.2149726Z   Feb 20 11:42:08.214: INFO: Waiting for amount of service:nodeportsvclocal endpoints to be 3
2024-02-20T11:42:08.2168457Z   Feb 20 11:42:08.216: INFO: Unexpected number of Endpoints, got 2, expected 3
2024-02-20T11:42:08.2171224Z   Feb 20 11:42:08.216: INFO: Unexpected error: failed to validate endpoints for service nodeportsvclocal in namespace: nodeport-ingress-test-5102: 
2024-02-20T11:42:08.2172680Z       <wait.errInterrupted>: 
2024-02-20T11:42:08.2173301Z       timed out waiting for the condition
2024-02-20T11:42:08.2173845Z       {
2024-02-20T11:42:08.2174467Z           cause: <*errors.errorString | 0xc0001bdb60>{
2024-02-20T11:42:08.2175404Z               s: "timed out waiting for the condition",
2024-02-20T11:42:08.2176027Z           },
2024-02-20T11:42:08.2176385Z       }
2024-02-20T11:42:08.2177807Z   �[38;5;9m[FAILED]�[0m in [It] - /home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/e2e.go:1385 �[38;5;243m@ 02/20/24 11:42:08.216�[0m
2024-02-20T11:42:09.3897690Z   �[1mSTEP:�[0m dump namespace information after failure �[38;5;243m@ 02/20/24 11:42:09.389�[0m
2024-02-20T11:42:09.3899266Z   �[1mSTEP:�[0m Collecting events from namespace "nodeport-ingress-test-5102". �[38;5;243m@ 02/20/24 11:42:09.389�[0m
2024-02-20T11:42:09.3926823Z   �[1mSTEP:�[0m Found 8 events. �[38;5;243m@ 02/20/24 11:42:09.392�[0m

This is where it starts:

2024-02-20T11:18:28.9749224Z �[0mCreating a static pod on a node �[0m�[1mShould successfully create then remove a static pod�[0m
2024-02-20T11:18:28.9750596Z �[38;5;243m/home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/static_pods.go:82�[0m
2024-02-20T11:18:28.9751451Z   �[1mSTEP:�[0m Creating a kubernetes client �[38;5;243m@ 02/20/24 11:18:28.974�[0m
2024-02-20T11:18:28.9752242Z   Feb 20 11:18:28.974: INFO: >>> kubeConfig: /home/runner/ovn.conf
2024-02-20T11:18:28.9756751Z   �[1mSTEP:�[0m Building a namespace api object, basename staticpods �[38;5;243m@ 02/20/24 11:18:28.975�[0m
2024-02-20T11:18:28.9831947Z   Feb 20 11:18:28.982: INFO: Skipping waiting for service account
2024-02-20T11:18:28.9895631Z   �[1mSTEP:�[0m copying a pod.yaml file into the /etc/kubernetes/manifests dir of a node �[38;5;243m@ 02/20/24 11:18:28.989�[0m
2024-02-20T11:18:28.9897422Z   Feb 20 11:18:28.989: INFO: creating static-pod-ovn-control-plane on node ovn-control-plane
2024-02-20T11:18:28.9899033Z   Feb 20 11:18:28.989: INFO: Running command [docker cp static-pod.yaml ovn-control-plane:/etc/kubernetes/manifests/static-pod.yaml]
2024-02-20T11:18:59.0099733Z   �[38;5;9m[FAILED]�[0m in [It] - /home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/static_pods.go:106 �[38;5;243m@ 02/20/24 11:18:59.009�[0m
2024-02-20T11:18:59.7047798Z   �[1mSTEP:�[0m dump namespace information after failure �[38;5;243m@ 02/20/24 11:18:59.704�[0m
2024-02-20T11:18:59.7055099Z   �[1mSTEP:�[0m Collecting events from namespace "staticpods-3550". �[38;5;243m@ 02/20/24 11:18:59.704�[0m
2024-02-20T11:18:59.7071469Z   �[1mSTEP:�[0m Found 2 events. �[38;5;243m@ 02/20/24 11:18:59.706�[0m
2024-02-20T11:18:59.7075090Z   Feb 20 11:18:59.707: INFO: At 2024-02-20 11:18:29 +0000 UTC - event for static-pod-ovn-control-plane: {kubelet ovn-control-plane} FailedCreatePodSandBox: Failed to create pod sandbox: rpc error: code = Unknown desc = failed to setup network for sandbox "38bcd60cd8636de595ee2ded8d72e2f2a02b28d31cbbd71e1446b99dd621cbd9": plugin type="ovn-k8s-cni-overlay" name="ovn-kubernetes" failed (add): failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused

trozet
trozet previously approved these changes Feb 20, 2024
@trozet
Copy link
Contributor

trozet commented Feb 20, 2024

@tssurya maybe we should identify these flakes that keep happening and skip them as well for now?

@tssurya
Copy link
Member Author

tssurya commented Feb 20, 2024

@tssurya maybe we should identify these flakes that keep happening and skip them as well for now?

yeah I agree
Run No3: https://github.com/ovn-org/ovn-kubernetes/actions/runs/7962558670/job/21772254763?pr=4106

2024-02-20T15:55:40.6001263Z �[38;5;243m------------------------------�[0m
2024-02-20T15:55:40.6001658Z 
2024-02-20T15:55:40.6001944Z �[38;5;9m�[1mSummarizing 7 Failures:�[0m
2024-02-20T15:55:40.6005318Z   �[38;5;9m[FAIL]�[0m �[0me2e br-int flow monitoring export validation �[38;5;243mShould validate flow data of br-int is sent to an external gateway �[38;5;9m�[1m[It] with sflow�[0m
2024-02-20T15:55:40.6007365Z   �[38;5;243m/home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/util.go:1027�[0m
2024-02-20T15:55:40.6009305Z   �[38;5;9m[FAIL]�[0m �[0me2e network policy hairpinning validation �[38;5;9m�[1m[It] Should validate the hairpinned traffic is always allowed�[0m
2024-02-20T15:55:40.6011319Z   �[38;5;243m/home/runner/go/pkg/mod/k8s.io/kubernetes@v1.28.3/test/e2e/framework/pod/pod_client.go:106�[0m
2024-02-20T15:55:40.6013124Z   �[38;5;9m[FAIL]�[0m �[0mMulticast �[38;5;9m�[1m[It] should be able to send multicast UDP traffic between nodes�[0m
2024-02-20T15:55:40.6014907Z   �[38;5;243m/home/runner/go/pkg/mod/k8s.io/kubernetes@v1.28.3/test/e2e/framework/pod/pod_client.go:106�[0m
2024-02-20T15:55:40.6017317Z   �[38;5;9m[FAIL]�[0m �[0mServices �[38;5;9m�[1m[It] Creates a service with session-affinity, and ensures it works after backend deletion�[0m
2024-02-20T15:55:40.6019163Z   �[38;5;243m/home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/service.go:148�[0m
2024-02-20T15:55:40.6021997Z   �[38;5;9m[FAIL]�[0m �[0mServices �[38;5;243mwhen a nodePort service targeting a pod with hostNetwork:false is created �[0mwhen tests are run towards the agnhost echo service �[38;5;9m�[1m[It] queries to the nodePort service shall work for UDP�[0m
2024-02-20T15:55:40.6024581Z   �[38;5;243m/home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/service.go:469�[0m
2024-02-20T15:55:40.6027658Z   �[38;5;9m[FAIL]�[0m �[0mServices �[38;5;243mwhen a nodePort service targeting a pod with hostNetwork:true is created �[0mwhen tests are run towards the agnhost echo service �[38;5;9m�[1m[It] queries to the nodePort service shall work for UDP�[0m
2024-02-20T15:55:40.6030185Z   �[38;5;243m/home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/service.go:469�[0m
2024-02-20T15:55:40.6032438Z   �[38;5;9m[FAIL]�[0m �[0mtest e2e inter-node connectivity between worker nodes �[38;5;9m�[1m[It] Should validate connectivity within a namespace of pods on separate nodes�[0m
2024-02-20T15:55:40.6034344Z   �[38;5;243m/home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/e2e.go:1010�[0m
2024-02-20T15:55:40.6035096Z 
2024-02-20T15:55:40.6035473Z �[38;5;9m�[1mRan 54 of 257 Specs in 4857.886 seconds�[0m
2024-02-20T15:55:40.6037059Z �[38;5;9m�[1mFAIL!�[0m -- �[38;5;10m�[1m47 Passed�[0m | �[38;5;9m�[1m7 Failed�[0m | �[38;5;228m�[1m1 Flaked�[0m | �[38;5;11m�[1m0 Pending�[0m | �[38;5;14m�[1m203 Skipped�[0m
2024-02-20T15:55:40.6038376Z --- FAIL: TestE2E (4857.91s)
2024-02-20T15:55:40.6038811Z FAIL

This set has failed twice now, I am skipping them and opening cards..

@tssurya
Copy link
Member Author

tssurya commented Feb 21, 2024

Run No4: https://github.com/ovn-org/ovn-kubernetes/actions/runs/7962558670/job/21809935569?pr=4106

2024-02-21T12:53:02.8932996Z �[38;5;9m�[1mSummarizing 14 Failures:�[0m
2024-02-21T12:53:02.8935588Z   �[38;5;9m[FAIL]�[0m �[0me2e br-int flow monitoring export validation �[38;5;243mShould validate flow data of br-int is sent to an external gateway �[38;5;9m�[1m[It] with sflow�[0m
2024-02-21T12:53:02.8940082Z   �[38;5;243m/home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/util.go:1027�[0m
2024-02-21T12:53:02.8942361Z   �[38;5;9m[FAIL]�[0m �[0me2e ingress traffic validation �[38;5;243mValidating ingress traffic �[38;5;9m�[1m[It] Should be allowed by nodeport services�[0m
2024-02-21T12:53:02.8943812Z   �[38;5;243m/home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/e2e.go:1174�[0m
2024-02-21T12:53:02.8945713Z   �[38;5;9m[FAIL]�[0m �[0me2e ingress traffic validation �[38;5;243mValidating ingress traffic �[38;5;9m�[1m[It] Should be allowed to node local cluster-networked endpoints by nodeport services with externalTrafficPolicy=local�[0m
2024-02-21T12:53:02.8947077Z   �[38;5;243m/home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/e2e.go:1385�[0m
2024-02-21T12:53:02.8948064Z   �[38;5;9m[FAIL]�[0m �[0mCreating a static pod on a node �[38;5;9m�[1m[It] Should successfully create then remove a static pod�[0m
2024-02-21T12:53:02.8949046Z   �[38;5;243m/home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/static_pods.go:106�[0m
2024-02-21T12:53:02.8950223Z   �[38;5;9m[FAIL]�[0m �[38;5;9m�[1mStatus manager validation [BeforeEach] �[38;5;243mShould validate the egress firewall status when adding an unknown zone�[0m
2024-02-21T12:53:02.8951348Z   �[38;5;243m/home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/status_manager.go:116�[0m
2024-02-21T12:53:02.8952995Z   �[38;5;9m[FAIL]�[0m �[38;5;9m�[1mStatus manager validation [BeforeEach] �[38;5;243mShould validate the egress firewall status when adding a new zone�[0m
2024-02-21T12:53:02.8954317Z   �[38;5;243m/home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/status_manager.go:116�[0m
2024-02-21T12:53:02.8955885Z   �[38;5;9m[FAIL]�[0m �[0me2e EgressQoS validation �[38;5;243mShould validate correct DSCP value on EgressQoS resource changes �[38;5;9m�[1m[It] ipv6 pod before resource�[0m
2024-02-21T12:53:02.8957051Z   �[38;5;243m/home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/egressqos.go:321�[0m
2024-02-21T12:53:02.8958278Z   �[38;5;9m[FAIL]�[0m �[0me2e EgressQoS validation �[38;5;243mShould validate correct DSCP value on EgressQoS resource changes �[38;5;9m�[1m[It] ipv6 pod after resource�[0m
2024-02-21T12:53:02.8959413Z   �[38;5;243m/home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/egressqos.go:321�[0m
2024-02-21T12:53:02.8960656Z   �[38;5;9m[FAIL]�[0m �[0me2e EgressQoS validation �[38;5;243mShould validate correct DSCP value on pod labels changes �[38;5;9m�[1m[It] ipv6 pod�[0m
2024-02-21T12:53:02.8961690Z   �[38;5;243m/home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/egressqos.go:321�[0m
2024-02-21T12:53:02.8962932Z   �[38;5;9m[FAIL]�[0m �[0mServices �[38;5;9m�[1m[It] Creates a service with session-affinity, and ensures it works after backend deletion�[0m
2024-02-21T12:53:02.8964574Z   �[38;5;243m/home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/service.go:148�[0m
2024-02-21T12:53:02.8966393Z   �[38;5;9m[FAIL]�[0m �[0mServices �[38;5;9m�[1mwhen a nodePort service targeting a pod with hostNetwork:false is created [BeforeEach] �[0mwhen tests are run towards the agnhost echo service �[38;5;243mqueries to the nodePort service shall work for TCP�[0m
2024-02-21T12:53:02.8967848Z   �[38;5;243m/home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/service.go:274�[0m
2024-02-21T12:53:02.8969476Z   �[38;5;9m[FAIL]�[0m �[0mServices �[38;5;9m�[1mwhen a nodePort service targeting a pod with hostNetwork:false is created [BeforeEach] �[0mwhen tests are run towards the agnhost echo service �[38;5;243mqueries to the nodePort service shall work for UDP�[0m
2024-02-21T12:53:02.8970900Z   �[38;5;243m/home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/service.go:274�[0m
2024-02-21T12:53:02.8972469Z   �[38;5;9m[FAIL]�[0m �[0mServices �[38;5;243mwhen a nodePort service targeting a pod with hostNetwork:true is created �[0mwhen tests are run towards the agnhost echo service �[38;5;9m�[1m[It] queries to the nodePort service shall work for UDP�[0m
2024-02-21T12:53:02.8974318Z   �[38;5;243m/home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/service.go:469�[0m
2024-02-21T12:53:02.8975867Z   �[38;5;9m[FAIL]�[0m �[0mServices �[38;5;9m�[1m[It] All service features work when manually listening on a non-default address�[0m
2024-02-21T12:53:02.8977101Z   �[38;5;243m/home/runner/go/pkg/mod/k8s.io/kubernetes@v1.28.3/test/e2e/framework/pod/pod_client.go:106�[0m

OK its mostly the same things, EgressQoS on LGW: For sure something is broken here, these used to pass until last week
I am not proceeding to skip the combo of tests that failed this sprint and then the next

@tssurya
Copy link
Member Author

tssurya commented Feb 26, 2024

Run No5: https://github.com/ovn-org/ovn-kubernetes/actions/runs/7962558670/job/21958818526?pr=4106
2024-02-25T23:36:42.3996995Z ##[error]The action has timed out.
sometimes it fails due to timeout of the whole lane actually

"2024-02-25T23:22:55.4292253Z Feb 25 23:22:55.428: INFO: Unable to fetch unidling-6501/execpod-noendpointssx8qp/agnhost-container logs: the server rejected our request for an unknown reason (get pods execpod-noendpointssx8qp)"

@tssurya
Copy link
Member Author

tssurya commented Feb 26, 2024

latest run NO.6: https://github.com/ovn-org/ovn-kubernetes/actions/runs/7962558670/job/21973857074?pr=4106
failed due to

• [FAILED] [1204.888 seconds]
e2e br-int flow monitoring export validation Should validate flow data of br-int is sent to an external gateway [It] with sflow
/home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/e2e.go:1908

  [FAILED] timed out waiting for the condition
  In [It] at: /home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/util.go:1027 @ 02/26/24 10:14:16.298

this is already included in the next list of skips..

@tssurya
Copy link
Member Author

tssurya commented Feb 27, 2024

disabling ANP+ipv6 opened #4142 to track them

@tssurya
Copy link
Member Author

tssurya commented Feb 27, 2024

Run No: 7 https://github.com/ovn-org/ovn-kubernetes/actions/runs/8053130165/job/21995745660?pr=4106#logs

Summarizing 11 Failures:
  [FAIL] e2e br-int flow monitoring export validation Should validate flow data of br-int is sent to an external gateway [It] with sflow
  /home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/util.go:1027
  [FAIL] e2e ingress traffic validation Validating ingress traffic [It] Should be allowed by nodeport services
  /home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/e2e.go:1174
  [FAIL] e2e ingress traffic validation Validating ingress traffic [It] Should be allowed to node local cluster-networked endpoints by nodeport services with externalTrafficPolicy=local
  /home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/e2e.go:1385
  [FAIL] Services [It] Creates a service with session-affinity, and ensures it works after backend deletion
  /home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/service.go:148
  [FAIL] Services when a nodePort service targeting a pod with hostNetwork:false is created when tests are run towards the agnhost echo service [It] queries to the nodePort service shall work for TCP
  /home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/service.go:334
  [FAIL] Services when a nodePort service targeting a pod with hostNetwork:false is created when tests are run towards the agnhost echo service [It] queries to the nodePort service shall work for UDP
  /home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/service.go:449
  [FAIL] Services when a nodePort service targeting a pod with hostNetwork:true is created when tests are run towards the agnhost echo service [It] queries to the nodePort service shall work for UDP
  /home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/service.go:449
  [FAIL] test e2e pod connectivity to host addresses [It] Should validate connectivity from a pod to a non-node host address on same node
  /home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/e2e.go:962
  [FAIL] Status manager validation [BeforeEach] Should validate the egress firewall status when adding an unknown zone
  /home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/status_manager.go:116
  [FAIL] Status manager validation [BeforeEach] Should validate the egress firewall status when adding a new zone
  /home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/status_manager.go:116
  [FAIL] test e2e inter-node connectivity between worker nodes [It] Should validate connectivity within a namespace of pods on separate nodes
  /home/runner/work/ovn-kubernetes/ovn-kubernetes/test/e2e/e2e.go:189

This commit fixes things here and there
in e2e's to make control planes run on
v6 as well. Note that some tests are
skipped since they cannot be run on
github runners with v6 enabled.

NOTE: One of the multicast test is moved
from e2e.go to multicast.go where it belongs.

Signed-off-by: Surya Seetharaman <suryaseetharaman.9@gmail.com>
@tssurya
Copy link
Member Author

tssurya commented Feb 27, 2024

Last run was all green: https://github.com/ovn-org/ovn-kubernetes/actions/runs/8060967448/job/22018507506?pr=4106
Re-running it again to ensure second time also we pass.

@tssurya
Copy link
Member Author

tssurya commented Feb 27, 2024

all right second time green too!
@trozet let's merge this :D

@trozet trozet merged commit b4388c5 into ovn-org:master Feb 27, 2024
34 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
ci-ipv6 Add support for IPV6 e2e's to run upstream
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

4 participants