-
Notifications
You must be signed in to change notification settings - Fork 2.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
io: load closed segment 0000000000064324-0000000000064793: found 469 entries (expected 470) #1639
Comments
I have the same problem with my cluster on Rasp4B, so It became one-time usable. Any service restart or node reboot simply destroy cluster. It's probably dependencies issue: |
Similar issue here |
I reinstalled everything, |
@mikeccuk2005 Don't close the issue, the bug isnt fixed yet. |
I've manually recompiled k3s with go-dqlite v1.4.1 - will report back if it resolves the issue after a little while of uptime. |
In the sqlite directory there were snapshot files. /var/lib/rancher/k3s/server/db/ |
Receiving errors like this after a little while:
And @mikeccuk2005 - I think backup of storage and Kubernetes configuration is to be done at another level, storage you can backup from your storage class, if local path, find the path and copy files away. For configuration, just save/export all the yamls. Backing up and restoring using the db files seems unsafe. |
And |
Is there a way to stop k3s gracefully to avoid this problem? |
This specific error should be fixed when using latest go-dqlite, libdqlite and libraft. |
The way to avoid this problem is to upgrade these dependencies. If that still doesn't work for you, please let me know. |
So I've compiled 1 node
2 node
3 node
|
I'm not totally sure about this one, but sounds like either a deployment issue of some kind on your side or k3s/kine not forming the dqlite cluster correctly. Probably something k3s/kine devs should look at and possibly get back to me if they find dqlite bugs or have questions about how to correctly form a cluster. I'd suggest to open a separate issue for this tho, since it seems unrelated to the original one. |
Thanks for the clarification on the unrealted issue @freeekanayaka Regarding the original issue, we've bumped the necessary dependencies here: @mikeccuk2005 or anyone else who hit this issue. Can you attempt reproduce with v1.18.2 (either an RC or the real release once its out). Since dqlite support is still experimental, this won't be consider a blocker for 1.18.2 and thus we won't holding up the release to test it. |
@SteelCrow I believe your issue is more tied to #1391 |
...
Came here because I had the error message: load closed segment 0000000000100224-0000000000100489: found 265 entries (expected 266) Running on RPI-4 buster, Installed a fresh 3 node cluster. Was trying to debug why when node 0 was turned off, everything stopped (a la 1391), so was turning the service on and off. |
Try to remove the file I don't know exactly the internals of k3s, so there might be an issue with it creates the dqlite which makes it not really HA. |
If it's of interest, dqlite may be replaced by etcd in #1770 |
We would like to close out this issue. If you strongly object to this you can leave comments - or if you wish create a separate GH issue. dqlite and raft versions have been bumped. We are having a hard time reproducing this issue now - in fact we have not been able to reproduce the issue. Further, as @jamesorlakin mentioned we are indeed moving to embedded etcd in K3s. You will find in our upcoming v1.19.0 release of K3s that this feature has replaced dqlite. We are planning to release v1.19.0 K3s not later than September 1, 2020 (however this is subject to change). |
This is my 5 times that this happened to me
Random Failures after restart using 'systemctl restart k3s'. Happens after serval days of use, then restart using 'systemctl restart k3s'
Version:
k3s version v1.17.2+k3s1 (cdab19b)
Ubuntu 18.04.3 LTS
K3s arguments:
ExecStart=/usr/local/bin/k3s
server
'--cluster-init'
'--tls-san'
'10.7.55.77'
'K3S_KUBECONFIG_MODE=664' \
What was install on the k3s:
What was configured:
What was the Log:
The text was updated successfully, but these errors were encountered: