-
Notifications
You must be signed in to change notification settings - Fork 1.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Changing the default post-parallel-reads
value
#9074
Comments
I checked few values: |
Thanks for adding your input here @donkabat! @benjaminh83 which values where you using for the @magik6k Not sure if we should wait for more datapoints around this, or make some changes based on the values/tests we already have? |
@rjan90 I was running it at 32: 0 bad |
About my case: |
After yesterday's wd post, sectors: example of error: |
total 25.755 |
Checklist
Ideas
.Lotus component
Improvement Suggestion
Background
windowPoSt workers has a
post-parallel-reads
feature that allows a storage provider to set an upper boundary to how many challenges are read from the storage simultaneously when doing windowPoSt. Currently that value is set to 128.Issue
We have been getting some reports that when you have a full partition this setting can cause very short network timeouts which causes sectors to be skipped (marked as bad), since the challenges for those sectors can´t be read in the short period that the network timeouts. See #8957, especially this comment for the full context of this issue.
The issue with skipped sectors was mitigated by either tuning down the
post-parallel-reads
value or by disabling the windowPoSt PreChecks which are mostly redundant on windowPoSt-workers.Improvement suggestion
I suggest that we reduce the current
post-parallel-reads
value to a more conservative value to mitigate these issues, and rather let SPs fine tune this value higher if their architectures can handle it.A big thanks to @benjaminh83 for very detailed write up of the issues, and investiagtion of the fixes.
The text was updated successfully, but these errors were encountered: