-
Notifications
You must be signed in to change notification settings - Fork 965
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Many disks are shutdown due to heavy workload #4779
Comments
#4749 |
In this writing scenario, maybe turn on maxStageWrite is better |
When one disk cache store is shutdown. Will we lose staged files in that store? |
files still exist. |
Conclusion: There are two scenarios to consider. One is long-term high-load writing, for which writing directly to object storage is recommended (resource issue: disk-cache or object storage bandwidth). The other is short-term high-load writing, for which the concurrency limit can be lifted to redirect and distribute writes to the object storage. This issue will be temporarily closed for now. |
We should enable max-stage-write for this issue, maybe 10 as the default value. |
What happened:
During a heavy IO workload, we observed more than 280 pending write requests to the cache store, which overwhelmed the disk and caused slow IO. The bad disk dictator thus shuts it down forever even it recovers shortly. But worse, we lost the read cache forever.
What you expected to happen:
How to reproduce it (as minimally and precisely as possible):
Anything else we need to know?
Environment:
juicefs --version
) or Hadoop Java SDK version:1.2-beta1
The text was updated successfully, but these errors were encountered: