dafaframe overwrite failed due to _temporary clearance on Azure #1361
-
We run a data collection app on Ubuntu box, and do the following to save the result on distributed storage: Previously we saved the result to hdfs on local hadoop cluster, no problem. Recently we moved the storage to Azure Data Storage Gen2, mounted with Blobfuse2. We also set up a RAM disk for blobfuse2 file cache. The data collector runs every 5 minutes and overwrites the previous existing csv file. It works for several rounds, then we got the following excpetion: _Exception in thread xxx: We changed the blobfuse2 filecache timeout from default 120 seconds to 0, the app runs longer time, but eventually we run into the same issue, the _tempoarary under mounted ramdisk cache can not be cleared, triggering application to crash. The questions are: |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 5 replies
-
Hello,
If caching isn’t necessary for your use case, consider using block-cache instead. You can find a configuration example here: sampleBlockCacheConfig. |
Beta Was this translation helpful? Give feedback.
File-Cache: This maintains a local copy of the file and serves all read/write requests from local disk until the timeout expires. This means if application needs to process the same file over and over again this will give you a significant boost in performance as there are not network calls involved here.
Steaming: This has some issues right now so we might just deprecate this soon enough. Avoid using this.
Block-cache: This is our latest offering which provides significant boost over streaming. In file-cache model we maintain a local cache. If you do not have sufficient disk space available to hold the file locally (consider a case where you have 1TB blob and you want to read and process…