-
Notifications
You must be signed in to change notification settings - Fork 18
Time-out when creating large zip files #17
Comments
Another option would be to stream the file while zipping. |
i am running into some timeout path issues as well. just curious @ryanlovett i know you are not the maintainer of this but are you using this in your hub setup at Berkeley with success? i'm having some issues now!! |
@lwasser We used to have more trouble with this, but it lessened when user home directories shrunk. I believe this was mostly due to using sparse checkouts in nbgitpuller. Prior to this, ~/.git took up the most space in user home directories. I initially started to patch nbzip with support for streaming, but ended up rewriting it as jupyter-tree-download. We are using it in a few classes. jupyter-archive also has this feature with support for Jupyter Lab. |
@ryanlovett funny you mention nbgitpuller. If the pod is full - and i try to run gitpuller it won't launch. Can you clarify how you handled avoiding large git histories? i only need the files to load i don't need the full history for the hub (student demo notebooks). And i'm also curious if you'd had the issue with pods not loading when they are full because of git puller and how to circumvented that via a check somewhere (if you have done this). BTW thank you for jupyter-tree-download it is working PERFECTLY for the download.!! |
I could have sworn that we either did sparse checkouts, or were doing shallow clones via the depth setting. However our user environment isn't setting this. The repo containing the student materials only goes back to the beginning of the semester so it doesn't have a huge git history. Our original materials repo had grown to at least 500MB at one point. This would explain why ~/.git/ isn't as large as it used to be when we're not altering the nbgitpuller configuration. So one strategy would be to start with a new repo each term. Otherwise there are a number of online recipes for reducing the size of the history in an existing repo. I'm not confident in my git-fu to recommend one recipe over another. Glad to hear that jupyter-tree-download is working for you! Since jupyter-archive is being actively developed and has support for lab, I should probably create a PR for jupyter-archive to have a button in classic notebook, then retire nbzip and jupyter-tree-download.
Our user storage is shared via an nfs volume so we haven't ever had to deal with out-of-disk conditions. But it would make sense for the notebook server to fail to start if it couldn't write to ~/. |
On a hub, if the I/O is sufficiently slow such that the time to complete the zipping process exceeds proxy time-out intervals, the user will get a 504 Gateway Time-out.
One possible solution:
The text was updated successfully, but these errors were encountered: