-
Notifications
You must be signed in to change notification settings - Fork 308
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Handling large number of files #538
Comments
Thank you for opening your first issue in this project! Engagement like this is essential for open source projects! 🤗 |
I created a draft pull request #539 Together with my other commit cnydw/jupyterlab@6e615c0 on the JupyterLab frontend, it could open a folder with 100000 files without problem. The two commits I made are just POC, the API changes can certainly be improved. I think it makes sense to first make the backend API changes in jupyter_server, then propagate the frontend changes to JupyterLab and Jupyter Notebook accordingly. |
hi, any updates on getting this merged? |
At the moment, when a user opens a folder from notebook or jupyterlab, jupyter_server would read all the files inside the folder using
os.lstat
, which is very costly for large number of files.jupyter_server/jupyter_server/services/contents/filemanager.py
Lines 262 to 271 in 51e3ec3
This makes it basically impossible to open a folder with large number of files, the backend would freeze for a long time before being responsive again. And even when the backend returns the data, the frontend would crash due to the rendering of all the files. See jupyterlab/jupyterlab#8700
It would be nice to improve this architecture, using paging or other methods to partially read the files.
The text was updated successfully, but these errors were encountered: