This repository has been archived by the owner on Nov 19, 2021. It is now read-only.
-
Notifications
You must be signed in to change notification settings - Fork 31
Load runs out of memory #73
Comments
I observed that its the same issue if I run through BulkLoad api with put() |
My team ran into the same issue. I believe the leak is caused when iterating over elements in a tuple to convert. PySequence_GetItem is called which yields a new reference but only the last element in the list is getting deallocated. I ran a few tests using 1 million records, 2 integers each record, memory profiler snips are below:
And the second with the change
The change can be seen in the commit below. Trying to figure this one out is my first dive into C/C++ since school so any feedback or alternative or better solutions would be appreciated. |
Thanks @hiker77 ! |
Sign up for free
to subscribe to this conversation on GitHub.
Already have an account?
Sign in.
giraffez version 2.0.24.2
Teradata Load Utility Version 16.20.00.09 64-Bit
Ubuntu 16.04
4 cores, 16 GB Ram
If I run the giraffez load operation for large CSVs, the process runs out of memory. I see on the command line the usual message (Processed X Rows), and I see in top that memory usage is slowly creeping up. After a while the process maxes out on memory around 16 GB and the giraffez process is killed. Is this expected? Or is there a configuration I am missing?
The text was updated successfully, but these errors were encountered: