Replies: 2 comments
-
The code is fine so looks like something else is going on. Did you try running it without |
Beta Was this translation helpful? Give feedback.
0 replies
-
I've tried iterating without the chunked parameter - it doesn't seem to work at all. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
P.S. Please do not attach files as it's considered a security risk. Add code snippets directly in the message body as much as possible.
Hi,
I'm trying to use awswrangler(version 2.2.0) read_parquet function with chunked - but without any success.
I have a lambda on aws that is going to an s3 location to get a parquet file(with 1,000,00 records) and process the records incrementally.
The aws lambda is defined with 1024mb memory and a timeout of 10 minutes.
The lambda runs and gets to the timeout, but it doesn't seem to get into the records.
I've tried printing an index just to see if the loop is entered, but no prints from inside the loop.
I've also tried to use the next() function, reaching timeout and again with no prints from inside the loop:
Increasing the memory and timeout is less relevent cause eventualy I'm expecting bigger parquet files - so I'm trying to understand what is wrong with this basic code for a small file like this?
Any help/suggesstion would be appriciated.
Thanks
Eran
Beta Was this translation helpful? Give feedback.
All reactions