Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

pyexcelerate consumes lot of memory when huge data needs to be written #57

Open
Anuradha-26 opened this issue Jul 3, 2017 · 11 comments

Comments

@Anuradha-26
Copy link

I have a business need where I have huge data , more than 3 lac rows per worksheet. Writing to xlsx file using pyexcelerate is very fast , but it consumes lot of memory while writing . For me memory is a constraint. So i'm forced to use xlswriter which has constant_memory = True. It will be great if you can add a feature like that to pyexcelerate.
Or can i do something to optimize memory?
I use this call work_book.new_sheet(sheet_name, data=sheet_rows) to write to xlsx file. sheet_rows can be a very huge data structure.

@Anuradha-26 Anuradha-26 changed the title pyexcelerate consumes lot of memory when huge data needs to written pyexcelerate consumes lot of memory when huge data needs to be written Jul 3, 2017
@kevmo314
Copy link
Collaborator

kevmo314 commented Jul 3, 2017

If you're constrained on memory, have you considered writing to csv and then converting to xlsx? We don't have an easy way to bound the memory usage as it would cripple the write speed.

@Anuradha-26
Copy link
Author

I did not use csv as we have some headers with different colors etc. Are you aware how i can convert csv to xlsx in a pythonic way? Thanks for the suggestion. I can try

@kevmo314
Copy link
Collaborator

kevmo314 commented Jul 5, 2017

If you need header styles, then you're out of luck and cannot go via csv. How much memory does your machine have?

@Anuradha-26
Copy link
Author

6GB. Report generation is our last step. All before steps (there are lots of them) take 4GB. Report alone takes 2GB (for 3 lac records in 2 reports each) and just overshoots 6GB and we get Out of Memory error.

@kevmo314
Copy link
Collaborator

kevmo314 commented Jul 5, 2017

What sorts of data are you writing? The library shouldn't be using that much memory unless you're passing objects to it. If you're doing that, you might want to serialize everything to strings before passing it to PyExcelerate.

Also, could you try the latest dev branch? It includes some performance optimizations that may help here.

@Anuradha-26
Copy link
Author

I convert everything to strings before we write report. But we have lot of columns( from A to AB).
There are 4 xls sheets each with 2 lac records.
Sample row:
str1 str1 longstr str1 str2 longstr IntraFrequency 475 ? ? 0 0 0 0 0 0 0 0 0 0 (N/A) 0 (N/A) N/A N/A N/A N/A N/A N/A N/A

Overall report size is 84MB

@kevmo314
Copy link
Collaborator

kevmo314 commented Jul 5, 2017

Yeah I would say try the latest dev branch and see if it helps.

@kevmo314
Copy link
Collaborator

Just checking in, have you had a chance to try the dev branch?

@Anuradha-26
Copy link
Author

Anuradha-26 commented Jul 18, 2017 via email

@yunfan
Copy link

yunfan commented Oct 9, 2019

i'd have same issue here, under the same situation

i think constante memory is really a important option
because for such job, user could wait for extra seconds
but the server cannot provide too many memory than it have.

so i think its better to gave user the choice of which strategy they want
someone need fast, while others need constant memory or constant CPU ?

@emersonbferreira
Copy link

Did you find any solution to this "problem"?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants