472,107 Members | 1,280 Online
Bytes | Software Development & Data Engineering Community
Post +

Home Posts Topics Members FAQ

Join Bytes to post your question to a community of 472,107 software developers and data experts.

Zipfile content reading via an iterator?

I'm dealing with several large items that have been zipped up to
get quite impressive compression. However, uncompressed, they're
large enough to thrash my memory to swap and in general do bad
performance-related things. I'm trying to figure out how to
produce a file-like iterator out of the contents of such an item.
>>z = zipfile.zipFile("test.zip")
info = z.getinfo("data.txt")

I need to be able to access multiple files within it, but I can
iterate over each one, only seeing small slices of the file.
Using the read() method triggers the volumnous read. Thus what I
have to do currently:
>>content = z.read("data.txt") # ouch!
>>for row in content.splitlines(): process(row) # pain!
What I'm trying to figure out how to do is something like the
>>for row in z.file_iter("data.txt"): process(row) # aah
to more efficiently handle the huge stream of data.

Am I missing something obvious? It seems like iterating over zip
contents would be a common thing to do (especially when compared
to reading the whole contents...I mean, they're zipped because
they're big! :)

Thanks for any pointers,


Dec 11 '07 #1
0 930

This discussion thread is closed

Replies have been disabled for this discussion.

Similar topics

6 posts views Thread by Bennie | last post: by
2 posts views Thread by Renzo | last post: by
3 posts views Thread by rzed | last post: by
11 posts views Thread by Hari Sekhon | last post: by
8 posts views Thread by =?utf-8?B?5Lq66KiA6JC95pel5piv5aSp5rav77yM5pyb5p6B | last post: by
12 posts views Thread by xamdam | last post: by
5 posts views Thread by Neil Crighton | last post: by
2 posts views Thread by Alan G Isaac | last post: by

By using Bytes.com and it's services, you agree to our Privacy Policy and Terms of Use.

To disable or enable advertisements and analytics tracking please visit the manage ads & tracking page.