Sujet : Re: Help with Streaming and Chunk Processing for Large JSON Data (60 GB) from Kenna API
De : list1 (at) *nospam* tompassin.net (Thomas Passin)
Groupes : comp.lang.pythonDate : 30. Sep 2024, 19:57:05
Autres entêtes
Message-ID : <mailman.12.1727722015.3018.python-list@python.org>
References : 1 2 3 4 5
User-Agent : Mozilla Thunderbird
On 9/30/2024 1:00 PM, Chris Angelico via Python-list wrote:
On Tue, 1 Oct 2024 at 02:20, Thomas Passin via Python-list
<python-list@python.org> wrote:
>
On 9/30/2024 11:30 AM, Barry via Python-list wrote:
>
>
On 30 Sep 2024, at 06:52, Abdur-Rahmaan Janhangeer via Python-list <python-list@python.org> wrote:
>
>
import polars as pl
pl.read_json("file.json")
>
>
>
This is not going to work unless the computer has a lot more the 60GiB of RAM.
>
As later suggested a streaming parser is required.
>
Streaming won't work because the file is gzipped. You have to receive
the whole thing before you can unzip it. Once unzipped it will be even
larger, and all in memory.
Streaming gzip is perfectly possible. You may be thinking of PKZip
which has its EOCD at the end of the file (although it may still be
possible to stream-decompress if you work at it).
ChrisA
You're right, that's what I was thinking of.