Re: Help with Streaming and Chunk Processing for Large JSON Data (60 GB) from Kenna API

Liste des GroupesRevenir à cl python 
Sujet : Re: Help with Streaming and Chunk Processing for Large JSON Data (60 GB) from Kenna API
De : list1 (at) *nospam* tompassin.net (Thomas Passin)
Groupes : comp.lang.python
Date : 30. Sep 2024, 18:57:05
Autres entêtes
Message-ID : <mailman.12.1727722015.3018.python-list@python.org>
References : 1 2 3 4 5
User-Agent : Mozilla Thunderbird
On 9/30/2024 1:00 PM, Chris Angelico via Python-list wrote:
On Tue, 1 Oct 2024 at 02:20, Thomas Passin via Python-list
<python-list@python.org> wrote:
>
On 9/30/2024 11:30 AM, Barry via Python-list wrote:
>
>
On 30 Sep 2024, at 06:52, Abdur-Rahmaan Janhangeer via Python-list <python-list@python.org> wrote:
>
>
import polars as pl
pl.read_json("file.json")
>
>
>
This is not going to work unless the computer has a lot more the 60GiB of RAM.
>
As later suggested a streaming parser is required.
>
Streaming won't work because the file is gzipped.  You have to receive
the whole thing before you can unzip it. Once unzipped it will be even
larger, and all in memory.
 Streaming gzip is perfectly possible. You may be thinking of PKZip
which has its EOCD at the end of the file (although it may still be
possible to stream-decompress if you work at it).
 ChrisA
You're right, that's what I was thinking of.

Date Sujet#  Auteur
30 Sep 24 o Re: Help with Streaming and Chunk Processing for Large JSON Data (60 GB) from Kenna API1Thomas Passin

Haut de la page

Les messages affichés proviennent d'usenet.

NewsPortal