I’ve got a couple hundred GB to download with Google Takeout, so selected the 50GB file sizes but unfortunately the browser crashes at ~46GB. It actually crashes the whole machine (MacOS) with activity monitor showing firefox “using” 46GB of memory.

Is there some weird niche problem I’m running into here? I’d expect firefox to just be streaming the download into its .part file, so keeping the 46GB in memory is odd.

Is there some way to mimic the firefox download with all cookies as a wget/curl? Dev tools let you copy anything in the network console as a curl request, but since this goes straight to the download I don’t think the console sees it.

Honestly any ideas on how to move forward would be appreciated.

Edit:

I ended up using an extension called cliget that does all the “copy as wget” work for me. I added a -c to the wget so I could use the partially downloaded 40GB file and went from there. I think copying the download link would have worked because it’s from some random domain and probably uses a jwt-like auth protocol, but it’s unclear whether it would deny a wget without correct user-agent or other headers. YMMV

  • phatskat@sh.itjust.works
    link
    fedilink
    arrow-up
    1
    ·
    1 year ago

    You could copy a different GET request from the same domain and then grab the URL for the download and replace the one in the GET request with that one

    • this_name@fedia.ioOP
      link
      fedilink
      arrow-up
      1
      ·
      1 year ago

      That’s a neat trick, good call! I ended up using an extension called cliget which seems functional so far.