-
Notifications
You must be signed in to change notification settings - Fork 80
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[QST] CPU compression for decompression with high-level interface #86
Comments
This issue has been labeled |
Any update on this? I have a bunch of old data that needs to be compressed and it would be unfortunate to spin up GPUs only for compression |
Hi @technillogue, we're looking at producing a binary that would do this for you (i.e. produce an HLIF buffer using the CPU). Which formats are you interested in? If you have a significant amount of old data, why not do this on GPU? |
Mostly Gdeflate, though Zstd would be good too. Is the HLIF format
proprietary?
I have some terabytes of data that sometimes needs to be loaded very
quickly. Compression is cost-sensitive, not time-sensitive.
…On Tue, Aug 22, 2023, 8:53 PM eschmidt-nvidia ***@***.***> wrote:
Hi @technillogue <https://github.com/technillogue>, we're looking at
producing a binary that would do this for you (i.e. produce an HLIF buffer
using the CPU). Which formats are you interested in?
If you have a significant amount of old data, why not do this on GPU?
—
Reply to this email directly, view it on GitHub
<#86 (comment)>, or
unsubscribe
<https://github.com/notifications/unsubscribe-auth/AAHG4G2LAEAW5XTBPXQJNNDXWT52XANCNFSM6AAAAAAZMHPKC4>
.
You are receiving this because you were mentioned.Message ID:
***@***.***>
|
I see. If I understand correctly, based on my experiments GPU and CPU will provide similar throughput/$. I've tested this on H100 compared to Genoa / SPR CPUs. Given that you're not time sensitive, have you investigated the GDeflate high-compression mode? This could provide cost savings if you're storing the data for a long time. We're looking at adding a similar mode to ZSTD. The format isn't proprietary but we haven't had time to produce a public document that fully describes it. |
Interesting, I'm not sure what prices you have access to that make that work out but I can give it a shot and do compression a little more efficiently As of June, GDeflate high-compression mode was broken for HLIF #81 (comment). I'm mostly interested in compressing model finetunes, not datasets, so almost all gains are from entropy coding and not dictionary compression. When trying the LLIF benchmark entropy-only had about the same compression ratio as high compression. I imagine there's ways to tune specifically entropy coding to be higher compression though. |
I responded to the earlier issue. This should be fixed. Interesting regarding entropy-only being better. Have you tried our ANS, bitcomp, or Cascaded formats? |
This issue has been labeled |
This issue has been labeled |
How can I generate the metadata needed for decompression?
The text was updated successfully, but these errors were encountered: