-
Notifications
You must be signed in to change notification settings - Fork 218
Open
Labels
documentationImprovements or additions to documentationImprovements or additions to documentation
Description
Can you post a VRAM table for what's required to quant models?
On AutoAWQ I used to be able to quant 70B AWQ models with 2x3090 and 128GB RAM.
I have since upgraded to 3x3090 on the same system and llmcompressor is still OOM...
This isn't really a support request, so much as requesting more documentation on what is required to quant models.
Metadata
Metadata
Assignees
Labels
documentationImprovements or additions to documentationImprovements or additions to documentation