Skip to content

VRAM Table #1775

@frenzybiscuit

Description

@frenzybiscuit

Can you post a VRAM table for what's required to quant models?

On AutoAWQ I used to be able to quant 70B AWQ models with 2x3090 and 128GB RAM.

I have since upgraded to 3x3090 on the same system and llmcompressor is still OOM...

This isn't really a support request, so much as requesting more documentation on what is required to quant models.

Metadata

Metadata

Labels

documentationImprovements or additions to documentation

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions