EchoMimicV3 โ NF4 Quantized
This is a NF4 (NormalFloat4) quantized version of EchoMimicV3.
Quantization Details
| Parameter | Value |
|---|---|
| Quantization Type | NF4 (NormalFloat4) |
| Double Quantization | Yes |
| Compute Dtype | float16 |
| Library | bitsandbytes |
| VRAM Reduction | ~65% (5 GB โ ~1.7 GB) |
Usage
from transformers import BitsAndBytesConfig
import torch
nf4_config = BitsAndBytesConfig(
load_in_4bit=True,
bnb_4bit_quant_type="nf4",
bnb_4bit_use_double_quant=True,
bnb_4bit_compute_dtype=torch.float16
)
- Downloads last month
- -
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support
Model tree for siyah1/EchoMimicV3-NF4
Base model
BadToBest/EchoMimicV3