EchoMimicV3 โ€” NF4 Quantized

This is a NF4 (NormalFloat4) quantized version of EchoMimicV3.

Quantization Details

Parameter Value
Quantization Type NF4 (NormalFloat4)
Double Quantization Yes
Compute Dtype float16
Library bitsandbytes
VRAM Reduction ~65% (5 GB โ†’ ~1.7 GB)

Usage

from transformers import BitsAndBytesConfig
import torch

nf4_config = BitsAndBytesConfig(
    load_in_4bit=True,
    bnb_4bit_quant_type="nf4",
    bnb_4bit_use_double_quant=True,
    bnb_4bit_compute_dtype=torch.float16
)

Downloads last month
-
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for siyah1/EchoMimicV3-NF4

Finetuned
(1)
this model