VibeVoice 7B - Low VRAM Quantized Models

Pre-quantized versions of VibeVoice 7B for low VRAM GPUs.

Available Versions

  • 4bit/ - 4-bit quantized model (~8GB VRAM needed)
  • 8bit/ - 8-bit quantized model (~12GB VRAM needed) - NOTE: Removed 8 bit until I can test it again. I'll re-up it soon.

Usage

from vibevoice.modular.modeling_vibevoice_inference import VibeVoiceForConditionalGenerationInference
from vibevoice.processor.vibevoice_processor import VibeVoiceProcessor

# For 4-bit model
model = VibeVoiceForConditionalGenerationInference.from_pretrained(
    "Dannidee/VibeVoice7b-low-vram/4bit",
    device_map='cuda',
    torch_dtype=torch.bfloat16,
)
processor = VibeVoiceProcessor.from_pretrained("Dannidee/VibeVoice7b-low-vram/4bit")

VRAM Requirements

  • 4-bit: ~8 GB total VRAM
  • 8-bit: ~12 GB total VRAM
  • Original: ~19 GB total VRAM

See individual model folders for detailed information.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support