Skip to content

Quantization support #5

@starkdmi

Description

@starkdmi

There is a one line change which allows quantized mlx models in Python version of f5-tts-mlx - source

I tried to patch F5TTS.fromPretrained in Swift to accept quantized weights using quantize(model:,groupSize:,bits:,filter:) on f5tts module but did not succeed.

Any idea how to add support for quantized weights like f5-tts-mlx-4bit and f5-tts-mlx-8bit?

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions