-
Notifications
You must be signed in to change notification settings - Fork 16
Open
Description
There is a one line change which allows quantized mlx models in Python version of f5-tts-mlx - source
I tried to patch F5TTS.fromPretrained
in Swift to accept quantized weights using quantize(model:,groupSize:,bits:,filter:)
on f5tts module but did not succeed.
Any idea how to add support for quantized weights like f5-tts-mlx-4bit and f5-tts-mlx-8bit?
Metadata
Metadata
Assignees
Labels
No labels