Skip to content

Some problem when fine-tuning Infinity-2B + VAE_d32_reg models #114

@Bamboos2003

Description

@Bamboos2003

google/flan-t5-large, xl, xxl has 780M, 3B, 11B params.
The Infinity uses flan-t5-xl of 3B as vanilla. But My GPU is a single RTX4090 of 24 VRAM, which get OOM error when encoding the texts.
I want to use google/flan-t5-large as text encoder.
But there is an error that the tensor's size doesn't fit the Infinity-2B model's state_dict.

Image
Image

If Infinity-2B model is only compatible with flan-t5-xl and xxl of 2048 dimensions output?
How can I use Infinity-2B model and google/flan-t5-large at the same time?

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions