Skip to content

4-card A100 training display cuda out of memory #162

@julia-0105

Description

@julia-0105

Hello, I used 4-card A100 to train hunyuan model but cuda out of memory was displayed. I noticed that 8 cards were used in your source code. I would like to ask you:

  1. Can 4-card A100 with a total memory of 80G train hunyuan model in lora mode?
  2. Is the 40GB maximum memory mentioned in GitHub a single card or 8 cards overall memory?
    Appreciate and look forward to your reply 😊

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions