Skip to content

Default configuration for the provided checkpoint? #32

@Wuzimeng

Description

@Wuzimeng

Hi, I downloaded the checkpoint file you provided in README and tried to train one more epoch to observe something interesting. However, it raised error indicating that loaded state dict has a different number of parameter groups when calling optimizer.load_state_dict(ckpt["optimizer"]). I wonder whether there is some difference between the code here and that you used to generate model ?

Metadata

Metadata

Assignees

No one assigned

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions