Skip to content

Multi-GPU and DataParallel #5

@chenjy1

Description

@chenjy1

I have a trouble with multi-GPU training:

graph_map.shape in fn _train_batch file base: torch.Size([2, 5479]) cuda:0
graph_map.shape: in fn forward file solver torch.Size([1, 5479]) cuda:0
graph_map.shape: in fn forward file solver torch.Size([1, 5479]) cuda:1

It seems that the dataparallel divided the graph_map by a wrong way. But it does not work if you simply set dim = 1 for dataparallel obviously.
How can I train the model by multi-GPU?

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions