Skip to content

[DLRM/PyTorch] Should tables labelled "Batch size / GPU" be "Global batch size"? #1039

@rkindi

Description

@rkindi

Related to Model/Framework(s)
DLRM/PyTorch

Describe the bug
In the README for DLRM, the tables for many of the experiments say "Batch size/GPU" with 64k as the value. When I think of batch size / GPU, this makes me think of local batch size. However, this line and this line in combination suggest the global batch size is 64k (i.e. for 8 gpus, the Batch size / GPU should be 8192).

To Reproduce
N/A

Expected behavior
Expect that the batch size/gpu is actually 64k/number of gpus.

Environment
N/A

Metadata

Metadata

Assignees

No one assigned

    Labels

    bugSomething isn't working

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions