convert batch size to float before torch.std in params reporter #4828
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Summary:
Description
This diff converts the batch size to a float before calculating the standard deviation in the
params_reporter
module. This change is made to ensure accurate calculations, astorch.std
expects a floating-point input.Changes
bench_params_reporter.py
torch.std(torch.tensor([b for bs in batch_size_per_feature_per_rank for b in bs]))
totorch.std(torch.tensor([b for bs in batch_size_per_feature_per_rank for b in bs])).float()
.Reason
The
torch.std
function requires a floating-point tensor to calculate the standard deviation. By converting the batch size to a float, we ensure that the calculation is performed correctly.Differential Revision: D81809491