You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
ColossalAI/colossalai/nn/metric/accuracy_2d.py

23 lines
524 B

import torch
from colossalai.nn.layer.parallel_2d import reduce_by_batch_2d
from torch import nn
from ._utils import calc_acc
class Accuracy2D(nn.Module):
"""
Accuracy for 2D parallelism
:param logits: predicted labels
:param targets: true labels
"""
def __init__(self):
super().__init__()
def forward(self, logits, targets):
with torch.no_grad():
correct = calc_acc(logits, targets)
correct = reduce_by_batch_2d.apply(correct)
return correct