dim 必须为1, 这样才能满足softmax每一行相加为一。
import torch.nn.functional as F
# 4 samples, 2 output classes
logits = torch.randn(4
dim 必须为1, 这样才能满足softmax每一行相加为一。
import torch.nn.functional as F
# 4 samples, 2 output classes
logits = torch.randn(4