-
Notifications
You must be signed in to change notification settings - Fork 3
Expand file tree
/
Copy pathloss_utils.py
More file actions
58 lines (45 loc) · 1.78 KB
/
loss_utils.py
File metadata and controls
58 lines (45 loc) · 1.78 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
import torch
import torch.nn.functional as F
from torch import nn
def entropy(logit):
logit = logit.mean(dim=0)
logit_ = torch.clamp(logit, min=1e-9)
b = logit_ * torch.log(logit_)
return -b.sum()
def consistency_loss(anchors, neighbors):
b, n = anchors.size()
similarity = torch.bmm(anchors.view(b, 1, n), neighbors.view(b, n, 1)).squeeze()
ones = torch.ones_like(similarity)
consistency_loss = F.binary_cross_entropy(similarity, ones)
return consistency_loss
class DistillLoss(nn.Module):
def __init__(self, class_num, temperature):
super(DistillLoss, self).__init__()
self.class_num = class_num
self.temperature = temperature
self.mask = self.mask_correlated_clusters(class_num).cuda()
self.criterion = nn.CrossEntropyLoss(reduction="sum")
def mask_correlated_clusters(self, class_num):
N = 2 * class_num
mask = torch.ones((N, N))
mask = mask.fill_diagonal_(0)
for i in range(class_num):
mask[i, class_num + i] = 0
mask[class_num + i, i] = 0
mask = mask.bool()
return mask
def forward(self, c_i, c_j):
c_i = c_i.t()
c_j = c_j.t()
N = 2 * self.class_num
c = torch.cat((c_i, c_j), dim=0)
c = F.normalize(c, dim=1)
sim = c @ c.T / self.temperature
sim_i_j = torch.diag(sim, self.class_num)
sim_j_i = torch.diag(sim, -self.class_num)
positive_clusters = torch.cat((sim_i_j, sim_j_i), dim=0).reshape(N, 1)
negative_clusters = sim[self.mask].reshape(N, -1)
labels = torch.zeros(N).to(positive_clusters.device).long()
logits = torch.cat((positive_clusters, negative_clusters), dim=1)
loss = self.criterion(logits, labels) / N
return loss