This repository has been archived by the owner on Oct 9, 2021. It is now read-only.
-
Notifications
You must be signed in to change notification settings - Fork 18
/
torch_functions.py
128 lines (102 loc) · 4.04 KB
/
torch_functions.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
# TODO(lujiang): put it into the third-party
# MIT License
# Copyright (c) 2018 Nam Vo
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
# copies of the Software, and to permit persons to whom the Software is
# furnished to do so, subject to the following conditions:
# The above copyright notice and this permission notice shall be included in all
# copies or substantial portions of the Software.
"""Metric learning functions.
Codes are modified from:
https://github.com/lugiavn/generalization-dml/blob/master/nams.py
"""
import numpy as np
import torch
import torchvision
def pairwise_distances(x, y=None):
"""
Input: x is a Nxd matrix
y is an optional Mxd matirx
Output: dist is a NxM matrix where dist[i,j] is the square norm between
x[i,:] and y[j,:]
if y is not given then use 'y=x'.
i.e. dist[i,j] = ||x[i,:]-y[j,:]||^2
source:
https://discuss.pytorch.org/t/efficient-distance-matrix-computation/9065/2
"""
x_norm = (x**2).sum(1).view(-1, 1)
if y is not None:
y_t = torch.transpose(y, 0, 1)
y_norm = (y**2).sum(1).view(1, -1)
else:
y_t = torch.transpose(x, 0, 1)
y_norm = x_norm.view(1, -1)
dist = x_norm + y_norm - 2.0 * torch.mm(x, y_t)
# Ensure diagonal is zero if x=y
# if y is None:
# dist = dist - torch.diag(dist.diag)
return torch.clamp(dist, 0.0, np.inf)
class MyTripletLossFunc(torch.autograd.Function):
def __init__(self, triplets):
super(MyTripletLossFunc, self).__init__()
self.triplets = triplets
self.triplet_count = len(triplets)
def forward(self, features):
self.save_for_backward(features)
self.distances = pairwise_distances(features).cpu().numpy()
loss = 0.0
triplet_count = 0.0
correct_count = 0.0
for i, j, k in self.triplets:
w = 1.0
triplet_count += w
loss += w * np.log(1 +
np.exp(self.distances[i, j] - self.distances[i, k]))
if self.distances[i, j] < self.distances[i, k]:
correct_count += 1
loss /= triplet_count
return torch.FloatTensor((loss,))
def backward(self, grad_output):
features, = self.saved_tensors
features_np = features.cpu().numpy()
grad_features = features.clone() * 0.0
grad_features_np = grad_features.cpu().numpy()
for i, j, k in self.triplets:
w = 1.0
f = 1.0 - 1.0 / (
1.0 + np.exp(self.distances[i, j] - self.distances[i, k]))
grad_features_np[i, :] += w * f * (
features_np[i, :] - features_np[j, :]) / self.triplet_count
grad_features_np[j, :] += w * f * (
features_np[j, :] - features_np[i, :]) / self.triplet_count
grad_features_np[i, :] += -w * f * (
features_np[i, :] - features_np[k, :]) / self.triplet_count
grad_features_np[k, :] += -w * f * (
features_np[k, :] - features_np[i, :]) / self.triplet_count
for i in range(features_np.shape[0]):
grad_features[i, :] = torch.from_numpy(grad_features_np[i, :])
grad_features *= float(grad_output.data[0])
return grad_features
class TripletLoss(torch.nn.Module):
"""Class for the triplet loss."""
def __init__(self, pre_layer=None):
super(TripletLoss, self).__init__()
self.pre_layer = pre_layer
def forward(self, x, triplets):
if self.pre_layer is not None:
x = self.pre_layer(x)
loss = MyTripletLossFunc(triplets)(x)
return loss
class NormalizationLayer(torch.nn.Module):
"""Class for normalization layer."""
def __init__(self, normalize_scale=1.0, learn_scale=True):
super(NormalizationLayer, self).__init__()
self.norm_s = float(normalize_scale)
if learn_scale:
self.norm_s = torch.nn.Parameter(torch.FloatTensor((self.norm_s,)))
def forward(self, x):
features = self.norm_s * x / torch.norm(x, dim=1, keepdim=True).expand_as(x)
return features