-
Notifications
You must be signed in to change notification settings - Fork 17
/
NormStabilizer.lua
103 lines (82 loc) · 3.21 KB
/
NormStabilizer.lua
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
------------------------------------------------------------------------
--[[ Norm Stabilization]]
-- Regularizing RNNs by Stabilizing Activations
-- Ref. A: http://arxiv.org/abs/1511.08400
-- For training, this module only works in batch mode.
------------------------------------------------------------------------
local NS, parent = torch.class("nn.NormStabilizer", "nn.AbstractRecurrent")
function NS:__init(beta)
parent.__init(self, nn.CopyGrad())
self.beta = beta or 1
end
function NS:_accGradParameters(input, gradOutput, scale)
-- No parameters to update
end
function NS:_updateOutput(input)
assert(input:dim() == 2)
local output
if self.train ~= false then
local rm = self:getStepModule(self.step)
output = rm:updateOutput(input)
-- in training mode, we also calculate norm of hidden state
rm.norm = rm.norm or output.new()
rm.norm:norm(output, 2, 2)
else
output = self.modules[1]:updateOutput(input)
end
return output
end
-- returns norm-stabilizer loss as defined in ref. A
function NS:updateLoss()
self.loss = 0
self._normsum = self._normsum or self.output.new()
for step=2,self.step-1 do
local rm1 = self:getStepModule(step-1)
local rm2 = self:getStepModule(step)
self._normsum:add(rm1.norm, rm2.norm)
self._normsum:pow(2)
local steploss = self._normsum:mean() -- sizeAverage
self.loss = self.loss + steploss
end
-- the loss is divided by the number of time-steps (but not the gradients)
self.loss = self.beta * self.loss / (self.step-1)
return self.loss
end
function NS:_updateGradInput(input, gradOutput)
-- First grab h[t] :
-- backward propagate through this step
local curStep = self.updateGradInputStep-1
local hiddenModule = self:getStepModule(curStep)
local gradInput = hiddenModule:updateGradInput(input, gradOutput)
assert(curStep < self.step)
-- buffers
self._normsum = self._normsum or self.output.new()
self._gradInput = self._gradInput or self.output.new()
local batchSize = hiddenModule.output:size(1)
-- Add gradient of norm stabilizer cost function directly to respective CopyGrad.gradInput tensors
if curStep > 1 then
-- then grab h[t-1]
local prevHiddenModule = self:getStepModule(curStep - 1)
self._normsum:resizeAs(hiddenModule.norm):copy(hiddenModule.norm)
self._normsum:add(-1, prevHiddenModule.norm)
self._normsum:mul(self.beta*2)
self._normsum:cdiv(hiddenModule.norm)
self._gradInput:mul(hiddenModule.output, 1/batchSize)
self._gradInput:cmul(self._normsum:expandAs(self._gradInput))
hiddenModule.gradInput:add(self._gradInput)
end
if curStep < self.step-1 then
local nextHiddenModule = self:getStepModule(curStep + 1)
self._normsum:resizeAs(hiddenModule.norm):copy(hiddenModule.norm)
self._normsum:add(-1, nextHiddenModule.norm)
self._normsum:mul(self.beta*2)
self._normsum:cdiv(hiddenModule.norm)
self._gradInput:mul(hiddenModule.output, 1/batchSize)
self._gradInput:cmul(self._normsum:expandAs(self._gradInput))
hiddenModule.gradInput:add(self._gradInput)
end
return hiddenModule.gradInput
end
function NS:__tostring__()
return "nn.NormStabilizer"
end