-
Notifications
You must be signed in to change notification settings - Fork 6
/
nnlr.lua
104 lines (96 loc) · 3.57 KB
/
nnlr.lua
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
require 'nn'
local m = require 'moses'
-----
-- :learningRate() - set relative learning rate for a module
-- Module must have parameters (weight and/or bias)
-----
nn.Module.learningRate = function(self, accessor, value)
assert(type(accessor) == 'string', 'accessor must be a string')
assert(type(value) == 'number', 'value must be a number')
if accessor == 'weight' then
if self.weight then
self.__weightLearningRate = value
else
error('Tried to assign \'weight\' learningRate, but module has no weight')
end
elseif accessor == 'bias' then
if self.bias then
self.__biasLearningRate = value
else
error('Tried to assign \'bias\' learningRate, but module has no bias')
end
else
error('Unknown accessor type (should be \'weight\' or \'bias\')')
end
return self
end
-- Seperate :learningRate() method for containers
nn.Container.learningRate = function(self, accessor, value)
for i, module in ipairs(self.modules) do
if module[accessor] ~= nil then
module:learningRate(accessor, value)
end
end
return self
end
-----
-- :weightDecay() - set relative weight decay for a module
-- Module must have parameters (weight and/or bias)
-----
nn.Module.weightDecay = function(self, accessor, value)
if accessor == 'weight' then
if self.weight then
self.__weightWeightDecay = value
else
error('Tried to assign \'weight\' weightDecay, but module has no weight')
end
elseif accessor == 'bias' then
if self.bias then
self.__biasWeightDecay = value
else
error('Tried to assign \'bias\' weightDecay, but module has no bias')
end
else
error('Unknown accessor type (should be \'weight\' or \'bias\')')
end
return self
end
-- Seperate :weightDecay() method for containers
nn.Container.weightDecay = function(self, accessor, value)
for i, module in ipairs(self.modules) do
if (module[accessor] ~= nil) then
module:weightDecay(accessor, value)
end
end
return self
end
-----
-- :optimConfig() -- similar to :parameters(),
-- but for learningRates and weightDecays
-----
nn.Module.optimConfig = function(self, baseLearningRate, baseWeightDecay)
local weightLearningRates = self.weight and self.weight:clone():fill((self.__weightLearningRate or 1))
local biasLearningRates = self.bias and self.bias:clone():fill((self.__biasLearningRate or 1))
local weightWeightDecays = self.weight and self.weight:clone():fill((self.__weightWeightDecay or 1) * baseWeightDecay)
local biasWeightDecays = self.bias and self.bias:clone():fill((self.__biasWeightDecay or 1) * baseWeightDecay)
return m.compact({weightLearningRates, biasLearningRates}), m.compact({weightWeightDecays, biasWeightDecays})
end
-- Seperate :optimConfig() method for containers
nn.Container.optimConfig = function(self, baseLearningRate, baseWeightDecay)
local learningRates = {}
local weightDecays = {}
for i, module in ipairs(self.modules) do
local moduleLearningRates, moduleWeightDecays = module:optimConfig(baseLearningRate, baseWeightDecay)
table.insert(learningRates, moduleLearningRates)
table.insert(weightDecays, moduleWeightDecays)
end
return m.compact(m.flatten(learningRates)), m.compact(m.flatten(weightDecays))
end
-----
-- :getOptimConfig() -- similar to :getParameters()
-- but for learningRates and weightDecays
-----
nn.Module.getOptimConfig = function(self, baseLearningRate, baseWeightDecay)
local learningRates, weightDecays = self:optimConfig(baseLearningRate, baseWeightDecay)
return nn.Module.flatten(learningRates), nn.Module.flatten(weightDecays)
end