summaryrefslogtreecommitdiff
path: root/examples
diff options
context:
space:
mode:
authorCarl Doersch <cdoersch@cs.cmu.edu>2015-10-06 14:19:59 -0700
committerCarl Doersch <cdoersch@cs.cmu.edu>2015-10-22 08:00:13 -0700
commita52ee656a589313901560c87b65a570ee41c9fee (patch)
treef4b9292ef51c76466da49f8bf07ee44b2c0d8746 /examples
parent2f05b03371e5936a478c7ad2946d0cd7c013920c (diff)
downloadcaffeonacl-a52ee656a589313901560c87b65a570ee41c9fee.tar.gz
caffeonacl-a52ee656a589313901560c87b65a570ee41c9fee.tar.bz2
caffeonacl-a52ee656a589313901560c87b65a570ee41c9fee.zip
Cleanup batch norm layer, include global stats computation
Diffstat (limited to 'examples')
-rw-r--r--examples/cifar10/cifar10_full_sigmoid_train_test.prototxt4
-rw-r--r--examples/cifar10/cifar10_full_sigmoid_train_test_bn.prototxt90
2 files changed, 25 insertions, 69 deletions
diff --git a/examples/cifar10/cifar10_full_sigmoid_train_test.prototxt b/examples/cifar10/cifar10_full_sigmoid_train_test.prototxt
index 6f5bf26b..fba69b81 100644
--- a/examples/cifar10/cifar10_full_sigmoid_train_test.prototxt
+++ b/examples/cifar10/cifar10_full_sigmoid_train_test.prototxt
@@ -176,10 +176,10 @@ layer {
top: "ip1"
param {
lr_mult: 1
- decay_mult: 250
+ decay_mult: 0
}
param {
- lr_mult: 0.2
+ lr_mult: 2
decay_mult: 0
}
inner_product_param {
diff --git a/examples/cifar10/cifar10_full_sigmoid_train_test_bn.prototxt b/examples/cifar10/cifar10_full_sigmoid_train_test_bn.prototxt
index 85c2dffe..1a810751 100644
--- a/examples/cifar10/cifar10_full_sigmoid_train_test_bn.prototxt
+++ b/examples/cifar10/cifar10_full_sigmoid_train_test_bn.prototxt
@@ -12,7 +12,7 @@ layer {
}
data_param {
source: "examples/cifar10/cifar10_train_lmdb"
- batch_size: 111
+ batch_size: 100
backend: LMDB
}
}
@@ -41,21 +41,16 @@ layer {
param {
lr_mult: 1
}
- param {
- lr_mult: 2
- }
convolution_param {
num_output: 32
pad: 2
kernel_size: 5
stride: 1
+ bias_term: false
weight_filler {
type: "gaussian"
std: 0.0001
}
- bias_filler {
- type: "constant"
- }
}
}
layer {
@@ -75,23 +70,14 @@ layer {
type: "BatchNorm"
bottom: "pool1"
top: "bn1"
- bn_param {
- scale_filler {
- type: "constant"
- value: 1
- }
- shift_filler {
- type: "constant"
- value: 0.001
- }
+ param {
+ lr_mult: 0
}
param {
- lr_mult: 1.00001
- decay_mult: 0
+ lr_mult: 0
}
param {
- lr_mult: 1.00001
- decay_mult: 0
+ lr_mult: 0
}
}
@@ -110,50 +96,35 @@ layer {
param {
lr_mult: 1
}
- param {
- lr_mult: 2
- }
convolution_param {
num_output: 32
pad: 2
kernel_size: 5
stride: 1
+ bias_term: false
weight_filler {
type: "gaussian"
std: 0.01
}
- bias_filler {
- type: "constant"
- }
}
}
-
-
layer {
name: "bn2"
type: "BatchNorm"
bottom: "conv2"
top: "bn2"
- bn_param {
- scale_filler {
- type: "constant"
- value: 1
- }
- shift_filler {
- type: "constant"
- value: 0.001
- }
+ param {
+ lr_mult: 0
}
param {
- lr_mult: 1.00001
- decay_mult: 0
+ lr_mult: 0
}
param {
- lr_mult: 1.00001
- decay_mult: 0
+ lr_mult: 0
}
}
+
layer {
name: "Sigmoid2"
type: "Sigmoid"
@@ -176,53 +147,38 @@ layer {
type: "Convolution"
bottom: "pool2"
top: "conv3"
+ param {
+ lr_mult: 1
+ }
convolution_param {
num_output: 64
pad: 2
kernel_size: 5
stride: 1
+ bias_term: false
weight_filler {
type: "gaussian"
std: 0.01
}
- bias_filler {
- type: "constant"
- }
- }
- param {
- lr_mult: 1
}
- param {
- lr_mult: 1
- }
-
}
-
layer {
name: "bn3"
type: "BatchNorm"
bottom: "conv3"
top: "bn3"
- bn_param {
- scale_filler {
- type: "constant"
- value: 1
- }
- shift_filler {
- type: "constant"
- value: 0.001
- }
+ param {
+ lr_mult: 0
}
param {
- lr_mult: 1.00001
- decay_mult: 0
+ lr_mult: 0
}
param {
- lr_mult: 1.00001
- decay_mult: 0
+ lr_mult: 0
}
}
+
layer {
name: "Sigmoid3"
type: "Sigmoid"
@@ -248,10 +204,10 @@ layer {
top: "ip1"
param {
lr_mult: 1
- decay_mult: 250
+ decay_mult: 1
}
param {
- lr_mult: 0.2
+ lr_mult: 1
decay_mult: 0
}
inner_product_param {