diff options
Diffstat (limited to 'src/caffe/layers')
38 files changed, 74 insertions, 45 deletions
diff --git a/src/caffe/layers/absval_layer.cpp b/src/caffe/layers/absval_layer.cpp index 0d054ee5..5ce28c9e 100644 --- a/src/caffe/layers/absval_layer.cpp +++ b/src/caffe/layers/absval_layer.cpp @@ -10,7 +10,7 @@ template <typename Dtype> void AbsValLayer<Dtype>::LayerSetUp(const vector<Blob<Dtype>*>& bottom, const vector<Blob<Dtype>*>& top) { NeuronLayer<Dtype>::LayerSetUp(bottom, top); - CHECK_NE(top[0], bottom[0]) << this->type_name() << " Layer does not " + CHECK_NE(top[0], bottom[0]) << this->type() << " Layer does not " "allow in-place computation."; } @@ -40,5 +40,6 @@ STUB_GPU(AbsValLayer); #endif INSTANTIATE_CLASS(AbsValLayer); -REGISTER_LAYER_CLASS(ABSVAL, AbsValLayer); +REGISTER_LAYER_CLASS(AbsVal); + } // namespace caffe diff --git a/src/caffe/layers/accuracy_layer.cpp b/src/caffe/layers/accuracy_layer.cpp index 800b848f..3e8df34c 100644 --- a/src/caffe/layers/accuracy_layer.cpp +++ b/src/caffe/layers/accuracy_layer.cpp @@ -64,5 +64,6 @@ void AccuracyLayer<Dtype>::Forward_cpu(const vector<Blob<Dtype>*>& bottom, } INSTANTIATE_CLASS(AccuracyLayer); -REGISTER_LAYER_CLASS(ACCURACY, AccuracyLayer); +REGISTER_LAYER_CLASS(Accuracy); + } // namespace caffe diff --git a/src/caffe/layers/argmax_layer.cpp b/src/caffe/layers/argmax_layer.cpp index 15e199eb..c4040cdc 100644 --- a/src/caffe/layers/argmax_layer.cpp +++ b/src/caffe/layers/argmax_layer.cpp @@ -58,6 +58,6 @@ void ArgMaxLayer<Dtype>::Forward_cpu(const vector<Blob<Dtype>*>& bottom, } INSTANTIATE_CLASS(ArgMaxLayer); -REGISTER_LAYER_CLASS(ARGMAX, ArgMaxLayer); +REGISTER_LAYER_CLASS(ArgMax); } // namespace caffe diff --git a/src/caffe/layers/bnll_layer.cpp b/src/caffe/layers/bnll_layer.cpp index cb3583ae..9ba0ea9a 100644 --- a/src/caffe/layers/bnll_layer.cpp +++ b/src/caffe/layers/bnll_layer.cpp @@ -43,5 +43,6 @@ STUB_GPU(BNLLLayer); #endif INSTANTIATE_CLASS(BNLLLayer); -REGISTER_LAYER_CLASS(BNLL, BNLLLayer); +REGISTER_LAYER_CLASS(BNLL); + } // namespace caffe diff --git a/src/caffe/layers/concat_layer.cpp b/src/caffe/layers/concat_layer.cpp index 42082195..fc88433c 100644 --- a/src/caffe/layers/concat_layer.cpp +++ b/src/caffe/layers/concat_layer.cpp @@ -105,5 +105,6 @@ STUB_GPU(ConcatLayer); #endif INSTANTIATE_CLASS(ConcatLayer); -REGISTER_LAYER_CLASS(CONCAT, ConcatLayer); +REGISTER_LAYER_CLASS(Concat); + } // namespace caffe diff --git a/src/caffe/layers/contrastive_loss_layer.cpp b/src/caffe/layers/contrastive_loss_layer.cpp index 0d0b443b..0692c11c 100644 --- a/src/caffe/layers/contrastive_loss_layer.cpp +++ b/src/caffe/layers/contrastive_loss_layer.cpp @@ -97,5 +97,6 @@ STUB_GPU(ContrastiveLossLayer); #endif INSTANTIATE_CLASS(ContrastiveLossLayer); -REGISTER_LAYER_CLASS(CONTRASTIVE_LOSS, ContrastiveLossLayer); +REGISTER_LAYER_CLASS(ContrastiveLoss); + } // namespace caffe diff --git a/src/caffe/layers/conv_layer.cpp b/src/caffe/layers/conv_layer.cpp index 9fd2fc6a..c0c9f6f3 100644 --- a/src/caffe/layers/conv_layer.cpp +++ b/src/caffe/layers/conv_layer.cpp @@ -79,4 +79,5 @@ STUB_GPU(ConvolutionLayer); #endif INSTANTIATE_CLASS(ConvolutionLayer); + } // namespace caffe diff --git a/src/caffe/layers/data_layer.cpp b/src/caffe/layers/data_layer.cpp index 96964566..227db201 100644 --- a/src/caffe/layers/data_layer.cpp +++ b/src/caffe/layers/data_layer.cpp @@ -129,6 +129,6 @@ void DataLayer<Dtype>::InternalThreadEntry() { } INSTANTIATE_CLASS(DataLayer); -REGISTER_LAYER_CLASS(DATA, DataLayer); +REGISTER_LAYER_CLASS(Data); } // namespace caffe diff --git a/src/caffe/layers/deconv_layer.cpp b/src/caffe/layers/deconv_layer.cpp index 59114f01..e6d65ab5 100644 --- a/src/caffe/layers/deconv_layer.cpp +++ b/src/caffe/layers/deconv_layer.cpp @@ -81,5 +81,6 @@ STUB_GPU(DeconvolutionLayer); #endif INSTANTIATE_CLASS(DeconvolutionLayer); -REGISTER_LAYER_CLASS(DECONVOLUTION, DeconvolutionLayer); +REGISTER_LAYER_CLASS(Deconvolution); + } // namespace caffe diff --git a/src/caffe/layers/dropout_layer.cpp b/src/caffe/layers/dropout_layer.cpp index 8c8936a7..5f81cc1c 100644 --- a/src/caffe/layers/dropout_layer.cpp +++ b/src/caffe/layers/dropout_layer.cpp @@ -73,5 +73,6 @@ STUB_GPU(DropoutLayer); #endif INSTANTIATE_CLASS(DropoutLayer); -REGISTER_LAYER_CLASS(DROPOUT, DropoutLayer); +REGISTER_LAYER_CLASS(Dropout); + } // namespace caffe diff --git a/src/caffe/layers/dummy_data_layer.cpp b/src/caffe/layers/dummy_data_layer.cpp index 15cf5a58..d254eb1f 100644 --- a/src/caffe/layers/dummy_data_layer.cpp +++ b/src/caffe/layers/dummy_data_layer.cpp @@ -93,5 +93,6 @@ void DummyDataLayer<Dtype>::Forward_cpu(const vector<Blob<Dtype>*>& bottom, } INSTANTIATE_CLASS(DummyDataLayer); -REGISTER_LAYER_CLASS(DUMMY_DATA, DummyDataLayer); +REGISTER_LAYER_CLASS(DummyData); + } // namespace caffe diff --git a/src/caffe/layers/eltwise_layer.cpp b/src/caffe/layers/eltwise_layer.cpp index 0c239f4a..bbc34449 100644 --- a/src/caffe/layers/eltwise_layer.cpp +++ b/src/caffe/layers/eltwise_layer.cpp @@ -163,5 +163,6 @@ STUB_GPU(EltwiseLayer); #endif INSTANTIATE_CLASS(EltwiseLayer); -REGISTER_LAYER_CLASS(ELTWISE, EltwiseLayer); +REGISTER_LAYER_CLASS(Eltwise); + } // namespace caffe diff --git a/src/caffe/layers/euclidean_loss_layer.cpp b/src/caffe/layers/euclidean_loss_layer.cpp index d965027f..b539d348 100644 --- a/src/caffe/layers/euclidean_loss_layer.cpp +++ b/src/caffe/layers/euclidean_loss_layer.cpp @@ -54,5 +54,6 @@ STUB_GPU(EuclideanLossLayer); #endif INSTANTIATE_CLASS(EuclideanLossLayer); -REGISTER_LAYER_CLASS(EUCLIDEAN_LOSS, EuclideanLossLayer); +REGISTER_LAYER_CLASS(EuclideanLoss); + } // namespace caffe diff --git a/src/caffe/layers/exp_layer.cpp b/src/caffe/layers/exp_layer.cpp index 92cb5deb..c7e7c60c 100644 --- a/src/caffe/layers/exp_layer.cpp +++ b/src/caffe/layers/exp_layer.cpp @@ -64,5 +64,6 @@ STUB_GPU(ExpLayer); #endif INSTANTIATE_CLASS(ExpLayer); -REGISTER_LAYER_CLASS(EXP, ExpLayer); +REGISTER_LAYER_CLASS(Exp); + } // namespace caffe diff --git a/src/caffe/layers/flatten_layer.cpp b/src/caffe/layers/flatten_layer.cpp index ec43caba..eb7b42bc 100644 --- a/src/caffe/layers/flatten_layer.cpp +++ b/src/caffe/layers/flatten_layer.cpp @@ -34,5 +34,6 @@ STUB_GPU(FlattenLayer); #endif INSTANTIATE_CLASS(FlattenLayer); -REGISTER_LAYER_CLASS(FLATTEN, FlattenLayer); +REGISTER_LAYER_CLASS(Flatten); + } // namespace caffe diff --git a/src/caffe/layers/hdf5_data_layer.cpp b/src/caffe/layers/hdf5_data_layer.cpp index 706c4a18..77555211 100644 --- a/src/caffe/layers/hdf5_data_layer.cpp +++ b/src/caffe/layers/hdf5_data_layer.cpp @@ -121,5 +121,6 @@ STUB_GPU_FORWARD(HDF5DataLayer, Forward); #endif INSTANTIATE_CLASS(HDF5DataLayer); -REGISTER_LAYER_CLASS(HDF5_DATA, HDF5DataLayer); +REGISTER_LAYER_CLASS(HDF5Data); + } // namespace caffe diff --git a/src/caffe/layers/hdf5_output_layer.cpp b/src/caffe/layers/hdf5_output_layer.cpp index 4a72a18a..d2fdeffc 100644 --- a/src/caffe/layers/hdf5_output_layer.cpp +++ b/src/caffe/layers/hdf5_output_layer.cpp @@ -70,5 +70,6 @@ STUB_GPU(HDF5OutputLayer); #endif INSTANTIATE_CLASS(HDF5OutputLayer); -REGISTER_LAYER_CLASS(HDF5_OUTPUT, HDF5OutputLayer); +REGISTER_LAYER_CLASS(HDF5Output); + } // namespace caffe diff --git a/src/caffe/layers/hinge_loss_layer.cpp b/src/caffe/layers/hinge_loss_layer.cpp index 4dfafcc8..a2fb2a18 100644 --- a/src/caffe/layers/hinge_loss_layer.cpp +++ b/src/caffe/layers/hinge_loss_layer.cpp @@ -47,7 +47,7 @@ template <typename Dtype> void HingeLossLayer<Dtype>::Backward_cpu(const vector<Blob<Dtype>*>& top, const vector<bool>& propagate_down, const vector<Blob<Dtype>*>& bottom) { if (propagate_down[1]) { - LOG(FATAL) << this->type_name() + LOG(FATAL) << this->type() << " Layer cannot backpropagate to label inputs."; } if (propagate_down[0]) { @@ -77,5 +77,6 @@ void HingeLossLayer<Dtype>::Backward_cpu(const vector<Blob<Dtype>*>& top, } INSTANTIATE_CLASS(HingeLossLayer); -REGISTER_LAYER_CLASS(HINGE_LOSS, HingeLossLayer); +REGISTER_LAYER_CLASS(HingeLoss); + } // namespace caffe diff --git a/src/caffe/layers/im2col_layer.cpp b/src/caffe/layers/im2col_layer.cpp index 2c4bb902..11222611 100644 --- a/src/caffe/layers/im2col_layer.cpp +++ b/src/caffe/layers/im2col_layer.cpp @@ -88,5 +88,6 @@ STUB_GPU(Im2colLayer); #endif INSTANTIATE_CLASS(Im2colLayer); -REGISTER_LAYER_CLASS(IM2COL, Im2colLayer); +REGISTER_LAYER_CLASS(Im2col); + } // namespace caffe diff --git a/src/caffe/layers/image_data_layer.cpp b/src/caffe/layers/image_data_layer.cpp index ef6a342c..b96be6ad 100644 --- a/src/caffe/layers/image_data_layer.cpp +++ b/src/caffe/layers/image_data_layer.cpp @@ -149,5 +149,6 @@ void ImageDataLayer<Dtype>::InternalThreadEntry() { } INSTANTIATE_CLASS(ImageDataLayer); -REGISTER_LAYER_CLASS(IMAGE_DATA, ImageDataLayer); +REGISTER_LAYER_CLASS(ImageData); + } // namespace caffe diff --git a/src/caffe/layers/infogain_loss_layer.cpp b/src/caffe/layers/infogain_loss_layer.cpp index 8910431d..a1e0b40d 100644 --- a/src/caffe/layers/infogain_loss_layer.cpp +++ b/src/caffe/layers/infogain_loss_layer.cpp @@ -75,11 +75,11 @@ void InfogainLossLayer<Dtype>::Backward_cpu(const vector<Blob<Dtype>*>& top, const vector<bool>& propagate_down, const vector<Blob<Dtype>*>& bottom) { if (propagate_down[1]) { - LOG(FATAL) << this->type_name() + LOG(FATAL) << this->type() << " Layer cannot backpropagate to label inputs."; } if (propagate_down.size() > 2 && propagate_down[2]) { - LOG(FATAL) << this->type_name() + LOG(FATAL) << this->type() << " Layer cannot backpropagate to infogain inputs."; } if (propagate_down[0]) { @@ -106,5 +106,5 @@ void InfogainLossLayer<Dtype>::Backward_cpu(const vector<Blob<Dtype>*>& top, } INSTANTIATE_CLASS(InfogainLossLayer); -REGISTER_LAYER_CLASS(INFOGAIN_LOSS, InfogainLossLayer); +REGISTER_LAYER_CLASS(InfogainLoss); } // namespace caffe diff --git a/src/caffe/layers/inner_product_layer.cpp b/src/caffe/layers/inner_product_layer.cpp index ecb849ba..b1ec6cb2 100644 --- a/src/caffe/layers/inner_product_layer.cpp +++ b/src/caffe/layers/inner_product_layer.cpp @@ -104,5 +104,6 @@ STUB_GPU(InnerProductLayer); #endif INSTANTIATE_CLASS(InnerProductLayer); -REGISTER_LAYER_CLASS(INNER_PRODUCT, InnerProductLayer); +REGISTER_LAYER_CLASS(InnerProduct); + } // namespace caffe diff --git a/src/caffe/layers/lrn_layer.cpp b/src/caffe/layers/lrn_layer.cpp index a878cf84..5e3e7c42 100644 --- a/src/caffe/layers/lrn_layer.cpp +++ b/src/caffe/layers/lrn_layer.cpp @@ -252,5 +252,6 @@ STUB_GPU_BACKWARD(LRNLayer, CrossChannelBackward); #endif INSTANTIATE_CLASS(LRNLayer); -REGISTER_LAYER_CLASS(LRN, LRNLayer); +REGISTER_LAYER_CLASS(LRN); + } // namespace caffe diff --git a/src/caffe/layers/memory_data_layer.cpp b/src/caffe/layers/memory_data_layer.cpp index 613ca2d4..3272b66a 100644 --- a/src/caffe/layers/memory_data_layer.cpp +++ b/src/caffe/layers/memory_data_layer.cpp @@ -72,5 +72,6 @@ void MemoryDataLayer<Dtype>::Forward_cpu(const vector<Blob<Dtype>*>& bottom, } INSTANTIATE_CLASS(MemoryDataLayer); -REGISTER_LAYER_CLASS(MEMORY_DATA, MemoryDataLayer); +REGISTER_LAYER_CLASS(MemoryData); + } // namespace caffe diff --git a/src/caffe/layers/multinomial_logistic_loss_layer.cpp b/src/caffe/layers/multinomial_logistic_loss_layer.cpp index 78a1f60f..4267a594 100644 --- a/src/caffe/layers/multinomial_logistic_loss_layer.cpp +++ b/src/caffe/layers/multinomial_logistic_loss_layer.cpp @@ -41,7 +41,7 @@ void MultinomialLogisticLossLayer<Dtype>::Backward_cpu( const vector<Blob<Dtype>*>& top, const vector<bool>& propagate_down, const vector<Blob<Dtype>*>& bottom) { if (propagate_down[1]) { - LOG(FATAL) << this->type_name() + LOG(FATAL) << this->type() << " Layer cannot backpropagate to label inputs."; } if (propagate_down[0]) { @@ -62,5 +62,6 @@ void MultinomialLogisticLossLayer<Dtype>::Backward_cpu( } INSTANTIATE_CLASS(MultinomialLogisticLossLayer); -REGISTER_LAYER_CLASS(MULTINOMIAL_LOGISTIC_LOSS, MultinomialLogisticLossLayer); +REGISTER_LAYER_CLASS(MultinomialLogisticLoss); + } // namespace caffe diff --git a/src/caffe/layers/mvn_layer.cpp b/src/caffe/layers/mvn_layer.cpp index 104ad95c..b74d7b4f 100644 --- a/src/caffe/layers/mvn_layer.cpp +++ b/src/caffe/layers/mvn_layer.cpp @@ -159,5 +159,6 @@ STUB_GPU(MVNLayer); #endif INSTANTIATE_CLASS(MVNLayer); -REGISTER_LAYER_CLASS(MVN, MVNLayer); +REGISTER_LAYER_CLASS(MVN); + } // namespace caffe diff --git a/src/caffe/layers/pooling_layer.cpp b/src/caffe/layers/pooling_layer.cpp index 2bfbb01f..6f4c69c8 100644 --- a/src/caffe/layers/pooling_layer.cpp +++ b/src/caffe/layers/pooling_layer.cpp @@ -314,5 +314,4 @@ STUB_GPU(PoolingLayer); INSTANTIATE_CLASS(PoolingLayer); - } // namespace caffe diff --git a/src/caffe/layers/power_layer.cpp b/src/caffe/layers/power_layer.cpp index 69bd120e..4fe34c49 100644 --- a/src/caffe/layers/power_layer.cpp +++ b/src/caffe/layers/power_layer.cpp @@ -99,5 +99,6 @@ STUB_GPU(PowerLayer); #endif INSTANTIATE_CLASS(PowerLayer); -REGISTER_LAYER_CLASS(POWER, PowerLayer); +REGISTER_LAYER_CLASS(Power); + } // namespace caffe diff --git a/src/caffe/layers/relu_layer.cpp b/src/caffe/layers/relu_layer.cpp index 7d5e6034..cc00319a 100644 --- a/src/caffe/layers/relu_layer.cpp +++ b/src/caffe/layers/relu_layer.cpp @@ -43,5 +43,4 @@ STUB_GPU(ReLULayer); INSTANTIATE_CLASS(ReLULayer); - } // namespace caffe diff --git a/src/caffe/layers/sigmoid_cross_entropy_loss_layer.cpp b/src/caffe/layers/sigmoid_cross_entropy_loss_layer.cpp index d1e327a5..077d9499 100644 --- a/src/caffe/layers/sigmoid_cross_entropy_loss_layer.cpp +++ b/src/caffe/layers/sigmoid_cross_entropy_loss_layer.cpp @@ -53,7 +53,7 @@ void SigmoidCrossEntropyLossLayer<Dtype>::Backward_cpu( const vector<Blob<Dtype>*>& top, const vector<bool>& propagate_down, const vector<Blob<Dtype>*>& bottom) { if (propagate_down[1]) { - LOG(FATAL) << this->type_name() + LOG(FATAL) << this->type() << " Layer cannot backpropagate to label inputs."; } if (propagate_down[0]) { @@ -75,5 +75,6 @@ STUB_GPU(SigmoidCrossEntropyLossLayer); #endif INSTANTIATE_CLASS(SigmoidCrossEntropyLossLayer); -REGISTER_LAYER_CLASS(SIGMOID_CROSS_ENTROPY_LOSS, SigmoidCrossEntropyLossLayer); +REGISTER_LAYER_CLASS(SigmoidCrossEntropyLoss); + } // namespace caffe diff --git a/src/caffe/layers/sigmoid_cross_entropy_loss_layer.cu b/src/caffe/layers/sigmoid_cross_entropy_loss_layer.cu index d9db4af6..08f7f492 100644 --- a/src/caffe/layers/sigmoid_cross_entropy_loss_layer.cu +++ b/src/caffe/layers/sigmoid_cross_entropy_loss_layer.cu @@ -33,7 +33,7 @@ void SigmoidCrossEntropyLossLayer<Dtype>::Backward_gpu( const vector<Blob<Dtype>*>& top, const vector<bool>& propagate_down, const vector<Blob<Dtype>*>& bottom) { if (propagate_down[1]) { - LOG(FATAL) << this->type_name() + LOG(FATAL) << this->type() << " Layer cannot backpropagate to label inputs."; } if (propagate_down[0]) { diff --git a/src/caffe/layers/silence_layer.cpp b/src/caffe/layers/silence_layer.cpp index 9bd20574..4abf9eff 100644 --- a/src/caffe/layers/silence_layer.cpp +++ b/src/caffe/layers/silence_layer.cpp @@ -22,5 +22,6 @@ STUB_GPU(SilenceLayer); #endif INSTANTIATE_CLASS(SilenceLayer); -REGISTER_LAYER_CLASS(SILENCE, SilenceLayer); +REGISTER_LAYER_CLASS(Silence); + } // namespace caffe diff --git a/src/caffe/layers/slice_layer.cpp b/src/caffe/layers/slice_layer.cpp index 60a5ecfa..46c3acd6 100644 --- a/src/caffe/layers/slice_layer.cpp +++ b/src/caffe/layers/slice_layer.cpp @@ -137,5 +137,6 @@ STUB_GPU(SliceLayer); #endif INSTANTIATE_CLASS(SliceLayer); -REGISTER_LAYER_CLASS(SLICE, SliceLayer); +REGISTER_LAYER_CLASS(Slice); + } // namespace caffe diff --git a/src/caffe/layers/softmax_layer.cpp b/src/caffe/layers/softmax_layer.cpp index c7b09fff..25142fde 100644 --- a/src/caffe/layers/softmax_layer.cpp +++ b/src/caffe/layers/softmax_layer.cpp @@ -92,4 +92,5 @@ STUB_GPU(SoftmaxLayer); #endif INSTANTIATE_CLASS(SoftmaxLayer); + } // namespace caffe diff --git a/src/caffe/layers/softmax_loss_layer.cpp b/src/caffe/layers/softmax_loss_layer.cpp index 14bd3837..bf20b605 100644 --- a/src/caffe/layers/softmax_loss_layer.cpp +++ b/src/caffe/layers/softmax_loss_layer.cpp @@ -14,7 +14,7 @@ void SoftmaxWithLossLayer<Dtype>::LayerSetUp( const vector<Blob<Dtype>*>& bottom, const vector<Blob<Dtype>*>& top) { LossLayer<Dtype>::LayerSetUp(bottom, top); LayerParameter softmax_param(this->layer_param_); - softmax_param.set_type(LayerParameter_LayerType_SOFTMAX); + softmax_param.set_type("Softmax"); softmax_layer_.reset(LayerRegistry<Dtype>::CreateLayer(softmax_param)); softmax_bottom_vec_.clear(); softmax_bottom_vec_.push_back(bottom[0]); @@ -80,7 +80,7 @@ template <typename Dtype> void SoftmaxWithLossLayer<Dtype>::Backward_cpu(const vector<Blob<Dtype>*>& top, const vector<bool>& propagate_down, const vector<Blob<Dtype>*>& bottom) { if (propagate_down[1]) { - LOG(FATAL) << this->type_name() + LOG(FATAL) << this->type() << " Layer cannot backpropagate to label inputs."; } if (propagate_down[0]) { @@ -116,6 +116,6 @@ void SoftmaxWithLossLayer<Dtype>::Backward_cpu(const vector<Blob<Dtype>*>& top, } INSTANTIATE_CLASS(SoftmaxWithLossLayer); -REGISTER_LAYER_CLASS(SOFTMAX_LOSS, SoftmaxWithLossLayer); +REGISTER_LAYER_CLASS(SoftmaxWithLoss); } // namespace caffe diff --git a/src/caffe/layers/split_layer.cpp b/src/caffe/layers/split_layer.cpp index 51ac61f4..d6929b99 100644 --- a/src/caffe/layers/split_layer.cpp +++ b/src/caffe/layers/split_layer.cpp @@ -16,7 +16,7 @@ void SplitLayer<Dtype>::Reshape(const vector<Blob<Dtype>*>& bottom, // the backward pass. (Technically, it should be possible to share the diff // blob of the first split output with the input, but this seems to cause // some strange effects in practice...) - CHECK_NE(top[i], bottom[0]) << this->type_name() << " Layer does not " + CHECK_NE(top[i], bottom[0]) << this->type() << " Layer does not " "allow in-place computation."; top[i]->Reshape(bottom[0]->num(), bottom[0]->channels(), bottom[0]->height(), bottom[0]->width()); @@ -56,5 +56,6 @@ STUB_GPU(SplitLayer); #endif INSTANTIATE_CLASS(SplitLayer); -REGISTER_LAYER_CLASS(SPLIT, SplitLayer); +REGISTER_LAYER_CLASS(Split); + } // namespace caffe diff --git a/src/caffe/layers/threshold_layer.cpp b/src/caffe/layers/threshold_layer.cpp index 9e68c32d..2365e7b9 100644 --- a/src/caffe/layers/threshold_layer.cpp +++ b/src/caffe/layers/threshold_layer.cpp @@ -29,5 +29,6 @@ STUB_GPU_FORWARD(ThresholdLayer, Forward); #endif INSTANTIATE_CLASS(ThresholdLayer); -REGISTER_LAYER_CLASS(THRESHOLD, ThresholdLayer); +REGISTER_LAYER_CLASS(Threshold); + } // namespace caffe diff --git a/src/caffe/layers/window_data_layer.cpp b/src/caffe/layers/window_data_layer.cpp index 83372e07..23ec83d1 100644 --- a/src/caffe/layers/window_data_layer.cpp +++ b/src/caffe/layers/window_data_layer.cpp @@ -463,5 +463,6 @@ void WindowDataLayer<Dtype>::InternalThreadEntry() { } INSTANTIATE_CLASS(WindowDataLayer); -REGISTER_LAYER_CLASS(WINDOW_DATA, WindowDataLayer); +REGISTER_LAYER_CLASS(WindowData); + } // namespace caffe |