diff options
author | honggui <hongguiyao@msn.com> | 2017-10-11 22:43:04 +0800 |
---|---|---|
committer | honggui <hongguiyao@msn.com> | 2017-10-11 22:43:04 +0800 |
commit | c298116bd650a8f256a9b8ec8ee927029a53bee9 (patch) | |
tree | 9520bb71d0f93b5b345993f5d277c59773ca7627 | |
parent | a7de0cf2fdfea97c8c961ee096d2ae456f984191 (diff) | |
download | caffeonacl-c298116bd650a8f256a9b8ec8ee927029a53bee9.tar.gz caffeonacl-c298116bd650a8f256a9b8ec8ee927029a53bee9.tar.bz2 caffeonacl-c298116bd650a8f256a9b8ec8ee927029a53bee9.zip |
revised an compatibility issue with the new ACL 17.09 version. Documents will be updated later on
-rw-r--r-- | src/caffe/acl_layer.cpp | 2 | ||||
-rw-r--r-- | src/caffe/layers/acl_batch_norm_layer.cpp | 6 | ||||
-rw-r--r-- | src/caffe/layers/acl_concat_layer.cpp | 1 |
3 files changed, 5 insertions, 4 deletions
diff --git a/src/caffe/acl_layer.cpp b/src/caffe/acl_layer.cpp index 84f201d3..879b6701 100644 --- a/src/caffe/acl_layer.cpp +++ b/src/caffe/acl_layer.cpp @@ -116,7 +116,7 @@ int BaseTensor<ACLTensor>::tensor_copy(void * mem,bool toTensor) #endif //USE_PROFILING arm_compute::Window window; ACLTensor* tensor=this; - window.use_tensor_dimensions(tensor->info(), /* first_dimension =*/Window::DimY); // Iterate through the rows (not each element) + window.use_tensor_dimensions(tensor->info()->tensor_shape(), /* first_dimension =*/Window::DimY); // Iterate through the rows (not each element) int width = tensor->info()->tensor_shape()[0]; //->dimension(0); //window.x().end() - window.x().start(); // + 1; int height = tensor->info()->tensor_shape()[1]; //->dimension(1); //window.y().end() - window.y().start(); // + 1; int deepth = tensor->info()->tensor_shape()[2]; diff --git a/src/caffe/layers/acl_batch_norm_layer.cpp b/src/caffe/layers/acl_batch_norm_layer.cpp index b117343c..a6bc16d1 100644 --- a/src/caffe/layers/acl_batch_norm_layer.cpp +++ b/src/caffe/layers/acl_batch_norm_layer.cpp @@ -115,13 +115,13 @@ void ACLBatchNormLayer<Dtype>::Forward_cpu(const vector<Blob<Dtype>*>& bottom, Forward_gpu(bottom, top); return; } +#ifdef USE_PROFILING + logtime_util log_time(ACL_BN_INFO); +#endif //USE_PROFILING if (this->force_bypass_acl_path_||!this->use_global_stats_) { BatchNormLayer<Dtype>::Forward_cpu(bottom,top); return; } -#ifdef USE_PROFILING - logtime_util log_time(ACL_BN_INFO); -#endif //USE_PROFILING const Dtype* bottom_data = bottom[0]->cpu_data(); Dtype* top_data = top[0]->mutable_cpu_data(); SetupACLLayer(bottom,top); diff --git a/src/caffe/layers/acl_concat_layer.cpp b/src/caffe/layers/acl_concat_layer.cpp index 4750cb4b..57a14126 100644 --- a/src/caffe/layers/acl_concat_layer.cpp +++ b/src/caffe/layers/acl_concat_layer.cpp @@ -113,6 +113,7 @@ void ACLConcatLayer<Dtype>::Forward_gpu(const vector<Blob<Dtype>*>& bottom, template <typename Dtype> ACLConcatLayer<Dtype>::~ACLConcatLayer() { + if(this->force_bypass_acl_path_)return; for (int i =0; i < cpu_vectors.size(); i ++) { delete cpu_vectors[i]; } |