diff options
author | Chunseok Lee <chunseok.lee@samsung.com> | 2020-07-30 11:32:26 +0900 |
---|---|---|
committer | Chunseok Lee <chunseok.lee@samsung.com> | 2020-07-30 11:32:26 +0900 |
commit | 05e0ec30a632339a8533082476f27bda31ccde16 (patch) | |
tree | 5f220ac83084fe133ffb08a6a17e99f9bb36ec1c /runtime/onert/backend/cpu/ops/LogSoftMaxLayer.cc | |
parent | e2ef8438a24f7c56a0744eb579a6e293ee2fbf8e (diff) | |
download | nnfw-05e0ec30a632339a8533082476f27bda31ccde16.tar.gz nnfw-05e0ec30a632339a8533082476f27bda31ccde16.tar.bz2 nnfw-05e0ec30a632339a8533082476f27bda31ccde16.zip |
Imported Upstream version 1.7.0upstream/1.7.0
Diffstat (limited to 'runtime/onert/backend/cpu/ops/LogSoftMaxLayer.cc')
-rw-r--r-- | runtime/onert/backend/cpu/ops/LogSoftMaxLayer.cc | 80 |
1 files changed, 80 insertions, 0 deletions
diff --git a/runtime/onert/backend/cpu/ops/LogSoftMaxLayer.cc b/runtime/onert/backend/cpu/ops/LogSoftMaxLayer.cc new file mode 100644 index 000000000..d71e325ac --- /dev/null +++ b/runtime/onert/backend/cpu/ops/LogSoftMaxLayer.cc @@ -0,0 +1,80 @@ +/* + * Copyright (c) 2020 Samsung Electronics Co., Ltd. All Rights Reserved + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +#include "LogSoftMaxLayer.h" + +#include "OperationUtils.h" + +#include <cker/operation/LogSoftMax.h> + +namespace onert +{ +namespace backend +{ +namespace cpu +{ +namespace ops +{ + +LogSoftMaxLayer::LogSoftMaxLayer() : _input(nullptr), _output(nullptr), _beta(0.0), _axis(0) +{ + // DO NOTHING +} + +void LogSoftMaxLayer::logsoftmaxFloat32() +{ + nnfw::cker::SoftmaxParams op_params; + op_params.beta = _beta; + op_params.axis = _axis; + nnfw::cker::LogSoftmax(op_params, getTensorShape(_input), + reinterpret_cast<const float *>(_input->buffer()), getTensorShape(_output), + reinterpret_cast<float *>(_output->buffer())); +} + +void LogSoftMaxLayer::logsoftmaxQuant8() +{ + // NYI +} + +void LogSoftMaxLayer::configure(const Tensor *input, const float beta, const int axis, + Tensor *output) +{ + _input = input; + _output = output; + _beta = beta; + _axis = axis; +} + +void LogSoftMaxLayer::run() +{ + if (_input->data_type() == OperandType::FLOAT32) + { + logsoftmaxFloat32(); + } + else if (_input->data_type() == OperandType::QUANT_UINT8_ASYMM) + { + throw std::runtime_error{"LogSoftmax : NYI"}; + } + else + { + throw std::runtime_error{"LogSoftmax : unsupported data type"}; + } +} + +} // namespace ops +} // namespace cpu +} // namespace backend +} // namespace onert |