blob: 114d3588e19cb72a38363f2a4fe03a349806b318 (
plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
|
/*
* Copyright (c) 2018 Samsung Electronics Co., Ltd. All Rights Reserved
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
/**
* @file TensorSource.h
* @brief This file contains TensorSource class which is inherited from Source class
* @ingroup COM_AI_RUNTIME
*/
#ifndef __INTERNAL_TENSOR_SOURCE_H__
#define __INTERNAL_TENSOR_SOURCE_H__
#include <misc/tensor/Shape.h>
#include <misc/tensor/IndexIterator.h>
#include "internal/Source.h"
#include "internal/Swizzle.h"
#include "internal/nnapi/tensor/Reader.h"
#include "internal/arm_compute/tensor/View.h"
// NOTE TensorSource is much slower than specialized Source(s)
/**
* @brief Class to define constructor and push function
*/
template <typename T> class TensorSource final : public Source
{
public:
/**
* @brief Construct a new TensorSource object with params
* @param [in] shape Shape of tensor
* @param [in] base Base address
* @param [in] size Size of tensor
*/
TensorSource(const nnfw::misc::tensor::Shape &shape, const T *base, const size_t size)
: _shape{shape}, _base{base}, _size{size}
{
// DO NOTHING
}
public:
/**
* @brief Function for pushing tensor
* @param [in] tensor Tensor to be pushed
* @return N/A
*/
void push(::arm_compute::ITensor &tensor) const override
{
const ::internal::nnapi::tensor::Reader<T> from{_shape, _base, _size};
::internal::arm_compute::tensor::View<T> into{&tensor};
::nnfw::misc::tensor::iterate(_shape) << [&](const nnfw::misc::tensor::Index &index_nnapi) {
const auto rank = index_nnapi.rank();
nnfw::misc::tensor::Index index_ACL(rank);
for (uint32_t axis = 0; axis < rank; ++axis)
{
index_ACL.at(ToARMComputeAxis(rank, axis).value()) = index_nnapi.at(axis);
}
into.at(index_ACL) = from.at(index_nnapi);
};
}
private:
const nnfw::misc::tensor::Shape _shape;
const T *const _base;
const size_t _size;
};
#endif // __INTERNAL_TENSOR_SOURCE_H__
|