24.02
|
Go to the documentation of this file.
20 template <
typename Functor,
typename ParentDescriptor,
typename armnn::StringMapping::Id DebugString>
22 const ParentDescriptor& desc,
28 template <
typename Functor,
typename ParentDescriptor,
typename armnn::StringMapping::Id DebugString>
31 Execute(m_Data.m_Inputs, m_Data.m_Outputs);
34 template <
typename Functor,
typename ParentDescriptor,
typename armnn::StringMapping::Id DebugString>
42 template <
typename Functor,
typename ParentDescriptor,
typename armnn::StringMapping::Id DebugString>
44 std::vector<ITensorHandle*> inputs, std::vector<ITensorHandle*> outputs)
const
55 std::unique_ptr<Decoder<InType>> input0 = MakeDecoder<InType>(inputInfo0, inputs[0]->
Map());
56 std::unique_ptr<Decoder<InType>> input1 = MakeDecoder<InType>(inputInfo1, inputs[1]->
Map());
57 std::unique_ptr<Encoder<OutType>> output= MakeEncoder<OutType>(outputInfo, outputs[0]->
Map());
@ RefMaximumWorkload_Execute
void ExecuteAsync(ExecutionData &executionData) override
@ RefDivisionWorkload_Execute
void Execute() const override
@ RefAdditionWorkload_Execute
#define ARMNN_SCOPED_PROFILING_EVENT_REF_NAME_GUID(label)
Creates a profiling event that uses GetGuid() and GetName() from the calling class.
@ RefMultiplicationWorkload_Execute
@ RefSubtractionWorkload_Execute
RefElementwiseWorkload(const ParentDescriptor &descriptor, const WorkloadInfo &info)
@ RefMinimumWorkload_Execute
Contains information about TensorInfos of a layer.
const TensorInfo & GetTensorInfo(const ITensorHandle *tensorHandle)
float32 helpers
const TensorShape & GetShape() const
static const StringMapping & Instance()
std::vector< ITensorHandle * > m_Inputs
Copyright (c) 2021 ARM Limited and Contributors.
std::vector< ITensorHandle * > m_Outputs