ArmNN
 25.02
All Classes Namespaces Files Functions Variables Typedefs Enumerations Enumerator Friends Macros Pages
RefDequantizeWorkload.cpp
Go to the documentation of this file.
1 //
2 // Copyright © 2019-2024 Arm Ltd and Contributors. All rights reserved.
3 // SPDX-License-Identifier: MIT
4 //
5 
7 #include "RefWorkloadUtils.hpp"
8 #include "Encoders.hpp"
9 #include "Decoders.hpp"
10 #include "Dequantize.hpp"
11 
12 namespace armnn
13 {
14 
16 {
18 }
19 
20 void RefDequantizeWorkload::Execute(std::vector<ITensorHandle*> inputs, std::vector<ITensorHandle*> outputs) const
21 {
22  ARMNN_SCOPED_PROFILING_EVENT_REF_NAME_GUID("RefDequantizeWorkload_Execute");
23 
24  const TensorInfo& inputInfo = GetTensorInfo(inputs[0]);
25  const TensorInfo& outputInfo = GetTensorInfo(outputs[0]);
26 
27  auto inputDecoder = MakeDecoder<float>(inputInfo, inputs[0]->Map());
28  auto outputEncoder = MakeEncoder<float>(outputInfo, outputs[0]->Map());
29 
30  Dequantize(*inputDecoder, *outputEncoder, inputInfo, outputInfo);
31 }
32 
33 } // namespace armnn
#define ARMNN_SCOPED_PROFILING_EVENT_REF_NAME_GUID(label)
Creates a profiling event that uses GetGuid() and GetName() from the calling class.
QueueDescriptor m_Data
Definition: Workload.hpp:74
Copyright (c) 2021 ARM Limited and Contributors.
float Dequantize(QuantizedType value, float scale, int32_t offset)
Dequantize an 8-bit data type into a floating point data type.
Definition: TypesUtils.cpp:48
const TensorInfo & GetTensorInfo(const ITensorHandle *tensorHandle)
float32 helpers
std::vector< ITensorHandle * > m_Inputs
std::vector< ITensorHandle * > m_Outputs