17 void FakeQuantization(
const float* inputData,
float* outputData, uint32_t numElements,
float min,
float max)
19 float scale = (max - min) / 255.f;
20 int32_t offset = armnn::numeric_cast<int32_t>((-min * 255.f) / (max - min));
22 for (uint32_t i = 0; i < numElements; i++)
24 outputData[i] =
static_cast<float>(armnn::Quantize<uint8_t>(inputData[i], scale, offset));
35 std::vector<ITensorHandle*> outputs)
const
41 const float* inputData =
reinterpret_cast<const float*
>(inputs[0]->Map());
42 float* outputData =
reinterpret_cast<float*
>(outputs[0]->Map());
45 m_Data.m_Parameters.m_Max);
#define ARMNN_SCOPED_PROFILING_EVENT_REF_NAME_GUID(label)
Creates a profiling event that uses GetGuid() and GetName() from the calling class.
void Execute() const override
unsigned int GetNumElements() const
Copyright (c) 2021 ARM Limited and Contributors.
const TensorInfo & GetTensorInfo(const ITensorHandle *tensorHandle)
float32 helpers
void FakeQuantization(const float *inputData, float *outputData, uint32_t numElements, float min, float max)
std::vector< ITensorHandle * > m_Inputs
std::vector< ITensorHandle * > m_Outputs