21.02
|
Neon kernel used to quantize down the int32 accumulator values of GEMMLowp to QASYMM8/QASYMM8_SIGNED. More...
#include <NEGEMMLowpQuantizeDownInt32ScaleKernel.h>
Public Member Functions | |
const char * | name () const override |
Name of the kernel. More... | |
NEGEMMLowpQuantizeDownInt32ScaleKernel () | |
Constructor. More... | |
NEGEMMLowpQuantizeDownInt32ScaleKernel (const NEGEMMLowpQuantizeDownInt32ScaleKernel &)=delete | |
Prevent instances of this class from being copied (As this class contains pointers) More... | |
NEGEMMLowpQuantizeDownInt32ScaleKernel & | operator= (const NEGEMMLowpQuantizeDownInt32ScaleKernel &)=delete |
Prevent instances of this class from being copied (As this class contains pointers) More... | |
NEGEMMLowpQuantizeDownInt32ScaleKernel (NEGEMMLowpQuantizeDownInt32ScaleKernel &&)=default | |
Allow instances of this class to be moved. More... | |
NEGEMMLowpQuantizeDownInt32ScaleKernel & | operator= (NEGEMMLowpQuantizeDownInt32ScaleKernel &&)=default |
Allow instances of this class to be moved. More... | |
~NEGEMMLowpQuantizeDownInt32ScaleKernel ()=default | |
Default destructor. More... | |
void | configure (const ITensor *input, const ITensor *bias, ITensor *output, const GEMMLowpOutputStageInfo *output_stage) |
Initialise the kernel's input and output. More... | |
void | run (const Window &window, const ThreadInfo &info) override |
Execute the kernel on the passed window. More... | |
Public Member Functions inherited from ICPPKernel | |
virtual | ~ICPPKernel ()=default |
Default destructor. More... | |
virtual void | run_nd (const Window &window, const ThreadInfo &info, const Window &thread_locator) |
legacy compatibility layer for implemantions which do not support thread_locator In these cases we simply narrow the interface down the legacy version More... | |
virtual void | run_op (ITensorPack &tensors, const Window &window, const ThreadInfo &info) |
Execute the kernel on the passed window. More... | |
Public Member Functions inherited from IKernel | |
IKernel () | |
Constructor. More... | |
virtual | ~IKernel ()=default |
Destructor. More... | |
virtual bool | is_parallelisable () const |
Indicates whether or not the kernel is parallelisable. More... | |
virtual BorderSize | border_size () const |
The size of the border for that kernel. More... | |
const Window & | window () const |
The maximum window the kernel can be executed on. More... | |
Static Public Member Functions | |
static Status | validate (const ITensorInfo *input, const ITensorInfo *bias, const ITensorInfo *output, const GEMMLowpOutputStageInfo *output_stage) |
Static function to check if given info will lead to a valid configuration of NEGEMMLowpQuantizeDownInt32ScaleKernel. More... | |
Neon kernel used to quantize down the int32 accumulator values of GEMMLowp to QASYMM8/QASYMM8_SIGNED.
This kernel takes a final int32 accumulator value (the output of NEGEMMLowpMatrixMultiplyKernel), and processes it to obtain the final QASYMM8/QASYMM8_SIGNED value. The following computations will be performed by the kernel:
Definition at line 48 of file NEGEMMLowpQuantizeDownInt32ScaleKernel.h.
Constructor.
Definition at line 258 of file NEGEMMLowpQuantizeDownInt32ScaleKernel.cpp.
Referenced by NEGEMMLowpQuantizeDownInt32ScaleKernel::name().
|
delete |
Prevent instances of this class from being copied (As this class contains pointers)
Allow instances of this class to be moved.
|
default |
Default destructor.
Referenced by NEGEMMLowpQuantizeDownInt32ScaleKernel::name().
void configure | ( | const ITensor * | input, |
const ITensor * | bias, | ||
ITensor * | output, | ||
const GEMMLowpOutputStageInfo * | output_stage | ||
) |
Initialise the kernel's input and output.
[in] | input | Input tensor. Data type supported: S32 |
[in] | bias | Biases tensor. Only shared biases supported and it can be a nullptr if the biases addition is not required. Biases are 1D tensor with dimensions [OFM]. Data type supported: Same as input . |
[out] | output | Output tensor. Data type supported: Data type supported: QASYMM8/QASYMM8_SIGNED |
[out] | output_stage | GEMMLowp output stage metadata. |
Definition at line 263 of file NEGEMMLowpQuantizeDownInt32ScaleKernel.cpp.
References ARM_COMPUTE_ERROR_ON_NULLPTR, ARM_COMPUTE_ERROR_THROW_ON, arm_compute::auto_init_if_empty(), ICloneable< T >::clone(), ITensor::info(), arm_compute::test::validation::input, GEMMLowpOutputStageInfo::output_data_type, and arm_compute::validate_arguments().
Referenced by NEGEMMLowpQuantizeDownInt32ScaleKernel::name().
|
inlineoverridevirtual |
Name of the kernel.
Implements ICPPKernel.
Definition at line 51 of file NEGEMMLowpQuantizeDownInt32ScaleKernel.h.
References NEGEMMLowpQuantizeDownInt32ScaleKernel::configure(), arm_compute::test::validation::info, arm_compute::test::validation::input, NEGEMMLowpQuantizeDownInt32ScaleKernel::NEGEMMLowpQuantizeDownInt32ScaleKernel(), NEGEMMLowpQuantizeDownInt32ScaleKernel::operator=(), NEGEMMLowpQuantizeDownInt32ScaleKernel::run(), NEGEMMLowpQuantizeDownInt32ScaleKernel::validate(), IKernel::window(), and NEGEMMLowpQuantizeDownInt32ScaleKernel::~NEGEMMLowpQuantizeDownInt32ScaleKernel().
|
delete |
Prevent instances of this class from being copied (As this class contains pointers)
Referenced by NEGEMMLowpQuantizeDownInt32ScaleKernel::name().
|
default |
Allow instances of this class to be moved.
|
overridevirtual |
Execute the kernel on the passed window.
[in] | window | Region on which to execute the kernel. (Must be a region of the window returned by window()) |
[in] | info | Info about executing thread and CPU. |
Reimplemented from ICPPKernel.
Definition at line 315 of file NEGEMMLowpQuantizeDownInt32ScaleKernel.cpp.
References ARM_COMPUTE_ERROR_ON_INVALID_SUBWINDOW, ARM_COMPUTE_ERROR_ON_UNCONFIGURED_KERNEL, ARM_COMPUTE_UNUSED, and IKernel::window().
Referenced by arm_compute::finalize_quantization(), and NEGEMMLowpQuantizeDownInt32ScaleKernel::name().
|
static |
Static function to check if given info will lead to a valid configuration of NEGEMMLowpQuantizeDownInt32ScaleKernel.
[in] | input | Input tensor. Data type supported: S32 |
[in] | bias | Biases tensor. Only shared biases supported and it can be a nullptr if the biases addition is not required. Biases are 1D tensor with dimensions [OFM]. Data type supported: Same as input . |
[in] | output | Output tensor. Data type supported: Data type supported: QASYMM8/QASYMM8_SIGNED |
[out] | output_stage | GEMMLowp output stage metadata. |
Definition at line 307 of file NEGEMMLowpQuantizeDownInt32ScaleKernel.cpp.
References ARM_COMPUTE_ERROR_ON_NULLPTR, ARM_COMPUTE_RETURN_ON_ERROR, and arm_compute::validate_arguments().
Referenced by NEGEMMLowpQuantizeDownInt32ScaleKernel::name(), and NEGEMMLowpOutputStage::validate().