16 #include <arm_compute/runtime/NEON/functions/NEPoolingLayer.h>
20 using namespace armcomputetensorutils;
26 const arm_compute::TensorInfo aclInputInfo =
27 BuildArmComputeTensorInfo(input, descriptor.
m_DataLayout);
28 const arm_compute::TensorInfo aclOutputInfo =
29 BuildArmComputeTensorInfo(output, descriptor.
m_DataLayout);
31 arm_compute::PoolingLayerInfo layerInfo = BuildArmComputePoolingLayerInfo(descriptor);
33 return arm_compute::NEPoolingLayer::validate(&aclInputInfo, &aclOutputInfo, layerInfo);
48 arm_compute::ITensor& input = PolymorphicDowncast<IAclTensorHandle*>(
m_Data.
m_Inputs[0])->GetTensor();
49 arm_compute::ITensor& output = PolymorphicDowncast<IAclTensorHandle*>(
m_Data.
m_Outputs[0])->GetTensor();
52 input.info()->set_data_layout(aclDataLayout);
53 output.info()->set_data_layout(aclDataLayout);
55 arm_compute::PoolingLayerInfo layerInfo = BuildArmComputePoolingLayerInfo(
m_Data.m_Parameters);
57 auto layer = std::make_unique<arm_compute::NEPoolingLayer>();
58 layer->configure(&input, &output, layerInfo);
59 m_PoolingLayer.reset(layer.release());
65 m_PoolingLayer->run();
#define ARMNN_SCOPED_PROFILING_EVENT_NEON_NAME_GUID(label)
Creates a profiling event that uses GetGuid() and GetName() from the calling class.
#define ARMNN_REPORT_PROFILING_WORKLOAD_DESC(name, desc, infos, guid)
void Execute() const override
NeonPooling2dWorkload(const Pooling2dQueueDescriptor &descriptor, const WorkloadInfo &info)
Copyright (c) 2021 ARM Limited and Contributors.
arm_compute::Status NeonPooling2dWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const Pooling2dDescriptor &descriptor)
A Pooling2dDescriptor for the Pooling2dLayer.
DataLayout m_DataLayout
The data layout to be used (NCHW, NHWC).
std::vector< ITensorHandle * > m_Inputs
std::vector< ITensorHandle * > m_Outputs
void ValidateInputsOutputs(const std::string &descName, unsigned int numExpectedIn, unsigned int numExpectedOut) const
LayerDescriptor m_Parameters
Contains information about TensorInfos of a layer.