ArmNN
 25.11
Loading...
Searching...
No Matches
NeonBatchNormalizationWorkload Class Reference

#include <NeonBatchNormalizationWorkload.hpp>

Inheritance diagram for NeonBatchNormalizationWorkload:
[legend]
Collaboration diagram for NeonBatchNormalizationWorkload:
[legend]

Public Member Functions

 NeonBatchNormalizationWorkload (const BatchNormalizationQueueDescriptor &descriptor, const WorkloadInfo &info)
virtual void Execute () const override
Public Member Functions inherited from NeonBaseWorkload< BatchNormalizationQueueDescriptor >
 NeonBaseWorkload (const BatchNormalizationQueueDescriptor &descriptor, const WorkloadInfo &info)
void ReplaceInputTensorHandle (ITensorHandle *tensorHandle, unsigned int slot) override
void ReplaceOutputTensorHandle (ITensorHandle *tensorHandle, unsigned int slot) override
Public Member Functions inherited from BaseWorkload< BatchNormalizationQueueDescriptor >
 BaseWorkload (const BatchNormalizationQueueDescriptor &descriptor, const WorkloadInfo &info)
virtual const std::string & GetName () const override
void PostAllocationConfigure () override
const BatchNormalizationQueueDescriptorGetData () const
arm::pipe::ProfilingGuid GetGuid () const final
virtual bool SupportsTensorHandleReplacement () const override
Public Member Functions inherited from IWorkload
virtual ~IWorkload ()
virtual void RegisterDebugCallback (const DebugCallbackFunction &)
virtual armnn::Optional< armnn::MemoryRequirementsGetMemoryRequirements ()

Additional Inherited Members

Protected Member Functions inherited from NeonBaseWorkload< BatchNormalizationQueueDescriptor >
virtual void Reconfigure ()
Protected Attributes inherited from BaseWorkload< BatchNormalizationQueueDescriptor >
BatchNormalizationQueueDescriptor m_Data
const arm::pipe::ProfilingGuid m_Guid
const std::string m_Name

Detailed Description

Definition at line 27 of file NeonBatchNormalizationWorkload.hpp.

Constructor & Destructor Documentation

◆ NeonBatchNormalizationWorkload()

Definition at line 59 of file NeonBatchNormalizationWorkload.cpp.

61 : NeonBaseWorkload<BatchNormalizationQueueDescriptor>(descriptor, info)
62{
63 // Report Profiling Details
64 ARMNN_REPORT_PROFILING_WORKLOAD_DESC("NeonBatchNormalizationWorkload_Construct",
65 descriptor.m_Parameters,
66 info,
67 this->GetGuid());
68
69 m_Data.ValidateInputsOutputs("NeonBatchNormalizationWorkload", 1, 1);
70
71 arm_compute::ITensor& input = PolymorphicDowncast<IAclTensorHandle*>(m_Data.m_Inputs[0])->GetTensor();
72 arm_compute::ITensor& output = PolymorphicDowncast<IAclTensorHandle*>(m_Data.m_Outputs[0])->GetTensor();
73
74 arm_compute::DataLayout aclDataLayout = ConvertDataLayout(m_Data.m_Parameters.m_DataLayout);
75 input.info()->set_data_layout(aclDataLayout);
76 output.info()->set_data_layout(aclDataLayout);
77
78 m_Mean = std::make_unique<arm_compute::Tensor>();
79 BuildArmComputeTensor(*m_Mean, m_Data.m_Mean->GetTensorInfo());
80
81 m_Variance = std::make_unique<arm_compute::Tensor>();
82 BuildArmComputeTensor(*m_Variance, m_Data.m_Variance->GetTensorInfo());
83
84 m_Gamma = std::make_unique<arm_compute::Tensor>();
85 BuildArmComputeTensor(*m_Gamma, m_Data.m_Gamma->GetTensorInfo());
86
87 m_Beta = std::make_unique<arm_compute::Tensor>();
88 BuildArmComputeTensor(*m_Beta, m_Data.m_Beta->GetTensorInfo());
89
90 const arm_compute::ActivationLayerInfo activationInfo = ConvertAdditionalInfoToAclActivationLayerInfo(descriptor);
91
92 auto layer = std::make_unique<arm_compute::NEBatchNormalizationLayer>();
93 layer->configure(&input,
94 &output,
95 m_Mean.get(),
96 m_Variance.get(),
97 m_Beta.get(),
98 m_Gamma.get(),
99 m_Data.m_Parameters.m_Eps,
100 activationInfo);
101 m_Layer.reset(layer.release());
102
103 InitializeArmComputeTensorData(*m_Mean, m_Data.m_Mean);
104 InitializeArmComputeTensorData(*m_Variance, m_Data.m_Variance);
105 InitializeArmComputeTensorData(*m_Gamma, m_Data.m_Gamma);
106 InitializeArmComputeTensorData(*m_Beta, m_Data.m_Beta);
107
108 // Force Compute Library to perform the necessary copying and reshaping, after which
109 // delete all the input tensors that will no longer be needed
110 m_Layer->prepare();
111 FreeUnusedTensors();
112}
#define ARMNN_REPORT_PROFILING_WORKLOAD_DESC(name, desc, infos, guid)
arm_compute::ActivationLayerInfo ConvertAdditionalInfoToAclActivationLayerInfo(const QueueDescriptor &queueDescriptor)
void InitializeArmComputeTensorData(arm_compute::Tensor &tensor, TensorInfo tensorInfo, const ITensorHandle *handle)

References ARMNN_REPORT_PROFILING_WORKLOAD_DESC, armnn::ConvertAdditionalInfoToAclActivationLayerInfo(), armnn::info, armnn::InitializeArmComputeTensorData(), BaseWorkload< BatchNormalizationQueueDescriptor >::m_Data, QueueDescriptorWithParameters< LayerDescriptor >::m_Parameters, NeonBaseWorkload< BatchNormalizationQueueDescriptor >::NeonBaseWorkload(), and armnn::PolymorphicDowncast().

Member Function Documentation

◆ Execute()

void Execute ( ) const
overridevirtual

Implements IWorkload.

Definition at line 114 of file NeonBatchNormalizationWorkload.cpp.

115{
116 ARMNN_SCOPED_PROFILING_EVENT_NEON_NAME_GUID("NeonBatchNormalizationWorkload_Execute");
117 m_Layer->run();
118}
#define ARMNN_SCOPED_PROFILING_EVENT_NEON_NAME_GUID(label)
Creates a profiling event that uses GetGuid() and GetName() from the calling class.

References ARMNN_SCOPED_PROFILING_EVENT_NEON_NAME_GUID.


The documentation for this class was generated from the following files: