ArmNN
 26.01
Loading...
Searching...
No Matches
NeonConvolution3dWorkload Class Reference

#include <NeonConvolution3dWorkload.hpp>

Inheritance diagram for NeonConvolution3dWorkload:
[legend]
Collaboration diagram for NeonConvolution3dWorkload:
[legend]

Public Member Functions

 NeonConvolution3dWorkload (const Convolution3dQueueDescriptor &descriptor, const WorkloadInfo &info, std::shared_ptr< arm_compute::MemoryManagerOnDemand > &memoryManager, const bool isFastMathENabled=false)
 
void Execute () const override
 
- Public Member Functions inherited from NeonBaseWorkload< Convolution3dQueueDescriptor >
 NeonBaseWorkload (const Convolution3dQueueDescriptor &descriptor, const WorkloadInfo &info)
 
void ReplaceInputTensorHandle (ITensorHandle *tensorHandle, unsigned int slot) override
 
void ReplaceOutputTensorHandle (ITensorHandle *tensorHandle, unsigned int slot) override
 
- Public Member Functions inherited from BaseWorkload< QueueDescriptor >
 BaseWorkload (const QueueDescriptor &descriptor, const WorkloadInfo &info)
 
virtual const std::string & GetName () const override
 
void PostAllocationConfigure () override
 
const QueueDescriptorGetData () const
 
arm::pipe::ProfilingGuid GetGuid () const final
 
virtual bool SupportsTensorHandleReplacement () const override
 
- Public Member Functions inherited from IWorkload
virtual ~IWorkload ()
 
virtual void RegisterDebugCallback (const DebugCallbackFunction &)
 
virtual armnn::Optional< armnn::MemoryRequirementsGetMemoryRequirements ()
 

Additional Inherited Members

- Protected Member Functions inherited from NeonBaseWorkload< Convolution3dQueueDescriptor >
virtual void Reconfigure ()
 
- Protected Attributes inherited from BaseWorkload< QueueDescriptor >
QueueDescriptor m_Data
 
const arm::pipe::ProfilingGuid m_Guid
 
const std::string m_Name
 

Detailed Description

Definition at line 27 of file NeonConvolution3dWorkload.hpp.

Constructor & Destructor Documentation

◆ NeonConvolution3dWorkload()

NeonConvolution3dWorkload ( const Convolution3dQueueDescriptor & descriptor,
const WorkloadInfo & info,
std::shared_ptr< arm_compute::MemoryManagerOnDemand > & memoryManager,
const bool isFastMathENabled = false )

Definition at line 60 of file NeonConvolution3dWorkload.cpp.

64 : NeonBaseWorkload<Convolution3dQueueDescriptor>(descriptor, info)
65{
66 IgnoreUnused(memoryManager);
67
68 using arm_compute::NEConv3D;
69 uint32_t numInputs = m_Data.m_Parameters.m_BiasEnabled ? 3: 2;
70 m_Data.ValidateInputsOutputs("NeonConvolution3dWorkload", numInputs, 1);
71
72 arm_compute::ITensor& input = PolymorphicDowncast<IAclTensorHandle*>(m_Data.m_Inputs[0])->GetTensor();
73 arm_compute::ITensor& weights = PolymorphicDowncast<IAclTensorHandle*>(m_Data.m_Inputs[1])->GetTensor();
74 arm_compute::ITensor* biasesPtr = nullptr;
75 if (m_Data.m_Parameters.m_BiasEnabled)
76 {
77 biasesPtr = &PolymorphicDowncast<IAclTensorHandle *>(m_Data.m_Inputs[2])->GetTensor();
78 }
79 arm_compute::ITensor& output = PolymorphicDowncast<IAclTensorHandle*>(m_Data.m_Outputs[0])->GetTensor();
80
81 arm_compute::DataLayout aclDataLayout = ConvertDataLayout(m_Data.m_Parameters.m_DataLayout);
82 input.info()->set_data_layout(aclDataLayout);
83 weights.info()->set_data_layout(aclDataLayout);
84 output.info()->set_data_layout(aclDataLayout);
85
86 const arm_compute::Conv3dInfo aclConv3DInfo = ComputeConv3DInfo(descriptor, isFastMathEnabled);
87
88 auto convolutionLayer = std::make_unique<arm_compute::NEConv3D>();
89 convolutionLayer->configure(&input,
90 &weights,
91 biasesPtr,
92 &output,
93 aclConv3DInfo);
94
95 // Add details for profiling output
96 WorkloadInfo detailsInfo;
97
98 detailsInfo.m_InputTensorInfos = info.m_InputTensorInfos;
99 detailsInfo.m_OutputTensorInfos = info.m_OutputTensorInfos;
100
101 // Report Profiling Details
102 ARMNN_REPORT_PROFILING_WORKLOAD_DESC("NeonConvolution3dWorkload_Construct",
103 descriptor.m_Parameters,
104 detailsInfo,
105 this->GetGuid());
106
107 m_ConvolutionLayer.reset(convolutionLayer.release());
108 m_ConvolutionLayer->prepare();
109}
#define ARMNN_REPORT_PROFILING_WORKLOAD_DESC(name, desc, infos, guid)
QueueDescriptor m_Data
Definition Workload.hpp:74
arm_compute::Conv3dInfo ComputeConv3DInfo(const armnn::Convolution3dDescriptor descriptor, bool isFastMathEnabled, const ActivationDescriptor *activationDescriptor)
Utility function used to setup an arm_compute::Conv3dInfo object from convolution3d descriptor.
void IgnoreUnused(Ts &&...)
std::vector< ITensorHandle * > m_Inputs
std::vector< ITensorHandle * > m_Outputs
void ValidateInputsOutputs(const std::string &descName, unsigned int numExpectedIn, unsigned int numExpectedOut) const

References ARMNN_REPORT_PROFILING_WORKLOAD_DESC, armnn::ComputeConv3DInfo(), armnn::IgnoreUnused(), armnn::info, BaseWorkload< QueueDescriptor >::m_Data, QueueDescriptor::m_Inputs, WorkloadInfo::m_InputTensorInfos, QueueDescriptor::m_Outputs, WorkloadInfo::m_OutputTensorInfos, QueueDescriptorWithParameters< LayerDescriptor >::m_Parameters, and QueueDescriptor::ValidateInputsOutputs().

Member Function Documentation

◆ Execute()

void Execute ( ) const
overridevirtual

Implements IWorkload.

Definition at line 111 of file NeonConvolution3dWorkload.cpp.

112{
113 ARMNN_SCOPED_PROFILING_EVENT_NEON_NAME_GUID("NeonConvolution3dWorkload_Execute");
114 m_ConvolutionLayer->run();
115}
#define ARMNN_SCOPED_PROFILING_EVENT_NEON_NAME_GUID(label)
Creates a profiling event that uses GetGuid() and GetName() from the calling class.

References ARMNN_SCOPED_PROFILING_EVENT_NEON_NAME_GUID.


The documentation for this class was generated from the following files: