17 #include <arm_compute/runtime/NEON/functions/NEFullyConnectedLayer.h> 21 using namespace armcomputetensorutils;
31 const arm_compute::TensorInfo aclInput = BuildArmComputeTensorInfo(input);
32 const arm_compute::TensorInfo aclOutput = BuildArmComputeTensorInfo(output);
33 const arm_compute::TensorInfo aclWeights = BuildArmComputeTensorInfo(weights);
35 arm_compute::TensorInfo aclBiases;
36 arm_compute::TensorInfo* optionalAclBiases =
nullptr;
39 aclBiases = BuildArmComputeTensorInfo(biases);
40 optionalAclBiases = &aclBiases;
43 const arm_compute::FullyConnectedLayerInfo fullyConnectedLayerInfo =
46 return arm_compute::NEFullyConnectedLayer::validate(&aclInput,
50 fullyConnectedLayerInfo);
60 arm_compute::ITensor& input = PolymorphicDowncast<IAclTensorHandle*>(
m_Data.
m_Inputs[0])->GetTensor();
61 arm_compute::ITensor& output = PolymorphicDowncast<IAclTensorHandle*>(
m_Data.
m_Outputs[0])->GetTensor();
63 m_WeightsTensor = std::make_unique<arm_compute::Tensor>();
68 m_BiasesTensor = std::make_unique<arm_compute::Tensor>();
74 arm_compute::FullyConnectedLayerInfo fc_info =
77 auto layer = std::make_unique<arm_compute::NEFullyConnectedLayer>(memoryManager);
78 layer->configure(&input, m_WeightsTensor.get(), m_BiasesTensor.get(), &output, fc_info);
79 m_FullyConnectedLayer.reset(layer.release());
122 m_FullyConnectedLayer->prepare();
129 m_FullyConnectedLayer->run();
132 void NeonFullyConnectedWorkload::FreeUnusedTensors()
134 FreeTensorIfUnused(m_WeightsTensor);
135 FreeTensorIfUnused(m_BiasesTensor);
arm_compute::ActivationLayerInfo ConvertAdditionalInfoToAclActivationLayerInfo(const QueueDescriptor &queueDescriptor)
NeonFullyConnectedWorkload(const FullyConnectedQueueDescriptor &descriptor, const WorkloadInfo &info, std::shared_ptr< arm_compute::MemoryManagerOnDemand > &memoryManager)
arm_compute::Status NeonFullyConnectedWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const TensorInfo &weights, const TensorInfo &biases, const FullyConnectedDescriptor &descriptor, const ActivationDescriptor *activationDescriptor)
void ValidateInputsOutputs(const std::string &descName, unsigned int numExpectedIn, unsigned int numExpectedOut) const
Copyright (c) 2021 ARM Limited and Contributors.
const ConstTensorHandle * m_Bias
LayerDescriptor m_Parameters
std::shared_ptr< arm_compute::MemoryManagerOnDemand > ACLMemManagerOnDemand
const TensorInfo & GetTensorInfo() const
std::vector< TensorInfo > m_InputTensorInfos
FullyConnectedQueueDescriptor m_Data
DataType GetDataType() const
A FullyConnectedDescriptor for the FullyConnectedLayer.
bool m_BiasEnabled
Enable/disable bias.
arm_compute::FullyConnectedLayerInfo ConvertFullyConnectedDescriptorToAclFullyConnectedLayerInfo(const FullyConnectedDescriptor &fullyConnectedDesc, const ActivationDescriptor *activationDesc)
std::vector< TensorInfo > m_OutputTensorInfos
const ConstTensorHandle * m_Weight
An ActivationDescriptor for the ActivationLayer.
profiling::ProfilingGuid GetGuid() const final
Optional< TensorInfo > m_BiasTensorInfo
std::vector< ITensorHandle * > m_Outputs
#define ARMNN_REPORT_PROFILING_WORKLOAD_DESC(name, desc, infos, guid)
void InitializeArmComputeTensorData(arm_compute::Tensor &tensor, const ConstTensorHandle *handle)
Contains information about TensorInfos of a layer.
std::vector< ITensorHandle * > m_Inputs
virtual void Execute() const override
#define ARMNN_SCOPED_PROFILING_EVENT_NEON_GUID(name, guid)
Optional< TensorInfo > m_WeightsTensorInfo