ArmNN
 25.11
Loading...
Searching...
No Matches
NeonConvertFp16ToFp32Workload.hpp
Go to the documentation of this file.
1//
2// Copyright © 2017 Arm Ltd and Contributors. All rights reserved.
3// SPDX-License-Identifier: MIT
4//
5
6#pragma once
7
8#include <arm_compute/runtime/NEON/functions/NECast.h>
11#include <memory>
13
14
15namespace armnn
16{
17
18arm_compute::Status NeonConvertFp16ToFp32WorkloadValidate(const TensorInfo& input, const TensorInfo& output);
19
20class NeonConvertFp16ToFp32Workload : public Float16ToFloat32Workload<ConvertFp16ToFp32QueueDescriptor>
21{
22public:
24 virtual void Execute() const override;
25 // Replace input tensor handle with the given TensorHandle
26 void ReplaceInputTensorHandle(ITensorHandle* tensorHandle, unsigned int slot) override;
27
28 // Replace output tensor handle with the given TensorHandle
29 void ReplaceOutputTensorHandle(ITensorHandle* tensorHandle, unsigned int slot) override;
30private:
31 using TensorHandlePair = std::pair<const ITensorHandle*, ITensorHandle*>;
32 std::vector<TensorHandlePair> m_TensorHandlePairs;
33 virtual void Reconfigure();
34 mutable std::unique_ptr<arm_compute::NECast> m_Cast;
35};
36
37} //namespace armnn
NeonConvertFp16ToFp32Workload(const ConvertFp16ToFp32QueueDescriptor &descriptor, const WorkloadInfo &info)
void ReplaceInputTensorHandle(ITensorHandle *tensorHandle, unsigned int slot) override
void ReplaceOutputTensorHandle(ITensorHandle *tensorHandle, unsigned int slot) override
Copyright (c) 2021 ARM Limited and Contributors.
arm_compute::Status NeonConvertFp16ToFp32WorkloadValidate(const TensorInfo &input, const TensorInfo &output)
MultiTypedWorkload< QueueDescriptor, armnn::DataType::Float16, armnn::DataType::Float32 > Float16ToFloat32Workload
Definition Workload.hpp:232
Contains information about TensorInfos of a layer.