ArmNN
 25.11
Loading...
Searching...
No Matches
NeonConvertFp32ToFp16Workload.hpp
Go to the documentation of this file.
1//
2// Copyright © 2017 Arm Ltd and Contributors. All rights reserved.
3// SPDX-License-Identifier: MIT
4//
5
6#pragma once
7
8#include <arm_compute/runtime/NEON/functions/NECast.h>
11#include <memory>
13
14namespace armnn
15{
16
17arm_compute::Status NeonConvertFp32ToFp16WorkloadValidate(const TensorInfo& input, const TensorInfo& output);
18
19class NeonConvertFp32ToFp16Workload : public Float32ToFloat16Workload<ConvertFp32ToFp16QueueDescriptor>
20{
21public:
23 virtual void Execute() const override;
24 // Replace input tensor handle with the given TensorHandle
25 void ReplaceInputTensorHandle(ITensorHandle* tensorHandle, unsigned int slot) override;
26
27 // Replace output tensor handle with the given TensorHandle
28 void ReplaceOutputTensorHandle(ITensorHandle* tensorHandle, unsigned int slot) override;
29private:
30 virtual void Reconfigure();
31 using TensorHandlePair = std::pair<const ITensorHandle*, ITensorHandle*>;
32 std::vector<TensorHandlePair> m_TensorHandlePairs;
33 mutable std::unique_ptr<arm_compute::NECast> m_Cast;
34};
35
36} //namespace armnn
NeonConvertFp32ToFp16Workload(const ConvertFp32ToFp16QueueDescriptor &descriptor, const WorkloadInfo &info)
void ReplaceInputTensorHandle(ITensorHandle *tensorHandle, unsigned int slot) override
void ReplaceOutputTensorHandle(ITensorHandle *tensorHandle, unsigned int slot) override
Copyright (c) 2021 ARM Limited and Contributors.
arm_compute::Status NeonConvertFp32ToFp16WorkloadValidate(const TensorInfo &input, const TensorInfo &output)
MultiTypedWorkload< QueueDescriptor, armnn::DataType::Float32, armnn::DataType::Float16 > Float32ToFloat16Workload
Definition Workload.hpp:237
Contains information about TensorInfos of a layer.