15 #include <arm_compute/core/CL/CLCompileContext.h>
35 std::string& outReasonIfUnsupported);
39 std::string& outReasonIfUnsupported,
47 unsigned int const* subTensorOrigin) const override;
51 const
bool IsMemoryManaged = true) const override;
56 const
bool IsMemoryManaged = true) const override;
63 static
std::unique_ptr<
IWorkload> MakeWorkload(const QueueDescriptorType& descriptor,
67 template <typename Workload, typename QueueDescriptorType, typename... Args>
68 static
std::unique_ptr<
IWorkload> MakeWorkload(const QueueDescriptorType& descriptor,
72 void InitializeCLCompileContext();
76 arm_compute::CLCompileContext m_CLCompileContext;
#define ARMNN_DEPRECATED_MSG(message)
std::unique_ptr< ITensorHandle > CreateTensorHandle(const TensorInfo &tensorInfo, const bool IsMemoryManaged=true) const override
std::unique_ptr< IWorkload > CreateWorkload(LayerType type, const QueueDescriptor &descriptor, const WorkloadInfo &info) const override
Backends should implement their own CreateWorkload function with a switch statement.
static bool IsLayerSupported(const Layer &layer, Optional< DataType > dataType, std::string &outReasonIfUnsupported)
std::unique_ptr< ITensorHandle > CreateSubTensorHandle(ITensorHandle &parent, TensorShape const &subTensorShape, unsigned int const *subTensorOrigin) const override
const BackendId & GetBackendId() const override
bool SupportsSubTensors() const override
ClWorkloadFactory(const std::shared_ptr< ClMemoryManager > &memoryManager)
void AfterWorkloadsCreated() override
std::shared_ptr< IBackendModelContext > IBackendSpecificModelContextPtr
Interface for a layer that is connectable to other layers via InputSlots and OutputSlots.
Workload interface to enqueue a layer computation.
Copyright (c) 2021 ARM Limited and Contributors.
LayerType
When adding a new layer, adapt also the LastLayer enum value in the enum class LayerType below.
std::vector< BackendOptions > ModelOptions
Contains information about TensorInfos of a layer.