24.05
|
#include <RefLayerSupport.hpp>
|
| bool | IsLayerSupported (const LayerType &type, const std::vector< TensorInfo > &infos, const BaseDescriptor &descriptor, const Optional< LstmInputParamsInfo > &lstmParamsInfo, const Optional< QuantizedLstmInputParamsInfo > &, Optional< std::string & > reasonIfUnsupported) const override |
| | Default implementation of the ILayerSupport interface, Backends should implement this as a switch statement for each of their LayerTypes calling their specific backend implementation of IsXXXLayerSupported. More...
|
| |
| bool | IsActivationSupported (const TensorInfo &input, const TensorInfo &output, const ActivationDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsAdditionSupported (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsArgMinMaxSupported (const TensorInfo &input, const TensorInfo &output, const ArgMinMaxDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsBatchMatMulSupported (const TensorInfo &inputX, const TensorInfo &inputY, const TensorInfo &output, const BatchMatMulDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsBatchNormalizationSupported (const TensorInfo &input, const TensorInfo &output, const TensorInfo &mean, const TensorInfo &var, const TensorInfo &beta, const TensorInfo &gamma, const BatchNormalizationDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsBatchToSpaceNdSupported (const TensorInfo &input, const TensorInfo &output, const BatchToSpaceNdDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsBroadcastToSupported (const TensorInfo &input, const TensorInfo &output, const BroadcastToDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsCastSupported (const TensorInfo &input, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsChannelShuffleSupported (const TensorInfo &input, const TensorInfo &output, const ChannelShuffleDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsComparisonSupported (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ComparisonDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsConcatSupported (const std::vector< const TensorInfo * > inputs, const TensorInfo &output, const OriginsDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsConstantSupported (const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsConvertFp16ToFp32Supported (const TensorInfo &input, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsConvertFp32ToFp16Supported (const TensorInfo &input, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsConvolution2dSupported (const TensorInfo &input, const TensorInfo &output, const Convolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsConvolution3dSupported (const TensorInfo &input, const TensorInfo &output, const Convolution3dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsDebugSupported (const TensorInfo &input, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsDepthToSpaceSupported (const TensorInfo &input, const TensorInfo &output, const DepthToSpaceDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsDepthwiseConvolutionSupported (const TensorInfo &input, const TensorInfo &output, const DepthwiseConvolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsDequantizeSupported (const TensorInfo &input, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsDetectionPostProcessSupported (const TensorInfo &boxEncodings, const TensorInfo &scores, const TensorInfo &anchors, const TensorInfo &detectionBoxes, const TensorInfo &detectionClasses, const TensorInfo &detectionScores, const TensorInfo &numDetections, const DetectionPostProcessDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsDilatedDepthwiseConvolutionSupported (const TensorInfo &input, const TensorInfo &output, const DepthwiseConvolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsDivisionSupported (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsElementwiseUnarySupported (const TensorInfo &input, const TensorInfo &output, const ElementwiseUnaryDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsFakeQuantizationSupported (const TensorInfo &input, const FakeQuantizationDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsFillSupported (const TensorInfo &input, const TensorInfo &output, const FillDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsFloorSupported (const TensorInfo &input, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsFullyConnectedSupported (const TensorInfo &input, const TensorInfo &output, const TensorInfo &weights, const TensorInfo &biases, const FullyConnectedDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsGatherNdSupported (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsGatherSupported (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const GatherDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsInputSupported (const TensorInfo &input, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsInstanceNormalizationSupported (const TensorInfo &input, const TensorInfo &output, const InstanceNormalizationDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsL2NormalizationSupported (const TensorInfo &input, const TensorInfo &output, const L2NormalizationDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsLogicalBinarySupported (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const LogicalBinaryDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported) const |
| |
| bool | IsLogSoftmaxSupported (const TensorInfo &input, const TensorInfo &output, const LogSoftmaxDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported) const |
| |
| bool | IsLstmSupported (const TensorInfo &input, const TensorInfo &outputStateIn, const TensorInfo &cellStateIn, const TensorInfo &scratchBuffer, const TensorInfo &outputStateOut, const TensorInfo &cellStateOut, const TensorInfo &output, const LstmDescriptor &descriptor, const LstmInputParamsInfo ¶msInfo, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsMaximumSupported (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsMeanSupported (const TensorInfo &input, const TensorInfo &output, const MeanDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsMemCopySupported (const TensorInfo &input, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsMinimumSupported (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsMultiplicationSupported (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsNormalizationSupported (const TensorInfo &input, const TensorInfo &output, const NormalizationDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsOutputSupported (const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsPadSupported (const TensorInfo &input, const TensorInfo &output, const PadDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsPermuteSupported (const TensorInfo &input, const TensorInfo &output, const PermuteDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsPooling2dSupported (const TensorInfo &input, const TensorInfo &output, const Pooling2dDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsPooling3dSupported (const TensorInfo &input, const TensorInfo &output, const Pooling3dDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsQuantizeSupported (const TensorInfo &input, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsQLstmSupported (const TensorInfo &input, const TensorInfo &previousOutputIn, const TensorInfo &previousCellStateIn, const TensorInfo &outputStateOut, const TensorInfo &cellStateOut, const TensorInfo &output, const QLstmDescriptor &descriptor, const LstmInputParamsInfo ¶msInfo, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsRankSupported (const TensorInfo &input, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsReduceSupported (const TensorInfo &input, const TensorInfo &output, const ReduceDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsReshapeSupported (const TensorInfo &input, const TensorInfo &output, const ReshapeDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsResizeSupported (const TensorInfo &input, const TensorInfo &output, const ResizeDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsReverseV2Supported (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsScatterNdSupported (const TensorInfo &input, const TensorInfo &indices, const TensorInfo &updates, const TensorInfo &output, const ScatterNdDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsShapeSupported (const TensorInfo &input, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsSliceSupported (const TensorInfo &input, const TensorInfo &output, const SliceDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsSoftmaxSupported (const TensorInfo &input, const TensorInfo &output, const SoftmaxDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsSpaceToBatchNdSupported (const TensorInfo &input, const TensorInfo &output, const SpaceToBatchNdDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsSpaceToDepthSupported (const TensorInfo &input, const TensorInfo &output, const SpaceToDepthDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsSplitterSupported (const TensorInfo &input, const std::vector< std::reference_wrapper< TensorInfo >> &outputs, const ViewsDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsStackSupported (const std::vector< const TensorInfo * > &inputs, const TensorInfo &output, const StackDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsStridedSliceSupported (const TensorInfo &input, const TensorInfo &output, const StridedSliceDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsSubtractionSupported (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsPreluSupported (const TensorInfo &input, const TensorInfo &alpha, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsTileSupported (const TensorInfo &input, const TensorInfo &output, const TileDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsTransposeConvolution2dSupported (const TensorInfo &input, const TensorInfo &output, const TransposeConvolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsTransposeSupported (const TensorInfo &input, const TensorInfo &output, const TransposeDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsUnidirectionalSequenceLstmSupported (const TensorInfo &input, const TensorInfo &outputStateIn, const TensorInfo &cellStateIn, const TensorInfo &outputStateOut, const TensorInfo &cellStateOut, const TensorInfo &output, const UnidirectionalSequenceLstmDescriptor &descriptor, const LstmInputParamsInfo ¶msInfo, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsDetectionPostProcessSupported (const TensorInfo &boxEncodings, const TensorInfo &scores, const TensorInfo &anchors, const TensorInfo &detectionBoxes, const TensorInfo &detectionClasses, const TensorInfo &detectionScores, const TensorInfo &numDetections, const DetectionPostProcessDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsMemCopySupported (const TensorInfo &input, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsMemImportSupported (const TensorInfo &input, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsMergeSupported (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsQuantizedLstmSupported (const TensorInfo &input, const TensorInfo &previousCellStateIn, const TensorInfo &previousOutputIn, const TensorInfo &cellStateOut, const TensorInfo &output, const QuantizedLstmInputParamsInfo ¶msInfo, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsShapeSupported (const TensorInfo &input, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
| bool | IsStandInSupported (const std::vector< const TensorInfo * > &inputs, const std::vector< const TensorInfo * > &outputs, const StandInDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const |
| |
Definition at line 12 of file RefLayerSupport.hpp.
◆ IsActivationSupported()
Definition at line 560 of file RefLayerSupport.cpp.
565 bool supported =
true;
568 std::array<DataType,6> supportedTypes = {
576 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
577 "Reference activation: input type not supported.");
579 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
580 "Reference activation: output type not supported.");
582 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
583 "Reference activation: input and output types mismatched.");
585 supported &=
CheckSupportRule(ShapesAreSameRank(input, output), reasonIfUnsupported,
586 "Reference activation: input and output shapes are of different rank.");
589 struct ActivationFunctionSupported :
public Rule
591 ActivationFunctionSupported(
const ActivationDescriptor& desc)
593 switch(desc.m_Function)
622 supported &=
CheckSupportRule(ActivationFunctionSupported(descriptor), reasonIfUnsupported,
623 "Reference activation: function not supported.");
References armnn::Abs, armnn::BoundedReLu, armnn::CheckSupportRule(), armnn::Elu, armnn::Float16, armnn::Float32, armnn::Gelu, armnn::HardSwish, armnn::LeakyReLu, armnn::Linear, ActivationDescriptor::m_Function, armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, armnn::ReLu, armnn::Sigmoid, armnn::SoftReLu, armnn::Sqrt, armnn::Square, and armnn::TanH.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsAdditionSupported()
Definition at line 628 of file RefLayerSupport.cpp.
633 bool supported =
true;
635 std::array<DataType,7> supportedTypes = {
644 supported &=
CheckSupportRule(TypeAnyOf(input0, supportedTypes), reasonIfUnsupported,
645 "Reference addition: input 0 is not a supported type.");
647 supported &=
CheckSupportRule(TypeAnyOf(input1, supportedTypes), reasonIfUnsupported,
648 "Reference addition: input 1 is not a supported type.");
650 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
651 "Reference addition: output is not a supported type.");
653 supported &=
CheckSupportRule(TypesAreEqual(input0, input1), reasonIfUnsupported,
654 "Reference addition: input 0 and Input 1 types are mismatched");
656 supported &=
CheckSupportRule(TypesAreEqual(input0, output), reasonIfUnsupported,
657 "Reference addition: input and output types are mismatched");
659 supported &=
CheckSupportRule(ShapesAreBroadcastCompatible(input0, input1, output), reasonIfUnsupported,
660 "Reference addition: shapes are not suitable for implicit broadcast.");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, and armnn::Signed32.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsArgMinMaxSupported()
Definition at line 665 of file RefLayerSupport.cpp.
671 std::array<DataType, 8> supportedInputTypes =
682 std::array<DataType,2> supportedOutputTypes = {
687 bool supported =
true;
689 supported &=
CheckSupportRule(TypeAnyOf(input, supportedInputTypes), reasonIfUnsupported,
690 "Reference ArgMinMax: input is not a supported type.");
691 supported &=
CheckSupportRule(TypeAnyOf(output, supportedOutputTypes), reasonIfUnsupported,
692 "Reference ArgMinMax: output type not supported");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::IgnoreUnused(), armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, armnn::Signed32, and armnn::Signed64.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsBatchMatMulSupported()
Definition at line 697 of file RefLayerSupport.cpp.
705 std::array<DataType, 6> supportedTypes =
714 bool supported =
true;
716 supported &=
CheckSupportRule(TypeAnyOf(inputX, supportedTypes), reasonIfUnsupported,
717 "Reference batch matrix multiplication: input X is not a supported type");
719 supported &=
CheckSupportRule(TypeAnyOf(inputY, supportedTypes), reasonIfUnsupported,
720 "Reference batch matrix multiplication: input Y is not a supported type");
722 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
723 "Reference batch matrix multiplication: output is not a supported type");
725 supported &=
CheckSupportRule(TypesAreEqual(inputX, inputY), reasonIfUnsupported,
726 "Reference batch matrix multiplication: input X and input Y types are mismatched");
728 supported &=
CheckSupportRule(TypesAreEqual(inputX, output), reasonIfUnsupported,
729 "Reference batch matrix multiplication: inputs and output types are mismatched");
731 supported &=
CheckSupportRule(TensorNumDimensionsAreGreaterOrEqualTo(inputX, 2),
733 "Reference batch matrix multiplication: input X is not of rank 2 or greater");
735 supported &=
CheckSupportRule(TensorNumDimensionsAreGreaterOrEqualTo(inputY, 2),
737 "Reference batch matrix multiplication: input Y is not of rank 2 or greater");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::IgnoreUnused(), armnn::QAsymmS8, armnn::QAsymmU8, and armnn::QSymmS16.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsBatchNormalizationSupported()
Definition at line 742 of file RefLayerSupport.cpp.
753 std::array<DataType, 6> supportedTypes =
762 bool supported =
true;
764 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
765 "Reference batch normalization: input is not a supported type.");
767 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
768 "Reference batch normalization: output is not a supported type.");
770 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
771 "Reference batch normalization: input and output types are mismatched");
773 supported &=
CheckSupportRule(TypeAnyOf(mean, supportedTypes), reasonIfUnsupported,
774 "Reference batch normalization: mean is not a supported type.");
776 supported &=
CheckSupportRule(TypeAnyOf(variance, supportedTypes), reasonIfUnsupported,
777 "Reference batch normalization: variance is not a supported type.");
779 supported &=
CheckSupportRule(TypeAnyOf(beta, supportedTypes), reasonIfUnsupported,
780 "Reference batch normalization: beta is not a supported type.");
782 supported &=
CheckSupportRule(TypeAnyOf(gamma, supportedTypes), reasonIfUnsupported,
783 "Reference batch normalization: gamma is not a supported type.");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::IgnoreUnused(), armnn::QAsymmS8, armnn::QAsymmU8, and armnn::QSymmS16.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsBatchToSpaceNdSupported()
Definition at line 788 of file RefLayerSupport.cpp.
795 bool supported =
true;
797 std::string batchToSpaceNdLayerStr =
"batchToSpaceNd";
798 std::string inputTensorStr =
"input";
799 std::string outputTensorStr =
"output";
802 std::array<DataType,6> supportedTypes =
811 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
812 "Reference BatchToSpaceNd: input type not supported.");
814 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
815 "Reference BatchToSpaceNd: output type not supported.");
817 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
818 "Reference BatchToSpaceNd: input and output types mismatched.");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::IgnoreUnused(), armnn::QAsymmS8, armnn::QAsymmU8, and armnn::QSymmS16.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsBroadcastToSupported()
Definition at line 823 of file RefLayerSupport.cpp.
830 bool supported =
true;
832 std::array<DataType, 8> supportedTypes
844 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
845 "BroadcastTo: input type not supported.");
847 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
848 "BroadcastTo: output type not supported");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::IgnoreUnused(), armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, armnn::QSymmS8, armnn::Signed32, and armnn::Signed64.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsCastSupported()
Definition at line 853 of file RefLayerSupport.cpp.
857 std::array<DataType, 10> supportedInputTypes =
869 bool supported =
true;
870 supported &=
CheckSupportRule(TypeAnyOf(input, supportedInputTypes), reasonIfUnsupported,
871 "Reference cast: input is not a supported type");
874 supported &=
CheckSupportRule(TypeAnyOf(output, supportedInputTypes), reasonIfUnsupported,
875 "Reference cast: output is not a supported type");
877 supported &=
CheckSupportRule(ShapesAreSameTotalSize(input, output), reasonIfUnsupported,
878 "Reference cast: input and output shapes have different number of total elements");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, armnn::QSymmS8, armnn::Signed32, and armnn::Signed64.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsChannelShuffleSupported()
Definition at line 883 of file RefLayerSupport.cpp.
889 bool supported =
true;
892 std::array<DataType, 7> supportedTypes =
902 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
903 "Reference ChannelShuffle: input is not a supported type.");
905 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
906 "Reference ChannelShuffle: output is not a supported type.");
908 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
909 "Reference ChannelShuffle: input and output types are mismatched.");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::IgnoreUnused(), armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, and armnn::QSymmS8.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsComparisonSupported()
Definition at line 915 of file RefLayerSupport.cpp.
922 std::array<DataType, 8> supportedInputTypes =
933 bool supported =
true;
934 supported &=
CheckSupportRule(TypeAnyOf(input0, supportedInputTypes), reasonIfUnsupported,
935 "Reference comparison: input 0 is not a supported type");
937 supported &=
CheckSupportRule(TypesAreEqual(input0, input1), reasonIfUnsupported,
938 "Reference comparison: input 0 and Input 1 types are mismatched");
941 "Reference comparison: output is not of type Boolean");
943 supported &=
CheckSupportRule(ShapesAreBroadcastCompatible(input0, input1, output), reasonIfUnsupported,
944 "Reference comparison: shapes are not suitable for implicit broadcast.");
References armnn::Boolean, armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::IgnoreUnused(), armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, and armnn::Signed32.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsConcatSupported()
Definition at line 949 of file RefLayerSupport.cpp.
956 bool supported =
true;
957 std::array<DataType,7> supportedTypes =
967 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
968 "Reference concatenation: output type not supported");
969 for (
const TensorInfo* input : inputs)
971 supported &=
CheckSupportRule(TypeAnyOf(*input, supportedTypes), reasonIfUnsupported,
972 "Reference concatenation: input type not supported");
974 supported &=
CheckSupportRule(TypesAreEqual(*input, output), reasonIfUnsupported,
975 "Reference concatenation: input and output types mismatched.");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::IgnoreUnused(), armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, and armnn::Signed32.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsConstantSupported()
Definition at line 981 of file RefLayerSupport.cpp.
984 std::array<DataType,8> supportedTypes =
995 return CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
996 "Reference constant: output is not a supported type.");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, armnn::QSymmS8, and armnn::Signed32.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsConvertFp16ToFp32Supported()
◆ IsConvertFp32ToFp16Supported()
◆ IsConvolution2dSupported()
Definition at line 1039 of file RefLayerSupport.cpp.
1046 bool supported =
true;
1049 std::array<DataType,7> supportedTypes =
1059 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
1060 "Reference Convolution2d: input is not a supported type.");
1062 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
1063 "Reference Convolution2d: output is not a supported type.");
1065 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
1066 "Reference Convolution2d: input and output types mismatched.");
1069 const DataType inputType = input.GetDataType();
1072 std::array<DataType, 3> supportedWeightTypes =
1079 supported &=
CheckSupportRule(TypeAnyOf(weights, supportedWeightTypes), reasonIfUnsupported,
1080 "Reference Convolution2d: weights type not supported for quantized input.");
1084 supported &=
CheckSupportRule(TypeAnyOf(weights, supportedTypes), reasonIfUnsupported,
1085 "Reference Convolution2d: weights is not a supported type.");
1087 supported &=
CheckSupportRule(TypesAreEqual(input, weights), reasonIfUnsupported,
1088 "Reference Convolution2d: input and weights types mismatched.");
1091 if (biases.has_value())
1093 std::array<DataType,4> biasesSupportedTypes =
1100 supported &=
CheckSupportRule(TypeAnyOf(biases.value(), biasesSupportedTypes), reasonIfUnsupported,
1101 "Reference Convolution2d: biases is not a supported type.");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, TensorInfo::GetDataType(), OptionalBase::has_value(), armnn::IgnoreUnused(), armnn::IsQuantized8BitType(), armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, armnn::QSymmS8, armnn::Signed32, and OptionalReferenceSwitch< std::is_reference< T >::value, T >::value().
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsConvolution3dSupported()
Definition at line 1108 of file RefLayerSupport.cpp.
1115 bool supported =
true;
1118 std::array<DataType,7> supportedTypes =
1128 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
1129 "Reference Convolution3d: input is not a supported type.");
1131 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
1132 "Reference Convolution3d: output is not a supported type.");
1134 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
1135 "Reference Convolution3d: input and output types mismatched.");
1137 const DataType inputType = input.GetDataType();
1140 std::array<DataType, 3> supportedWeightTypes =
1147 supported &=
CheckSupportRule(TypeAnyOf(weights, supportedWeightTypes), reasonIfUnsupported,
1148 "Reference Convolution3d: weights type not supported for quantized input.");
1152 supported &=
CheckSupportRule(TypeAnyOf(weights, supportedTypes), reasonIfUnsupported,
1153 "Reference Convolution3d: weights is not a supported type.");
1155 supported &=
CheckSupportRule(TypesAreEqual(input, weights), reasonIfUnsupported,
1156 "Reference Convolution3d: input and weights types mismatched.");
1159 if (biases.has_value())
1161 std::array<DataType,4> biasesSupportedTypes =
1168 supported &=
CheckSupportRule(TypeAnyOf(biases.value(), biasesSupportedTypes), reasonIfUnsupported,
1169 "Reference Convolution3d: biases is not a supported type.");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, TensorInfo::GetDataType(), OptionalBase::has_value(), armnn::IgnoreUnused(), armnn::IsQuantized8BitType(), armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, armnn::QSymmS8, armnn::Signed32, and OptionalReferenceSwitch< std::is_reference< T >::value, T >::value().
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsDebugSupported()
Definition at line 1176 of file RefLayerSupport.cpp.
1180 bool supported =
true;
1182 std::array<DataType, 8> supportedTypes =
1194 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
1195 "Reference for Debug layer: input type not supported");
1197 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
1198 "Reference for Debug layer: output type not supported");
1200 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
1201 "Reference for Debug layer: input and output types are mismatched");
References armnn::BFloat16, armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, armnn::QSymmS8, and armnn::Signed32.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsDepthToSpaceSupported()
Definition at line 1206 of file RefLayerSupport.cpp.
1212 bool supported =
true;
1214 std::array<DataType,6> supportedTypes =
1223 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
1224 "Reference DepthToSpace: input type not supported");
1226 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
1227 "Reference DepthToSpace: output type not supported");
1229 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
1230 "Reference DepthToSpace: input and output types are mismatched");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::IgnoreUnused(), armnn::QAsymmS8, armnn::QAsymmU8, and armnn::QSymmS16.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsDepthwiseConvolutionSupported()
Definition at line 1235 of file RefLayerSupport.cpp.
1243 bool supported =
true;
1246 std::array<DataType,7> supportedTypes =
1256 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
1257 "Reference DepthwiseConvolution2d: input is not a supported type.");
1259 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
1260 "Reference DepthwiseConvolution2d: output is not a supported type.");
1262 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
1263 "Reference DepthwiseConvolution2d: input and output types mismatched.");
1265 const DataType inputType = input.GetDataType();
1268 std::array<DataType, 3> supportedWeightTypes =
1275 supported &=
CheckSupportRule(TypeAnyOf(weights, supportedWeightTypes), reasonIfUnsupported,
1276 "Reference DepthwiseConvolution2d: weights type not supported for "
1277 "quantized input.");
1281 supported &=
CheckSupportRule(TypeAnyOf(weights, supportedTypes), reasonIfUnsupported,
1282 "Reference DepthwiseConvolution2d: weights is not a supported type.");
1284 supported &=
CheckSupportRule(TypesAreEqual(input, weights), reasonIfUnsupported,
1285 "Reference DepthwiseConvolution2d: input and weights types mismatched.");
1288 if (biases.has_value())
1290 std::array<DataType,4> biasesSupportedTypes =
1296 supported &=
CheckSupportRule(TypeAnyOf(biases.value(), biasesSupportedTypes), reasonIfUnsupported,
1297 "Reference DepthwiseConvolution2d: biases is not a supported type.");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, TensorInfo::GetDataType(), OptionalBase::has_value(), armnn::IgnoreUnused(), armnn::IsQuantized8BitType(), armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, armnn::QSymmS8, armnn::Signed32, and OptionalReferenceSwitch< std::is_reference< T >::value, T >::value().
Referenced by RefLayerSupport::IsDilatedDepthwiseConvolutionSupported(), and RefLayerSupport::IsLayerSupported().
◆ IsDequantizeSupported()
Definition at line 1304 of file RefLayerSupport.cpp.
1308 bool supported =
true;
1310 std::array<DataType,5> supportedInputTypes = {
1318 supported &=
CheckSupportRule(TypeAnyOf(input, supportedInputTypes), reasonIfUnsupported,
1319 "Reference for Dequantize layer: input type not supported.");
1321 supported &=
CheckSupportRule(TypeNotPerAxisQuantized(input), reasonIfUnsupported,
1322 "Reference for Dequantize layer: per-axis quantized input not supported.");
1324 std::array<DataType,3> supportedOutputTypes = {
1329 supported &=
CheckSupportRule(TypeAnyOf(output, supportedOutputTypes), reasonIfUnsupported,
1330 "Reference for Dequantize layer: output type not supported.");
1332 supported &=
CheckSupportRule(ShapesAreSameTotalSize(input, output), reasonIfUnsupported,
1333 "Reference for Dequantize layer: input/output shapes have different num total "
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, and armnn::QSymmS8.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsDetectionPostProcessSupported()
Definition at line 1339 of file RefLayerSupport.cpp.
1349 IgnoreUnused(anchors, detectionBoxes, detectionClasses, detectionScores, numDetections, descriptor);
1351 bool supported =
true;
1353 std::array<DataType,6> supportedInputTypes =
1362 supported &=
CheckSupportRule(TypeAnyOf(boxEncodings, supportedInputTypes), reasonIfUnsupported,
1363 "Reference DetectionPostProcess: input 0 is not a supported type.");
1365 supported &=
CheckSupportRule(TypeAnyOf(scores, supportedInputTypes), reasonIfUnsupported,
1366 "Reference DetectionPostProcess: input 1 is not a supported type.");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::IgnoreUnused(), armnn::QAsymmS8, armnn::QAsymmU8, and armnn::QSymmS16.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsDilatedDepthwiseConvolutionSupported()
◆ IsDivisionSupported()
Definition at line 1381 of file RefLayerSupport.cpp.
1386 bool supported =
true;
1388 std::array<DataType,7> supportedTypes = {
1397 supported &=
CheckSupportRule(TypeAnyOf(input0, supportedTypes), reasonIfUnsupported,
1398 "Reference division: input 0 is not a supported type.");
1400 supported &=
CheckSupportRule(TypeAnyOf(input1, supportedTypes), reasonIfUnsupported,
1401 "Reference division: input 1 is not a supported type.");
1403 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
1404 "Reference division: output is not a supported type.");
1406 supported &=
CheckSupportRule(TypesAreEqual(input0, input1), reasonIfUnsupported,
1407 "Reference division: input 0 and Input 1 types are mismatched");
1409 supported &=
CheckSupportRule(TypesAreEqual(input0, output), reasonIfUnsupported,
1410 "Reference division: input and output types are mismatched");
1412 supported &=
CheckSupportRule(ShapesAreBroadcastCompatible(input0, input1, output), reasonIfUnsupported,
1413 "Reference division: shapes are not suitable for implicit broadcast.");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, and armnn::Signed32.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsElementwiseUnarySupported()
Definition at line 1418 of file RefLayerSupport.cpp.
1425 std::array<DataType, 7> supportedTypes =
1435 std::array<DataType, 1> logicalSupportedTypes =
1440 bool supported =
true;
1444 supported &=
CheckSupportRule(TypeAnyOf(input, logicalSupportedTypes), reasonIfUnsupported,
1445 "Reference elementwise unary: input type not supported");
1447 supported &=
CheckSupportRule(TypeAnyOf(output, logicalSupportedTypes), reasonIfUnsupported,
1448 "Reference elementwise unary: output type not supported");
1452 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
1453 "Reference elementwise unary: input type not supported");
1455 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
1456 "Reference elementwise unary: output type not supported");
1459 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
1460 "Reference elementwise unary: input and output types not matching");
1462 supported &=
CheckSupportRule(ShapesAreSameTotalSize(input, output), reasonIfUnsupported,
1463 "Reference elementwise unary: input and output shapes"
1464 "have different number of total elements");
References armnn::Boolean, armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::IgnoreUnused(), armnn::LogicalNot, ElementwiseUnaryDescriptor::m_Operation, armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, and armnn::Signed32.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsFakeQuantizationSupported()
◆ IsFillSupported()
◆ IsFloorSupported()
Definition at line 1512 of file RefLayerSupport.cpp.
1517 bool supported =
true;
1519 std::array<DataType,3> supportedTypes =
1525 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
1526 "Reference Floor: input type not supported.");
1528 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
1529 "Reference Floor: output type not supported.");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, and armnn::IgnoreUnused().
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsFullyConnectedSupported()
Definition at line 1534 of file RefLayerSupport.cpp.
1541 bool supported =
true;
1544 std::array<DataType,6> supportedTypes =
1553 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
1554 "Reference Fully Connected: input type not supported.");
1556 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
1557 "Reference Fully Connected: output type not supported.");
1559 supported &=
CheckSupportRule(TypeAnyOf(weights, supportedTypes), reasonIfUnsupported,
1560 "Reference Fully Connected: weights type not supported.");
1562 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
1563 "Reference Fully Connected: input and output types mismatched.");
1565 supported &=
CheckSupportRule(TypeAnyOf(weights, supportedTypes), reasonIfUnsupported,
1566 "Reference Fully Connected: weights is not a supported type.");
1568 supported &=
CheckSupportRule(TypesAreEqual(input, weights), reasonIfUnsupported,
1569 "Reference Fully Connected: input and weights types mismatched.");
1571 if (descriptor.m_BiasEnabled)
1574 std::array<DataType, 5>
1575 supportedBiasTypes =
1583 supported &=
CheckSupportRule(TypeAnyOf(biases, supportedBiasTypes), reasonIfUnsupported,
1584 "Reference Fully Connected: bias type not supported.");
1586 supported &=
CheckSupportRule(BiasAndWeightsTypesMatch(biases, weights), reasonIfUnsupported,
1587 "Reference Fully Connected: bias and weight types mismatch.");
1589 supported &=
CheckSupportRule(BiasAndWeightsTypesCompatible(weights, supportedBiasTypes), reasonIfUnsupported,
1590 "Reference Fully Connected: bias type inferred from weights is incompatible.");
1592 supported &=
CheckSupportRule(TensorNumDimensionsAreCorrect(biases, 1U), reasonIfUnsupported,
1593 "Reference Fully Connected: bias must have 1 dimension.");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, FullyConnectedDescriptor::m_BiasEnabled, armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, and armnn::Signed32.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsGatherNdSupported()
Definition at line 1600 of file RefLayerSupport.cpp.
1605 bool supported =
true;
1606 std::array<DataType,7> supportedTypes =
1616 supported &=
CheckSupportRule(TypeAnyOf(input0, supportedTypes), reasonIfUnsupported,
1617 "Reference GatherNd: input type not supported");
1619 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
1620 "Reference GatherNd: output type not supported");
1623 "Reference GatherNd: indices (input1) type not supported");
1625 supported &=
CheckSupportRule(TypesAreEqual(input0, output), reasonIfUnsupported,
1626 "Reference GatherNd: input and output types not matching");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, and armnn::Signed32.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsGatherSupported()
Definition at line 1631 of file RefLayerSupport.cpp.
1637 bool supported =
true;
1638 std::array<DataType,7> supportedTypes =
1649 supported &=
CheckSupportRule(TypeAnyOf(input0, supportedTypes), reasonIfUnsupported,
1650 "Reference Gather: input type not supported");
1652 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
1653 "Reference Gather: output type not supported");
1656 "Reference Gather: indices (input1) type not supported");
1658 supported &=
CheckSupportRule(TypesAreEqual(input0, output), reasonIfUnsupported,
1659 "Reference Gather: input and output types not matching");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::IgnoreUnused(), armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, and armnn::Signed32.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsInputSupported()
◆ IsInstanceNormalizationSupported()
Definition at line 1670 of file RefLayerSupport.cpp.
1677 std::array<DataType, 3> supportedTypes =
1683 bool supported =
true;
1685 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
1686 "Reference Instance Normalization: input type not supported.");
1688 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
1689 "Reference Instance Normalization: output type not supported.");
1691 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
1692 "Reference Instance Normalization: input and output types mismatched.");
1694 supported &=
CheckSupportRule(ShapesAreSameTotalSize(input, output), reasonIfUnsupported,
1695 "Reference Instance Normalization: input and output shapes have different "
1696 "num total elements.");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, and armnn::IgnoreUnused().
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsL2NormalizationSupported()
Definition at line 1701 of file RefLayerSupport.cpp.
1708 std::array<DataType, 6> supportedTypes =
1717 bool supported =
true;
1719 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
1720 "Reference L2normalization: input type not supported.");
1722 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
1723 "Reference L2normalization: output type not supported.");
1725 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
1726 "Reference L2normalization: input and output types mismatched.");
1728 supported &=
CheckSupportRule(ShapesAreSameTotalSize(input, output), reasonIfUnsupported,
1729 "Reference L2normalization: input and output shapes have different "
1730 "num total elements.");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::IgnoreUnused(), armnn::QAsymmS8, armnn::QAsymmU8, and armnn::QSymmS16.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsLayerSupported()
Default implementation of the ILayerSupport interface, Backends should implement this as a switch statement for each of their LayerTypes calling their specific backend implementation of IsXXXLayerSupported.
Reimplemented from LayerSupportBase.
Definition at line 61 of file RefLayerSupport.cpp.
73 *(PolymorphicDowncast<const ActivationDescriptor*>(&descriptor)),
80 *(PolymorphicDowncast<const ArgMinMaxDescriptor*>(&descriptor)),
86 *(PolymorphicDowncast<const BatchMatMulDescriptor*>(&descriptor)),
95 *(PolymorphicDowncast<const BatchNormalizationDescriptor*>
101 *(PolymorphicDowncast<const BatchToSpaceNdDescriptor*>(&descriptor)),
102 reasonIfUnsupported);
106 *(PolymorphicDowncast<const BroadcastToDescriptor*>(&descriptor)),
107 reasonIfUnsupported);
112 *(PolymorphicDowncast<const ComparisonDescriptor*>(&descriptor)),
113 reasonIfUnsupported);
116 std::vector<const TensorInfo*> inputInfos;
117 for (uint32_t i = 0; i < (infos.size() - 1); i++)
119 inputInfos.push_back(&infos[i]);
122 infos[infos.size() - 1],
123 *(PolymorphicDowncast<const OriginsDescriptor*>(&descriptor)),
124 reasonIfUnsupported);
134 if (infos.size() != 4)
136 throw InvalidArgumentException(
"Invalid number of Convolution2d TensorInfos. "
137 "TensorInfos should be of format: {input, output, weights, biases}.");
140 auto desc = *(PolymorphicDowncast<const Convolution2dDescriptor*>(&descriptor));
141 if (infos[3] == TensorInfo())
148 reasonIfUnsupported);
157 reasonIfUnsupported);
163 *(PolymorphicDowncast<const DepthToSpaceDescriptor*>(&descriptor)),
164 reasonIfUnsupported);
167 if (infos.size() != 4)
169 throw InvalidArgumentException(
"Invalid number of DepthwiseConvolution2d TensorInfos. "
170 "TensorInfos should be of format: {input, output, weights, biases}.");
173 auto desc = *(PolymorphicDowncast<const DepthwiseConvolution2dDescriptor*>(&descriptor));
174 if (infos[3] == TensorInfo())
181 reasonIfUnsupported);
190 reasonIfUnsupported);
199 std::array<DataType, 7> supportedTypes =
209 bool supported =
true;
210 supported &=
CheckSupportRule(TypeAnyOf(infos[0], supportedTypes), reasonIfUnsupported,
211 "Reference elementwise unary: input type not supported");
213 supported &=
CheckSupportRule(TypeAnyOf(infos[1], supportedTypes), reasonIfUnsupported,
214 "Reference elementwise unary: input type not supported");
216 supported &=
CheckSupportRule(TypeAnyOf(infos[2], supportedTypes), reasonIfUnsupported,
217 "Reference elementwise unary: output type not supported");
219 supported &=
CheckSupportRule(TypesAreEqual(infos[0], infos[1]), reasonIfUnsupported,
220 "Reference elementwise unary: input types not matching");
222 supported &=
CheckSupportRule(TypesAreEqual(infos[0], infos[2]), reasonIfUnsupported,
223 "Reference elementwise unary: input and output types not matching");
230 *(PolymorphicDowncast<const ElementwiseUnaryDescriptor*>(&descriptor)),
231 reasonIfUnsupported);
235 *(PolymorphicDowncast<const FillDescriptor*>(&descriptor)),
236 reasonIfUnsupported);
244 *(PolymorphicDowncast<const FullyConnectedDescriptor*>(&descriptor)),
245 reasonIfUnsupported);
250 *(PolymorphicDowncast<const GatherDescriptor*>(&descriptor)),
251 reasonIfUnsupported);
256 reasonIfUnsupported);
262 *(PolymorphicDowncast<const InstanceNormalizationDescriptor*>
264 reasonIfUnsupported);
268 *(PolymorphicDowncast<const L2NormalizationDescriptor*>(&descriptor)),
269 reasonIfUnsupported);
274 *(PolymorphicDowncast<const LogicalBinaryDescriptor*>(&descriptor)),
275 reasonIfUnsupported);
279 *(PolymorphicDowncast<const LogSoftmaxDescriptor*>(&descriptor)),
280 reasonIfUnsupported);
289 *(PolymorphicDowncast<const LstmDescriptor*>(&descriptor)),
290 lstmParamsInfo.value(),
291 reasonIfUnsupported);
299 *(PolymorphicDowncast<const QLstmDescriptor*>(&descriptor)),
300 lstmParamsInfo.value(),
301 reasonIfUnsupported);
307 *(PolymorphicDowncast<const MeanDescriptor*>(&descriptor)),
308 reasonIfUnsupported);
316 *(PolymorphicDowncast<const NormalizationDescriptor*>(&descriptor)),
317 reasonIfUnsupported);
323 *(PolymorphicDowncast<const PadDescriptor*>(&descriptor)),
324 reasonIfUnsupported);
328 *(PolymorphicDowncast<const PermuteDescriptor*>(&descriptor)),
329 reasonIfUnsupported);
333 *(PolymorphicDowncast<const Pooling2dDescriptor*>(&descriptor)),
334 reasonIfUnsupported);
342 *(PolymorphicDowncast<const ReshapeDescriptor*>(&descriptor)),
343 reasonIfUnsupported);
347 *(PolymorphicDowncast<const ResizeDescriptor*>(&descriptor)),
348 reasonIfUnsupported);
353 reasonIfUnsupported);
357 *(PolymorphicDowncast<const ReduceDescriptor*>(&descriptor)),
358 reasonIfUnsupported);
364 *(PolymorphicDowncast<const ScatterNdDescriptor*>(&descriptor)),
365 reasonIfUnsupported);
369 *(PolymorphicDowncast<const SliceDescriptor*>(&descriptor)),
370 reasonIfUnsupported);
374 *(PolymorphicDowncast<const SoftmaxDescriptor*>(&descriptor)),
375 reasonIfUnsupported);
379 *(PolymorphicDowncast<const SpaceToBatchNdDescriptor*>(&descriptor)),
380 reasonIfUnsupported);
384 *(PolymorphicDowncast<const SpaceToDepthDescriptor*>(&descriptor)),
385 reasonIfUnsupported);
388 std::vector<TensorInfo> outputInfos;
389 for (uint32_t i = 1; i < infos.size(); i++)
391 outputInfos.push_back(infos[i]);
394 {outputInfos.begin(), outputInfos.end()},
395 *(PolymorphicDowncast<const ViewsDescriptor*>(&descriptor)),
396 reasonIfUnsupported);
400 std::vector<const TensorInfo*> inputInfos;
401 for (uint32_t i = 0; i < infos.size() - 1; i++)
403 inputInfos.push_back(&infos[i]);
406 infos[infos.size() - 1],
407 *(PolymorphicDowncast<const StackDescriptor*>(&descriptor)),
408 reasonIfUnsupported);
413 *(PolymorphicDowncast<const StridedSliceDescriptor*>(&descriptor)),
414 reasonIfUnsupported);
420 *(PolymorphicDowncast<const TileDescriptor*>(&descriptor)),
421 reasonIfUnsupported);
425 *(PolymorphicDowncast<const TransposeDescriptor*>(&descriptor)),
426 reasonIfUnsupported);
429 if (infos.size() != 4)
431 throw InvalidArgumentException(
"Invalid number of TransposeConvolution2d TensorInfos. "
432 "TensorInfos should be of format: {input, output, weights, biases}.");
435 auto desc = *(PolymorphicDowncast<const TransposeConvolution2dDescriptor*>(&descriptor));
436 if (infos[3] == TensorInfo())
443 reasonIfUnsupported);
452 reasonIfUnsupported);
460 *(PolymorphicDowncast<const ChannelShuffleDescriptor*>(&descriptor)),
461 reasonIfUnsupported);
464 if (infos.size() != 4)
466 throw InvalidArgumentException(
"Invalid number of Convolution3d TensorInfos. "
467 "TensorInfos should be of format: {input, output, weights, biases}.");
470 auto desc = *(PolymorphicDowncast<const Convolution3dDescriptor*>(&descriptor));
471 if (infos[3] == TensorInfo())
478 reasonIfUnsupported);
487 reasonIfUnsupported);
500 *(PolymorphicDowncast<const DetectionPostProcessDescriptor*>
502 reasonIfUnsupported);
505 *(PolymorphicDowncast<const FakeQuantizationDescriptor*>(&descriptor)),
506 reasonIfUnsupported);
515 if (infos.size() != 6)
517 throw InvalidArgumentException(
"Invalid number of UnidirectionalSequenceLstm TensorInfos. TensorInfos "
518 "should be of format: {input, outputStateIn, cellStateIn, "
519 "hiddenStateOutputVal, cellStateOutputVal, output}");
521 auto desc = *(PolymorphicDowncast<const UnidirectionalSequenceLstmDescriptor*>(&descriptor));
529 lstmParamsInfo.value(),
530 reasonIfUnsupported);
535 *(PolymorphicDowncast<const Pooling3dDescriptor*>(&descriptor)),
536 reasonIfUnsupported);
551 quantizedLstmInputParamsInfo.value(),
552 reasonIfUnsupported);
References armnn::Activation, armnn::Addition, armnn::ArgMinMax, armnn::BatchMatMul, armnn::BatchNormalization, armnn::BatchToSpaceNd, armnn::BroadcastTo, armnn::Cast, armnn::ChannelShuffle, armnn::CheckSupportRule(), armnn::Comparison, armnn::Concat, armnn::Constant, armnn::ConvertFp16ToFp32, armnn::ConvertFp32ToFp16, armnn::Convolution2d, armnn::Convolution3d, armnn::Debug, armnn::DepthToSpace, armnn::DepthwiseConvolution2d, armnn::Dequantize, armnn::DetectionPostProcess, armnn::Division, armnn::ElementwiseBinary, armnn::ElementwiseUnary, armnn::FakeQuantization, armnn::Fill, armnn::Float16, armnn::Float32, armnn::Floor, armnn::FullyConnected, armnn::Gather, armnn::GatherNd, armnn::Input, armnn::InstanceNormalization, RefLayerSupport::IsActivationSupported(), RefLayerSupport::IsAdditionSupported(), RefLayerSupport::IsArgMinMaxSupported(), RefLayerSupport::IsBatchMatMulSupported(), RefLayerSupport::IsBatchNormalizationSupported(), RefLayerSupport::IsBatchToSpaceNdSupported(), RefLayerSupport::IsBroadcastToSupported(), RefLayerSupport::IsCastSupported(), RefLayerSupport::IsChannelShuffleSupported(), RefLayerSupport::IsComparisonSupported(), RefLayerSupport::IsConcatSupported(), RefLayerSupport::IsConstantSupported(), RefLayerSupport::IsConvertFp16ToFp32Supported(), RefLayerSupport::IsConvertFp32ToFp16Supported(), RefLayerSupport::IsConvolution2dSupported(), RefLayerSupport::IsConvolution3dSupported(), RefLayerSupport::IsDebugSupported(), RefLayerSupport::IsDepthToSpaceSupported(), RefLayerSupport::IsDepthwiseConvolutionSupported(), RefLayerSupport::IsDequantizeSupported(), RefLayerSupport::IsDetectionPostProcessSupported(), RefLayerSupport::IsDivisionSupported(), RefLayerSupport::IsElementwiseUnarySupported(), RefLayerSupport::IsFakeQuantizationSupported(), RefLayerSupport::IsFillSupported(), RefLayerSupport::IsFloorSupported(), RefLayerSupport::IsFullyConnectedSupported(), RefLayerSupport::IsGatherNdSupported(), RefLayerSupport::IsGatherSupported(), RefLayerSupport::IsInputSupported(), RefLayerSupport::IsInstanceNormalizationSupported(), RefLayerSupport::IsL2NormalizationSupported(), RefLayerSupport::IsLogicalBinarySupported(), RefLayerSupport::IsLogSoftmaxSupported(), RefLayerSupport::IsLstmSupported(), RefLayerSupport::IsMaximumSupported(), RefLayerSupport::IsMeanSupported(), RefLayerSupport::IsMemCopySupported(), LayerSupportBase::IsMemImportSupported(), LayerSupportBase::IsMergeSupported(), RefLayerSupport::IsMinimumSupported(), RefLayerSupport::IsMultiplicationSupported(), RefLayerSupport::IsNormalizationSupported(), RefLayerSupport::IsOutputSupported(), RefLayerSupport::IsPadSupported(), RefLayerSupport::IsPermuteSupported(), RefLayerSupport::IsPooling2dSupported(), RefLayerSupport::IsPooling3dSupported(), RefLayerSupport::IsPreluSupported(), RefLayerSupport::IsQLstmSupported(), LayerSupportBase::IsQuantizedLstmSupported(), RefLayerSupport::IsQuantizeSupported(), RefLayerSupport::IsRankSupported(), RefLayerSupport::IsReduceSupported(), RefLayerSupport::IsReshapeSupported(), RefLayerSupport::IsResizeSupported(), RefLayerSupport::IsReverseV2Supported(), RefLayerSupport::IsScatterNdSupported(), RefLayerSupport::IsShapeSupported(), RefLayerSupport::IsSliceSupported(), RefLayerSupport::IsSoftmaxSupported(), RefLayerSupport::IsSpaceToBatchNdSupported(), RefLayerSupport::IsSpaceToDepthSupported(), RefLayerSupport::IsSplitterSupported(), RefLayerSupport::IsStackSupported(), RefLayerSupport::IsStridedSliceSupported(), RefLayerSupport::IsSubtractionSupported(), RefLayerSupport::IsTileSupported(), RefLayerSupport::IsTransposeConvolution2dSupported(), RefLayerSupport::IsTransposeSupported(), RefLayerSupport::IsUnidirectionalSequenceLstmSupported(), armnn::L2Normalization, armnn::LogicalBinary, armnn::LogSoftmax, armnn::Lstm, armnn::Map, armnn::Maximum, armnn::Mean, armnn::MemCopy, armnn::MemImport, armnn::Merge, armnn::Minimum, armnn::Multiplication, armnn::Normalization, armnn::Output, armnn::Pad, armnn::Permute, armnn::Pooling2d, armnn::Pooling3d, armnn::Prelu, armnn::QAsymmS8, armnn::QAsymmU8, armnn::QLstm, armnn::QSymmS16, armnn::Quantize, armnn::QuantizedLstm, armnn::Rank, armnn::Reduce, armnn::Reshape, armnn::Resize, armnn::ReverseV2, armnn::ScatterNd, armnn::Shape, armnn::Signed32, armnn::Slice, armnn::Softmax, armnn::SpaceToBatchNd, armnn::SpaceToDepth, armnn::Splitter, armnn::Stack, armnn::StridedSlice, armnn::Subtraction, armnn::Tile, armnn::Transpose, armnn::TransposeConvolution2d, armnn::UnidirectionalSequenceLstm, armnn::Unmap, and OptionalReferenceSwitch< std::is_reference< T >::value, T >::value().
◆ IsLogicalBinarySupported()
Definition at line 1735 of file RefLayerSupport.cpp.
1743 std::array<DataType, 1> supportedTypes =
1748 bool supported =
true;
1749 supported &=
CheckSupportRule(TypeAnyOf(input0, supportedTypes), reasonIfUnsupported,
1750 "Reference LogicalBinary: input 0 type not supported");
1751 supported &=
CheckSupportRule(TypeAnyOf(input1, supportedTypes), reasonIfUnsupported,
1752 "Reference LogicalBinary: input 1 type not supported");
1754 supported &=
CheckSupportRule(TypesAreEqual(input0, output), reasonIfUnsupported,
1755 "Reference LogicalBinary: input and output types do not match");
1757 supported &=
CheckSupportRule(ShapesAreBroadcastCompatible(input0, input1, output), reasonIfUnsupported,
1758 "Reference LogicalBinary: shapes are not suitable for implicit broadcast.");
References armnn::Boolean, armnn::CheckSupportRule(), and armnn::IgnoreUnused().
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsLogSoftmaxSupported()
Definition at line 1763 of file RefLayerSupport.cpp.
1770 std::array<DataType, 3> supportedTypes =
1776 bool supported =
true;
1777 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
1778 "Reference LogSoftmax: input type not supported");
1780 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
1781 "Reference LogSoftmax: output type not supported");
1783 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
1784 "Reference LogSoftmax: input and output types do not match");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, and armnn::IgnoreUnused().
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsLstmSupported()
Definition at line 1789 of file RefLayerSupport.cpp.
1803 bool supported =
true;
1805 std::array<DataType,3> supportedTypes = {
1811 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
1812 "Reference Lstm: input is not a supported type.");
1813 supported &=
CheckSupportRule(TypesAreEqual(input, outputStateIn), reasonIfUnsupported,
1814 "Reference Lstm: input and outputStateIn types are mismatched");
1815 supported &=
CheckSupportRule(TypesAreEqual(input, cellStateIn), reasonIfUnsupported,
1816 "Reference Lstm: input and cellStateIn types are mismatched");
1817 supported &=
CheckSupportRule(TypesAreEqual(input, scratchBuffer), reasonIfUnsupported,
1818 "Reference Lstm: input and scratchBuffer types are mismatched");
1819 supported &=
CheckSupportRule(TypesAreEqual(input, outputStateOut), reasonIfUnsupported,
1820 "Reference Lstm: input and outputStateOut types are mismatched");
1821 supported &=
CheckSupportRule(TypesAreEqual(input, cellStateOut), reasonIfUnsupported,
1822 "Reference Lstm: input and cellStateOut types are mismatched");
1824 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
1825 "Reference Lstm: input and output types are mismatched");
1827 supported &=
CheckSupportRule(TypesAreEqual(input, paramsInfo.GetInputToForgetWeights()), reasonIfUnsupported,
1828 "Reference Lstm: input and InputToForgetWeights types are mismatched");
1829 supported &=
CheckSupportRule(TypesAreEqual(input, paramsInfo.GetInputToCellWeights()), reasonIfUnsupported,
1830 "Reference Lstm: input and InputToCellWeights types are mismatched");
1831 supported &=
CheckSupportRule(TypesAreEqual(input, paramsInfo.GetInputToOutputWeights()), reasonIfUnsupported,
1832 "Reference Lstm: input and InputToOutputWeights types are mismatched");
1833 supported &=
CheckSupportRule(TypesAreEqual(input, paramsInfo.GetRecurrentToForgetWeights()), reasonIfUnsupported,
1834 "Reference Lstm: input and RecurrentToForgetWeights types are mismatched");
1835 supported &=
CheckSupportRule(TypesAreEqual(input, paramsInfo.GetRecurrentToCellWeights()), reasonIfUnsupported,
1836 "Reference Lstm: input and RecurrentToCellWeights types are mismatched");
1837 supported &=
CheckSupportRule(TypesAreEqual(input, paramsInfo.GetRecurrentToOutputWeights()), reasonIfUnsupported,
1838 "Reference Lstm: input and RecurrentToOutputWeights types are mismatched");
1839 supported &=
CheckSupportRule(TypesAreEqual(input, paramsInfo.GetForgetGateBias()), reasonIfUnsupported,
1840 "Reference Lstm: input and ForgetGateBias types are mismatched");
1841 supported &=
CheckSupportRule(TypesAreEqual(input, paramsInfo.GetCellBias()), reasonIfUnsupported,
1842 "Reference Lstm: input and CellBias types are mismatched");
1843 supported &=
CheckSupportRule(TypesAreEqual(input, paramsInfo.GetOutputGateBias()), reasonIfUnsupported,
1844 "Reference Lstm: input and OutputGateBias types are mismatched");
1845 if (!descriptor.m_CifgEnabled)
1847 supported &=
CheckSupportRule(TypesAreEqual(input, paramsInfo.GetInputToInputWeights()), reasonIfUnsupported,
1848 "Reference Lstm: input and InputToInputWeights types are mismatched");
1849 supported &=
CheckSupportRule(TypesAreEqual(input, paramsInfo.GetRecurrentToInputWeights()),
1850 reasonIfUnsupported,
1851 "Reference Lstm: input and RecurrentToInputWeights types are mismatched");
1852 supported &=
CheckSupportRule(TypesAreEqual(input, paramsInfo.GetInputGateBias()), reasonIfUnsupported,
1853 "Reference Lstm: input and InputGateBias types are mismatched");
1854 if (descriptor.m_PeepholeEnabled)
1856 supported &=
CheckSupportRule(TypesAreEqual(input, paramsInfo.GetCellToInputWeights()),
1857 reasonIfUnsupported,
1858 "Reference Lstm: input and CellToInputWeights types are mismatched");
1861 if (descriptor.m_PeepholeEnabled)
1863 supported &=
CheckSupportRule(TypesAreEqual(input, paramsInfo.GetCellToForgetWeights()), reasonIfUnsupported,
1864 "Reference Lstm: input and CellToForgetWeights types are mismatched");
1865 supported &=
CheckSupportRule(TypesAreEqual(input, paramsInfo.GetCellToOutputWeights()), reasonIfUnsupported,
1866 "Reference Lstm: input and CellToOutputWeights types are mismatched");
1868 if (descriptor.m_ProjectionEnabled)
1870 supported &=
CheckSupportRule(TypesAreEqual(input, paramsInfo.GetProjectionWeights()), reasonIfUnsupported,
1871 "Reference Lstm: input and mProjectionWeights types are mismatched");
1872 if (paramsInfo.m_ProjectionBias !=
nullptr)
1874 supported &=
CheckSupportRule(TypesAreEqual(input, paramsInfo.GetProjectionBias()), reasonIfUnsupported,
1875 "Reference Lstm: input and ProjectionBias types are mismatched");
1878 if (descriptor.m_LayerNormEnabled)
1880 if (!descriptor.m_CifgEnabled)
1882 supported &=
CheckSupportRule(TypesAreEqual(input, paramsInfo.GetInputLayerNormWeights()),
1883 reasonIfUnsupported,
1884 "Reference Lstm: input and InputLayerNormWeights types are mismatched");
1886 supported &=
CheckSupportRule(TypesAreEqual(input, paramsInfo.GetForgetLayerNormWeights()),
1887 reasonIfUnsupported,
1888 "Reference Lstm: input and ForgetLayerNormWeights types are mismatched");
1889 supported &=
CheckSupportRule(TypesAreEqual(input, paramsInfo.GetCellLayerNormWeights()),
1890 reasonIfUnsupported,
1891 "Reference Lstm: input and CellLayerNormWeights types are mismatched");
1892 supported &=
CheckSupportRule(TypesAreEqual(input, paramsInfo.GetOutputLayerNormWeights()),
1893 reasonIfUnsupported,
1894 "Reference Lstm: input and OutputLayerNormWeights types are mismatched");
References armnn::CheckSupportRule(), armnn::Float32, LstmInputParamsInfo::GetCellBias(), LstmInputParamsInfo::GetCellLayerNormWeights(), LstmInputParamsInfo::GetCellToForgetWeights(), LstmInputParamsInfo::GetCellToInputWeights(), LstmInputParamsInfo::GetCellToOutputWeights(), LstmInputParamsInfo::GetForgetGateBias(), LstmInputParamsInfo::GetForgetLayerNormWeights(), LstmInputParamsInfo::GetInputGateBias(), LstmInputParamsInfo::GetInputLayerNormWeights(), LstmInputParamsInfo::GetInputToCellWeights(), LstmInputParamsInfo::GetInputToForgetWeights(), LstmInputParamsInfo::GetInputToInputWeights(), LstmInputParamsInfo::GetInputToOutputWeights(), LstmInputParamsInfo::GetOutputGateBias(), LstmInputParamsInfo::GetOutputLayerNormWeights(), LstmInputParamsInfo::GetProjectionBias(), LstmInputParamsInfo::GetProjectionWeights(), LstmInputParamsInfo::GetRecurrentToCellWeights(), LstmInputParamsInfo::GetRecurrentToForgetWeights(), LstmInputParamsInfo::GetRecurrentToInputWeights(), LstmInputParamsInfo::GetRecurrentToOutputWeights(), armnn::IgnoreUnused(), LstmDescriptor::m_CifgEnabled, LstmDescriptor::m_LayerNormEnabled, LstmDescriptor::m_PeepholeEnabled, LstmInputParamsInfo::m_ProjectionBias, LstmDescriptor::m_ProjectionEnabled, and armnn::QSymmS16.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsMaximumSupported()
Definition at line 1900 of file RefLayerSupport.cpp.
1905 bool supported =
true;
1907 std::array<DataType,7> supportedTypes = {
1916 supported &=
CheckSupportRule(TypeAnyOf(input0, supportedTypes), reasonIfUnsupported,
1917 "Reference maximum: input 0 is not a supported type.");
1919 supported &=
CheckSupportRule(TypeAnyOf(input1, supportedTypes), reasonIfUnsupported,
1920 "Reference maximum: input 1 is not a supported type.");
1922 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
1923 "Reference maximum: output is not a supported type.");
1925 supported &=
CheckSupportRule(TypesAreEqual(input0, input1), reasonIfUnsupported,
1926 "Reference maximum: input 0 and Input 1 types are mismatched");
1928 supported &=
CheckSupportRule(TypesAreEqual(input0, output), reasonIfUnsupported,
1929 "Reference maximum: input and output types are mismatched");
1931 supported &=
CheckSupportRule(ShapesAreBroadcastCompatible(input0, input1, output), reasonIfUnsupported,
1932 "Reference maximum: shapes are not suitable for implicit broadcast.");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, and armnn::Signed32.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsMeanSupported()
Definition at line 1937 of file RefLayerSupport.cpp.
1942 bool supported =
true;
1943 std::string meanLayerStr =
"Mean";
1944 std::string outputTensorStr =
"output";
1946 std::array<DataType,6> supportedTypes =
1955 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
1956 "Reference Mean: input type not supported.");
1958 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
1959 "Reference Mean: input and output types are mismatched");
1961 if (descriptor.m_KeepDims)
1963 supported &=
CheckSupportRule(TensorNumDimensionsAreCorrect(output, input.GetNumDimensions()),
1964 reasonIfUnsupported,
1965 CreateIncorrectDimensionsErrorMsg(input.GetNumDimensions(),
1966 output.GetNumDimensions(),
1967 meanLayerStr, outputTensorStr).data());
1969 else if (descriptor.m_Axis.empty())
1972 reasonIfUnsupported,
1973 CreateIncorrectDimensionsErrorMsg(1, output.GetNumDimensions(),
1974 meanLayerStr, outputTensorStr).data());
1978 auto outputDim = input.GetNumDimensions() - armnn::numeric_cast<unsigned int>(descriptor.m_Axis.size());
1982 supported &=
CheckSupportRule(TensorNumDimensionsAreCorrect(output, outputDim),
1983 reasonIfUnsupported,
1984 CreateIncorrectDimensionsErrorMsg(outputDim, output.GetNumDimensions(),
1985 meanLayerStr, outputTensorStr).data());
1990 reasonIfUnsupported,
1991 CreateIncorrectDimensionsErrorMsg(1, output.GetNumDimensions(),
1992 meanLayerStr, outputTensorStr).data());
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, TensorInfo::GetNumDimensions(), MeanDescriptor::m_Axis, MeanDescriptor::m_KeepDims, armnn::QAsymmS8, armnn::QAsymmU8, and armnn::QSymmS16.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsMemCopySupported()
Definition at line 1999 of file RefLayerSupport.cpp.
2003 bool supported =
true;
2005 std::array<DataType,7> supportedTypes =
2016 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
2017 "Reference MemCopy: input type not supported");
2019 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
2020 "Reference MemCopy: output type not supported");
2022 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
2023 "Reference MemCopy: input and output types are mismatched");
References armnn::BFloat16, armnn::Boolean, armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::QAsymmS8, armnn::QAsymmU8, and armnn::QSymmS16.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsMinimumSupported()
Definition at line 2028 of file RefLayerSupport.cpp.
2033 bool supported =
true;
2035 std::array<DataType,7> supportedTypes = {
2044 supported &=
CheckSupportRule(TypeAnyOf(input0, supportedTypes), reasonIfUnsupported,
2045 "Reference minimum: input 0 is not a supported type.");
2047 supported &=
CheckSupportRule(TypeAnyOf(input1, supportedTypes), reasonIfUnsupported,
2048 "Reference minimum: input 1 is not a supported type.");
2050 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
2051 "Reference minimum: output is not a supported type.");
2053 supported &=
CheckSupportRule(TypesAreEqual(input0, input1), reasonIfUnsupported,
2054 "Reference minimum: input 0 and Input 1 types are mismatched");
2056 supported &=
CheckSupportRule(TypesAreEqual(input0, output), reasonIfUnsupported,
2057 "Reference minimum: input and output types are mismatched");
2059 supported &=
CheckSupportRule(ShapesAreBroadcastCompatible(input0, input1, output), reasonIfUnsupported,
2060 "Reference minimum: shapes are not suitable for implicit broadcast.");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, and armnn::Signed32.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsMultiplicationSupported()
Definition at line 2065 of file RefLayerSupport.cpp.
2070 bool supported =
true;
2072 std::array<DataType,7> supportedTypes = {
2081 supported &=
CheckSupportRule(TypeAnyOf(input0, supportedTypes), reasonIfUnsupported,
2082 "Reference multiplication: input 0 is not a supported type.");
2084 supported &=
CheckSupportRule(TypeAnyOf(input1, supportedTypes), reasonIfUnsupported,
2085 "Reference multiplication: input 1 is not a supported type.");
2087 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
2088 "Reference multiplication: output is not a supported type.");
2090 supported &=
CheckSupportRule(TypesAreEqual(input0, input1), reasonIfUnsupported,
2091 "Reference multiplication: input 0 and Input 1 types are mismatched");
2093 supported &=
CheckSupportRule(TypesAreEqual(input0, output), reasonIfUnsupported,
2094 "Reference multiplication: input and output types are mismatched");
2096 supported &=
CheckSupportRule(ShapesAreBroadcastCompatible(input0, input1, output), reasonIfUnsupported,
2097 "Reference multiplication: shapes are not suitable for implicit broadcast.");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, and armnn::Signed32.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsNormalizationSupported()
Definition at line 2102 of file RefLayerSupport.cpp.
2110 std::array<DataType, 6> supportedTypes =
2119 bool supported =
true;
2121 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
2122 "Reference normalization: input type not supported.");
2124 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
2125 "Reference normalization: output type not supported.");
2127 supported &=
CheckSupportRule(ShapesAreSameTotalSize(input, output), reasonIfUnsupported,
2128 "Reference normalization: input and output shapes have different "
2129 "num total elements.");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::IgnoreUnused(), armnn::QAsymmS8, armnn::QAsymmU8, and armnn::QSymmS16.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsOutputSupported()
◆ IsPadSupported()
Definition at line 2140 of file RefLayerSupport.cpp.
2146 bool supported =
true;
2149 std::array<DataType,6> supportedTypes =
2158 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
2159 "Reference pad: input is not a supported type.");
2161 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
2162 "Reference pad: output is not a supported type.");
2164 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
2165 "Reference pad: input and output types are mismatched.");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::IgnoreUnused(), armnn::QAsymmS8, armnn::QAsymmU8, and armnn::QSymmS16.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsPermuteSupported()
Definition at line 2170 of file RefLayerSupport.cpp.
2176 bool supported =
true;
2179 std::array<DataType, 6> supportedTypes =
2189 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
2190 "Reference permute: input is not a supported type.");
2192 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
2193 "Reference permute: output is not a supported type.");
2195 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
2196 "Reference permute: input and output types are mismatched.");
References armnn::BFloat16, armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::IgnoreUnused(), armnn::QAsymmS8, armnn::QAsymmU8, and armnn::QSymmS16.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsPooling2dSupported()
Definition at line 2201 of file RefLayerSupport.cpp.
2207 bool supported =
true;
2210 std::array<DataType,6> supportedTypes =
2219 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
2220 "Reference poolind2d: input is not a supported type.");
2222 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
2223 "Reference poolind2d: output is not a supported type.");
2225 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
2226 "Reference poolind2d: input and output types are mismatched.");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::IgnoreUnused(), armnn::QAsymmS8, armnn::QAsymmU8, and armnn::QSymmS16.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsPooling3dSupported()
Definition at line 2231 of file RefLayerSupport.cpp.
2237 bool supported =
true;
2240 std::array<DataType,6> supportedTypes =
2249 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
2250 "Reference poolind3d: input is not a supported type.");
2252 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
2253 "Reference poolind3d: output is not a supported type.");
2255 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
2256 "Reference poolind3d: input and output types are mismatched.");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::IgnoreUnused(), armnn::QAsymmS8, armnn::QAsymmU8, and armnn::QSymmS16.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsPreluSupported()
Definition at line 2777 of file RefLayerSupport.cpp.
2782 bool supported =
true;
2784 std::array<DataType, 6> supportedTypes
2793 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
2794 "PReLU: input is not a supported type.");
2796 supported &=
CheckSupportRule(TypeAnyOf(alpha, supportedTypes), reasonIfUnsupported,
2797 "PReLU: alpha is not a supported type.");
2799 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
2800 "PReLU: output is not a supported type.");
2802 supported &=
CheckSupportRule(TypesAreEqual(input, alpha, output), reasonIfUnsupported,
2803 "PReLU: input, alpha and output types are mismatched");
2805 supported &=
CheckSupportRule(ShapesAreBroadcastCompatible(input, alpha, output), reasonIfUnsupported,
2806 "PReLU: shapes are not suitable for implicit broadcast");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::QAsymmS8, armnn::QAsymmU8, and armnn::QSymmS16.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsQLstmSupported()
◆ IsQuantizeSupported()
Definition at line 2286 of file RefLayerSupport.cpp.
2290 bool supported =
true;
2293 std::array<DataType,7> supportedInputTypes = {
2302 supported &=
CheckSupportRule(TypeAnyOf(input, supportedInputTypes), reasonIfUnsupported,
2303 "Reference quantize: input type not supported.");
2306 std::array<DataType,4> supportedOutputTypes = {
2312 supported &=
CheckSupportRule(TypeAnyOf(output, supportedOutputTypes), reasonIfUnsupported,
2313 "Reference quantize: output type not supported.");
2315 supported &=
CheckSupportRule(ShapesAreSameTotalSize(input, output), reasonIfUnsupported,
2316 "Reference quantize: input and output shapes have different num total elements.");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, and armnn::QSymmS8.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsRankSupported()
◆ IsReduceSupported()
Definition at line 2336 of file RefLayerSupport.cpp.
2342 bool supported =
true;
2343 std::array<DataType,7> supportedTypes =
2353 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
2354 "Reference Reduce: input type not supported");
2356 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
2357 "Reference Reduce: output type not supported");
2359 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
2360 "Reference Reduce: input and output types not matching");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::IgnoreUnused(), armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, and armnn::Signed32.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsReshapeSupported()
Definition at line 2365 of file RefLayerSupport.cpp.
2373 std::array<DataType,8> supportedOutputTypes =
2385 return CheckSupportRule(TypeAnyOf(input, supportedOutputTypes), reasonIfUnsupported,
2386 "Reference reshape: input type not supported.");
References armnn::BFloat16, armnn::Boolean, armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::IgnoreUnused(), armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, and armnn::Signed32.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsResizeSupported()
Definition at line 2389 of file RefLayerSupport.cpp.
2395 bool supported =
true;
2396 std::array<DataType,7> supportedTypes =
2407 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
2408 "Reference Resize: input type not supported");
2410 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
2411 "Reference Resize: output type not supported");
2413 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
2414 "Reference Resize: input and output types not matching");
References armnn::BFloat16, armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::IgnoreUnused(), armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, and armnn::QSymmS8.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsReverseV2Supported()
Definition at line 2419 of file RefLayerSupport.cpp.
2424 bool supported =
true;
2426 std::array<DataType,8> supportedTypes =
2438 supported &=
CheckSupportRule(TypeAnyOf(input0, supportedTypes), reasonIfUnsupported,
2439 "Reference ReverseV2: input0 type not supported");
2441 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
2442 "Reference ReverseV2: output type not supported");
2444 supported &=
CheckSupportRule(TypesAreEqual(input0, output), reasonIfUnsupported,
2445 "Reference ReverseV2: input0 and output types not matching");
2447 std::array<DataType,6> input2SupportedTypes =
2452 supported &=
CheckSupportRule(TypeAnyOf(input1, input2SupportedTypes), reasonIfUnsupported,
2453 "Reference ReverseV2: input1 type not supported");
References armnn::BFloat16, armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, armnn::QSymmS8, and armnn::Signed32.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsScatterNdSupported()
Definition at line 2458 of file RefLayerSupport.cpp.
2467 bool supported =
true;
2469 std::array<DataType, 7> supportedTypes
2480 std::array<DataType, 1> indicesSupportedTypes =
2485 supported &=
CheckSupportRule(TypeAnyOf(indices, indicesSupportedTypes), reasonIfUnsupported,
2486 "ScatterNd: indices type not supported.");
2488 supported &=
CheckSupportRule(TypeAnyOf(updates, supportedTypes), reasonIfUnsupported,
2489 "ScatterNd: updates type not supported.");
2491 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
2492 "ScatterNd: output type not supported");
2494 supported &=
CheckSupportRule(TypesAreEqual(updates, output), reasonIfUnsupported,
2495 "ScatterNd: input and updates types are mismatched");
2497 if (descriptor.m_InputEnabled)
2500 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
2501 "ScatterNd: input type not supported.");
2503 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
2504 "ScatterNd: input and output types are mismatched");
2509 supported &=
CheckSupportRule(TypeAnyOf(input, indicesSupportedTypes), reasonIfUnsupported,
2510 "ScatterNd: shape type not supported.");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::IgnoreUnused(), ScatterNdDescriptor::m_InputEnabled, armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, armnn::QSymmS8, and armnn::Signed32.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsShapeSupported()
◆ IsSliceSupported()
Definition at line 2534 of file RefLayerSupport.cpp.
2540 bool supported =
true;
2542 std::array<DataType, 5> supportedTypes =
2551 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
2552 "Reference Slice: input type not supported");
2554 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
2555 "Reference Slice: output type not supported");
2557 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
2558 "Reference Slice: input and output types are mismatched");
References armnn::CheckSupportRule(), armnn::Float32, armnn::IgnoreUnused(), armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, and armnn::Signed32.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsSoftmaxSupported()
Definition at line 2563 of file RefLayerSupport.cpp.
2569 bool supported =
true;
2570 std::array<DataType,7> supportedTypes =
2580 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
2581 "Reference Softmax: output type not supported");
2583 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
2584 "Reference Softmax: input type not supported");
2586 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
2587 "Reference Softmax: input type not supported");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::IgnoreUnused(), armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, and armnn::QSymmS8.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsSpaceToBatchNdSupported()
Definition at line 2592 of file RefLayerSupport.cpp.
2598 bool supported =
true;
2599 std::array<DataType,6> supportedTypes =
2608 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
2609 "Reference SpaceToBatchNd: input type not supported");
2611 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
2612 "Reference SpaceToBatchNd: output type not supported");
2614 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
2615 "Reference SpaceToBatchNd: input and output types are mismatched");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::IgnoreUnused(), armnn::QAsymmS8, armnn::QAsymmU8, and armnn::QSymmS16.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsSpaceToDepthSupported()
Definition at line 2620 of file RefLayerSupport.cpp.
2627 bool supported =
true;
2629 std::array<DataType,6> supportedTypes =
2638 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
2639 "Reference SpaceToDepth: input type not supported");
2641 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
2642 "Reference SpaceToDepth: output type not supported");
2644 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
2645 "Reference SpaceToDepth: input and output types are mismatched");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::IgnoreUnused(), armnn::QAsymmS8, armnn::QAsymmU8, and armnn::QSymmS16.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsSplitterSupported()
Definition at line 2650 of file RefLayerSupport.cpp.
2656 bool supported =
true;
2657 std::array<DataType,6> supportedTypes =
2666 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
2667 "Reference splitter: output type not supported");
2668 for (
const TensorInfo& output : outputs)
2670 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
2671 "Reference splitter: input type not supported");
2673 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
2674 "Reference splitter: input and output types mismatched.");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::IgnoreUnused(), armnn::QAsymmS8, armnn::QAsymmU8, and armnn::QSymmS16.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsStackSupported()
Definition at line 2680 of file RefLayerSupport.cpp.
2687 bool supported =
true;
2688 std::array<DataType,7> supportedTypes =
2698 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
2699 "Reference stack: output type not supported");
2700 for (
const TensorInfo* input : inputs)
2702 supported &=
CheckSupportRule(TypeAnyOf(*input, supportedTypes), reasonIfUnsupported,
2703 "Reference stack: input type not supported");
2705 supported &=
CheckSupportRule(TypesAreEqual(*input, output), reasonIfUnsupported,
2706 "Reference stack: input and output types mismatched.");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::IgnoreUnused(), armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, and armnn::Signed32.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsStridedSliceSupported()
Definition at line 2712 of file RefLayerSupport.cpp.
2718 bool supported =
true;
2720 std::array<DataType,5> supportedTypes =
2728 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
2729 "Reference StridedSlice: input type not supported");
2731 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
2732 "Reference StridedSlice: output type not supported");
2734 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
2735 "Reference StridedSlice: input and output types are mismatched");
References armnn::CheckSupportRule(), armnn::Float32, armnn::IgnoreUnused(), armnn::QAsymmS8, armnn::QAsymmU8, and armnn::QSymmS16.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsSubtractionSupported()
Definition at line 2740 of file RefLayerSupport.cpp.
2745 bool supported =
true;
2747 std::array<DataType,7> supportedTypes = {
2756 supported &=
CheckSupportRule(TypeAnyOf(input0, supportedTypes), reasonIfUnsupported,
2757 "Reference subtraction: input 0 is not a supported type.");
2759 supported &=
CheckSupportRule(TypeAnyOf(input1, supportedTypes), reasonIfUnsupported,
2760 "Reference subtraction: input 1 is not a supported type.");
2762 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
2763 "Reference subtraction: output is not a supported type.");
2765 supported &=
CheckSupportRule(TypesAreEqual(input0, input1), reasonIfUnsupported,
2766 "Reference subtraction: input 0 and Input 1 types are mismatched");
2768 supported &=
CheckSupportRule(TypesAreEqual(input0, output), reasonIfUnsupported,
2769 "Reference subtraction: input and output types are mismatched");
2771 supported &=
CheckSupportRule(ShapesAreBroadcastCompatible(input0, input1, output), reasonIfUnsupported,
2772 "Reference subtraction: shapes are not suitable for implicit broadcast.");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, and armnn::Signed32.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsTileSupported()
Definition at line 2811 of file RefLayerSupport.cpp.
2818 bool supported =
true;
2820 std::array<DataType, 7> supportedTypes
2831 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
2832 "Tile: input type not supported.");
2834 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
2835 "Tile: output type not supported");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::IgnoreUnused(), armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, armnn::QSymmS8, and armnn::Signed32.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsTransposeConvolution2dSupported()
Definition at line 2840 of file RefLayerSupport.cpp.
2848 bool supported =
true;
2850 std::array<DataType,7> supportedTypes =
2860 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
2861 "Reference TransposeConvolution2d: input is not a supported type.");
2863 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
2864 "Reference TransposeConvolution2d: output is not a supported type.");
2866 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
2867 "Reference TransposeConvolution2d: input and output types mismatched.");
2870 const DataType inputType = input.GetDataType();
2873 std::array<DataType, 3> supportedWeightTypes =
2880 supported &=
CheckSupportRule(TypeAnyOf(weights, supportedWeightTypes), reasonIfUnsupported,
2881 "Reference TransposeConvolution2d: weights type not supported for "
2882 "quantized input.");
2886 supported &=
CheckSupportRule(TypeAnyOf(weights, supportedTypes), reasonIfUnsupported,
2887 "Reference TransposeConvolution2d: weights is not a supported type.");
2889 supported &=
CheckSupportRule(TypesAreEqual(input, weights), reasonIfUnsupported,
2890 "Reference TransposeConvolution2d: input and weights types mismatched.");
2893 if (biases.has_value())
2895 std::array<DataType,4> biasesSupportedTypes =
2901 supported &=
CheckSupportRule(TypeAnyOf(biases.value(), biasesSupportedTypes), reasonIfUnsupported,
2902 "Reference TransposeConvolution2d: biases is not a supported type.");
References armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, TensorInfo::GetDataType(), OptionalBase::has_value(), armnn::IgnoreUnused(), armnn::IsQuantized8BitType(), armnn::QAsymmS8, armnn::QAsymmU8, armnn::QSymmS16, armnn::QSymmS8, armnn::Signed32, and OptionalReferenceSwitch< std::is_reference< T >::value, T >::value().
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsTransposeSupported()
Definition at line 2908 of file RefLayerSupport.cpp.
2914 bool supported =
true;
2917 std::array<DataType, 6> supportedTypes =
2927 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
2928 "Reference transpose: input is not a supported type.");
2930 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
2931 "Reference transpose: output is not a supported type.");
2933 supported &=
CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
2934 "Reference transpose: input and output types are mismatched.");
References armnn::BFloat16, armnn::CheckSupportRule(), armnn::Float16, armnn::Float32, armnn::IgnoreUnused(), armnn::QAsymmS8, armnn::QAsymmU8, and armnn::QSymmS16.
Referenced by RefLayerSupport::IsLayerSupported().
◆ IsUnidirectionalSequenceLstmSupported()
Definition at line 2939 of file RefLayerSupport.cpp.
2956 bool supported =
true;
2958 std::array<DataType, 2> supportedTypes =
2964 std::array<DataType, 2> supportedWeightTypes =
2970 std::array<DataType, 3> supportedBiasTypes =
2978 supported &=
CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
2979 "Reference UnidirectionalSequenceLstm: input is not a supported type.");
2980 supported &=
CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
2981 "Reference UnidirectionalSequenceLstm: output is not a supported type.");
2984 supported &=
CheckSupportRule(TypeAnyOf(paramsInfo.GetInputToForgetWeights(), supportedWeightTypes),
2985 reasonIfUnsupported,
2986 "Reference UnidirectionalSequenceLstm: InputToForgetWeights "
2987 "is not a supported type.");
2988 supported &=
CheckSupportRule(TypeAnyOf(paramsInfo.GetInputToCellWeights(), supportedWeightTypes),
2989 reasonIfUnsupported,
2990 "Reference UnidirectionalSequenceLstm: InputToCellWeights is not a supported type.");
2991 supported &=
CheckSupportRule(TypeAnyOf(paramsInfo.GetInputToOutputWeights(), supportedWeightTypes),
2992 reasonIfUnsupported,
2993 "Reference UnidirectionalSequenceLstm: InputToOutputWeights "
2994 "is not a supported type.");
2995 supported &=
CheckSupportRule(TypeAnyOf(paramsInfo.GetRecurrentToForgetWeights(), supportedWeightTypes),
2996 reasonIfUnsupported,
2997 "Reference UnidirectionalSequenceLstm: RecurrentToForgetWeights "
2998 "is not a supported type.");
2999 supported &=
CheckSupportRule(TypeAnyOf(paramsInfo.GetRecurrentToCellWeights(), supportedWeightTypes),
3000 reasonIfUnsupported,
3001 "Reference UnidirectionalSequenceLstm: RecurrentToCellWeights "
3002 "is not a supported type.");
3003 supported &=
CheckSupportRule(TypeAnyOf(paramsInfo.GetRecurrentToOutputWeights(), supportedWeightTypes),
3004 reasonIfUnsupported,
3005 "Reference UnidirectionalSequenceLstm: RecurrentToOutputWeights "
3006 "is not a supported type.");
3008 supported &=
CheckSupportRule(TypeAnyOf(paramsInfo.GetForgetGateBias(), supportedBiasTypes), reasonIfUnsupported,
3009 "Reference UnidirectionalSequenceLstm: ForgetGateBias is not a supported type.");
3010 supported &=
CheckSupportRule(TypeAnyOf(paramsInfo.GetCellBias(), supportedBiasTypes), reasonIfUnsupported,
3011 "Reference UnidirectionalSequenceLstm: CellBias is not a supported type.");
3012 supported &=
CheckSupportRule(TypeAnyOf(paramsInfo.GetOutputGateBias(), supportedBiasTypes), reasonIfUnsupported,
3013 "Reference UnidirectionalSequenceLstm: OutputGateBias is not a supported type.");
3014 if (!descriptor.m_CifgEnabled)
3016 supported &=
CheckSupportRule(TypeAnyOf(paramsInfo.GetInputToInputWeights(), supportedWeightTypes),
3017 reasonIfUnsupported,
3018 "Reference UnidirectionalSequenceLstm: InputToInputWeights "
3019 "is not a supported type.");
3020 supported &=
CheckSupportRule(TypeAnyOf(paramsInfo.GetRecurrentToInputWeights(), supportedWeightTypes),
3021 reasonIfUnsupported,
3022 "Reference UnidirectionalSequenceLstm: RecurrentToInputWeights "
3023 "is not a supported type.");
3024 supported &=
CheckSupportRule(TypeAnyOf(paramsInfo.GetInputGateBias(), supportedBiasTypes), reasonIfUnsupported,
3025 "Reference UnidirectionalSequenceLstm: InputGateBias is not a supported type.");
3026 if (descriptor.m_PeepholeEnabled)
3028 supported &=
CheckSupportRule(TypeAnyOf(paramsInfo.GetCellToInputWeights(), supportedWeightTypes),
3029 reasonIfUnsupported,
3030 "Reference UnidirectionalSequenceLstm: CellToInputWeights "
3031 "is not a supported type.");
3034 if (descriptor.m_PeepholeEnabled)
3036 supported &=
CheckSupportRule(TypeAnyOf(paramsInfo.GetCellToForgetWeights(), supportedWeightTypes),
3037 reasonIfUnsupported,
3038 "Reference UnidirectionalSequenceLstm: CellToForgetWeights "
3039 "is not a supported type.");
3040 supported &=
CheckSupportRule(TypeAnyOf(paramsInfo.GetCellToOutputWeights(), supportedWeightTypes),
3041 reasonIfUnsupported,
3042 "Reference UnidirectionalSequenceLstm: CellToOutputWeights "
3043 "is not a supported type.");
3045 if (descriptor.m_ProjectionEnabled)
3047 supported &=
CheckSupportRule(TypeAnyOf(paramsInfo.GetProjectionWeights(), supportedWeightTypes),
3048 reasonIfUnsupported,
3049 "Reference UnidirectionalSequenceLstm: ProjectionWeights "
3050 "is not a supported type.");
3051 if (paramsInfo.m_ProjectionBias !=
nullptr)
3053 supported &=
CheckSupportRule(TypesAreEqual(input, paramsInfo.GetProjectionBias()), reasonIfUnsupported,
3054 "Reference UnidirectionalSequenceLstm: input and ProjectionBias types "
3058 if (descriptor.m_LayerNormEnabled)
3060 if (!descriptor.m_CifgEnabled)
3062 supported &=
CheckSupportRule(TypeAnyOf(paramsInfo.GetInputLayerNormWeights(), supportedWeightTypes),
3063 reasonIfUnsupported,
3064 "Reference UnidirectionalSequenceLstm: InputLayerNormWeights "
3065 "is not a supported type.");
3067 supported &=
CheckSupportRule(TypeAnyOf(paramsInfo.GetForgetLayerNormWeights(), supportedWeightTypes),
3068 reasonIfUnsupported,
3069 "Reference UnidirectionalSequenceLstm: ForgetLayerNormWeights "
3070 "is not a supported type.");
3071 supported &=
CheckSupportRule(TypeAnyOf(paramsInfo.GetCellLayerNormWeights(), supportedWeightTypes),
3072 reasonIfUnsupported,
3073 "Reference UnidirectionalSequenceLstm: CellLayerNormWeights "
3074 "is not a supported type.");
3075 supported &=
CheckSupportRule(TypeAnyOf(paramsInfo.GetOutputLayerNormWeights(), supportedWeightTypes),
3076 reasonIfUnsupported,
3077 "Reference UnidirectionalSequenceLstm: OutputLayerNormWeights "
3078 "is not a supported type.");
References armnn::CheckSupportRule(), armnn::Float32, LstmInputParamsInfo::GetCellBias(), LstmInputParamsInfo::GetCellLayerNormWeights(), LstmInputParamsInfo::GetCellToForgetWeights(), LstmInputParamsInfo::GetCellToInputWeights(), LstmInputParamsInfo::GetCellToOutputWeights(), LstmInputParamsInfo::GetForgetGateBias(), LstmInputParamsInfo::GetForgetLayerNormWeights(), LstmInputParamsInfo::GetInputGateBias(), LstmInputParamsInfo::GetInputLayerNormWeights(), LstmInputParamsInfo::GetInputToCellWeights(), LstmInputParamsInfo::GetInputToForgetWeights(), LstmInputParamsInfo::GetInputToInputWeights(), LstmInputParamsInfo::GetInputToOutputWeights(), LstmInputParamsInfo::GetOutputGateBias(), LstmInputParamsInfo::GetOutputLayerNormWeights(), LstmInputParamsInfo::GetProjectionBias(), LstmInputParamsInfo::GetProjectionWeights(), LstmInputParamsInfo::GetRecurrentToCellWeights(), LstmInputParamsInfo::GetRecurrentToForgetWeights(), LstmInputParamsInfo::GetRecurrentToInputWeights(), LstmInputParamsInfo::GetRecurrentToOutputWeights(), armnn::IgnoreUnused(), LstmDescriptor::m_CifgEnabled, LstmDescriptor::m_LayerNormEnabled, LstmDescriptor::m_PeepholeEnabled, LstmInputParamsInfo::m_ProjectionBias, LstmDescriptor::m_ProjectionEnabled, armnn::QAsymmS8, and armnn::Signed32.
Referenced by RefLayerSupport::IsLayerSupported().
The documentation for this class was generated from the following files:
bool IsLstmSupported(const TensorInfo &input, const TensorInfo &outputStateIn, const TensorInfo &cellStateIn, const TensorInfo &scratchBuffer, const TensorInfo &outputStateOut, const TensorInfo &cellStateOut, const TensorInfo &output, const LstmDescriptor &descriptor, const LstmInputParamsInfo ¶msInfo, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsDepthwiseConvolutionSupported(const TensorInfo &input, const TensorInfo &output, const DepthwiseConvolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsPooling3dSupported(const TensorInfo &input, const TensorInfo &output, const Pooling3dDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsReverseV2Supported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsQLstmSupported(const TensorInfo &input, const TensorInfo &previousOutputIn, const TensorInfo &previousCellStateIn, const TensorInfo &outputStateOut, const TensorInfo &cellStateOut, const TensorInfo &output, const QLstmDescriptor &descriptor, const LstmInputParamsInfo ¶msInfo, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
constexpr bool IsQuantized8BitType(DataType dataType)
bool IsTileSupported(const TensorInfo &input, const TensorInfo &output, const TileDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsConvolution2dSupported(const TensorInfo &input, const TensorInfo &output, const Convolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsFloorSupported(const TensorInfo &input, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsDequantizeSupported(const TensorInfo &input, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsPooling2dSupported(const TensorInfo &input, const TensorInfo &output, const Pooling2dDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsDetectionPostProcessSupported(const TensorInfo &boxEncodings, const TensorInfo &scores, const TensorInfo &anchors, const TensorInfo &detectionBoxes, const TensorInfo &detectionClasses, const TensorInfo &detectionScores, const TensorInfo &numDetections, const DetectionPostProcessDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsAdditionSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsInputSupported(const TensorInfo &input, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsMeanSupported(const TensorInfo &input, const TensorInfo &output, const MeanDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsTransposeConvolution2dSupported(const TensorInfo &input, const TensorInfo &output, const TransposeConvolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsRankSupported(const TensorInfo &input, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
@ BoundedReLu
min(a, max(b, input)) ReLu1 & ReLu6.
bool IsSplitterSupported(const TensorInfo &input, const std::vector< std::reference_wrapper< TensorInfo >> &outputs, const ViewsDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsConvolution3dSupported(const TensorInfo &input, const TensorInfo &output, const Convolution3dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsBatchMatMulSupported(const TensorInfo &inputX, const TensorInfo &inputY, const TensorInfo &output, const BatchMatMulDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsSubtractionSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsQuantizeSupported(const TensorInfo &input, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsArgMinMaxSupported(const TensorInfo &input, const TensorInfo &output, const ArgMinMaxDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsPermuteSupported(const TensorInfo &input, const TensorInfo &output, const PermuteDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsComparisonSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ComparisonDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsBatchNormalizationSupported(const TensorInfo &input, const TensorInfo &output, const TensorInfo &mean, const TensorInfo &var, const TensorInfo &beta, const TensorInfo &gamma, const BatchNormalizationDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsSpaceToDepthSupported(const TensorInfo &input, const TensorInfo &output, const SpaceToDepthDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsActivationSupported(const TensorInfo &input, const TensorInfo &output, const ActivationDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsFakeQuantizationSupported(const TensorInfo &input, const FakeQuantizationDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsQuantizedLstmSupported(const TensorInfo &input, const TensorInfo &previousCellStateIn, const TensorInfo &previousOutputIn, const TensorInfo &cellStateOut, const TensorInfo &output, const QuantizedLstmInputParamsInfo ¶msInfo, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsElementwiseUnarySupported(const TensorInfo &input, const TensorInfo &output, const ElementwiseUnaryDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsConcatSupported(const std::vector< const TensorInfo * > inputs, const TensorInfo &output, const OriginsDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsReduceSupported(const TensorInfo &input, const TensorInfo &output, const ReduceDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsResizeSupported(const TensorInfo &input, const TensorInfo &output, const ResizeDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsDepthToSpaceSupported(const TensorInfo &input, const TensorInfo &output, const DepthToSpaceDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsBatchToSpaceNdSupported(const TensorInfo &input, const TensorInfo &output, const BatchToSpaceNdDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsBroadcastToSupported(const TensorInfo &input, const TensorInfo &output, const BroadcastToDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsSpaceToBatchNdSupported(const TensorInfo &input, const TensorInfo &output, const SpaceToBatchNdDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsMergeSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsL2NormalizationSupported(const TensorInfo &input, const TensorInfo &output, const L2NormalizationDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsMaximumSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsMemImportSupported(const TensorInfo &input, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsReshapeSupported(const TensorInfo &input, const TensorInfo &output, const ReshapeDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsCastSupported(const TensorInfo &input, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsFullyConnectedSupported(const TensorInfo &input, const TensorInfo &output, const TensorInfo &weights, const TensorInfo &biases, const FullyConnectedDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsOutputSupported(const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsDebugSupported(const TensorInfo &input, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsConvertFp16ToFp32Supported(const TensorInfo &input, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsLogicalBinarySupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const LogicalBinaryDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported) const
bool IsDivisionSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsNormalizationSupported(const TensorInfo &input, const TensorInfo &output, const NormalizationDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsPreluSupported(const TensorInfo &input, const TensorInfo &alpha, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsFillSupported(const TensorInfo &input, const TensorInfo &output, const FillDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
void IgnoreUnused(Ts &&...)
bool IsMultiplicationSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsSupportedForDataTypeGeneric(Optional< std::string & > reasonIfUnsupported, DataType dataType, Float16Func float16FuncPtr, Float32Func float32FuncPtr, Uint8Func uint8FuncPtr, Int32Func int32FuncPtr, BooleanFunc booleanFuncPtr, Params &&... params)
@ UnidirectionalSequenceLstm
bool IsStridedSliceSupported(const TensorInfo &input, const TensorInfo &output, const StridedSliceDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsMemCopySupported(const TensorInfo &input, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsScatterNdSupported(const TensorInfo &input, const TensorInfo &indices, const TensorInfo &updates, const TensorInfo &output, const ScatterNdDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsConstantSupported(const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsChannelShuffleSupported(const TensorInfo &input, const TensorInfo &output, const ChannelShuffleDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsShapeSupported(const TensorInfo &input, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsTransposeSupported(const TensorInfo &input, const TensorInfo &output, const TransposeDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsConvertFp32ToFp16Supported(const TensorInfo &input, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsGatherSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const GatherDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsStackSupported(const std::vector< const TensorInfo * > &inputs, const TensorInfo &output, const StackDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool CheckSupportRule(F rule, Optional< std::string & > reasonIfUnsupported, const char *reason)
bool IsLogSoftmaxSupported(const TensorInfo &input, const TensorInfo &output, const LogSoftmaxDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported) const
bool IsSoftmaxSupported(const TensorInfo &input, const TensorInfo &output, const SoftmaxDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsSliceSupported(const TensorInfo &input, const TensorInfo &output, const SliceDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsPadSupported(const TensorInfo &input, const TensorInfo &output, const PadDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsGatherNdSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsUnidirectionalSequenceLstmSupported(const TensorInfo &input, const TensorInfo &outputStateIn, const TensorInfo &cellStateIn, const TensorInfo &outputStateOut, const TensorInfo &cellStateOut, const TensorInfo &output, const UnidirectionalSequenceLstmDescriptor &descriptor, const LstmInputParamsInfo ¶msInfo, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsInstanceNormalizationSupported(const TensorInfo &input, const TensorInfo &output, const InstanceNormalizationDescriptor &descriptor, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const
bool IsMinimumSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string & > reasonIfUnsupported=EmptyOptional()) const