|
25.11
|
Copyright (c) 2021 ARM Limited and Contributors. More...
Namespaces | |
| namespace | FoldPadConstraints |
| namespace | optimizations |
| namespace | profiling |
| namespace | stringUtils |
| namespace | timelinedecoder |
| namespace | utility |
Enumerations | |
| enum class | Compute { Undefined = 0 , CpuRef = 1 , CpuAcc = 2 , GpuAcc = 3 , TosaRef = 4 } |
| The Compute enum is now deprecated and it is now being replaced by BackendId. More... | |
| enum class | CapabilityClass { PaddingRequired = 1 , FallbackImportDisabled = 2 , CapabilityClassMax = 254 } |
| Capability class to calculate in the GetCapabilities function so that only the capability in the scope can be choose to calculate. More... | |
| enum class | EdgeStrategy { Undefined , DirectCompatibility , ExportToTarget , CopyToTarget } |
| enum class | BoostLogSeverityMapping { trace , debug , info , warning , error , fatal } |
| enum class | Status { Success = 0 , Failure = 1 } |
| enumeration More... | |
| enum class | DataType { Float16 = 0 , Float32 = 1 , QAsymmU8 = 2 , Signed32 = 3 , Boolean = 4 , QSymmS16 = 5 , QSymmS8 = 6 , QAsymmS8 = 7 , BFloat16 = 8 , Signed64 = 9 } |
| enum class | DataLayout { NCHW = 1 , NHWC = 2 , NDHWC = 3 , NCDHW = 4 } |
| enum class | ProfilingDetailsMethod { Undefined = 0 , DetailsWithEvents = 1 , DetailsOnly = 2 } |
| Define the behaviour of the internal profiler when outputting network details. More... | |
| enum class | QosExecPriority { Low = 0 , Medium = 1 , High = 2 } |
| enum class | ActivationFunction { Sigmoid = 0 , TanH = 1 , Linear = 2 , ReLu = 3 , BoundedReLu = 4 , SoftReLu = 5 , LeakyReLu = 6 , Abs = 7 , Sqrt = 8 , Square = 9 , Elu = 10 , HardSwish = 11 , Gelu = 12 } |
| enum class | ArgMinMaxFunction { Min = 0 , Max = 1 } |
| enum class | ComparisonOperation { Equal = 0 , Greater = 1 , GreaterOrEqual = 2 , Less = 3 , LessOrEqual = 4 , NotEqual = 5 } |
| enum class | LogicalBinaryOperation { LogicalAnd = 0 , LogicalOr = 1 } |
| enum class | UnaryOperation { Abs = 0 , Exp = 1 , Sqrt = 2 , Rsqrt = 3 , Neg = 4 , LogicalNot = 5 , Log = 6 , Sin = 7 , Ceil = 8 } |
| enum class | BinaryOperation { Add = 0 , Div = 1 , Maximum = 2 , Minimum = 3 , Mul = 4 , Sub = 5 , SqDiff = 6 , Power = 7 , FloorDiv = 8 } |
| enum class | PoolingAlgorithm { Max = 0 , Average = 1 , L2 = 2 } |
| enum class | ReduceOperation { Sum = 0 , Max = 1 , Mean = 2 , Min = 3 , Prod = 4 } |
| enum class | ResizeMethod { Bilinear = 0 , NearestNeighbor = 1 } |
| enum class | Dimensionality { NotSpecified = 0 , Specified = 1 , Scalar = 2 } |
| enum class | PaddingMethod { IgnoreValue = 0 , Exclude = 1 } |
| The padding method modifies the output of pooling layers. More... | |
| enum class | PaddingMode { Constant = 0 , Reflect = 1 , Symmetric = 2 } |
| The padding mode controls whether the padding should be filled with constant values (Constant), or reflect the input, either including the border values (Symmetric) or not (Reflect). More... | |
| enum class | NormalizationAlgorithmChannel { Across = 0 , Within = 1 } |
| enum class | NormalizationAlgorithmMethod { LocalBrightness = 0 , LocalContrast = 1 } |
| enum class | OutputShapeRounding { Floor = 0 , Ceiling = 1 } |
| enum class | ShapeInferenceMethod { ValidateOnly = 0 , InferAndValidate = 1 } |
| The ShapeInferenceMethod modify how the output shapes are treated. More... | |
| enum class | MemorySource : uint32_t { Undefined = 0 , Malloc = 1 , DmaBuf = 2 , DmaBufProtected = 4 , Gralloc = 8 } |
| Define the Memory Source to reduce copies. More... | |
| enum class | MemBlockStrategyType { SingleAxisPacking = 0 , MultiAxisPacking = 1 } |
| enum class | FusedKernelType { AddMulAdd = 0 } |
| enum class | BackendCapability : uint32_t { NonConstWeights , AllOrNothing } |
| BackendCapability class. More... | |
| enum class | LayerType { Activation , Addition , ArgMinMax , BatchNormalization , BatchToSpaceNd , Comparison , Concat , Constant , ConvertFp16ToFp32 , ConvertFp32ToFp16 , Convolution2d , Debug , DepthToSpace , DepthwiseConvolution2d , Dequantize , DetectionPostProcess , Division , ElementwiseUnary , FakeQuantization , Fill , Floor , FullyConnected , Gather , Input , InstanceNormalization , L2Normalization , LogicalBinary , LogSoftmax , Lstm , QLstm , Map , Maximum , Mean , MemCopy , MemImport , Merge , Minimum , Multiplication , Normalization , Output , Pad , Permute , Pooling2d , PreCompiled , Prelu , Quantize , QuantizedLstm , Reshape , Rank , Resize , Reduce , Slice , Softmax , SpaceToBatchNd , SpaceToDepth , Splitter , Stack , StandIn , StridedSlice , Subtraction , Switch , Transpose , TransposeConvolution2d , Unmap , Cast , Shape , UnidirectionalSequenceLstm , ChannelShuffle , Convolution3d , Pooling3d , GatherNd , BatchMatMul , ElementwiseBinary , ReverseV2 , Tile , Fused , BroadcastTo , ScatterNd , FirstLayer = Activation , LastLayer = ScatterNd } |
| When adding a new layer, adapt also the LastLayer enum value in the enum class LayerType below. More... | |
| enum class | ScatterNdFunction { Update = 0 , Add = 1 , Sub = 2 , Max = 3 , Min = 4 , Mul = 5 } |
| enum class | LogSeverity { Trace , Debug , Info , Warning , Error , Fatal } |
| enum class | GraphEvent { LayerAdded , LayerErased } |
| enum class | JsonObjectType { Measurement , Event , ExecObjectDesc } |
| enum class | TuningLevel { None , Rapid , Normal , Exhaustive } |
Functions | |
| LayerSupportHandle | GetILayerSupportByBackendId (const armnn::BackendId &backend) |
| Convenience function to retrieve the ILayerSupportHandle for a backend. | |
| bool | HasCapability (const std::string &name, const BackendCapabilities &capabilities) |
| Convenience function to check if a capability exists in a BackendCapabilites struct. | |
| bool | HasCapability (const std::string &name, const armnn::BackendId &backend) |
| Convenience function to check if a capability exists in a backend. | |
| bool | HasCapability (const BackendOptions::BackendOption &capability, const BackendCapabilities &capabilities) |
| Convenience function to check if a given capability matches a capability in a BackendCapabilities struct. | |
| bool | HasCapability (const BackendOptions::BackendOption &backendOption, const armnn::BackendId &backend) |
| Convenience function to check if a given capability matches a capability in a backend. | |
| bool | HasMatchingCapability (const BackendOptions::BackendOption &capability, const BackendCapabilities &capabilities) |
| Convenience function to check if a given capability matches a capability in a BackendCapabilities struct. | |
| bool | HasMatchingCapability (const BackendOptions::BackendOption &backendOption, const armnn::BackendId &backend) |
| Convenience function to check if a given capability matches a capability in a backend. | |
| Optional< const BackendOptions::BackendOption > | GetCapability (const std::string &backendCapabilityName, const BackendCapabilities &capabilities) |
| Returns a BackendCapability if the backend lists the capability The BackendCapability must then be inspected to check whether or not that BackendCapability is supported Otherwise returns an EmptyOptional if the BackendCapability is unlisted. | |
| Optional< const BackendOptions::BackendOption > | GetCapability (const std::string &backendCapabilityName, const armnn::BackendId &backend) |
| Returns a BackendCapability if the backend lists the capability The BackendCapability must then be inspected to check whether or not that BackendCapability is supported Otherwise returns an EmptyOptional if the BackendCapability is unlisted. | |
| unsigned int | GetNumberOfCacheFiles (const armnn::BackendId &backend) |
| Returns the number of cached files if backend supports caching. | |
| constexpr char const * | GetComputeDeviceAsCString (Compute compute) |
| Deprecated function that will be removed together with the Compute enum. | |
| std::ostream & | operator<< (std::ostream &os, const std::vector< Compute > &compute) |
| Deprecated function that will be removed together with the Compute enum. | |
| std::ostream & | operator<< (std::ostream &os, const std::set< Compute > &compute) |
| Deprecated function that will be removed together with the Compute enum. | |
| std::ostream & | operator<< (std::ostream &os, const Compute &compute) |
| Deprecated function that will be removed together with the Compute enum. | |
| std::ostream & | operator<< (std::ostream &os, const BackendId &id) |
| template<template< typename... > class TContainer, typename... TContainerTemplateArgs> | |
| std::ostream & | operator<< (std::ostream &os, const TContainer< BackendId, TContainerTemplateArgs... > &ids) |
| template<typename F> | |
| void | ParseOptions (const std::vector< BackendOptions > &options, BackendId backend, F f) |
| bool | ParseBooleanBackendOption (const armnn::BackendOptions::Var &value, bool defaultValue) |
| std::string | ParseStringBackendOption (const armnn::BackendOptions::Var &value, std::string defaultValue) |
| int | ParseIntBackendOption (const armnn::BackendOptions::Var &value, int defaultValue) |
| BackendRegistry & | BackendRegistryInstance () |
| std::ostream & | operator<< (std::ostream &os, const BackendVersion &backendVersion) |
| TensorShape | GetUnpaddedTensorStrides (const TensorInfo &tensorInfo) |
| DataType | GetBiasDataType (DataType inputDataType) |
| template<typename TensorShapeIt> | |
| OriginsDescriptor | CreateDescriptorForConcatenation (TensorShapeIt first, TensorShapeIt last, unsigned int concatenationDimension) |
| Convenience template to create an OriginsDescriptor to use when creating a ConcatLayer for performing concatenation of a number of input tensors. | |
| template<typename ExceptionType> | |
| void | ConditionalThrow (bool condition, const std::string &message) |
| template<typename ExceptionType> | |
| void | ConditionalThrow (bool condition) |
| template<typename ExceptionType, typename ComparedType> | |
| void | ConditionalThrowIfNotEqual (const std::string &message, const ComparedType &leftHandSide, const ComparedType &rightHandSide) |
| ComparedType must support: operator==(const ComparedType&) operator<<(ostream&, const ComparedType&) | |
| IOptimizedNetworkPtr | Optimize (const INetwork &network, const std::vector< BackendId > &backendPreferences, const IDeviceSpec &deviceSpec, const OptimizerOptionsOpaque &options=OptimizerOptionsOpaque(), Optional< std::vector< std::string > & > messages=EmptyOptional()) |
| Create an optimized version of the network. | |
| IOptimizedNetworkPtr | Optimize (const Graph &inGraph, const std::vector< BackendId > &backendPreferences, const IDeviceSpec &deviceSpec, const OptimizerOptionsOpaque &options, Optional< std::vector< std::string > & > messages=EmptyOptional()) |
| Create an optimized version of the network. | |
| IOptimizedNetworkPtr | Optimize (const Graph &inGraph, const std::vector< BackendId > &backendPreferences, const IDeviceSpec &deviceSpec, const OptimizerOptions &options, Optional< std::vector< std::string > & > messages=EmptyOptional()) |
| Accept legacy OptimizerOptions. | |
| IOptimizedNetworkPtr | Optimize (const INetwork &network, const std::vector< BackendId > &backendPreferences, const IDeviceSpec &deviceSpec, const OptimizerOptions &options, Optional< std::vector< std::string > & > messages=EmptyOptional()) |
| Accept legacy OptimizerOptions. | |
| std::string | LevelToString (LogSeverity level) |
| LogSeverity | StringToLogLevel (std::string level) |
| void | SetLogFilter (LogSeverity level) |
| void | SetAllLoggingSinks (bool standardOut, bool debugOut, bool coloured) |
| constexpr LogSeverity | ConvertLogSeverity (BoostLogSeverityMapping severity) |
| template<typename Arg, typename std::enable_if< IsMemorySource< Arg >::value >::type * = nullptr> | |
| MemorySourceFlags | Combine (Arg sourceA, Arg sourceB) |
| template<typename Arg, typename ... Args, typename std::enable_if< IsMemorySource< Arg >::value >::type * = nullptr> | |
| MemorySourceFlags | Combine (Arg source, Args... rest) |
| bool | CheckFlag (MemorySourceFlags flags, MemorySource source) |
| template<typename T, class... Args> | |
| Optional< T > | MakeOptional (Args &&... args) |
| Utility template that constructs an object of type T in-place and wraps it inside an Optional<T> object. | |
| const char * | GetLayerTypeAsCString (LayerType type) |
| constexpr char const * | GetStatusAsCString (Status status) |
| constexpr char const * | GetActivationFunctionAsCString (ActivationFunction activation) |
| constexpr char const * | GetArgMinMaxFunctionAsCString (ArgMinMaxFunction function) |
| constexpr char const * | GetComparisonOperationAsCString (ComparisonOperation operation) |
| constexpr char const * | GetBinaryOperationAsCString (BinaryOperation operation) |
| constexpr char const * | GetUnaryOperationAsCString (UnaryOperation operation) |
| constexpr char const * | GetLogicalBinaryOperationAsCString (LogicalBinaryOperation operation) |
| constexpr char const * | GetFusedTypeAsCString (FusedKernelType type) |
| constexpr char const * | GetPoolingAlgorithmAsCString (PoolingAlgorithm pooling) |
| constexpr char const * | GetOutputShapeRoundingAsCString (OutputShapeRounding rounding) |
| constexpr char const * | GetPaddingMethodAsCString (PaddingMethod method) |
| constexpr char const * | GetPaddingModeAsCString (PaddingMode mode) |
| constexpr char const * | GetReduceOperationAsCString (ReduceOperation reduce_operation) |
| constexpr unsigned int | GetDataTypeSize (DataType dataType) |
| template<unsigned N> | |
| constexpr bool | StrEqual (const char *strA, const char(&strB)[N]) |
| constexpr armnn::Compute | ParseComputeDevice (const char *str) |
| Deprecated function that will be removed together with the Compute enum. | |
| constexpr const char * | GetDataTypeName (DataType dataType) |
| constexpr const char * | GetDataLayoutName (DataLayout dataLayout) |
| constexpr const char * | GetNormalizationAlgorithmChannelAsCString (NormalizationAlgorithmChannel channel) |
| constexpr const char * | GetNormalizationAlgorithmMethodAsCString (NormalizationAlgorithmMethod method) |
| constexpr const char * | GetResizeMethodAsCString (ResizeMethod method) |
| constexpr const char * | GetMemBlockStrategyTypeName (MemBlockStrategyType memBlockStrategyType) |
| template<typename T> | |
| constexpr bool | IsQuantizedType () |
| constexpr bool | IsQuantized8BitType (DataType dataType) |
| constexpr bool | IsQuantizedType (DataType dataType) |
| std::ostream & | operator<< (std::ostream &os, Status stat) |
| std::ostream & | operator<< (std::ostream &os, const armnn::TensorShape &shape) |
| template<typename QuantizedType> | |
| QuantizedType | Quantize (float value, float scale, int32_t offset) |
| Quantize a floating point data type into an 8-bit data type. | |
| template<typename QuantizedType> | |
| float | Dequantize (QuantizedType value, float scale, int32_t offset) |
| Dequantize an 8-bit data type into a floating point data type. | |
| void | VerifyTensorInfoDataType (const armnn::TensorInfo &info, armnn::DataType dataType) |
| template<typename ... Ts> | |
| void | IgnoreUnused (Ts &&...) |
| template<typename Dest, typename Source> | |
| std::enable_if_t< std::is_unsigned< Source >::value &&std::is_unsigned< Dest >::value, Dest > | numeric_cast (Source source) |
| template<typename Dest, typename Source> | |
| std::enable_if_t< std::is_signed< Source >::value &&std::is_integral< Source >::value &&std::is_signed< Dest >::value &&std::is_integral< Dest >::value, Dest > | numeric_cast (Source source) |
| template<typename Dest, typename Source> | |
| std::enable_if_t< std::is_floating_point< Source >::value &&std::is_floating_point< Dest >::value, Dest > | numeric_cast (Source source) |
| template<typename Dest, typename Source> | |
| std::enable_if_t< std::is_floating_point< Source >::value &&std::is_signed< Dest >::value &&std::is_integral< Dest >::value, Dest > | numeric_cast (Source source) |
| template<typename Dest, typename Source> | |
| std::enable_if_t< std::is_signed< Source >::value &&std::is_integral< Source >::value &&std::is_floating_point< Dest >::value, Dest > | numeric_cast (Source source) |
| template<typename Dest, typename Source> | |
| std::enable_if_t< std::is_signed< Dest >::value &&std::is_integral< Dest >::value &&std::is_unsigned< Source >::value, Dest > | numeric_cast (Source sValue) |
| template<typename Dest, typename Source> | |
| std::enable_if_t< std::is_floating_point< Dest >::value &&std::is_unsigned< Source >::value, Dest > | numeric_cast (Source sValue) |
| template<typename Dest, typename Source> | |
| std::enable_if_t< std::is_unsigned< Dest >::value &&std::is_signed< Source >::value &&std::is_integral< Source >::value, Dest > | numeric_cast (Source sValue) |
| template<typename Dest, typename Source> | |
| std::enable_if_t< std::is_unsigned< Dest >::value &&std::is_floating_point< Source >::value, Dest > | numeric_cast (Source sValue) |
| template<typename DestType, typename SourceType> | |
| DestType | PolymorphicDowncast (SourceType *value) |
| Polymorphic downcast for build in pointers only. | |
| template<typename DestType, typename SourceType> | |
| auto | PolymorphicPointerDowncast (const SourceType &value) |
| Polymorphic downcast for shared pointers and build in pointers. | |
| std::chrono::high_resolution_clock::time_point | GetTimeNow () |
| std::chrono::duration< double, std::milli > | GetTimeDuration (std::chrono::high_resolution_clock::time_point start_time) |
| template<typename Function, typename Iterator> | |
| constexpr TransformIterator< Function, Iterator > | MakeTransformIterator (Iterator i, Function f) |
| void | ConfigureLogging (bool printToStandardOutput, bool printToDebugOutput, LogSeverity severity) |
| Configures the logging behaviour of the ARMNN library. | |
| bool | NeonDetected () |
| const std::string | GetVersion () |
| float | roundf (float value) |
| void | swap (OriginsDescriptor &first, OriginsDescriptor &second) |
| void | swap (ViewsDescriptor &first, ViewsDescriptor &second) |
| uint32_t | GetNumInputs (bool biasEnabled) |
| void | AssertNumberOfInputSlots (Layer &layer) |
| template<typename T> | |
| constexpr LayerType | LayerEnumOf (const T *=nullptr) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const ActivationLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const AdditionLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const ArgMinMaxLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const BatchMatMulLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const BatchNormalizationLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const BatchToSpaceNdLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const BroadcastToLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const CastLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const ChannelShuffleLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const ComparisonLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const ConcatLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const ConstantLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const ConvertFp16ToFp32Layer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const ConvertFp32ToFp16Layer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const Convolution2dLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const Convolution3dLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const DebugLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const DepthToSpaceLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const DepthwiseConvolution2dLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const DequantizeLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const DetectionPostProcessLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const DivisionLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const ElementwiseBinaryLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const ElementwiseUnaryLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const FakeQuantizationLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const FillLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const FloorLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const FullyConnectedLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const FusedLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const GatherLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const GatherNdLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const InputLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const InstanceNormalizationLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const L2NormalizationLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const LogicalBinaryLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const LogSoftmaxLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const LstmLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const MapLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const MaximumLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const MeanLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const MemCopyLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const MemImportLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const MergeLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const MinimumLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const MultiplicationLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const NormalizationLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const OutputLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const PadLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const PermuteLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const Pooling2dLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const Pooling3dLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const PreCompiledLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const PreluLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const QuantizeLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const QLstmLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const QuantizedLstmLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const RankLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const ReduceLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const ReshapeLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const ResizeLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const ReverseV2Layer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const ScatterNdLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const ShapeLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const SliceLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const SoftmaxLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const SpaceToBatchNdLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const SpaceToDepthLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const SplitterLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const StackLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const StandInLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const StridedSliceLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const SubtractionLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const SwitchLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const TileLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const TransposeLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const TransposeConvolution2dLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const UnidirectionalSequenceLstmLayer *) |
| template<> | |
| constexpr LayerType | LayerEnumOf (const UnmapLayer *) |
| template<typename T, typename V> | |
| void | SetValueChecked (Optional< T & > optionalRef, V &&val) |
| template<typename Float16Func, typename Float32Func, typename Uint8Func, typename Int32Func, typename BooleanFunc, typename ... Params> | |
| bool | IsSupportedForDataTypeGeneric (Optional< std::string & > reasonIfUnsupported, DataType dataType, Float16Func float16FuncPtr, Float32Func float32FuncPtr, Uint8Func uint8FuncPtr, Int32Func int32FuncPtr, BooleanFunc booleanFuncPtr, Params &&... params) |
| template<typename ... Params> | |
| bool | TrueFunc (Optional< std::string & > reasonIfUnsupported, Params &&... params) |
| template<typename ... Params> | |
| bool | FalseFunc (Optional< std::string & > reasonIfUnsupported, Params &&... params) |
| template<typename ... Params> | |
| bool | FalseFuncF16 (Optional< std::string & > reasonIfUnsupported, Params &&... params) |
| template<typename ... Params> | |
| bool | FalseFuncF32 (Optional< std::string & > reasonIfUnsupported, Params &&... params) |
| template<typename ... Params> | |
| bool | FalseFuncU8 (Optional< std::string & > reasonIfUnsupported, Params &&... params) |
| template<typename ... Params> | |
| bool | FalseFuncI32 (Optional< std::string & > reasonIfUnsupported, Params &&... params) |
| template<typename ... Params> | |
| bool | FalseInputFuncF32 (Optional< std::string & > reasonIfUnsupported, Params &&... params) |
| template<typename ... Params> | |
| bool | FalseInputFuncF16 (Optional< std::string & > reasonIfUnsupported, Params &&... params) |
| template<typename ... Params> | |
| bool | FalseOutputFuncF32 (Optional< std::string & > reasonIfUnsupported, Params &&... params) |
| template<typename ... Params> | |
| bool | FalseOutputFuncF16 (Optional< std::string & > reasonIfUnsupported, Params &&... params) |
| void | ValidateSourcesMatchOptimizedNetwork (std::vector< BackendOptions > optimizedOptions, const INetworkProperties &networkProperties) |
| This function performs a sanity check to ensure that the combination of input and output memory source matches the values for importEnabled and exportEnabled that were specified during optimization. | |
| void | CopyToOutputTensor (const Tensor &outputTensor, ITensorHandle *outputTensorHandle) |
| const armnn::ConstTensor | GetInputTensor (const LayerBindingId layerId, const InputTensors &inputTensors) |
| const armnn::Tensor | GetOutputTensor (const LayerBindingId layerId, const OutputTensors &outputTensors) |
| template<LogSeverity Level> | |
| void | SetLoggingSinks (bool standardOut, bool debugOut, bool coloured) |
| void | ReportError (const std::string &errorMessage, Optional< std::vector< std::string > & > errorMessages) |
| void | ReportWarning (const std::string &warningMessage, Optional< std::vector< std::string > & > warningMessages) |
| OptimizationResult | ReturnWithError (OptimizationResult res, const Layer *layer, const BackendSettings &backendSettings, Optional< std::vector< std::string > & > errMessages) |
| bool | CheckScaleSetOnQuantizedType (Layer *layer, Optional< std::vector< std::string > & > errMessages) |
| OptimizationResult | AttemptBackendAssignment (BackendSettings &backendSettings, Graph &graph, Layer *layer, BackendId backend, DataType dataTypeIn, DataType dataTypeOut, const std::vector< BackendId > &availablePreferredBackends, std::string &reasonIfUnsupported, Optional< std::vector< std::string > & > messages) |
| std::vector< DataType > | GetLayerInOutDatatype (const Layer *layer) |
| bool | CheckFp16Support (BackendsMap &backends, const std::vector< BackendId > &availablePreferredBackends) |
| void | AssignBackendsIConnectable (OptimizedNetworkImpl *optNetObjPtr, IConnectableLayer *it, Optional< std::vector< std::string > & > errMessages, OptimizationResult &result, BackendSettings &backendSettings, std::vector< BackendId > &availablePreferredBackends, bool &restart) |
| OptimizationResult | AssignBackends (OptimizedNetworkImpl *optNetObjPtr, BackendSettings &backendSettings, Graph::Iterator &firstLayer, Graph::Iterator &lastLayer, Optional< std::vector< std::string > & > errMessages) |
| OptimizationResult | AssignBackends (OptimizedNetworkImpl *optNetObjPtr, BackendSettings &backendSettings, SubgraphView::IConnectableLayerIterator &firstLayer, SubgraphView::IConnectableLayerIterator &lastLayer, Optional< std::vector< std::string > & > errMessages) |
| OptimizationResult | AssignBackends (OptimizedNetworkImpl *optNetObjPtr, BackendSettings &backendSettings, SubgraphView &subgraph, Optional< std::vector< std::string > & > errMessages) |
| BackendsMap | CreateSupportedBackends (TensorHandleFactoryRegistry &handleFactoryRegistry, BackendSettings &backendSettings) |
| OptimizationResult | ApplyBackendOptimizations (OptimizedNetworkImpl *optNetObjPtr, BackendSettings &backendSettings, BackendsMap &backends, const ModelOptions &modelOptions, Optional< std::vector< std::string > & > errMessages) |
| bool | RequiresCopy (ITensorHandleFactory::FactoryId src, ITensorHandleFactory::FactoryId dst, TensorHandleFactoryRegistry ®istry) |
| ITensorHandleFactory::FactoryId | CalculateSlotOptionForInput (BackendsMap &backends, OutputSlot &slot, TensorHandleFactoryRegistry ®istry, bool importEnabled) |
| ITensorHandleFactory::FactoryId | CalculateSlotOptionForOutput (BackendsMap &backends, OutputSlot &slot, TensorHandleFactoryRegistry ®istry) |
| ITensorHandleFactory::FactoryId | CalculateSlotOption (BackendsMap &backends, OutputSlot &outputSlot, TensorHandleFactoryRegistry ®istry, bool exportEnabled) |
| EdgeStrategy | CalculateEdgeStrategy (BackendsMap &backends, ITensorHandleFactory::FactoryId srcFactoryId, const Layer &layer, const Layer &connectedLayer, TensorHandleFactoryRegistry ®istry, bool importEnabled) |
| OptimizationResult | SelectTensorHandleStrategy (Graph &optGraph, BackendsMap &backends, TensorHandleFactoryRegistry ®istry, bool importEnabled, bool exportEnabled, Optional< std::vector< std::string > & > errMessages) |
| bool | CheckFastMathSupport (const std::vector< BackendId > &availablePreferredBackends, const ModelOptions &modelOptions) |
| bool | IsTfLiteTurboModel (const Graph &optGraph) |
| std::vector< ConvertFp16ToFp32Layer * > | InsertConvertFp16ToFp32LayersBefore (Graph &graph, Layer &layer, bool expectCorrectInputType) |
| std::vector< ConvertFp32ToFp16Layer * > | InsertConvertFp32ToFp16LayersAfter (Graph &graph, Layer &layer) |
| std::vector< DebugLayer * > | InsertDebugLayerAfter (Graph &graph, Layer &layer, bool toFile) |
| bool | RevertConstantWeightsToFP32 (Layer *layer) |
| template<typename T> | |
| void | Append (Optimizer::Optimizations &optimizations, T &&optimization) |
| template<typename Front, typename... Others> | |
| void | Append (Optimizer::Optimizations &optimizations, Front &&front, Others &&... others) |
| template<typename... Args> | |
| Optimizer::Optimizations | MakeOptimizations (Args &&... args) |
| Measurement | FindMeasurement (const std::string &name, const Event *event) |
| std::vector< Measurement > | FindKernelMeasurements (const Event *event) |
| const Event * | GetEventPtr (const Event *ptr) |
| const Event * | GetEventPtr (const std::unique_ptr< Event > &ptr) |
| int | CalcLevel (const Event *eventPtr) |
| void | ConfigureDetailsObject (JsonChildObject &detailsObject, std::string layerDetailsStr) |
| void | ExtractJsonObjects (unsigned int inferenceIndex, const Event *parentEvent, JsonChildObject &parentObject, std::map< const Event *, std::vector< const Event * > > descendantsMap) |
| template<typename DescriptorType> | |
| void | ProfilingUpdateDescriptions (const std::string &name, const DescriptorType &desc, const WorkloadInfo &infos, const arm::pipe::ProfilingGuid guid) |
| template<typename Delegate> | |
| void | ForEachLayerInput (LayerSelectionInfo::LayerInfoContainer &layerInfos, LayerSelectionInfo &layerInfo, Delegate function) |
| template<typename Delegate> | |
| void | ForEachLayerOutput (LayerSelectionInfo::LayerInfoContainer &layerInfos, LayerSelectionInfo &layerInfo, Delegate function) |
| void | AssignSplitId (LayerSelectionInfo::LayerInfoContainer &layerInfos, LayerSelectionInfo &layerInfo) |
| bool | IsReadyForSplitAssignment (LayerSelectionInfo::LayerInfoContainer &layerInfos, LayerSelectionInfo &layerInfo) |
| bool | IsLayerSupported (const armnn::Layer *layer) |
| bool | IsLayerSupported (const armnn::Layer &layer) |
| bool | IsLayerOptimizable (const armnn::Layer *layer) |
| bool | IsLayerOptimizable (const armnn::Layer &layer) |
| constexpr const char * | MockTensorHandleFactoryId () |
| Graph & | GetGraphForTesting (IOptimizedNetwork *optNet) |
| ModelOptions & | GetModelOptionsForTesting (IOptimizedNetwork *optNet) |
| arm::pipe::IProfilingService & | GetProfilingService (armnn::RuntimeImpl *runtime) |
| std::ostream & | operator<< (std::ostream &os, const BFloat16 &b) |
| template<typename LayerType> | |
| LayerType * | FuseLayer (OptimizationViews &optimizationViews, LayerType *baseLayer, LayerType *replacementLayer, ActivationLayer *activationLayer, ActivationDescriptor &activationDesc) |
| template<typename LayerType> | |
| LayerType * | FuseAdditionLayer (OptimizationViews &optimizationViews, LayerType *baseLayer, ActivationLayer *activationLayer, ActivationDescriptor &activationDesc, std::string name) |
| template<typename LayerType> | |
| LayerType * | FuseSubtractionLayer (OptimizationViews &optimizationViews, LayerType *baseLayer, ActivationLayer *activationLayer, ActivationDescriptor &activationDesc, std::string name) |
| template<typename LayerType> | |
| LayerType * | FuseDivisionLayer (OptimizationViews &optimizationViews, LayerType *baseLayer, ActivationLayer *activationLayer, ActivationDescriptor &activationDesc, std::string name) |
| template<typename LayerType> | |
| LayerType * | FuseMultiplicationLayer (OptimizationViews &optimizationViews, LayerType *baseLayer, ActivationLayer *activationLayer, ActivationDescriptor &activationDesc, std::string name) |
| template<typename LayerType> | |
| LayerType * | FuseElementwiseBinaryLayer (OptimizationViews &optimizationViews, LayerType *baseLayer, ActivationLayer *activationLayer, ActivationDescriptor &activationDesc, BinaryOperation operation, std::string name) |
| template<typename LayerType> | |
| LayerType * | FuseBatchNormalizationLayer (OptimizationViews &optimizationViews, LayerType *baseLayer, ActivationLayer *activationLayer, ActivationDescriptor &activationDesc, std::string name) |
| template<typename LayerType> | |
| LayerType * | FuseConvolution2dLayer (OptimizationViews &optimizationViews, LayerType *baseLayer, ActivationLayer *activationLayer, ActivationDescriptor &activationDesc, std::string name) |
| template<typename LayerType> | |
| LayerType * | FuseDepthwiseConvolution2dLayer (OptimizationViews &optimizationViews, LayerType *baseLayer, ActivationLayer *activationLayer, ActivationDescriptor &activationDesc, std::string name) |
| template<typename LayerType> | |
| LayerType * | FuseFullyConnectedLayer (OptimizationViews &optimizationViews, LayerType *baseLayer, ActivationLayer *activationLayer, ActivationDescriptor &activationDesc, std::string name) |
| template<typename LayerType> | |
| std::vector< IConnectableLayer * > | ChainReduceLayers (OptimizationViews &optimizationViews, LayerType *baseLayer, ReduceDescriptor &desc) |
| template<typename LayerType> | |
| void | ReplaceLayers (OptimizationViews &optimizationViews, LayerType *baseLayer, std::vector< IConnectableLayer * > &layers) |
| template<typename LayerType> | |
| void | ReplaceMultipleLayers (OptimizationViews &optimizationViews, std::vector< IConnectableLayer * > &originalLayers, LayerType *baseLayer, const std::vector< SlotList > inputLayersSlotLists, const std::vector< SlotList > outputLayersSlotLists) |
| TuningLevel | ParseTuningLevel (const BackendOptions::Var &value, TuningLevel defaultValue) |
| void | ConfigureTuner (arm_compute::CLTuner &tuner, TuningLevel level) |
| arm_compute::NormalizationLayerInfo | CreateAclNormalizationLayerInfoForL2Normalization (const armnn::TensorInfo &tensorInfo, armnn::DataLayout dataLayout) |
| arm_compute::ActivationLayerInfo::ActivationFunction | ConvertActivationFunctionToAclActivationFunction (ActivationFunction armnnFunction) |
| arm_compute::ActivationLayerInfo | ConvertActivationDescriptorToAclActivationLayerInfo (const ActivationDescriptor &actDesc) |
| arm_compute::ActivationLayerInfo | ConvertActivationDescriptorToAclActivationLayerInfo (const ActivationDescriptor *activationDescPtr) |
| arm_compute::ActivationLayerInfo | ConvertAdditionalInfoToAclActivationLayerInfo (const QueueDescriptor &queueDescriptor) |
| arm_compute::ActivationLayerInfo | ConvertLstmActivationFuncToAclLayerInfo (uint32_t activationFunction) |
| arm_compute::ComparisonOperation | ConvertComparisonOperationToAcl (const ComparisonDescriptor &descriptor) |
| arm_compute::PoolingType | ConvertPoolingAlgorithmToAclPoolingType (PoolingAlgorithm poolingAlgorithm) |
| arm_compute::DimensionRoundingType | ConvertOutputShapeRoundingToAclDimensionRoundingType (OutputShapeRounding rounding) |
| arm_compute::NormType | ConvertNormalizationAlgorithmChannelToAclNormType (NormalizationAlgorithmChannel channelType) |
| arm_compute::FullyConnectedLayerInfo | ConvertFullyConnectedDescriptorToAclFullyConnectedLayerInfo (const FullyConnectedDescriptor &fullyConnectedDesc, const ActivationDescriptor *activationDesc) |
| arm_compute::FullyConnectedLayerInfo | ConvertFullyConnectedDescriptorToAclFullyConnectedLayerInfo (const FullyConnectedDescriptor &fullyConnectedDesc, arm_compute::ActivationLayerInfo activationLayerInfo) |
| arm_compute::InterpolationPolicy | ConvertResizeMethodToAclInterpolationPolicy (ResizeMethod resizeMethod) |
| template<typename T> | |
| T | ComputeSoftmaxAclAxis (const SoftmaxDescriptor &softmaxDesc, const armnn::TensorInfo &tensor) |
| int | ComputeAclAxis (const int &armnnAxis, const armnn::TensorInfo &tensor) |
| Function to convert ArmNN axis (left to right) to ACL axis (right to left) ranging from [-rank, rank) | |
| arm_compute::Conv3dInfo | ComputeConv3DInfo (const armnn::Convolution3dDescriptor descriptor, bool isFastMathEnabled, const ActivationDescriptor *activationDescriptor) |
| Utility function used to setup an arm_compute::Conv3dInfo object from convolution3d descriptor. | |
| arm_compute::Conv3dInfo | ComputeConv3DInfo (const armnn::Convolution3dQueueDescriptor queueDescriptor, bool isFastMathEnabled) |
| arm_compute::PaddingMode | ConvertPaddingModeToAcl (const PaddingMode &paddingMode) |
| arm_compute::ReductionOperation | ConvertReductionOperationToAcl (const ReduceDescriptor &descriptor) |
| const TensorInfo | ComputeReductionTensorShape (const armnn::TensorInfo &input, const std::vector< uint32_t > &vAxis, const bool keepDims) |
| Function to compute the output tensor shape based on the axes and if keepDims is set. | |
| void | IsMultiAxesReduceSupported (reduceValidateFunction func, const armnn::TensorInfo &input, const armnn::TensorInfo &output, const armnn::ReduceDescriptor &desc, arm_compute::Status &status) |
| Function to check if layer with multiple axes is supported on each backend. | |
| armnn::Optional< armnn::DataType > | GetBiasTypeFromWeightsType (armnn::Optional< armnn::DataType > weightsType) |
| template<typename F> | |
| bool | CheckSupportRule (F rule, Optional< std::string & > reasonIfUnsupported, const char *reason) |
| template<typename T> | |
| bool | AllTypesAreEqualImpl (T) |
| template<typename T, typename... Rest> | |
| bool | AllTypesAreEqualImpl (T t1, T t2, Rest... rest) |
| std::unique_ptr< IMemoryOptimizerStrategy > | GetMemoryOptimizerStrategy (const std::string &strategyName) |
| const std::vector< std::string > | GetMemoryOptimizerStrategyNames () |
| bool | IsNCHW (armnn::Layer &layer) |
| void | ReportUntouchedLayers (OptimizationViews &optimizationViews, std::map< LayerGuid, Layer * > untouched) |
| template<typename LayerType> | |
| LayerType * | ReplaceLayer (OptimizationViews &optimizationViews, LayerType *baseLayer, LayerType *replacementLayer) |
| template<typename LayerType> | |
| LayerType * | FoldPadLayer (OptimizationViews &optimizationViews, LayerType *baseLayer, LayerType *replacementLayer, PadLayer *padLayer) |
| bool | ConnectedToLayerWithNCHW (Layer *baseLayer) |
| Checks if the Layer is connected to any Layer that has an NCHW layout. | |
| bool | ConnectedToLayerType (Layer *baseLayer, LayerType layerType, unsigned int dimSize=0) |
| Checks the Layer's Connections to see if it's connected to a Layer with the provided layerType. | |
| void | RemoveReshapeLayer (ReshapeLayer *baseLayer, std::map< LayerGuid, Layer * > &untouched, OptimizationViews &optimizationViews) |
| template<typename LayerT, typename Descriptor> | |
| void | FoldPadLayer2d (OptimizationViews &optimizationViews, LayerT *baseLayer, Descriptor &descriptor, PadLayer *padLayer) |
| bool | IsSequenceLayerType (Layer &layer, LayerType type) |
| bool | IsSequenceLayerType (Layer &layer, BinaryOperation type) |
| template<typename TYPE> | |
| bool | IsLayerSequence (Layer ¤tLayer, TYPE first, TYPE second, TYPE third, Layer *layerList[4], bool handleValidActivates, const std::vector< ActivationFunction > &validActivates) |
| armnn::ConstTensor | PermuteTensor (const ConstTensorHandle *tensor, const PermutationVector &permutationVector, void *permuteBuffer) |
| void | ReshapeWeightsForAcl (TensorInfo &weightInfo, DataLayout dataLayout) |
| template<typename DataType> | |
| ConstTensor | ReorderWeightChannelsForAcl (const ConstTensor &weightHandle, DataLayout dataLayout, void *permuteBuffer) |
| TensorInfo | ConvertWeightTensorInfoFromArmnnToAcl (const TensorInfo &weightInfo, DataLayout dataLayout) |
| std::tuple< ConstTensor, unsigned int > | Convert1HWOTensorToAcl (const ConstTensorHandle *weightTensor, const TensorInfo &inputInfo, const DataLayout dataLayout, void *permuteBuffer) |
| Weights for depthwise have a datalayout of [1,H,W,O] = [1,H,W,I*M] This function coverts a ConstCpuTensorHandle from [1,H,W,I*M] to [1,I*M,H,W] (if NCHW) or keeps it at [1,H,W,I*M] (if NHWC) as required by the compute library. | |
| std::tuple< TensorInfo, unsigned int > | Convert1HWOTensorInfoToAcl (const TensorInfo &weightInfo, const TensorInfo &inputInfo, const DataLayout dataLayout) |
| Weights for depthwise have a datalayout of [1,H,W,O] = [1,H,W,I*M] This function coverts a TensorInfo from [1,H,W,I*M] to [1,I*M,H,W] (if NCHW) or keeps it at [1,H,W,I*M] (if NHWC) as required by the compute library Returns a tuple of converted weights tensor info and depth multiplier. | |
| std::tuple< ConstTensor, unsigned int > | Convert1HWOtoMIHW (const ConstTensorHandle *weightTensor, const TensorInfo &inputInfo, const DataLayout &dataLayout, void *permuteBuffer) |
| Converts a (weights) tensor from [1, H, W, I*M] = [1, H, W, O] to [M, I, H, W]. | |
| armnn::ConstTensor | ConvertWeightTensorFromArmnnToAcl (const ConstTensorHandle *weightTensor, DataLayout dataLayout, void *permuteBuffer) |
| int32_t | ConvertMaskToACLFormat (int32_t mask, int32_t numDim) |
| std::map< std::string, unsigned int > | CalculateGatherNdKeyIndices (TensorInfo inputInfo0, TensorInfo inputInfo1) |
| Calculates the key index values needed for GatherNd: N, ND, K, W, C (N is always 1) | |
| armnn::PermutationVector | GeneratePermutationVectorOnLastTwoDimensions (unsigned int rank) |
| Generates a permutation vector of size rank that permutes the 2 most right dimensions. | |
| std::set< unsigned int > | ComputeSplitAxis (const armnn::SplitterDescriptor &desc, const TensorShape &input) |
| Calculates the axis values for split operation. | |
| template<typename CopyFunc> | |
| void | CopyTensorContentsGeneric (const ITensorHandle *srcTensor, ITensorHandle *dstTensor, CopyFunc copy) |
| template<typename SrcTensorHandleType, typename DstTensorHandleType, typename DescriptorType> | |
| void | GatherTensorHandlePairs (const DescriptorType &descriptor, std::vector< std::pair< SrcTensorHandleType *, DstTensorHandleType * > > &tensorHandlePairs) |
| constexpr const char * | ClBackendId () |
| flatbuffers::Offset< ClContext > | CreateClContext (flatbuffers::FlatBufferBuilder &_fbb, flatbuffers::Offset< flatbuffers::Vector< flatbuffers::Offset< armnn::Program > > > programs=0) |
| flatbuffers::Offset< ClContext > | CreateClContextDirect (flatbuffers::FlatBufferBuilder &_fbb, const std::vector< flatbuffers::Offset< armnn::Program > > *programs=nullptr) |
| flatbuffers::Offset< Program > | CreateProgram (flatbuffers::FlatBufferBuilder &_fbb, flatbuffers::Offset< flatbuffers::String > name=0, flatbuffers::Offset< flatbuffers::Vector< uint8_t > > binary=0) |
| flatbuffers::Offset< Program > | CreateProgramDirect (flatbuffers::FlatBufferBuilder &_fbb, const char *name=nullptr, const std::vector< uint8_t > *binary=nullptr) |
| const armnn::ClContext * | GetClContext (const void *buf) |
| const armnn::ClContext * | GetSizePrefixedClContext (const void *buf) |
| const char * | ClContextIdentifier () |
| bool | ClContextBufferHasIdentifier (const void *buf) |
| bool | VerifyClContextBuffer (flatbuffers::Verifier &verifier) |
| bool | VerifySizePrefixedClContextBuffer (flatbuffers::Verifier &verifier) |
| const char * | ClContextExtension () |
| void | FinishClContextBuffer (flatbuffers::FlatBufferBuilder &fbb, flatbuffers::Offset< armnn::ClContext > root) |
| void | FinishSizePrefixedClContextBuffer (flatbuffers::FlatBufferBuilder &fbb, flatbuffers::Offset< armnn::ClContext > root) |
| constexpr const char * | ClImportTensorHandleFactoryId () |
| constexpr const char * | ClTensorHandleFactoryId () |
| arm_compute::Status | ClAbsWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
| arm_compute::Status | ClActivationWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const ActivationDescriptor &descriptor) |
| arm_compute::Status | ClAdditionValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ActivationDescriptor *activationDescriptor) |
| arm_compute::Status | ClArgMinMaxWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const ArgMinMaxDescriptor &descriptor) |
| arm_compute::Status | ClBatchMatMulValidate (const TensorInfo &inputInfoX, const TensorInfo &inputInfoY, const TensorInfo &outputInfo, const BatchMatMulDescriptor &descriptor, const ActivationDescriptor *activationDescriptor) |
| arm_compute::Status | ClBatchNormalizationValidate (const TensorInfo &input, const TensorInfo &output, const TensorInfo &mean, const TensorInfo &var, const TensorInfo &beta, const TensorInfo &gamma, const BatchNormalizationDescriptor &descriptor, const ActivationDescriptor *activationDescriptor) |
| arm_compute::Status | ClBatchToSpaceNdWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const BatchToSpaceNdDescriptor &descriptor) |
| arm_compute::Status | ClCastValidate (const TensorInfo &input, const TensorInfo &output) |
| arm_compute::Status | ClChannelShuffleValidate (const TensorInfo &input, const TensorInfo &output, const ChannelShuffleDescriptor &descriptor) |
| arm_compute::Status | ClComparisonWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ComparisonDescriptor &descriptor) |
| arm_compute::Status | ClConcatWorkloadValidate (const std::vector< const TensorInfo * > &inputs, const TensorInfo &output, const OriginsDescriptor &descriptor) |
| arm_compute::Status | ClConstantWorkloadValidate (const TensorInfo &output) |
| arm_compute::Status | ClConvertFp16ToFp32WorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
| arm_compute::Status | ClConvertFp32ToFp16WorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
| arm_compute::Status | ClConvolution2dWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const Convolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, bool isFastMathEnabled, const ActivationDescriptor *activationDescriptor) |
| arm_compute::Status | ClConvolution3dWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const Convolution3dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, bool isFastMathEnabled, const ActivationDescriptor *activationDescriptor) |
| arm_compute::Status | ClDepthToSpaceWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const DepthToSpaceDescriptor &descriptor) |
| arm_compute::Status | ClDepthwiseConvolutionWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const DepthwiseConvolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, const ActivationDescriptor *activationDescriptor) |
| arm_compute::Status | ClDequantizeWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
| arm_compute::Status | ClDivisionWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ActivationDescriptor *activationDescriptor) |
| arm_compute::Status | ClElementwiseBinaryValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ElementwiseBinaryDescriptor &descriptor, const ActivationDescriptor *activationDescriptor) |
| arm_compute::Status | ClExpWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
| TensorInfo | ConvertTensorToFloat32 (const TensorInfo &tensorInfo) |
| Utility function used for the two cast layer inputs to convert the output layer tensor types. | |
| bool | AreAllTensorsSigned32 (const std::vector< TensorInfo > &tensorInfos) |
| Utility function used to check if a vector of tensors are Signed32. | |
| bool | IsValidationPassing (const std::vector< arm_compute::Status > &statuses) |
| Utility function used to check if statuses are returning 'OK'. | |
| arm_compute::Status | ClFloorDivWorkloadValidate (const TensorInfo &input0Info, const TensorInfo &input1Info, const TensorInfo &outputInfo, const ActivationDescriptor *activationDescriptor) |
| arm_compute::Status | ClFloorWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
| arm_compute::Status | ClFullyConnectedWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const TensorInfo &weights, const Optional< TensorInfo > &biases, const FullyConnectedDescriptor &descriptor, const ActivationDescriptor *activationDescriptor) |
| arm_compute::Status | ClGatherNdWorkloadValidate (const TensorInfo ¶msInfo, const TensorInfo &indicesInfo, const TensorInfo &outputInfo) |
| arm_compute::Status | ClGatherWorkloadValidate (const TensorInfo &input, const TensorInfo &indices, const TensorInfo &output, const GatherDescriptor &descriptor) |
| arm_compute::Status | ClInstanceNormalizationWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const InstanceNormalizationDescriptor &descriptor) |
| arm_compute::Status | ClL2NormalizationWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const L2NormalizationDescriptor &descriptor) |
| arm_compute::Status | ClLogicalAndWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output) |
| arm_compute::Status | ClLogicalNotWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
| arm_compute::Status | ClLogicalOrWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output) |
| arm_compute::Status | ClLogSoftmaxWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const LogSoftmaxDescriptor &descriptor) |
| arm_compute::Status | ClLogWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
| arm_compute::Status | ClLstmFloatWorkloadValidate (const TensorInfo &input, const TensorInfo &outputStateIn, const TensorInfo &cellStateIn, const TensorInfo &scratchBuffer, const TensorInfo &outputStateOut, const TensorInfo &cellStateOut, const TensorInfo &output, const LstmDescriptor &descriptor, const LstmInputParamsInfo ¶msInfo) |
| arm_compute::Status | ClMaximumWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output) |
| arm_compute::Status | ClMeanValidate (const TensorInfo &input, const TensorInfo &output, const MeanDescriptor &descriptor) |
| arm_compute::Status | ClMinimumWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output) |
| arm_compute::Status | ClMultiplicationWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ActivationDescriptor *activationDescriptor) |
| arm_compute::Status | ClNegWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
| arm_compute::Status | ClNormalizationWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const NormalizationDescriptor &descriptor) |
| arm_compute::Status | ClPadValidate (const TensorInfo &input, const TensorInfo &output, const PadDescriptor &descriptor) |
| arm_compute::Status | ClPermuteWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const PermuteDescriptor &descriptor) |
| arm_compute::Status | ClPooling2dWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const Pooling2dDescriptor &descriptor) |
| arm_compute::Status | ClPooling3dWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const Pooling3dDescriptor &descriptor) |
| arm_compute::Status | ClPreluWorkloadValidate (const TensorInfo &input, const TensorInfo &alpha, const TensorInfo &output) |
| arm_compute::Status | ClQLstmWorkloadValidate (const TensorInfo &input, const TensorInfo &cellStateIn, const TensorInfo &outputStateIn, const TensorInfo &cellStateOut, const TensorInfo &outputStateOut, const TensorInfo &output, const QLstmDescriptor &descriptor, const LstmInputParamsInfo ¶msInfo) |
| arm_compute::Status | ClQuantizedLstmWorkloadValidate (const TensorInfo &input, const TensorInfo &previousCellStateIn, const TensorInfo &previousOutputIn, const TensorInfo &cellStateOut, const TensorInfo &output, const QuantizedLstmInputParamsInfo ¶msInfo) |
| arm_compute::Status | ClQuantizeWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
| arm_compute::Status | ClReduceWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const ReduceDescriptor &descriptor) |
| arm_compute::Status | ClReshapeWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
| arm_compute::Status | ClResizeWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const ResizeDescriptor &descriptor) |
| arm_compute::Status | ClReverseV2WorkloadValidate (const TensorInfo &input, const TensorInfo &axis, const TensorInfo &output) |
| arm_compute::Status | ClRsqrtWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
| arm_compute::Status | ClScatterNdWorkloadValidate (const TensorInfo &inputInfo, const TensorInfo &indicesInfo, const TensorInfo &updatesInfo, const TensorInfo &outputInfo, const ScatterNdDescriptor &descriptor) |
| arm_compute::Status | ClSinWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
| arm_compute::Status | ClSliceWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const SliceDescriptor &descriptor) |
| arm_compute::Status | ClSoftmaxWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const SoftmaxDescriptor &descriptor) |
| arm_compute::Status | ClSpaceToBatchNdWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const SpaceToBatchNdDescriptor &descriptor) |
| arm_compute::Status | ClSpaceToDepthWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const SpaceToDepthDescriptor &descriptor) |
| arm_compute::Status | ClSplitterWorkloadValidate (const TensorInfo &input, const std::vector< std::reference_wrapper< TensorInfo > > &outputs, unsigned int splitAxis) |
| arm_compute::Status | ClSqrtWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
| arm_compute::Status | ClStackWorkloadValidate (const std::vector< const TensorInfo * > &inputs, const TensorInfo &output, const StackDescriptor &descriptor) |
| arm_compute::Status | ClStridedSliceWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const StridedSliceDescriptor &descriptor) |
| arm_compute::Status | ClSubtractionValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ActivationDescriptor *activationDescriptor) |
| arm_compute::Status | ClTileWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const TileDescriptor &descriptor) |
| arm_compute::Status | ClTransposeConvolution2dWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const TransposeConvolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases) |
| arm_compute::Status | ClTransposeWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const TransposeDescriptor &descriptor) |
| arm_compute::Status | ClUnidirectionalSequenceLstmFloatWorkloadValidate (const TensorInfo &input, const TensorInfo &outputStateIn, const TensorInfo &cellStateIn, const TensorInfo &outputStateOut, const TensorInfo &cellStateOut, const TensorInfo &output, const UnidirectionalSequenceLstmDescriptor &descriptor, const LstmInputParamsInfo ¶msInfo) |
| std::string | GetConvolutionMethodString (arm_compute::ConvolutionMethod &convolutionMethod) |
| template<typename T> | |
| void | CopyArmComputeClTensorData (arm_compute::CLTensor &dstTensor, const T *srcData) |
| auto | SetClStridedSliceData (const std::vector< int > &m_begin, const std::vector< int > &m_end, const std::vector< int > &m_stride) |
| auto | SetClSliceData (const std::vector< unsigned int > &m_begin, const std::vector< unsigned int > &m_size) |
| void | InitializeArmComputeClTensorData (arm_compute::CLTensor &clTensor, const ConstTensorHandle *handle) |
| RuntimeException | WrapClError (const cl::Error &clError, const CheckLocation &location) |
| void | RunClFunction (arm_compute::IFunction &function, const CheckLocation &location) |
| template<typename DataType, typename PayloadType> | |
| DataType * | GetOutputTensorData (unsigned int idx, const PayloadType &data) |
| constexpr const char * | NeonBackendId () |
| bool | CollapseLeadingUnitDimensions (const TensorInfo &in, TensorInfo &out) |
| template<typename SlotListType> | |
| void | BuildAddMulAddSlotLists (bool handleReLu, bool multipleOutputs, std::vector< SlotListType > &inputLayersSlotLists, std::vector< SlotListType > &outputLayersSlotLists) |
| void | GetFusedName (Layer *layerList[4], std::string &fusedName) |
| template<typename Type> | |
| bool | BuildAddMulAddTensorInfoLists (Type *layerList[4], unsigned int &numInputs, unsigned int &numOutputs, std::vector< TensorInfo > &inputInfos, std::vector< TensorInfo > &outputInfos, const ActivationDescriptor *&activationDescriptor, bool &fuseReLu) |
| bool | IsLayerTypeSupported (const LayerType &type, const std::vector< TensorInfo > &infos, const BaseDescriptor &descriptor, const Optional< LstmInputParamsInfo > &lstmParamsInfo, const Optional< QuantizedLstmInputParamsInfo > &quantizedLstmParamsInfo, Optional< std::string & > reasonIfUnsupported, const NeonLayerSupport &support) |
| constexpr const char * | NeonTensorHandleFactoryId () |
| arm_compute::Status | NeonAbsWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
| arm_compute::Status | NeonActivationWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const ActivationDescriptor &descriptor) |
| arm_compute::Status | NeonAdditionWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ActivationDescriptor *activationDescriptor) |
| arm_compute::Status | NeonArgMinMaxWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const ArgMinMaxDescriptor &descriptor) |
| arm_compute::Status | NeonBatchMatMulValidate (const TensorInfo &inputInfoX, const TensorInfo &inputInfoY, const TensorInfo &outputInfo, const BatchMatMulDescriptor &descriptor, const bool isFastMathEnabled, const ActivationDescriptor *activationDescriptor) |
| arm_compute::Status | NeonBatchNormalizationValidate (const TensorInfo &input, const TensorInfo &output, const TensorInfo &mean, const TensorInfo &var, const TensorInfo &beta, const TensorInfo &gamma, const BatchNormalizationDescriptor &descriptor, const ActivationDescriptor *activationDescriptor) |
| arm_compute::Status | NeonBatchToSpaceNdWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const BatchToSpaceNdDescriptor &descriptor) |
| arm_compute::Status | NeonCastValidate (const TensorInfo &input, const TensorInfo &output) |
| arm_compute::Status | NeonChannelShuffleValidate (const TensorInfo &input, const TensorInfo &output, const ChannelShuffleDescriptor &descriptor) |
| arm_compute::Status | NeonComparisonWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ComparisonDescriptor &descriptor) |
| arm_compute::Status | NeonConcatWorkloadValidate (const std::vector< const TensorInfo * > &inputs, const TensorInfo &output, const OriginsDescriptor &descriptor) |
| arm_compute::Status | NeonConstantWorkloadValidate (const TensorInfo &output) |
| arm_compute::Status | NeonConvertFp16ToFp32WorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
| arm_compute::Status | NeonConvertFp32ToFp16WorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
| arm_compute::Status | NeonConvolution2dWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const Convolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, bool isFastMathEnabled, const ActivationDescriptor *activationDescriptor) |
| arm_compute::Status | NeonConvolution3dWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const Convolution3dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, bool isFastMathEnabled, const ActivationDescriptor *activationDescriptor) |
| arm_compute::Status | NeonDepthToSpaceWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const DepthToSpaceDescriptor &descriptor) |
| arm_compute::Status | NeonDepthwiseConvolutionWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const DepthwiseConvolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, const ActivationDescriptor *activationDescriptor) |
| arm_compute::Status | NeonDequantizeWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
| arm_compute::DetectionPostProcessLayerInfo | MakeInfo (const DetectionPostProcessDescriptor &descriptor) |
| arm_compute::Status | NeonDetectionPostProcessValidate (const TensorInfo &boxEncodings, const TensorInfo &scores, const TensorInfo &anchors, const TensorInfo &detectionBoxes, const TensorInfo &detectionClasses, const TensorInfo &detectionScores, const TensorInfo &numDetections, const DetectionPostProcessDescriptor &descriptor) |
| arm_compute::Status | NeonDivisionWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ActivationDescriptor *activationDescriptor) |
| arm_compute::Status | NeonElementwiseBinaryWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ElementwiseBinaryDescriptor &descriptor, const ActivationDescriptor *activationDescriptor) |
| arm_compute::Status | NeonExpWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
| bool | IsAllTensorsSigned32 (const std::vector< TensorInfo > &tensorInfos) |
| Utility function used to check if a vector of tensors are Signed32. | |
| arm_compute::Status | NeonFloorDivWorkloadValidate (const TensorInfo &input0Info, const TensorInfo &input1Info, const TensorInfo &outputInfo, const ActivationDescriptor *activationDescriptor) |
| Validation for the Floor Div Workload. | |
| arm_compute::Status | NeonFullyConnectedWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const TensorInfo &weights, const Optional< TensorInfo > &biases, const FullyConnectedDescriptor &descriptor, const ActivationDescriptor *activationDescriptor) |
| arm_compute::Status | NeonFusedWorkloadValidate (const std::vector< std::reference_wrapper< TensorInfo > > &inputInfos, const std::vector< std::reference_wrapper< TensorInfo > > &outputInfos, const FusedDescriptor &fusedDescriptor, const ActivationDescriptor *activationDescriptor) |
| arm_compute::Status | NeonGatherNdWorkloadValidate (const TensorInfo ¶msInfo, const TensorInfo &indicesInfo, const TensorInfo &outputInfo) |
| arm_compute::Status | NeonGatherWorkloadValidate (const TensorInfo &input, const TensorInfo &indices, const TensorInfo &output, const GatherDescriptor &descriptor) |
| arm_compute::Status | NeonInstanceNormalizationWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const InstanceNormalizationDescriptor &descriptor) |
| arm_compute::Status | NeonL2NormalizationWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const L2NormalizationDescriptor &descriptor) |
| arm_compute::Status | NeonLogicalAndWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output) |
| arm_compute::Status | NeonLogicalNotWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
| arm_compute::Status | NeonLogicalOrWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output) |
| arm_compute::Status | NeonLogSoftmaxWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const LogSoftmaxDescriptor &descriptor) |
| arm_compute::Status | NeonLogWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
| arm_compute::Status | NeonLstmFloatWorkloadValidate (const TensorInfo &input, const TensorInfo &outputStateIn, const TensorInfo &cellStateIn, const TensorInfo &scratchBuffer, const TensorInfo &outputStateOut, const TensorInfo &cellStateOut, const TensorInfo &output, const LstmDescriptor &descriptor, const LstmInputParamsInfo ¶msInfo) |
| arm_compute::Status | NeonMaximumWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output) |
| arm_compute::Status | NeonMeanWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const MeanDescriptor &descriptor) |
| arm_compute::Status | NeonMinimumWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output) |
| Validate function for validating the inputs and output. | |
| arm_compute::Status | NeonMultiplicationWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ActivationDescriptor *activationDescriptor) |
| arm_compute::Status | NeonNegWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
| arm_compute::Status | NeonNormalizationWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const NormalizationDescriptor &descriptor) |
| arm_compute::Status | NeonPadWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const PadDescriptor &descriptor) |
| arm_compute::Status | NeonPermuteWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const PermuteDescriptor &descriptor) |
| arm_compute::Status | NeonPooling2dWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const Pooling2dDescriptor &descriptor) |
| arm_compute::Status | NeonPooling3dWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const Pooling3dDescriptor &descriptor) |
| arm_compute::Status | NeonPreluWorkloadValidate (const TensorInfo &input, const TensorInfo &alpha, const TensorInfo &output) |
| arm_compute::Status | NeonQLstmWorkloadValidate (const TensorInfo &input, const TensorInfo &cellStateIn, const TensorInfo &outputStateIn, const TensorInfo &cellStateOut, const TensorInfo &outputStateOut, const TensorInfo &output, const QLstmDescriptor &descriptor, const LstmInputParamsInfo ¶msInfo) |
| arm_compute::Status | NeonQuantizedLstmWorkloadValidate (const TensorInfo &input, const TensorInfo &cellStateIn, const TensorInfo &outputStateIn, const TensorInfo &cellStateOut, const TensorInfo &outputStateOut, const QuantizedLstmInputParamsInfo ¶msInfo) |
| arm_compute::Status | NeonQuantizeWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
| arm_compute::Status | NeonReduceWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const ReduceDescriptor &descriptor) |
| arm_compute::Status | NeonReshapeWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
| arm_compute::Status | NeonResizeWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const ResizeDescriptor &descriptor) |
| arm_compute::Status | NeonReverseV2WorkloadValidate (const TensorInfo &input, const TensorInfo &axis, const TensorInfo &output) |
| arm_compute::Status | NeonRsqrtWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
| arm_compute::Status | NeonSinWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
| arm_compute::Status | NeonSliceWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const SliceDescriptor &descriptor) |
| arm_compute::Status | NeonSoftmaxWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const SoftmaxDescriptor &descriptor) |
| arm_compute::Status | NeonSpaceToBatchNdWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const SpaceToBatchNdDescriptor &descriptor) |
| arm_compute::Status | NeonSpaceToDepthWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const SpaceToDepthDescriptor &descriptor) |
| arm_compute::Status | NeonSplitterWorkloadValidate (const TensorInfo &input, const std::vector< std::reference_wrapper< TensorInfo > > &outputs, unsigned int splitAxis) |
| arm_compute::Status | NeonSqrtWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
| arm_compute::Status | NeonStackWorkloadValidate (const std::vector< const TensorInfo * > &inputs, const TensorInfo &output, const StackDescriptor &descriptor) |
| arm_compute::Status | NeonStridedSliceWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const StridedSliceDescriptor &descriptor) |
| arm_compute::Status | NeonSubtractionWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ActivationDescriptor *activationDescriptor) |
| arm_compute::Status | NeonTileWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const TileDescriptor &descriptor) |
| arm_compute::Status | NeonTransposeConvolution2dWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const TransposeConvolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases) |
| arm_compute::Status | NeonTransposeWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const TransposeDescriptor &descriptor) |
| arm_compute::Status | NeonUnidirectionalSequenceLstmFloatWorkloadValidate (const TensorInfo &input, const TensorInfo &outputStateIn, const TensorInfo &cellStateIn, const TensorInfo &outputStateOut, const TensorInfo &cellStateOut, const TensorInfo &output, const UnidirectionalSequenceLstmDescriptor &descriptor, const LstmInputParamsInfo ¶msInfo) |
| arm_compute::Status | NeonUnidirectionalSequenceLstmWorkloadValidate (const TensorInfo &input, const TensorInfo &outputStateIn, const TensorInfo &cellStateIn, const TensorInfo &outputStateOut, const TensorInfo &cellStateOut, const TensorInfo &output, const UnidirectionalSequenceLstmDescriptor &descriptor, const LstmInputParamsInfo ¶msInfo) |
| template<typename T> | |
| void | CopyArmComputeTensorData (arm_compute::Tensor &dstTensor, const T *srcData) |
| void | InitializeArmComputeTensorData (arm_compute::Tensor &tensor, TensorInfo tensorInfo, const ITensorHandle *handle) |
| void | InitializeArmComputeTensorData (arm_compute::Tensor &tensor, const ConstTensorHandle *handle) |
| auto | SetNeonStridedSliceData (const std::vector< int > &m_begin, const std::vector< int > &m_end, const std::vector< int > &m_stride) |
| auto | SetNeonSliceData (const std::vector< unsigned int > &m_begin, const std::vector< unsigned int > &m_size) |
| constexpr const char * | RefBackendId () |
| constexpr const char * | RefTensorHandleFactoryId () |
| template<DataType ArmnnType> | |
| bool | IsDataType (const WorkloadInfo &info) |
| bool | IsSigned64 (const WorkloadInfo &info) |
| bool | IsSigned32 (const WorkloadInfo &info) |
| bool | IsBFloat16 (const WorkloadInfo &info) |
| bool | IsFloat16 (const WorkloadInfo &info) |
| bool | IsQSymmS16 (const WorkloadInfo &info) |
| bool | IsQSymmS8 (const WorkloadInfo &info) |
| bool | IsQAsymmS8 (const WorkloadInfo &info) |
| bool | IsQAsymmU8 (const WorkloadInfo &info) |
| bool | IsBoolean (const WorkloadInfo &info) |
| template<typename QueueDescriptorType> | |
| constexpr bool | IsOperationQueueDescriptor (const QueueDescriptorType &) |
| template<> | |
| constexpr bool | IsOperationQueueDescriptor (const MemCopyQueueDescriptor &) |
| template<> | |
| constexpr bool | IsOperationQueueDescriptor (const ConstantQueueDescriptor &) |
| template<> | |
| constexpr bool | IsOperationQueueDescriptor (const PermuteQueueDescriptor &) |
| float | Activation (float in, ActivationFunction function, float a, float b) |
| void | Activation (Decoder< float > &in, Encoder< float > &out, const TensorInfo &tensorInfo, ActivationFunction function, float a, float b) |
| template<typename OUT> | |
| void | ArgMinMax (Decoder< float > &in, OUT *out, const TensorInfo &inputTensorInfo, const TensorInfo &outputTensorInfo, ArgMinMaxFunction function, int axis) |
| template void | ArgMinMax (Decoder< float > &in, int32_t *out, const TensorInfo &inputTensorInfo, const TensorInfo &outputTensorInfo, ArgMinMaxFunction function, int axis) |
| template void | ArgMinMax (Decoder< float > &in, int64_t *out, const TensorInfo &inputTensorInfo, const TensorInfo &outputTensorInfo, ArgMinMaxFunction function, int axis) |
| void | BatchNormImpl (const BatchNormalizationQueueDescriptor &data, Decoder< float > &meanDecoder, Decoder< float > &varianceDecoder, Decoder< float > &betaDecoder, Decoder< float > &gammaDecoder, Decoder< float > &inputDecoder, Encoder< float > &outputEncoder) |
| unsigned int | Offset (const TensorShape &shape, unsigned int batch, unsigned int height, unsigned int width, unsigned int channels, const DataLayoutIndexed &dataLayout) |
| void | BatchToSpaceNd (const TensorInfo &inputInfo, const TensorInfo &outputInfo, const BatchToSpaceNdDescriptor ¶ms, Decoder< float > &inputData, Encoder< float > &outputData) |
| void | Concatenate (const ConcatQueueDescriptor &data, std::vector< ITensorHandle * > inputs, std::vector< ITensorHandle * > outputs) |
| void | Convolve3d (const TensorShape &rInputShape, Decoder< float > &rInputDecoder, const TensorShape &rOutputShape, Encoder< float > &rOutputEncoder, const TensorShape &rFilterShape, Decoder< float > &rFilterDecoder, bool biasEnabled, Decoder< float > *pBiasDecoder, DataLayout dataLayout, unsigned int paddingTop, unsigned int paddingLeft, unsigned int paddingFront, unsigned int xStride, unsigned int yStride, unsigned int zStride, unsigned int xDilation, unsigned int yDilation, unsigned int zDilation) |
| void | Convolve (const TensorShape &rInputShape, Decoder< float > &rInputDecoder, const TensorShape &rOutputShape, Encoder< float > &rOutputEncoder, const TensorShape &rFilterShape, Decoder< float > &rFilterDecoder, bool biasEnabled, Decoder< float > *pBiasDecoder, DataLayout dataLayout, unsigned int paddingTop, unsigned int paddingLeft, unsigned int xStride, unsigned int yStride, unsigned int xDilation, unsigned int yDilation, bool depthwise) |
| template<typename T> | |
| void | PrintOutput (const TensorInfo &inputInfo, const T *inputData, LayerGuid guid, const std::string &layerName, unsigned int slotIndex, std::ostream &os) |
| template<typename T> | |
| void | Debug (const TensorInfo &inputInfo, const T *inputData, LayerGuid guid, const std::string &layerName, unsigned int slotIndex, bool outputsToFile) |
| template void | Debug< BFloat16 > (const TensorInfo &inputInfo, const BFloat16 *inputData, LayerGuid guid, const std::string &layerName, unsigned int slotIndex, bool outputsToFile) |
| template void | Debug< Half > (const TensorInfo &inputInfo, const Half *inputData, LayerGuid guid, const std::string &layerName, unsigned int slotIndex, bool outputsToFile) |
| template void | Debug< float > (const TensorInfo &inputInfo, const float *inputData, LayerGuid guid, const std::string &layerName, unsigned int slotIndex, bool outputsToFile) |
| template void | Debug< uint8_t > (const TensorInfo &inputInfo, const uint8_t *inputData, LayerGuid guid, const std::string &layerName, unsigned int slotIndex, bool outputsToFile) |
| template void | Debug< int8_t > (const TensorInfo &inputInfo, const int8_t *inputData, LayerGuid guid, const std::string &layerName, unsigned int slotIndex, bool outputsToFile) |
| template void | Debug< int16_t > (const TensorInfo &inputInfo, const int16_t *inputData, LayerGuid guid, const std::string &layerName, unsigned int slotIndex, bool outputsToFile) |
| template void | Debug< int32_t > (const TensorInfo &inputInfo, const int32_t *inputData, LayerGuid guid, const std::string &layerName, unsigned int slotIndex, bool outputsToFile) |
| template void | Debug< int64_t > (const TensorInfo &inputInfo, const int64_t *inputData, LayerGuid guid, const std::string &layerName, unsigned int slotIndex, bool outputsToFile) |
| template<typename T> | |
| std::unique_ptr< Decoder< T > > | MakeDecoder (const TensorInfo &info, const void *data=nullptr) |
| template<> | |
| std::unique_ptr< Decoder< float > > | MakeDecoder (const TensorInfo &info, const void *data) |
| template<> | |
| std::unique_ptr< Decoder< double_t > > | MakeDecoder (const TensorInfo &info, const void *data) |
| template<> | |
| std::unique_ptr< Decoder< bool > > | MakeDecoder (const TensorInfo &info, const void *data) |
| template<> | |
| std::unique_ptr< Decoder< int32_t > > | MakeDecoder (const TensorInfo &info, const void *data) |
| void | DepthToSpace (const TensorInfo &inputInfo, const DepthToSpaceDescriptor &descriptor, const void *inputData, void *outputData, unsigned int dataTypeSize) |
| void | Dequantize (Decoder< float > &inputDecoder, Encoder< float > &outputEncoder, const TensorInfo &inputInfo, const TensorInfo &outputInfo) |
| std::vector< unsigned int > | GenerateRangeK (unsigned int k) |
| void | TopKSort (unsigned int k, unsigned int *indices, const float *values, unsigned int numElement) |
| float | IntersectionOverUnion (const float *boxI, const float *boxJ) |
| std::vector< unsigned int > | NonMaxSuppression (unsigned int numBoxes, const std::vector< float > &boxCorners, const std::vector< float > &scores, float nmsScoreThreshold, unsigned int maxDetection, float nmsIouThreshold) |
| void | AllocateOutputData (unsigned int numOutput, unsigned int numSelected, const std::vector< float > &boxCorners, const std::vector< unsigned int > &outputIndices, const std::vector< unsigned int > &selectedBoxes, const std::vector< unsigned int > &selectedClasses, const std::vector< float > &selectedScores, float *detectionBoxes, float *detectionScores, float *detectionClasses, float *numDetections) |
| void | DetectionPostProcess (const TensorInfo &boxEncodingsInfo, const TensorInfo &scoresInfo, const TensorInfo &, const TensorInfo &detectionBoxesInfo, const TensorInfo &, const TensorInfo &, const TensorInfo &, const DetectionPostProcessDescriptor &desc, Decoder< float > &boxEncodings, Decoder< float > &scores, Decoder< float > &anchors, float *detectionBoxes, float *detectionClasses, float *detectionScores, float *numDetections) |
| template<typename T> | |
| std::unique_ptr< Encoder< T > > | MakeEncoder (const TensorInfo &info, void *data=nullptr) |
| template<> | |
| std::unique_ptr< Encoder< float > > | MakeEncoder (const TensorInfo &info, void *data) |
| template<> | |
| std::unique_ptr< Encoder< double_t > > | MakeEncoder (const TensorInfo &info, void *data) |
| template<> | |
| std::unique_ptr< Encoder< bool > > | MakeEncoder (const TensorInfo &info, void *data) |
| template<> | |
| std::unique_ptr< Encoder< int32_t > > | MakeEncoder (const TensorInfo &info, void *data) |
| void | Fill (Encoder< float > &output, const TensorShape &desiredOutputShape, const float value) |
| Creates a tensor and fills it with a scalar value. | |
| void | FullyConnected (const TensorShape &rInputShape, Decoder< float > &rInputDecoder, const TensorShape &rOutputShape, Encoder< float > &rOutputEncoder, const TensorShape &rWeightsShape, Decoder< float > &rWeightDecoder, Decoder< float > *rBiasDecoder, bool biasEnabled, unsigned int K, bool transposeWeights) |
| Performs a matrix multiplication and optionally adds a bias. | |
| template<typename I, typename O> | |
| void | Gather (const TensorInfo ¶msInfo, const TensorInfo &indicesInfo, const TensorInfo &outputInfo, Decoder< I > ¶ms, const int32_t *indices, Encoder< O > &output, const int32_t axis_int) |
| template void | Gather (const TensorInfo ¶msInfo, const TensorInfo &indicesInfo, const TensorInfo &outputInfo, Decoder< float > ¶ms, const int32_t *indices, Encoder< float > &output, const int32_t axis_int) |
| template void | Gather (const TensorInfo ¶msInfo, const TensorInfo &indicesInfo, const TensorInfo &outputInfo, Decoder< double_t > ¶ms, const int32_t *indices, Encoder< double_t > &output, const int32_t axis_int) |
| void | InstanceNorm (const InstanceNormalizationQueueDescriptor &data, const TensorInfo &inputInfo, Decoder< float > &inputDecoder, Encoder< float > &outputEncoder) |
| void | LogSoftmax (Decoder< float > &input, Encoder< float > &output, const TensorInfo &inputInfo, const LogSoftmaxDescriptor &descriptor) |
| void | LstmImpl (const LstmDescriptor &descriptor, const TensorInfo &inputInfo, const TensorInfo &outputInfo, const TensorShape &inputToOutputWeightsShape, const TensorShape &recurrentToOutputWeightsShape, std::unique_ptr< Decoder< float > > &inputData, std::unique_ptr< Decoder< float > > &outputStateIn, std::unique_ptr< Decoder< float > > &cellStateIn, std::unique_ptr< Encoder< float > > &outputStateOut, std::unique_ptr< Encoder< float > > &cellStateOut, std::unique_ptr< Encoder< float > > &output, std::unique_ptr< Decoder< float > > &cellStateOutDecoder, std::unique_ptr< Decoder< float > > &outputDecoder, std::unique_ptr< Decoder< float > > &inputToInputWeightsTensor, std::unique_ptr< Decoder< float > > &inputToForgetWeightsTensor, std::unique_ptr< Decoder< float > > &inputToCellWeightsTensor, std::unique_ptr< Decoder< float > > &inputToOutputWeightsTensor, std::unique_ptr< Decoder< float > > &recurrentToInputWeightsTensor, std::unique_ptr< Decoder< float > > &recurrentToForgetWeightsTensor, std::unique_ptr< Decoder< float > > &recurrentToCellWeightsTensor, std::unique_ptr< Decoder< float > > &recurrentToOutputWeightsTensor, std::unique_ptr< Decoder< float > > &cellToInputWeightsTensor, std::unique_ptr< Decoder< float > > &cellToForgetWeightsTensor, std::unique_ptr< Decoder< float > > &cellToOutputWeightsTensor, std::unique_ptr< Decoder< float > > &inputGateBiasTensor, std::unique_ptr< Decoder< float > > &forgetGateBiasTensor, std::unique_ptr< Decoder< float > > &cellBiasTensor, std::unique_ptr< Decoder< float > > &outputGateBiasTensor, std::unique_ptr< Decoder< float > > &projectionWeightsTensor, std::unique_ptr< Decoder< float > > &projectionBiasTensor, std::unique_ptr< Decoder< float > > &inputLayerNormWeights, std::unique_ptr< Decoder< float > > &forgetLayerNormWeights, std::unique_ptr< Decoder< float > > &cellLayerNormWeights, std::unique_ptr< Decoder< float > > &outputLayerNormWeights, std::unique_ptr< Encoder< float > > &inputGateScratch, std::unique_ptr< Encoder< float > > &cellScratch, std::unique_ptr< Encoder< float > > &forgetGateScratch, std::unique_ptr< Encoder< float > > &outputGateScratch, std::unique_ptr< Decoder< float > > &inputGateScratchDecoder, std::unique_ptr< Decoder< float > > &cellScratchDecoder, std::unique_ptr< Decoder< float > > &forgetGateScratchDecoder, std::unique_ptr< Decoder< float > > &outputGateScratchDecoder, float layerNormEpsilon) |
| void | MirrorPad (const TensorInfo &inputInfo, const TensorInfo &outputInfo, const ITensorHandle *inputHandle, ITensorHandle *outputHandle, const PadQueueDescriptor &data) |
| void | Pad (const TensorInfo &inputInfo, const TensorInfo &outputInfo, const ITensorHandle *inputHandle, ITensorHandle *outputHandle, const PadQueueDescriptor &data) |
| void | Pooling2d (Decoder< float > &rInputDecoder, Encoder< float > &rOutputEncoder, const TensorInfo &inputInfo, const TensorInfo &outputInfo, const Pooling2dDescriptor ¶ms) |
| Computes the Pooling2d operation. | |
| void | Pooling3d (Decoder< float > &rInputDecoder, Encoder< float > &rOutputEncoder, const TensorInfo &inputInfo, const TensorInfo &outputInfo, const Pooling3dDescriptor ¶ms) |
| Computes the Pooling3d operation. | |
| void | PreluImpl (const TensorInfo &inputInfo, const TensorInfo &alphaInfo, const TensorInfo &outputInfo, Decoder< float > &inputData, Decoder< float > &alphaData, Encoder< float > &outputData) |
| bool | NextIndex (const unsigned int numDims, const armnn::TensorShape &dims, std::vector< unsigned int > ¤t) |
| unsigned int | ReducedOutputOffset (const unsigned int numDims, const armnn::TensorShape &dims, std::vector< unsigned int > &index, const unsigned int numAxis, const std::vector< unsigned int > &axis) |
| void | Reduce (const TensorInfo &inputInfo, const TensorInfo &outputInfo, Decoder< float > &input, Encoder< float > &output, const std::vector< uint32_t > axis, const ReduceOperation reduceOperation) |
| template<typename DataType> | |
| void | ExecuteFunction (std::vector< ITensorHandle * > inputs, std::vector< ITensorHandle * > outputs, BinaryOperation operation) |
| void | FakeQuantization (const float *inputData, float *outputData, uint32_t numElements, float min, float max) |
| unsigned int | GetNumActivations (const TensorInfo &inputInfo) |
| template<typename TensorHandleType = RefTensorHandle> | |
| const TensorInfo & | GetTensorInfo (const ITensorHandle *tensorHandle) |
| float32 helpers | |
| template<typename DataType, typename PayloadType> | |
| const DataType * | GetInputTensorData (unsigned int idx, const PayloadType &data) |
| template<typename DataType> | |
| DataType * | GetOutputTensorData (ITensorHandle *tensorHandle) |
| template<typename PayloadType> | |
| const float * | GetInputTensorDataFloat (unsigned int idx, const PayloadType &data) |
| template<typename PayloadType> | |
| float * | GetOutputTensorDataFloat (unsigned int idx, const PayloadType &data) |
| template<typename PayloadType> | |
| const Half * | GetInputTensorDataHalf (unsigned int idx, const PayloadType &data) |
| template<typename PayloadType> | |
| Half * | GetOutputTensorDataHalf (unsigned int idx, const PayloadType &data) |
| template<typename PayloadType> | |
| const BFloat16 * | GetInputTensorDataBFloat16 (unsigned int idx, const PayloadType &data) |
| template<typename PayloadType> | |
| BFloat16 * | GetOutputTensorDataBFloat16 (unsigned int idx, const PayloadType &data) |
| template<typename T> | |
| std::vector< float > | Dequantize (const T *quant, const TensorInfo &info) |
| u8 helpers | |
| template<typename T> | |
| void | Dequantize (const T *inputData, float *outputData, const TensorInfo &info) |
| void | Quantize (uint8_t *quant, const float *dequant, const TensorInfo &info) |
| void | Resize (Decoder< float > &in, const TensorInfo &inputInfo, Encoder< float > &out, const TensorInfo &outputInfo, DataLayoutIndexed dataLayout, ResizeMethod resizeMethod, bool alignCorners, bool halfPixelCenters) |
| std::vector< unsigned int > | ReverseGetMultIdx (const unsigned int idx, unsigned int inputRank, std::vector< unsigned int > &elementNumInner) |
| unsigned int | ReverseGetFlatIdx (const std::vector< unsigned int > &idxList, unsigned int inputRank, std::vector< unsigned int > &elementNumInner) |
| unsigned int | ReverseRelocateIdx (unsigned int idx, unsigned int inputRank, std::vector< bool > &axisFlag, std::vector< unsigned int > &dimSize, std::vector< unsigned int > &elementNumInner) |
| void | ReverseV2 (const TensorInfo &inputInfo, const TensorInfo &axisInfo, Decoder< float > &inputDecoder, Decoder< int > &axisDecoder, Encoder< float > &outputEncoder) |
| float | ScatterOperation (ScatterNdFunction operation, float input, float update) |
| void | ScatterNd (const TensorInfo &inputInfo, const TensorInfo &indicesInfo, const TensorInfo &updatesInfo, Decoder< float > &input, Decoder< int > &indices, Decoder< float > &updates, Encoder< float > &output, ScatterNdDescriptor descriptor) |
| void | ScatterNd (const TensorInfo &indicesInfo, const TensorInfo &updatesInfo, const TensorInfo &shapeInfo, Decoder< int > &indices, Decoder< float > &updates, Decoder< int > &shape, Encoder< float > &output, ScatterNdDescriptor descriptor) |
| void | Slice (const TensorInfo &inputInfo, const SliceDescriptor &descriptor, const void *inputData, void *outputData, unsigned int dataTypeSize) |
| void | Softmax (Decoder< float > &in, Encoder< float > &out, const TensorInfo &inputTensorInfo, float beta, int axis) |
| Computes the softmax function on some inputs, into outputs, with a shape given by tensorInfo. | |
| unsigned int | GetOffset (const TensorShape &shape, unsigned int b, unsigned int h, unsigned int w, unsigned int c, const DataLayoutIndexed &dataLayout) |
| void | SpaceToBatchNd (const TensorInfo &inputInfo, const TensorInfo &outputInfo, const SpaceToBatchNdDescriptor ¶ms, Decoder< float > &inputData, Encoder< float > &outputData) |
| void | SpaceToDepth (const TensorInfo &inputInfo, const TensorInfo &outputInfo, const SpaceToDepthDescriptor ¶ms, Decoder< float > &inputData, Encoder< float > &outputData) |
| void | Split (const SplitterQueueDescriptor &data, std::vector< ITensorHandle * > inputs, std::vector< ITensorHandle * > outputs) |
| template<typename DataType> | |
| void | Splitter (const SplitterQueueDescriptor &data, std::vector< ITensorHandle * > inputs, std::vector< ITensorHandle * > outputs) |
| void | Stack (const StackQueueDescriptor &data, std::vector< std::unique_ptr< Decoder< float > > > &inputs, Encoder< float > &output, const TensorInfo &inputInfo, const TensorInfo &outputInfo) |
| void | StridedSlice (const TensorInfo &inputInfo, const StridedSliceDescriptor ¶ms, const void *inputData, void *outputData, unsigned int dataTypeSize) |
| std::vector< uint32_t > | IndexToCoordinates (std::vector< uint32_t > &shape, uint32_t index) |
| uint32_t | CoordinatesToIndex (TensorShape &shape, std::vector< uint32_t > &coordinates) |
| template<typename I, typename O> | |
| void | Tile (const TileDescriptor ¶ms, const TensorInfo &inputInfo, Decoder< I > &inputDecoder, Encoder< O > &outputEncoder) |
| template void | Tile (const TileDescriptor ¶ms, const TensorInfo &inputInfo, Decoder< float > &inputDecoder, Encoder< float > &outputEncoder) |
| template void | Tile (const TileDescriptor ¶ms, const TensorInfo &inputInfo, Decoder< double_t > &inputDecoder, Encoder< double_t > &outputEncoder) |
| void | TransposeConvolution2dImpl (const TransposeConvolution2dDescriptor &descriptor, const TensorShape &inputShape, Decoder< float > &inputDecoder, const TensorShape &outputShape, Encoder< float > &outputEncoder, const TensorShape &weightsShape, Decoder< float > &weightsDecoder, Decoder< float > *biasesDecoder) |
| template<typename T> | |
| void | DeleteAsType (const void *const blob) |
| constexpr const char * | TosaRefBackendId () |
| constexpr const char * | TosaRefTensorHandleFactoryId () |
| bool | TosaRefPreCompiledWorkloadValidate (std::string *) |
Variables | |
| constexpr unsigned int | MaxNumOfTensorDimensions = 5U |
| constexpr unsigned int | LOWEST_CAPTURE_PERIOD = 10000u |
| The lowest performance data capture interval we support is 10 miliseconds. | |
| constexpr unsigned int | EXPIRE_RATE = 3U |
| Variable to control expire rate of priority queue. | |
| constexpr std::size_t | g_ProfilingEventCountHint = 1024 |
| constexpr bool | g_WriteProfilingEventSequence = true |
| constexpr bool | g_AggregateProfilingEventsByInference = true |
| constexpr bool | g_WriteReportToStdOutOnProfilerDestruction = false |
| thread_local IProfiler * | tl_Profiler = nullptr |
| constexpr size_t | wordSize = sizeof(size_t) * 8 |
| const BackendCapabilities | cpuAccCapabilities ("CpuAcc", { {"NonConstWeights", true}, {"ProtectedContentAllocation", false}, {"ConstantTensorsAsInputs", true}, {"PreImportIOTensors", false}, {"ExternallyManagedMemory", true}, {"MultiAxisPacking", false}, {"SingleAxisPacking", true}, {"HasFp16", arm_compute::CPUInfo::get().has_fp16()}, {"AllOrNothing", false} }) |
| const std::set< armnn::LayerType > | paddingRequiredLayers |
| const BackendCapabilities | cpuRefCapabilities ("CpuRef", { {"NonConstWeights", true}, {"ProtectedContentAllocation", false}, {"ConstantTensorsAsInputs", true}, {"PreImportIOTensors", true}, {"ExternallyManagedMemory", true}, {"MultiAxisPacking", false}, {"SingleAxisPacking", true}, {"HasFp16", true}, {"AllOrNothing", false} }) |
| const std::set< armnn::BackendCapability > | oldCpuRefCapabilities |
Copyright (c) 2021 ARM Limited and Contributors.
Copyright (c) 2021, 2024 ARM Limited and Contributors.
Optional is a drop in replacement for std::optional until we migrate to c++-17.
Copyright (c) 2022 ARM Limited and Contributors.
Copyright (c) 2022-2023 ARM Limited and Contributors.
Copyright (c) 2021-2023 ARM Limited and Contributors.
Copyright (c) 2022-2024 Arm Ltd and Contributors.
Copyright (c) 2022, 2024 ARM Limited and Contributors.
Create pages for each tool so they appear nicely in the doxygen tree-view.
Copyright (c) 2021-2024 ARM Limited and Contributors.
All rights reserved.
SPDX-License-Identifier: MIT
Subpages are not listed there. Also we can overwrite the page name this way.
Subpages are not listed there.
Note: The parser, serializer and deserializer pages are created in 01_parsers.dox or 02_deserializer_serializer.dox
Only a subset of the optional features are implemented that we intend to use in ArmNN. There are two distinct implementations here:
1, for normal constructable/destructable types and reference types 2, for reference types The std::optional features we support are:
| using ACLMemManagerOnDemand = std::shared_ptr<arm_compute::MemoryManagerOnDemand> |
Definition at line 22 of file NeonFullyConnectedWorkload.cpp.
| using AdditionalInfoObjectPtr = std::shared_ptr<void> |
| using BackendCapabilities = BackendOptions |
Definition at line 19 of file BackendOptions.hpp.
| using BackendIdSet = std::unordered_set<BackendId> |
Definition at line 196 of file BackendId.hpp.
| using BackendIdVector = std::vector<BackendId> |
Definition at line 195 of file BackendId.hpp.
| using BackendsMap = std::map<BackendId, std::unique_ptr<class IBackendInternal>> |
Definition at line 285 of file Network.hpp.
| using BaseFloat32ComparisonWorkload |
Definition at line 212 of file Workload.hpp.
| using BaseUint8ComparisonWorkload |
Definition at line 217 of file Workload.hpp.
| using BFloat16ToFloat32Workload |
Definition at line 222 of file Workload.hpp.
| using BindingPointInfo = std::pair<armnn::LayerBindingId, armnn::TensorInfo> |
Definition at line 276 of file Tensor.hpp.
| using BooleanWorkload = TypedWorkload<QueueDescriptor, armnn::DataType::Boolean> |
Definition at line 209 of file Workload.hpp.
| typedef std::function< void(const void *)> CompiledBlobDeleter = std::function<void(const void*)> |
Definition at line 342 of file INetwork.hpp.
| typedef std::unique_ptr< void, CompiledBlobDeleter > CompiledBlobPtr = std::unique_ptr<void, CompiledBlobDeleter> |
Definition at line 343 of file INetwork.hpp.
| using ConcatDescriptor = OriginsDescriptor |
Definition at line 60 of file DescriptorsFwd.hpp.
| using Coordinates = std::array<unsigned int, MaxNumOfTensorDimensions> |
Definition at line 15 of file InternalTypes.hpp.
| using CopyAndImportFactoryPairs = std::map<ITensorHandleFactory::FactoryId, ITensorHandleFactory::FactoryId> |
Definition at line 20 of file TensorHandleFactoryRegistry.hpp.
| using DebugCallbackFunction = std::function<void(LayerGuid guid, unsigned int slotIndex, ITensorHandle* tensorHandle)> |
Define the type of callback for the Debug layer to call.
| guid | - guid of layer connected to the input of the Debug layer |
| slotIndex | - index of the output slot connected to the input of the Debug layer |
| tensorHandle | - TensorHandle for the input tensor to the Debug layer |
A DepthToSpaceDescriptor for the DepthToSpaceLayer.
Definition at line 1099 of file Descriptors.hpp.
| using Dimensions = std::array<unsigned int, MaxNumOfTensorDimensions> |
Definition at line 16 of file InternalTypes.hpp.
| using DynamicBackendPtr = std::unique_ptr<DynamicBackend> |
Definition at line 54 of file DynamicBackend.hpp.
Definition at line 12 of file MockTensorHandleFactory.cpp.
| using Float16ToFloat32Workload |
Definition at line 232 of file Workload.hpp.
| using Float32ToBFloat16Workload |
Definition at line 227 of file Workload.hpp.
| using Float32ToFloat16Workload |
Definition at line 237 of file Workload.hpp.
| using Float32Workload = TypedWorkload<QueueDescriptor, armnn::DataType::Float32> |
Definition at line 200 of file Workload.hpp.
| using FloatWorkload |
Definition at line 195 of file Workload.hpp.
| using HighResolutionClock = std::chrono::high_resolution_clock::time_point |
| using IBackendContextUniquePtr = std::unique_ptr<IBackendContext> |
Definition at line 34 of file IBackendContext.hpp.
| typedef std::unique_ptr< IBackendInternal > IBackendInternalUniquePtr = std::unique_ptr<IBackendInternal> |
Definition at line 32 of file BackendRegistry.hpp.
| using IBackendSharedPtr = std::shared_ptr<IBackend> |
| using IBackendUniquePtr = std::unique_ptr<IBackend, void(*)(IBackend* backend)> |
| using IGpuAccTunedParametersPtr = std::shared_ptr<IGpuAccTunedParameters> |
The following API is replaced by the backend options API.
Definition at line 268 of file IRuntime.hpp.
| using IInitialiseProfilingService = arm::pipe::IInitialiseProfilingService |
Definition at line 28 of file Runtime.hpp.
| using ILayerSupportSharedPtr = std::shared_ptr<ILayerSupport> |
Definition at line 40 of file ILayerSupport.hpp.
| using IMemoryManagerUniquePtr = std::unique_ptr<IMemoryManager> |
Definition at line 24 of file IMemoryManager.hpp.
| using ImportedInputId = unsigned int |
| using ImportedOutputId = unsigned int |
| using INetworkPtr = std::unique_ptr<INetwork, void(*)(INetwork* network)> |
Definition at line 339 of file INetwork.hpp.
| using InferenceTimingPair = std::pair<HighResolutionClock, HighResolutionClock> |
Definition at line 91 of file WorkloadData.hpp.
| using InputTensors = std::vector<std::pair<LayerBindingId, class ConstTensor>> |
Definition at line 394 of file Tensor.hpp.
| using Int32Workload = TypedWorkload<QueueDescriptor, armnn::DataType::Signed32> |
Definition at line 206 of file Workload.hpp.
| using IOptimizedNetworkPtr = std::unique_ptr<IOptimizedNetwork, void(*)(IOptimizedNetwork* network)> |
Definition at line 340 of file INetwork.hpp.
| using IReportStructure = arm::pipe::IReportStructure |
Definition at line 27 of file Runtime.hpp.
| using IRuntimePtr = std::unique_ptr<IRuntime, void(*)(IRuntime* runtime)> |
Definition at line 39 of file IRuntime.hpp.
| using LayerBindingId = int |
| using LayerPriority = unsigned int |
| using LayerTypeOf = typename LayerTypeOfImpl<Type>::Type |
Definition at line 95 of file LayersFwd.hpp.
| using LoadedNetworks = std::unordered_map<NetworkId, std::unique_ptr<LoadedNetwork>> |
Definition at line 26 of file Runtime.hpp.
A LogSoftmaxDescriptor for the LogSoftmaxLayer.
Definition at line 196 of file Descriptors.hpp.
| using MemoryOptimizerStrategiesMapRef = std::unordered_map<BackendId, std::shared_ptr<IMemoryOptimizerStrategy>> |
Definition at line 33 of file BackendRegistry.hpp.
| using MemorySourceFlags = unsigned int |
Definition at line 15 of file MemorySources.hpp.
| using MergerDescriptor = OriginsDescriptor |
MergerDescriptor is deprecated, use ConcatDescriptor instead.
Definition at line 64 of file DescriptorsFwd.hpp.
Definition at line 149 of file WorkloadData.hpp.
| using ModelOptions = std::vector<BackendOptions> |
Definition at line 18 of file BackendOptions.hpp.
| using NetworkId = int |
Definition at line 33 of file IRuntime.hpp.
| using NetworkImplPtr = std::unique_ptr<NetworkImpl, void (*)(NetworkImpl* network)> |
Definition at line 29 of file Network.hpp.
| using NetworkOptions = std::vector<BackendOptions> |
Definition at line 16 of file BackendOptions.hpp.
Definition at line 92 of file WorkloadData.hpp.
| using OutputTensors = std::vector<std::pair<LayerBindingId, class Tensor>> |
Definition at line 395 of file Tensor.hpp.
| using ParameterStringifyFunction = std::function<void(const std::string& name, const std::string& value)> |
Definition at line 14 of file SerializeLayerParameters.hpp.
| using PreCompiledObjectDeleter = std::function<void(const void*)> |
Definition at line 19 of file PreCompiledLayer.hpp.
| using PreCompiledObjectPtr = std::unique_ptr<void, PreCompiledObjectDeleter> |
Definition at line 20 of file PreCompiledLayer.hpp.
| using reduceValidateFunction |
Function pointer type used in IsMultiAxesReduceSupported for readability.
Definition at line 373 of file ArmComputeUtils.hpp.
| using RefAdditionWorkload |
Definition at line 36 of file RefElementwiseWorkload.hpp.
Definition at line 41 of file RefDebugWorkload.hpp.
Definition at line 50 of file RefDebugWorkload.hpp.
Definition at line 42 of file RefDebugWorkload.hpp.
Definition at line 43 of file RefDebugWorkload.hpp.
Definition at line 45 of file RefDebugWorkload.hpp.
Definition at line 44 of file RefDebugWorkload.hpp.
Definition at line 46 of file RefDebugWorkload.hpp.
Definition at line 47 of file RefDebugWorkload.hpp.
Definition at line 48 of file RefDebugWorkload.hpp.
Definition at line 49 of file RefDebugWorkload.hpp.
| using RefDivisionWorkload |
Definition at line 54 of file RefElementwiseWorkload.hpp.
| using RefMaximumWorkload |
Definition at line 60 of file RefElementwiseWorkload.hpp.
| using RefMinimumWorkload |
Definition at line 66 of file RefElementwiseWorkload.hpp.
| using RefMultiplicationWorkload |
Definition at line 48 of file RefElementwiseWorkload.hpp.
Definition at line 27 of file RefPermuteWorkload.hpp.
Definition at line 28 of file RefPermuteWorkload.hpp.
Definition at line 29 of file RefPermuteWorkload.hpp.
Definition at line 31 of file RefPermuteWorkload.hpp.
Definition at line 30 of file RefPermuteWorkload.hpp.
Definition at line 32 of file RefPermuteWorkload.hpp.
| using RefSubtractionWorkload |
Definition at line 42 of file RefElementwiseWorkload.hpp.
Definition at line 26 of file RefTransposeWorkload.hpp.
Definition at line 27 of file RefTransposeWorkload.hpp.
Definition at line 28 of file RefTransposeWorkload.hpp.
Definition at line 30 of file RefTransposeWorkload.hpp.
Definition at line 29 of file RefTransposeWorkload.hpp.
Definition at line 31 of file RefTransposeWorkload.hpp.
| using ResolveType = typename ResolveTypeImpl<DT>::Type |
Definition at line 79 of file ResolveType.hpp.
| using SplitterDescriptor = ViewsDescriptor |
Definition at line 65 of file DescriptorsFwd.hpp.
| using TensorInfos = std::vector<TensorInfo> |
Definition at line 152 of file BackendHelper.cpp.
| using Uint8ToFloat32Workload |
Definition at line 242 of file Workload.hpp.
| using Uint8Workload = TypedWorkload<QueueDescriptor, armnn::DataType::QAsymmU8> |
Definition at line 203 of file Workload.hpp.
Definition at line 1169 of file Descriptors.hpp.
| using WorkloadQueue = std::vector< std::unique_ptr<IWorkload> > |
Definition at line 13 of file ExecutionFrame.hpp.
|
strong |
| Enumerator | |
|---|---|
| Sigmoid | |
| TanH | |
| Linear | |
| ReLu | |
| BoundedReLu | min(a, max(b, input)) ReLu1 & ReLu6. |
| SoftReLu | |
| LeakyReLu | |
| Abs | |
| Sqrt | |
| Square | |
| Elu | |
| HardSwish | |
| Gelu | |
|
strong |
|
strong |
BackendCapability class.
Definition at line 287 of file Types.hpp.
|
strong |
|
strong |
|
strong |
Capability class to calculate in the GetCapabilities function so that only the capability in the scope can be choose to calculate.
| Enumerator | |
|---|---|
| PaddingRequired | |
| FallbackImportDisabled | |
| CapabilityClassMax | |
Definition at line 24 of file ITensorHandleFactory.hpp.
|
strong |
| Enumerator | |
|---|---|
| Equal | |
| Greater | |
| GreaterOrEqual | |
| Less | |
| LessOrEqual | |
| NotEqual | |
Definition at line 109 of file Types.hpp.
|
strong |
The Compute enum is now deprecated and it is now being replaced by BackendId.
| Enumerator | |
|---|---|
| Undefined | |
| CpuRef | CPU Execution: Reference C++ kernels. |
| CpuAcc | CPU Execution: NEON: ArmCompute. |
| GpuAcc | GPU Execution: OpenCL: ArmCompute. |
| TosaRef | CPU Execution: TOSA Reference Model. |
Definition at line 21 of file BackendId.hpp.
|
strong |
|
strong |
| Enumerator | |
|---|---|
| Float16 | |
| Float32 | |
| QAsymmU8 | |
| Signed32 | |
| Boolean | |
| QSymmS16 | |
| QSymmS8 | |
| QAsymmS8 | |
| BFloat16 | |
| Signed64 | |
|
strong |
| Enumerator | |
|---|---|
| NotSpecified | |
| Specified | |
| Scalar | |
|
strong |
Definition at line 104 of file ITensorHandleFactory.hpp.
|
strong |
|
strong |
| Enumerator | |
|---|---|
| LayerAdded | |
| LayerErased | |
Definition at line 12 of file IGraphObservable.hpp.
|
strong |
| Enumerator | |
|---|---|
| Measurement | |
| Event | |
| ExecObjectDesc | |
Definition at line 20 of file JsonPrinter.hpp.
|
strong |
When adding a new layer, adapt also the LastLayer enum value in the enum class LayerType below.
Definition at line 493 of file Types.hpp.
|
strong |
| Enumerator | |
|---|---|
| LogicalAnd | |
| LogicalOr | |
|
strong |
| Enumerator | |
|---|---|
| Trace | |
| Debug | |
| Info | |
| Warning | |
| Error | |
| Fatal | |
|
strong |
| Enumerator | |
|---|---|
| SingleAxisPacking | |
| MultiAxisPacking | |
Definition at line 254 of file Types.hpp.
|
strong |
Define the Memory Source to reduce copies.
| Enumerator | |
|---|---|
| Undefined | |
| Malloc | |
| DmaBuf | |
| DmaBufProtected | |
| Gralloc | |
Definition at line 245 of file Types.hpp.
|
strong |
|
strong |
| Enumerator | |
|---|---|
| LocalBrightness | Krichevsky 2012: Local Brightness Normalization. |
| LocalContrast | Jarret 2009: Local Contrast Normalization. |
Definition at line 214 of file Types.hpp.
|
strong |
|
strong |
The padding method modifies the output of pooling layers.
In both supported methods, the values are ignored (they are not even zeroes, which would make a difference for max pooling a tensor with negative values). The difference between IgnoreValue and Exclude is that the former counts the padding fields in the divisor of Average and L2 pooling, while Exclude does not.
| Enumerator | |
|---|---|
| IgnoreValue | The padding fields count, but are ignored. |
| Exclude | The padding fields don't count and are ignored. |
Definition at line 189 of file Types.hpp.
|
strong |
The padding mode controls whether the padding should be filled with constant values (Constant), or reflect the input, either including the border values (Symmetric) or not (Reflect).
| Enumerator | |
|---|---|
| Constant | |
| Reflect | |
| Symmetric | |
|
strong |
|
strong |
Define the behaviour of the internal profiler when outputting network details.
| Enumerator | |
|---|---|
| Undefined | |
| DetailsWithEvents | |
| DetailsOnly | |
|
strong |
|
strong |
|
strong |
| Enumerator | |
|---|---|
| Bilinear | |
| NearestNeighbor | |
|
strong |
|
strong |
The ShapeInferenceMethod modify how the output shapes are treated.
When ValidateOnly is selected, the output shapes are inferred from the input parameters of the layer and any mismatch is reported. When InferAndValidate is selected 2 actions are performed: (1)infer output shape from inputs and (2)validate the shapes as in ValidateOnly. This option has been added to work with tensors which rank or dimension sizes are not specified explicitly, however this information can be calculated from the inputs.
| Enumerator | |
|---|---|
| ValidateOnly | Validate all output shapes. |
| InferAndValidate | Infer missing output shapes and validate all output shapes. |
Definition at line 236 of file Types.hpp.
|
strong |
|
strong |
| Enumerator | |
|---|---|
| None | |
| Rapid | |
| Normal | |
| Exhaustive | |
Definition at line 18 of file ArmComputeTuningUtils.hpp.
|
strong |
| void Activation | ( | Decoder< float > & | in, |
| Encoder< float > & | out, | ||
| const TensorInfo & | tensorInfo, | ||
| ActivationFunction | function, | ||
| float | a, | ||
| float | b ) |
Definition at line 102 of file Activation.cpp.
References Activation, Decoder< IType >::Get(), TensorInfo::GetNumElements(), and Encoder< IType >::Set().
| float Activation | ( | float | in, |
| ActivationFunction | function, | ||
| float | a, | ||
| float | b ) |
Definition at line 13 of file Activation.cpp.
References Abs, BoundedReLu, Elu, Gelu, HardSwish, LeakyReLu, Linear, ReLu, Sigmoid, SoftReLu, Sqrt, Square, and TanH.
Referenced by ActivationLayer::ActivationLayer().
| void AllocateOutputData | ( | unsigned int | numOutput, |
| unsigned int | numSelected, | ||
| const std::vector< float > & | boxCorners, | ||
| const std::vector< unsigned int > & | outputIndices, | ||
| const std::vector< unsigned int > & | selectedBoxes, | ||
| const std::vector< unsigned int > & | selectedClasses, | ||
| const std::vector< float > & | selectedScores, | ||
| float * | detectionBoxes, | ||
| float * | detectionScores, | ||
| float * | detectionClasses, | ||
| float * | numDetections ) |
Definition at line 101 of file DetectionPostProcess.cpp.
References numeric_cast().
Referenced by DetectionPostProcess().
| bool AllTypesAreEqualImpl | ( | T | t1, |
| T | t2, | ||
| Rest... | rest ) |
Definition at line 64 of file LayerSupportRules.hpp.
References AllTypesAreEqualImpl().
| bool AllTypesAreEqualImpl | ( | T | ) |
Definition at line 58 of file LayerSupportRules.hpp.
Referenced by AllTypesAreEqualImpl(), and TypesAreEqual::TypesAreEqual().
| void Append | ( | Optimizer::Optimizations & | optimizations, |
| Front && | front, | ||
| Others &&... | others ) |
Definition at line 36 of file Optimizer.hpp.
References Append().
| void Append | ( | Optimizer::Optimizations & | optimizations, |
| T && | optimization ) |
Definition at line 30 of file Optimizer.hpp.
Referenced by Append(), and MakeOptimizations().
| OptimizationResult ApplyBackendOptimizations | ( | OptimizedNetworkImpl * | optNetObjPtr, |
| BackendSettings & | backendSettings, | ||
| BackendsMap & | backends, | ||
| const ModelOptions & | modelOptions, | ||
| Optional< std::vector< std::string > & > | errMessages ) |
Definition at line 1372 of file Network.cpp.
References ARMNN_SCOPED_PROFILING_EVENT, AssignBackends(), OutputSlot::Connect(), CpuAcc, OutputSlot::Disconnect(), Graph::EraseLayer(), Layer::GetBackendId(), OutputSlot::GetConnection(), OptimizationViews::GetDeletedSubgraphs(), OptimizationViews::GetFailedSubgraphs(), OptimizedNetworkImpl::GetGraph(), SubgraphView::GetIConnectableLayers(), Layer::GetInputSlot(), OutputSlot::GetNumConnections(), Layer::GetNumInputSlots(), Layer::GetOutputSlot(), OptimizationViews::GetSubstitutions(), Layer::GetType(), GpuAcc, Input, OptimizationResult::m_Error, BackendSettings::m_IgnoredBackends, BackendSettings::m_SelectedBackends, MakeOptimizations(), Output, Optimizer::Pass(), PolymorphicDowncast(), ReportWarning(), SubgraphViewSelector::SelectSubgraphs(), Graph::SubstituteSubgraph(), Undefined, and OptimizationViews::Validate().
Referenced by IOptimizedNetwork::Optimize.
|
inline |
Utility function used to check if a vector of tensors are Signed32.
Definition at line 28 of file ClFloorDivWorkload.cpp.
References Signed32.
Referenced by ClFloorDivWorkloadValidate().
| template void ArgMinMax | ( | Decoder< float > & | in, |
| int32_t * | out, | ||
| const TensorInfo & | inputTensorInfo, | ||
| const TensorInfo & | outputTensorInfo, | ||
| ArgMinMaxFunction | function, | ||
| int | axis ) |
| template void ArgMinMax | ( | Decoder< float > & | in, |
| int64_t * | out, | ||
| const TensorInfo & | inputTensorInfo, | ||
| const TensorInfo & | outputTensorInfo, | ||
| ArgMinMaxFunction | function, | ||
| int | axis ) |
| void ArgMinMax | ( | Decoder< float > & | in, |
| OUT * | out, | ||
| const TensorInfo & | inputTensorInfo, | ||
| const TensorInfo & | outputTensorInfo, | ||
| ArgMinMaxFunction | function, | ||
| int | axis ) |
Definition at line 17 of file ArgMinMax.cpp.
References Decoder< IType >::Get(), TensorInfo::GetNumDimensions(), armnnUtils::GetNumElementsBetween(), TensorInfo::GetShape(), armnnUtils::GetUnsignedAxis(), IgnoreUnused(), Max, Min, and numeric_cast().
Referenced by ArgMinMaxLayer::ArgMinMaxLayer().
| void AssertNumberOfInputSlots | ( | Layer & | layer | ) |
Definition at line 28 of file Layer.cpp.
References Convolution2d, DepthwiseConvolution2d, FullyConnected, Layer::GetNumInputSlots(), and Layer::GetType().
Referenced by InputSlot::Insert().
| OptimizationResult AssignBackends | ( | OptimizedNetworkImpl * | optNetObjPtr, |
| BackendSettings & | backendSettings, | ||
| Graph::Iterator & | firstLayer, | ||
| Graph::Iterator & | lastLayer, | ||
| Optional< std::vector< std::string > & > | errMessages ) |
Definition at line 1211 of file Network.cpp.
References ARMNN_SCOPED_PROFILING_EVENT, AssignBackendsIConnectable(), Float16, BackendSettings::GetAvailablePreferredBackends(), GetLayerInOutDatatype(), Input, OptimizationResult::m_Error, PolymorphicDowncast(), ReportError(), and Undefined.
Referenced by ApplyBackendOptimizations(), AssignBackends(), and IOptimizedNetwork::Optimize.
| OptimizationResult AssignBackends | ( | OptimizedNetworkImpl * | optNetObjPtr, |
| BackendSettings & | backendSettings, | ||
| SubgraphView & | subgraph, | ||
| Optional< std::vector< std::string > & > | errMessages ) |
Definition at line 1340 of file Network.cpp.
References AssignBackends(), SubgraphView::begin(), and SubgraphView::end().
| OptimizationResult AssignBackends | ( | OptimizedNetworkImpl * | optNetObjPtr, |
| BackendSettings & | backendSettings, | ||
| SubgraphView::IConnectableLayerIterator & | firstLayer, | ||
| SubgraphView::IConnectableLayerIterator & | lastLayer, | ||
| Optional< std::vector< std::string > & > | errMessages ) |
Definition at line 1294 of file Network.cpp.
References ARMNN_SCOPED_PROFILING_EVENT, AssignBackendsIConnectable(), BackendSettings::GetAvailablePreferredBackends(), Input, OptimizationResult::m_Error, PolymorphicDowncast(), ReportError(), and Undefined.
| void AssignBackendsIConnectable | ( | OptimizedNetworkImpl * | optNetObjPtr, |
| IConnectableLayer * | it, | ||
| Optional< std::vector< std::string > & > | errMessages, | ||
| OptimizationResult & | result, | ||
| BackendSettings & | backendSettings, | ||
| std::vector< BackendId > & | availablePreferredBackends, | ||
| bool & | restart ) |
Definition at line 1092 of file Network.cpp.
References AttemptBackendAssignment(), CheckScaleSetOnQuantizedType(), Constant, CpuRef, OptimizedNetworkImpl::GetGraph(), GetLayerInOutDatatype(), Input, BackendSettings::IsBackendSupported(), BackendSettings::IsCpuRefUsed(), OptimizationResult::IsError(), OptimizationResult::IsOk(), OptimizationResult::IsWarningOnly(), OptimizationResult::m_Error, BackendSettings::m_IgnoredBackends, BackendSettings::m_SelectedBackends, MemCopy, Permute, PolymorphicDowncast(), and ReturnWithError().
Referenced by AssignBackends(), and AssignBackends().
| void AssignSplitId | ( | LayerSelectionInfo::LayerInfoContainer & | layerInfos, |
| LayerSelectionInfo & | layerInfo ) |
Definition at line 322 of file SubgraphViewSelector.cpp.
References ForEachLayerInput().
Referenced by SubgraphViewSelector::SelectSubgraphs().
| OptimizationResult AttemptBackendAssignment | ( | BackendSettings & | backendSettings, |
| Graph & | graph, | ||
| Layer * | layer, | ||
| BackendId | backend, | ||
| DataType | dataTypeIn, | ||
| DataType | dataTypeOut, | ||
| const std::vector< BackendId > & | availablePreferredBackends, | ||
| std::string & | reasonIfUnsupported, | ||
| Optional< std::vector< std::string > & > | messages ) |
Definition at line 844 of file Network.cpp.
References Constant, FloatingPointConverter::ConvertFloat16To32(), ConvertFp16ToFp32, ConvertFp32ToFp16, Float16, Float32, BackendId::Get(), Layer::GetBackendId(), GetCapability(), GetDataTypeName(), Layer::GetInputSlots(), GetLayerTypeAsCString(), Layer::GetOutputSlot(), Layer::GetType(), HasCapability(), info, InsertConvertFp16ToFp32LayersBefore(), InsertConvertFp32ToFp16LayersAfter(), IWorkloadFactory::IsLayerSupported(), ConstantLayer::m_LayerOutput, OptimizationResult::m_Warning, PolymorphicDowncast(), ReportWarning(), ReturnWithError(), Layer::SetBackendId(), TensorInfo::SetDataType(), and OutputSlot::SetTensorInfo().
Referenced by AssignBackendsIConnectable().
| BackendRegistry & BackendRegistryInstance | ( | ) |
Definition at line 15 of file BackendRegistry.cpp.
Referenced by CreateBackendObject(), CreateSupportedBackends(), DynamicBackendUtils::DeregisterDynamicBackends(), GetCapability(), GetILayerSupportByBackendId(), GetNumberOfCacheFiles(), HasMatchingCapability(), ArmNNProfilingServiceInitialiser::InitialiseProfilingService(), IOptimizedNetwork::Optimize, DynamicBackendUtils::RegisterDynamicBackends(), RuntimeImpl::RuntimeImpl(), and RuntimeImpl::~RuntimeImpl().
| void BatchNormImpl | ( | const BatchNormalizationQueueDescriptor & | data, |
| Decoder< float > & | meanDecoder, | ||
| Decoder< float > & | varianceDecoder, | ||
| Decoder< float > & | betaDecoder, | ||
| Decoder< float > & | gammaDecoder, | ||
| Decoder< float > & | inputDecoder, | ||
| Encoder< float > & | outputEncoder ) |
Definition at line 18 of file BatchNormImpl.cpp.
References Decoder< IType >::Get(), DataLayoutIndexed::GetChannelsIndex(), DataLayoutIndexed::GetHeightIndex(), DataLayoutIndexed::GetIndex(), TensorInfo::GetShape(), armnnUtils::GetTensorInfo(), DataLayoutIndexed::GetWidthIndex(), BatchNormalizationDescriptor::m_DataLayout, BatchNormalizationDescriptor::m_Eps, QueueDescriptor::m_Inputs, QueueDescriptorWithParameters< LayerDescriptor >::m_Parameters, and Encoder< IType >::Set().
| void BatchToSpaceNd | ( | const TensorInfo & | inputInfo, |
| const TensorInfo & | outputInfo, | ||
| const BatchToSpaceNdDescriptor & | params, | ||
| Decoder< float > & | inputData, | ||
| Encoder< float > & | outputData ) |
Definition at line 50 of file BatchToSpaceNd.cpp.
References CHECK_LOCATION, Decoder< IType >::Get(), DataLayoutIndexed::GetChannelsIndex(), DataLayoutIndexed::GetHeightIndex(), TensorInfo::GetNumDimensions(), TensorInfo::GetShape(), DataLayoutIndexed::GetWidthIndex(), BatchToSpaceNdDescriptor::m_BlockShape, BatchToSpaceNdDescriptor::m_Crops, BatchToSpaceNdDescriptor::m_DataLayout, NCHW, Offset(), and Encoder< IType >::Set().
Referenced by BatchToSpaceNdLayer::BatchToSpaceNdLayer().
| void BuildAddMulAddSlotLists | ( | bool | handleReLu, |
| bool | multipleOutputs, | ||
| std::vector< SlotListType > & | inputLayersSlotLists, | ||
| std::vector< SlotListType > & | outputLayersSlotLists ) |
Definition at line 36 of file NeonBackendOptimizationUtils.hpp.
Referenced by NeonBackend::OptimizeSubgraphView().
| bool BuildAddMulAddTensorInfoLists | ( | Type * | layerList[4], |
| unsigned int & | numInputs, | ||
| unsigned int & | numOutputs, | ||
| std::vector< TensorInfo > & | inputInfos, | ||
| std::vector< TensorInfo > & | outputInfos, | ||
| const ActivationDescriptor *& | activationDescriptor, | ||
| bool & | fuseReLu ) |
Definition at line 87 of file NeonBackendOptimizationUtils.hpp.
References Add, ARMNN_THROW_INVALIDARG_IF_FALSE, BoundedReLu, CollapseLeadingUnitDimensions(), Constant, InputSlot::GetConnectedOutputSlot(), OutputSlot::GetConnection(), Layer::GetInputSlot(), OutputSlot::GetNumConnections(), TensorInfo::GetNumDimensions(), Layer::GetOutputSlot(), InputSlot::GetOwningLayer(), OutputSlot::GetOwningLayer(), InputSlot::GetTensorInfo(), OutputSlot::GetTensorInfo(), Layer::GetType(), IsSequenceLayerType(), ActivationDescriptor::m_Function, ConstantLayer::m_LayerOutput, Mul, PolymorphicDowncast(), ReLu, and OutputSlot::SetTensorInfo().
Referenced by NeonBackend::OptimizeSubgraphView().
| int CalcLevel | ( | const Event * | eventPtr | ) |
Definition at line 257 of file Profiling.cpp.
References Event::GetParentEvent().
Referenced by ProfilerImpl::AnalyzeEventsAndWriteResults(), and ProfilerImpl::PopulateParent().
| EdgeStrategy CalculateEdgeStrategy | ( | BackendsMap & | backends, |
| ITensorHandleFactory::FactoryId | srcFactoryId, | ||
| const Layer & | layer, | ||
| const Layer & | connectedLayer, | ||
| TensorHandleFactoryRegistry & | registry, | ||
| bool | importEnabled ) |
Definition at line 1785 of file Network.cpp.
References CopyToTarget, DirectCompatibility, ExportToTarget, FallbackImportDisabled, Layer::GetBackendId(), ITensorHandleFactory::GetCapabilities(), ITensorHandleFactory::GetExportFlags(), TensorHandleFactoryRegistry::GetFactory(), ITensorHandleFactory::GetImportFlags(), Layer::GetType(), ITensorHandleFactory::LegacyFactoryId, Output, PaddingRequired, ITensorHandleFactory::SupportsMapUnmap(), and Undefined.
Referenced by SelectTensorHandleStrategy().
| std::map< std::string, unsigned int > CalculateGatherNdKeyIndices | ( | TensorInfo | inputInfo0, |
| TensorInfo | inputInfo1 ) |
Calculates the key index values needed for GatherNd: N, ND, K, W, C (N is always 1)
| inputInfo0 | - TensorInfo of the corresponding input tensor: params |
| inputInfo1 | - TensorInfo of the corresponding input tensor: indices |
Definition at line 313 of file WorkloadUtils.cpp.
References TensorInfo::GetNumDimensions(), and TensorInfo::GetShape().
Referenced by ClGatherNdWorkload::ClGatherNdWorkload(), ClGatherNdWorkloadValidate(), NeonGatherNdWorkload::NeonGatherNdWorkload(), and NeonGatherNdWorkloadValidate().
| ITensorHandleFactory::FactoryId CalculateSlotOption | ( | BackendsMap & | backends, |
| OutputSlot & | outputSlot, | ||
| TensorHandleFactoryRegistry & | registry, | ||
| bool | exportEnabled ) |
Definition at line 1632 of file Network.cpp.
References FallbackImportDisabled, Layer::GetBackendId(), ITensorHandleFactory::GetCapabilities(), OutputSlot::GetConnections(), ITensorHandleFactory::GetExportFlags(), TensorHandleFactoryRegistry::GetFactory(), IBackendInternal::GetHandleFactoryPreferences(), Layer::GetInputSlots(), OutputSlot::GetOwningLayer(), Layer::GetType(), ITensorHandleFactory::LegacyFactoryId, Output, RequiresCopy(), and ITensorHandleFactory::SupportsMapUnmap().
Referenced by SelectTensorHandleStrategy().
| ITensorHandleFactory::FactoryId CalculateSlotOptionForInput | ( | BackendsMap & | backends, |
| OutputSlot & | slot, | ||
| TensorHandleFactoryRegistry & | registry, | ||
| bool | importEnabled ) |
Definition at line 1530 of file Network.cpp.
References Layer::GetBackendId(), OutputSlot::GetConnections(), TensorHandleFactoryRegistry::GetFactory(), ITensorHandleFactory::GetImportFlags(), OutputSlot::GetOwningLayer(), Layer::GetType(), Input, ITensorHandleFactory::LegacyFactoryId, and ITensorHandleFactory::SupportsMapUnmap().
Referenced by SelectTensorHandleStrategy().
| ITensorHandleFactory::FactoryId CalculateSlotOptionForOutput | ( | BackendsMap & | backends, |
| OutputSlot & | slot, | ||
| TensorHandleFactoryRegistry & | registry ) |
Definition at line 1622 of file Network.cpp.
References ITensorHandleFactory::DeferredFactoryId, and IgnoreUnused().
Referenced by SelectTensorHandleStrategy().
| std::vector< IConnectableLayer * > ChainReduceLayers | ( | OptimizationViews & | optimizationViews, |
| LayerType * | baseLayer, | ||
| ReduceDescriptor & | desc ) |
Definition at line 278 of file ArmComputeSubgraphUtils.hpp.
References INetwork::AddReduceLayer(), ComputeReductionTensorShape(), OptimizationViews::GetINetwork(), Layer::GetInputSlot(), Layer::GetOutputSlot(), ReduceDescriptor::m_KeepDims, ReduceDescriptor::m_vAxis, PolymorphicDowncast(), and OutputSlot::SetTensorInfo().
Referenced by ClBackend::OptimizeSubgraphView(), and NeonBackend::OptimizeSubgraphView().
| bool CheckFastMathSupport | ( | const std::vector< BackendId > & | availablePreferredBackends, |
| const ModelOptions & | modelOptions ) |
Definition at line 1948 of file Network.cpp.
References ARMNN_LOG, BackendOptions::Var::AsBool(), debug, ParseOptions(), and warning.
Referenced by IOptimizedNetwork::Optimize.
|
inline |
Definition at line 41 of file MemorySources.hpp.
| bool CheckFp16Support | ( | BackendsMap & | backends, |
| const std::vector< BackendId > & | availablePreferredBackends ) |
Definition at line 1045 of file Network.cpp.
References ARMNN_LOG, debug, and HasMatchingCapability().
Referenced by IOptimizedNetwork::Optimize.
| bool CheckScaleSetOnQuantizedType | ( | Layer * | layer, |
| Optional< std::vector< std::string > & > | errMessages ) |
Definition at line 801 of file Network.cpp.
References ARMNN_LOG, Layer::GetNumOutputSlots(), Layer::GetOutputSlot(), OutputSlot::GetTensorInfo(), Layer::GetType(), info, QAsymmS8, QAsymmU8, QSymmS16, QSymmS8, ReportError(), OutputSlot::SetTensorInfo(), Softmax, and warning.
Referenced by AssignBackendsIConnectable().
| bool CheckSupportRule | ( | F | rule, |
| Optional< std::string & > | reasonIfUnsupported, | ||
| const char * | reason ) |
Definition at line 37 of file LayerSupportRules.hpp.
References OptionalReferenceSwitch< IsReference, T >::value().
Referenced by RefLayerSupport::IsActivationSupported(), RefLayerSupport::IsAdditionSupported(), RefLayerSupport::IsArgMinMaxSupported(), RefLayerSupport::IsBatchMatMulSupported(), RefLayerSupport::IsBatchNormalizationSupported(), RefLayerSupport::IsBatchToSpaceNdSupported(), RefLayerSupport::IsBroadcastToSupported(), RefLayerSupport::IsCastSupported(), RefLayerSupport::IsChannelShuffleSupported(), RefLayerSupport::IsComparisonSupported(), RefLayerSupport::IsConcatSupported(), RefLayerSupport::IsConstantSupported(), RefLayerSupport::IsConvolution2dSupported(), RefLayerSupport::IsConvolution3dSupported(), RefLayerSupport::IsDebugSupported(), RefLayerSupport::IsDepthToSpaceSupported(), RefLayerSupport::IsDepthwiseConvolutionSupported(), RefLayerSupport::IsDequantizeSupported(), RefLayerSupport::IsDetectionPostProcessSupported(), RefLayerSupport::IsDivisionSupported(), RefLayerSupport::IsElementwiseUnarySupported(), RefLayerSupport::IsFakeQuantizationSupported(), RefLayerSupport::IsFillSupported(), RefLayerSupport::IsFloorSupported(), RefLayerSupport::IsFullyConnectedSupported(), RefLayerSupport::IsGatherNdSupported(), RefLayerSupport::IsGatherSupported(), RefLayerSupport::IsInstanceNormalizationSupported(), RefLayerSupport::IsL2NormalizationSupported(), RefLayerSupport::IsLayerSupported(), RefLayerSupport::IsLogicalBinarySupported(), RefLayerSupport::IsLogSoftmaxSupported(), RefLayerSupport::IsLstmSupported(), RefLayerSupport::IsMaximumSupported(), RefLayerSupport::IsMeanSupported(), RefLayerSupport::IsMemCopySupported(), RefLayerSupport::IsMinimumSupported(), RefLayerSupport::IsMultiplicationSupported(), RefLayerSupport::IsNormalizationSupported(), RefLayerSupport::IsPadSupported(), RefLayerSupport::IsPermuteSupported(), RefLayerSupport::IsPooling2dSupported(), RefLayerSupport::IsPooling3dSupported(), RefLayerSupport::IsPreluSupported(), RefLayerSupport::IsQuantizeSupported(), RefLayerSupport::IsRankSupported(), RefLayerSupport::IsReduceSupported(), RefLayerSupport::IsReshapeSupported(), RefLayerSupport::IsResizeSupported(), RefLayerSupport::IsReverseV2Supported(), RefLayerSupport::IsScatterNdSupported(), RefLayerSupport::IsShapeSupported(), RefLayerSupport::IsSliceSupported(), RefLayerSupport::IsSoftmaxSupported(), RefLayerSupport::IsSpaceToBatchNdSupported(), RefLayerSupport::IsSpaceToDepthSupported(), RefLayerSupport::IsSplitterSupported(), RefLayerSupport::IsStackSupported(), RefLayerSupport::IsStridedSliceSupported(), RefLayerSupport::IsSubtractionSupported(), RefLayerSupport::IsTileSupported(), RefLayerSupport::IsTransposeConvolution2dSupported(), RefLayerSupport::IsTransposeSupported(), and RefLayerSupport::IsUnidirectionalSequenceLstmSupported().
| arm_compute::Status ClAbsWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output ) |
Definition at line 19 of file ClAbsWorkload.cpp.
Referenced by ClLayerSupport::IsElementwiseUnarySupported().
| arm_compute::Status ClActivationWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const ActivationDescriptor & | descriptor ) |
Definition at line 17 of file ClActivationWorkload.cpp.
References ConvertActivationDescriptorToAclActivationLayerInfo().
Referenced by ClLayerSupport::IsActivationSupported().
| arm_compute::Status ClAdditionValidate | ( | const TensorInfo & | input0, |
| const TensorInfo & | input1, | ||
| const TensorInfo & | output, | ||
| const ActivationDescriptor * | activationDescriptor ) |
Definition at line 45 of file ClAdditionWorkload.cpp.
References ConvertActivationDescriptorToAclActivationLayerInfo().
Referenced by ClLayerSupport::IsAdditionSupported(), ClLayerSupport::IsLayerSupported(), and ClBackend::OptimizeSubgraphView().
| arm_compute::Status ClArgMinMaxWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const ArgMinMaxDescriptor & | descriptor ) |
Definition at line 31 of file ClArgMinMaxWorkload.cpp.
References TensorInfo::GetNumDimensions(), armnnUtils::GetUnsignedAxis(), ArgMinMaxDescriptor::m_Axis, ArgMinMaxDescriptor::m_Function, Max, and numeric_cast().
Referenced by ClLayerSupport::IsArgMinMaxSupported().
|
constexpr |
Definition at line 10 of file ClBackendId.hpp.
Referenced by ClBackend::GetIdStatic().
| arm_compute::Status ClBatchMatMulValidate | ( | const TensorInfo & | inputInfoX, |
| const TensorInfo & | inputInfoY, | ||
| const TensorInfo & | outputInfo, | ||
| const BatchMatMulDescriptor & | descriptor, | ||
| const ActivationDescriptor * | activationDescriptor ) |
Definition at line 24 of file ClBatchMatMulWorkload.cpp.
References ConvertActivationDescriptorToAclActivationLayerInfo(), BatchMatMulDescriptor::m_AdjointX, BatchMatMulDescriptor::m_AdjointY, BatchMatMulDescriptor::m_DataLayoutX, BatchMatMulDescriptor::m_DataLayoutY, BatchMatMulDescriptor::m_TransposeX, BatchMatMulDescriptor::m_TransposeY, and NCHW.
Referenced by ClLayerSupport::IsBatchMatMulSupported().
| arm_compute::Status ClBatchNormalizationValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const TensorInfo & | mean, | ||
| const TensorInfo & | var, | ||
| const TensorInfo & | beta, | ||
| const TensorInfo & | gamma, | ||
| const BatchNormalizationDescriptor & | descriptor, | ||
| const ActivationDescriptor * | activationDescriptor ) |
Definition at line 19 of file ClBatchNormalizationFloatWorkload.cpp.
References ConvertActivationDescriptorToAclActivationLayerInfo(), BatchNormalizationDescriptor::m_DataLayout, and BatchNormalizationDescriptor::m_Eps.
Referenced by ClLayerSupport::IsBatchNormalizationSupported(), and ClBackend::OptimizeSubgraphView().
| arm_compute::Status ClBatchToSpaceNdWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const BatchToSpaceNdDescriptor & | descriptor ) |
Definition at line 17 of file ClBatchToSpaceNdWorkload.cpp.
References CHECK_LOCATION, TensorInfo::GetNumDimensions(), BatchToSpaceNdDescriptor::m_BlockShape, BatchToSpaceNdDescriptor::m_DataLayout, NCHW, NHWC, and numeric_cast().
Referenced by ClLayerSupport::IsBatchToSpaceNdSupported().
| arm_compute::Status ClCastValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output ) |
Definition at line 20 of file ClCastWorkload.cpp.
References Boolean, and TensorInfo::GetDataType().
Referenced by ClLayerSupport::IsCastSupported().
| arm_compute::Status ClChannelShuffleValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const ChannelShuffleDescriptor & | descriptor ) |
Definition at line 20 of file ClChannelShuffleWorkload.cpp.
References TensorInfo::GetNumDimensions(), ChannelShuffleDescriptor::m_Axis, ChannelShuffleDescriptor::m_NumGroups, NCHW, and NHWC.
Referenced by ClLayerSupport::IsChannelShuffleSupported().
| arm_compute::Status ClComparisonWorkloadValidate | ( | const TensorInfo & | input0, |
| const TensorInfo & | input1, | ||
| const TensorInfo & | output, | ||
| const ComparisonDescriptor & | descriptor ) |
Definition at line 24 of file ClComparisonWorkload.cpp.
References ConvertComparisonOperationToAcl().
Referenced by ClLayerSupport::IsComparisonSupported().
| arm_compute::Status ClConcatWorkloadValidate | ( | const std::vector< const TensorInfo * > & | inputs, |
| const TensorInfo & | output, | ||
| const OriginsDescriptor & | descriptor ) |
Definition at line 27 of file ClConcatWorkload.cpp.
References NCHW.
Referenced by ClLayerSupport::IsConcatSupported().
| arm_compute::Status ClConstantWorkloadValidate | ( | const TensorInfo & | output | ) |
Definition at line 18 of file ClConstantWorkload.cpp.
Referenced by ClLayerSupport::IsConstantSupported().
|
inline |
Definition at line 152 of file ClContextSchema_generated.h.
References ClContextIdentifier().
|
inline |
Definition at line 167 of file ClContextSchema_generated.h.
|
inline |
Definition at line 148 of file ClContextSchema_generated.h.
Referenced by ClContextBufferHasIdentifier(), FinishClContextBuffer(), FinishSizePrefixedClContextBuffer(), VerifyClContextBuffer(), and VerifySizePrefixedClContextBuffer().
| arm_compute::Status ClConvertFp16ToFp32WorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output ) |
Definition at line 44 of file ClConvertFp16ToFp32Workload.cpp.
References Float16, Float32, and TensorInfo::GetDataType().
Referenced by ClLayerSupport::IsConvertFp16ToFp32Supported().
| arm_compute::Status ClConvertFp32ToFp16WorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output ) |
Definition at line 44 of file ClConvertFp32ToFp16Workload.cpp.
References Float16, Float32, and TensorInfo::GetDataType().
Referenced by ClLayerSupport::IsConvertFp32ToFp16Supported().
| arm_compute::Status ClConvolution2dWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const Convolution2dDescriptor & | descriptor, | ||
| const TensorInfo & | weights, | ||
| const Optional< TensorInfo > & | biases, | ||
| bool | isFastMathEnabled, | ||
| const ActivationDescriptor * | activationDescriptor ) |
Definition at line 23 of file ClConvolution2dWorkload.cpp.
References ConvertActivationDescriptorToAclActivationLayerInfo(), OptionalBase::has_value(), TensorInfo::IsConstant(), Convolution2dDescriptor::m_BiasEnabled, Convolution2dDescriptor::m_DataLayout, Convolution2dDescriptor::m_DilationX, Convolution2dDescriptor::m_DilationY, and OptionalReferenceSwitch< IsReference, T >::value().
Referenced by ClLayerSupport::IsConvolution2dSupported(), and ClBackend::OptimizeSubgraphView().
| arm_compute::Status ClConvolution3dWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const Convolution3dDescriptor & | descriptor, | ||
| const TensorInfo & | weights, | ||
| const Optional< TensorInfo > & | biases, | ||
| bool | isFastMathEnabled, | ||
| const ActivationDescriptor * | activationDescriptor ) |
Definition at line 23 of file ClConvolution3dWorkload.cpp.
References ComputeConv3DInfo(), OptionalBase::has_value(), Convolution3dDescriptor::m_BiasEnabled, Convolution3dDescriptor::m_DataLayout, and OptionalReferenceSwitch< IsReference, T >::value().
Referenced by ClLayerSupport::IsConvolution3dSupported().
| arm_compute::Status ClDepthToSpaceWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const DepthToSpaceDescriptor & | descriptor ) |
Definition at line 22 of file ClDepthToSpaceWorkload.cpp.
References SpaceToDepthDescriptor::m_BlockSize, SpaceToDepthDescriptor::m_DataLayout, and numeric_cast().
Referenced by ClLayerSupport::IsDepthToSpaceSupported().
| arm_compute::Status ClDepthwiseConvolutionWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const DepthwiseConvolution2dDescriptor & | descriptor, | ||
| const TensorInfo & | weights, | ||
| const Optional< TensorInfo > & | biases, | ||
| const ActivationDescriptor * | activationDescriptor ) |
Definition at line 26 of file ClDepthwiseConvolutionWorkload.cpp.
References Convert1HWOTensorInfoToAcl(), ConvertActivationDescriptorToAclActivationLayerInfo(), OptionalBase::has_value(), TensorInfo::IsConstant(), DepthwiseConvolution2dDescriptor::m_BiasEnabled, DepthwiseConvolution2dDescriptor::m_DataLayout, DepthwiseConvolution2dDescriptor::m_DilationX, DepthwiseConvolution2dDescriptor::m_DilationY, and OptionalReferenceSwitch< IsReference, T >::value().
Referenced by ClLayerSupport::IsDepthwiseConvolutionSupported(), ClLayerSupport::IsDilatedDepthwiseConvolutionSupported(), and ClBackend::OptimizeSubgraphView().
| arm_compute::Status ClDequantizeWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output ) |
Definition at line 22 of file ClDequantizeWorkload.cpp.
Referenced by ClLayerSupport::IsDequantizeSupported().
| arm_compute::Status ClDivisionWorkloadValidate | ( | const TensorInfo & | input0, |
| const TensorInfo & | input1, | ||
| const TensorInfo & | output, | ||
| const ActivationDescriptor * | activationDescriptor ) |
Definition at line 18 of file ClDivisionWorkload.cpp.
References ConvertActivationDescriptorToAclActivationLayerInfo().
Referenced by ClLayerSupport::IsDivisionSupported(), ClLayerSupport::IsLayerSupported(), and ClBackend::OptimizeSubgraphView().
| arm_compute::Status ClElementwiseBinaryValidate | ( | const TensorInfo & | input0, |
| const TensorInfo & | input1, | ||
| const TensorInfo & | output, | ||
| const ElementwiseBinaryDescriptor & | descriptor, | ||
| const ActivationDescriptor * | activationDescriptor ) |
Definition at line 64 of file ClElementwiseBinaryWorkload.cpp.
References CHECK_LOCATION, ConvertActivationDescriptorToAclActivationLayerInfo(), ElementwiseBinaryDescriptor::m_Operation, Power, and SqDiff.
Referenced by ClLayerSupport::IsLayerSupported().
| arm_compute::Status ClExpWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output ) |
Definition at line 18 of file ClExpWorkload.cpp.
Referenced by ClLayerSupport::IsElementwiseUnarySupported().
| arm_compute::Status ClFloorDivWorkloadValidate | ( | const TensorInfo & | input0Info, |
| const TensorInfo & | input1Info, | ||
| const TensorInfo & | outputInfo, | ||
| const ActivationDescriptor * | activationDescriptor ) |
Definition at line 55 of file ClFloorDivWorkload.cpp.
References AreAllTensorsSigned32(), ConvertActivationDescriptorToAclActivationLayerInfo(), ConvertTensorToFloat32(), and IsValidationPassing().
Referenced by ClLayerSupport::IsLayerSupported().
| arm_compute::Status ClFloorWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output ) |
Definition at line 14 of file ClFloorFloatWorkload.cpp.
Referenced by ClLayerSupport::IsFloorSupported().
| arm_compute::Status ClFullyConnectedWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const TensorInfo & | weights, | ||
| const Optional< TensorInfo > & | biases, | ||
| const FullyConnectedDescriptor & | descriptor, | ||
| const ActivationDescriptor * | activationDescriptor ) |
Definition at line 19 of file ClFullyConnectedWorkload.cpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, ConvertFullyConnectedDescriptorToAclFullyConnectedLayerInfo(), OptionalBase::has_value(), TensorInfo::IsConstant(), FullyConnectedDescriptor::m_BiasEnabled, and OptionalReferenceSwitch< IsReference, T >::value().
Referenced by ClLayerSupport::IsFullyConnectedSupported(), and ClBackend::OptimizeSubgraphView().
| arm_compute::Status ClGatherNdWorkloadValidate | ( | const TensorInfo & | paramsInfo, |
| const TensorInfo & | indicesInfo, | ||
| const TensorInfo & | outputInfo ) |
Validate Mul
Validate ReduceSum
Validate Gather
Validate Reshape
Return OK if all the layers are valid
Definition at line 16 of file ClGatherNdWorkload.cpp.
References CalculateGatherNdKeyIndices(), ComputeAclAxis(), TensorInfo::GetNumDimensions(), and TensorInfo::SetShape().
Referenced by ClLayerSupport::IsGatherNdSupported().
| arm_compute::Status ClGatherWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | indices, | ||
| const TensorInfo & | output, | ||
| const GatherDescriptor & | descriptor ) |
Definition at line 15 of file ClGatherWorkload.cpp.
References ComputeAclAxis(), and GatherDescriptor::m_Axis.
Referenced by ClLayerSupport::IsGatherSupported().
|
constexpr |
Definition at line 14 of file ClImportTensorHandleFactory.hpp.
Referenced by ClImportTensorHandleFactory::GetIdStatic().
| arm_compute::Status ClInstanceNormalizationWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const InstanceNormalizationDescriptor & | descriptor ) |
Definition at line 18 of file ClInstanceNormalizationWorkload.cpp.
References InstanceNormalizationDescriptor::m_Beta, InstanceNormalizationDescriptor::m_DataLayout, InstanceNormalizationDescriptor::m_Eps, and InstanceNormalizationDescriptor::m_Gamma.
Referenced by ClLayerSupport::IsInstanceNormalizationSupported().
| arm_compute::Status ClL2NormalizationWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const L2NormalizationDescriptor & | descriptor ) |
Definition at line 17 of file ClL2NormalizationFloatWorkload.cpp.
References L2NormalizationDescriptor::m_DataLayout, L2NormalizationDescriptor::m_Eps, and NCHW.
Referenced by ClLayerSupport::IsL2NormalizationSupported().
| arm_compute::Status ClLogicalAndWorkloadValidate | ( | const TensorInfo & | input0, |
| const TensorInfo & | input1, | ||
| const TensorInfo & | output ) |
Definition at line 20 of file ClLogicalAndWorkload.cpp.
Referenced by ClLayerSupport::IsLogicalBinarySupported().
| arm_compute::Status ClLogicalNotWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output ) |
Definition at line 20 of file ClLogicalNotWorkload.cpp.
Referenced by ClLayerSupport::IsElementwiseUnarySupported().
| arm_compute::Status ClLogicalOrWorkloadValidate | ( | const TensorInfo & | input0, |
| const TensorInfo & | input1, | ||
| const TensorInfo & | output ) |
Definition at line 20 of file ClLogicalOrWorkload.cpp.
Referenced by ClLayerSupport::IsLogicalBinarySupported().
| arm_compute::Status ClLogSoftmaxWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const LogSoftmaxDescriptor & | descriptor ) |
Definition at line 17 of file ClLogSoftmaxWorkload.cpp.
References ComputeAclAxis(), SoftmaxDescriptor::m_Axis, and SoftmaxDescriptor::m_Beta.
Referenced by ClLayerSupport::IsLogSoftmaxSupported().
| arm_compute::Status ClLogWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output ) |
Definition at line 18 of file ClLogWorkload.cpp.
Referenced by ClLayerSupport::IsElementwiseUnarySupported().
| arm_compute::Status ClLstmFloatWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | outputStateIn, | ||
| const TensorInfo & | cellStateIn, | ||
| const TensorInfo & | scratchBuffer, | ||
| const TensorInfo & | outputStateOut, | ||
| const TensorInfo & | cellStateOut, | ||
| const TensorInfo & | output, | ||
| const LstmDescriptor & | descriptor, | ||
| const LstmInputParamsInfo & | paramsInfo ) |
Definition at line 244 of file ClLstmFloatWorkload.cpp.
References ConvertLstmActivationFuncToAclLayerInfo(), LstmInputParamsInfo::GetCellBias(), LstmInputParamsInfo::GetCellLayerNormWeights(), LstmInputParamsInfo::GetCellToForgetWeights(), LstmInputParamsInfo::GetCellToInputWeights(), LstmInputParamsInfo::GetCellToOutputWeights(), LstmInputParamsInfo::GetForgetGateBias(), LstmInputParamsInfo::GetForgetLayerNormWeights(), LstmInputParamsInfo::GetInputGateBias(), LstmInputParamsInfo::GetInputLayerNormWeights(), LstmInputParamsInfo::GetInputToCellWeights(), LstmInputParamsInfo::GetInputToForgetWeights(), LstmInputParamsInfo::GetInputToInputWeights(), LstmInputParamsInfo::GetInputToOutputWeights(), LstmInputParamsInfo::GetOutputGateBias(), LstmInputParamsInfo::GetOutputLayerNormWeights(), LstmInputParamsInfo::GetProjectionBias(), LstmInputParamsInfo::GetProjectionWeights(), LstmInputParamsInfo::GetRecurrentToCellWeights(), LstmInputParamsInfo::GetRecurrentToForgetWeights(), LstmInputParamsInfo::GetRecurrentToInputWeights(), LstmInputParamsInfo::GetRecurrentToOutputWeights(), LstmDescriptor::m_ActivationFunc, LstmInputParamsInfo::m_CellToInputWeights, LstmDescriptor::m_CifgEnabled, LstmDescriptor::m_ClippingThresCell, LstmDescriptor::m_ClippingThresProj, LstmDescriptor::m_LayerNormEnabled, LstmDescriptor::m_PeepholeEnabled, LstmInputParamsInfo::m_ProjectionBias, and LstmDescriptor::m_ProjectionEnabled.
Referenced by ClLayerSupport::IsLstmSupported().
| arm_compute::Status ClMaximumWorkloadValidate | ( | const TensorInfo & | input0, |
| const TensorInfo & | input1, | ||
| const TensorInfo & | output ) |
Definition at line 24 of file ClMaximumWorkload.cpp.
Referenced by ClLayerSupport::IsLayerSupported(), and ClLayerSupport::IsMaximumSupported().
| arm_compute::Status ClMeanValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const MeanDescriptor & | descriptor ) |
Definition at line 17 of file ClMeanWorkload.cpp.
References TensorInfo::GetNumDimensions(), MeanDescriptor::m_Axis, and MeanDescriptor::m_KeepDims.
Referenced by ClLayerSupport::IsMeanSupported().
| arm_compute::Status ClMinimumWorkloadValidate | ( | const TensorInfo & | input0, |
| const TensorInfo & | input1, | ||
| const TensorInfo & | output ) |
Definition at line 24 of file ClMinimumWorkload.cpp.
Referenced by ClLayerSupport::IsLayerSupported(), and ClLayerSupport::IsMinimumSupported().
| arm_compute::Status ClMultiplicationWorkloadValidate | ( | const TensorInfo & | input0, |
| const TensorInfo & | input1, | ||
| const TensorInfo & | output, | ||
| const ActivationDescriptor * | activationDescriptor ) |
Definition at line 18 of file ClMultiplicationWorkload.cpp.
References ConvertActivationDescriptorToAclActivationLayerInfo(), TensorInfo::GetDataType(), and IsQuantizedType().
Referenced by ClLayerSupport::IsLayerSupported(), ClLayerSupport::IsMultiplicationSupported(), and ClBackend::OptimizeSubgraphView().
| arm_compute::Status ClNegWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output ) |
Definition at line 18 of file ClNegWorkload.cpp.
Referenced by ClLayerSupport::IsElementwiseUnarySupported().
| arm_compute::Status ClNormalizationWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const NormalizationDescriptor & | descriptor ) |
Definition at line 19 of file ClNormalizationFloatWorkload.cpp.
References NormalizationDescriptor::m_DataLayout.
Referenced by ClLayerSupport::IsNormalizationSupported().
| arm_compute::Status ClPadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const PadDescriptor & | descriptor ) |
Definition at line 62 of file ClPadWorkload.cpp.
References ConvertPaddingModeToAcl(), PadDescriptor::m_PaddingMode, PadDescriptor::m_PadList, and PadDescriptor::m_PadValue.
Referenced by ClLayerSupport::IsPadSupported().
| arm_compute::Status ClPermuteWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const PermuteDescriptor & | descriptor ) |
Definition at line 17 of file ClPermuteWorkload.cpp.
References PermuteDescriptor::m_DimMappings.
Referenced by ClLayerSupport::IsPermuteSupported().
| arm_compute::Status ClPooling2dWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const Pooling2dDescriptor & | descriptor ) |
Definition at line 18 of file ClPooling2dWorkload.cpp.
References Pooling2dDescriptor::m_DataLayout.
Referenced by ClLayerSupport::IsPooling2dSupported().
| arm_compute::Status ClPooling3dWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const Pooling3dDescriptor & | descriptor ) |
Definition at line 18 of file ClPooling3dWorkload.cpp.
References Pooling3dDescriptor::m_DataLayout.
Referenced by ClLayerSupport::IsPooling3dSupported().
| arm_compute::Status ClPreluWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | alpha, | ||
| const TensorInfo & | output ) |
Definition at line 16 of file ClPreluWorkload.cpp.
Referenced by ClLayerSupport::IsPreluSupported().
| arm_compute::Status ClQLstmWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | cellStateIn, | ||
| const TensorInfo & | outputStateIn, | ||
| const TensorInfo & | cellStateOut, | ||
| const TensorInfo & | outputStateOut, | ||
| const TensorInfo & | output, | ||
| const QLstmDescriptor & | descriptor, | ||
| const LstmInputParamsInfo & | paramsInfo ) |
Definition at line 247 of file ClQLstmWorkload.cpp.
References LstmInputParamsInfo::GetCellBias(), LstmInputParamsInfo::GetCellLayerNormWeights(), LstmInputParamsInfo::GetCellToForgetWeights(), LstmInputParamsInfo::GetCellToInputWeights(), LstmInputParamsInfo::GetCellToOutputWeights(), LstmInputParamsInfo::GetForgetGateBias(), LstmInputParamsInfo::GetForgetLayerNormWeights(), LstmInputParamsInfo::GetInputGateBias(), LstmInputParamsInfo::GetInputLayerNormWeights(), LstmInputParamsInfo::GetInputToCellWeights(), LstmInputParamsInfo::GetInputToForgetWeights(), LstmInputParamsInfo::GetInputToInputWeights(), LstmInputParamsInfo::GetInputToOutputWeights(), LstmInputParamsInfo::GetOutputGateBias(), LstmInputParamsInfo::GetOutputLayerNormWeights(), LstmInputParamsInfo::GetProjectionBias(), LstmInputParamsInfo::GetProjectionWeights(), LstmInputParamsInfo::GetRecurrentToCellWeights(), LstmInputParamsInfo::GetRecurrentToForgetWeights(), LstmInputParamsInfo::GetRecurrentToInputWeights(), LstmInputParamsInfo::GetRecurrentToOutputWeights(), QLstmDescriptor::m_CellClip, QLstmDescriptor::m_CellIntermediateScale, LstmInputParamsInfo::m_CellToInputWeights, QLstmDescriptor::m_CifgEnabled, QLstmDescriptor::m_ForgetIntermediateScale, QLstmDescriptor::m_HiddenStateScale, QLstmDescriptor::m_HiddenStateZeroPoint, QLstmDescriptor::m_InputIntermediateScale, LstmInputParamsInfo::m_InputLayerNormWeights, QLstmDescriptor::m_LayerNormEnabled, QLstmDescriptor::m_OutputIntermediateScale, QLstmDescriptor::m_PeepholeEnabled, LstmInputParamsInfo::m_ProjectionBias, QLstmDescriptor::m_ProjectionClip, and QLstmDescriptor::m_ProjectionEnabled.
Referenced by ClLayerSupport::IsQLstmSupported().
| arm_compute::Status ClQuantizedLstmWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | previousCellStateIn, | ||
| const TensorInfo & | previousOutputIn, | ||
| const TensorInfo & | cellStateOut, | ||
| const TensorInfo & | output, | ||
| const QuantizedLstmInputParamsInfo & | paramsInfo ) |
Definition at line 18 of file ClQuantizedLstmWorkload.cpp.
References QuantizedLstmInputParamsInfo::GetCellBias(), QuantizedLstmInputParamsInfo::GetForgetGateBias(), QuantizedLstmInputParamsInfo::GetInputGateBias(), QuantizedLstmInputParamsInfo::GetInputToCellWeights(), QuantizedLstmInputParamsInfo::GetInputToForgetWeights(), QuantizedLstmInputParamsInfo::GetInputToInputWeights(), QuantizedLstmInputParamsInfo::GetInputToOutputWeights(), QuantizedLstmInputParamsInfo::GetOutputGateBias(), QuantizedLstmInputParamsInfo::GetRecurrentToCellWeights(), QuantizedLstmInputParamsInfo::GetRecurrentToForgetWeights(), QuantizedLstmInputParamsInfo::GetRecurrentToInputWeights(), and QuantizedLstmInputParamsInfo::GetRecurrentToOutputWeights().
Referenced by ClLayerSupport::IsQuantizedLstmSupported().
| arm_compute::Status ClQuantizeWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output ) |
Definition at line 22 of file ClQuantizeWorkload.cpp.
Referenced by ClLayerSupport::IsQuantizeSupported().
| arm_compute::Status ClReduceWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const ReduceDescriptor & | descriptor ) |
Definition at line 18 of file ClReduceWorkload.cpp.
References ClReduceWorkloadValidate(), ConvertReductionOperationToAcl(), TensorInfo::GetNumDimensions(), IsMultiAxesReduceSupported(), ReduceDescriptor::m_KeepDims, and ReduceDescriptor::m_vAxis.
Referenced by ClReduceWorkloadValidate(), and ClLayerSupport::IsReduceSupported().
| arm_compute::Status ClReshapeWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output ) |
Definition at line 15 of file ClReshapeWorkload.cpp.
Referenced by ClLayerSupport::IsReshapeSupported().
| arm_compute::Status ClResizeWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const ResizeDescriptor & | descriptor ) |
Definition at line 22 of file ClResizeWorkload.cpp.
References ConvertResizeMethodToAclInterpolationPolicy(), ResizeDescriptor::m_AlignCorners, ResizeDescriptor::m_DataLayout, ResizeDescriptor::m_HalfPixelCenters, and ResizeDescriptor::m_Method.
Referenced by ClLayerSupport::IsResizeSupported().
| arm_compute::Status ClReverseV2WorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | axis, | ||
| const TensorInfo & | output ) |
Definition at line 16 of file ClReverseV2Workload.cpp.
Referenced by ClLayerSupport::IsReverseV2Supported().
| arm_compute::Status ClRsqrtWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output ) |
Definition at line 18 of file ClRsqrtWorkload.cpp.
Referenced by ClLayerSupport::IsElementwiseUnarySupported().
| arm_compute::Status ClScatterNdWorkloadValidate | ( | const TensorInfo & | inputInfo, |
| const TensorInfo & | indicesInfo, | ||
| const TensorInfo & | updatesInfo, | ||
| const TensorInfo & | outputInfo, | ||
| const ScatterNdDescriptor & | descriptor ) |
Definition at line 20 of file ClScatterNdWorkload.cpp.
References ScatterNdDescriptor::m_InputEnabled.
Referenced by ClLayerSupport::IsScatterNdSupported().
| arm_compute::Status ClSinWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output ) |
Definition at line 18 of file ClSinWorkload.cpp.
Referenced by ClLayerSupport::IsElementwiseUnarySupported().
| arm_compute::Status ClSliceWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const SliceDescriptor & | descriptor ) |
Definition at line 18 of file ClSliceWorkload.cpp.
References SliceDescriptor::m_Begin, SliceDescriptor::m_Size, and SetClSliceData().
Referenced by ClLayerSupport::IsSliceSupported().
| arm_compute::Status ClSoftmaxWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const SoftmaxDescriptor & | descriptor ) |
Definition at line 17 of file ClSoftmaxWorkload.cpp.
References ComputeAclAxis(), SoftmaxDescriptor::m_Axis, and SoftmaxDescriptor::m_Beta.
Referenced by ClLayerSupport::IsSoftmaxSupported().
| arm_compute::Status ClSpaceToBatchNdWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const SpaceToBatchNdDescriptor & | descriptor ) |
Definition at line 16 of file ClSpaceToBatchNdWorkload.cpp.
References CHECK_LOCATION, TensorInfo::GetNumDimensions(), SpaceToBatchNdDescriptor::m_BlockShape, SpaceToBatchNdDescriptor::m_DataLayout, SpaceToBatchNdDescriptor::m_PadList, NCHW, NHWC, and numeric_cast().
Referenced by ClLayerSupport::IsSpaceToBatchNdSupported().
| arm_compute::Status ClSpaceToDepthWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const SpaceToDepthDescriptor & | descriptor ) |
Definition at line 54 of file ClSpaceToDepthWorkload.cpp.
References SpaceToDepthDescriptor::m_BlockSize, SpaceToDepthDescriptor::m_DataLayout, and numeric_cast().
Referenced by ClLayerSupport::IsSpaceToDepthSupported().
| arm_compute::Status ClSplitterWorkloadValidate | ( | const TensorInfo & | input, |
| const std::vector< std::reference_wrapper< TensorInfo > > & | outputs, | ||
| unsigned int | splitAxis ) |
Definition at line 32 of file ClSplitterWorkload.cpp.
References TensorInfo::GetNumDimensions().
Referenced by ClLayerSupport::IsSplitterSupported().
| arm_compute::Status ClSqrtWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output ) |
Definition at line 19 of file ClSqrtWorkload.cpp.
References ConvertActivationDescriptorToAclActivationLayerInfo(), ActivationDescriptor::m_Function, and Sqrt.
Referenced by ClLayerSupport::IsElementwiseUnarySupported().
| arm_compute::Status ClStackWorkloadValidate | ( | const std::vector< const TensorInfo * > & | inputs, |
| const TensorInfo & | output, | ||
| const StackDescriptor & | descriptor ) |
Definition at line 29 of file ClStackWorkload.cpp.
References TensorShape::GetNumDimensions(), StackDescriptor::m_Axis, and StackDescriptor::m_InputShape.
Referenced by ClLayerSupport::IsStackSupported().
| arm_compute::Status ClStridedSliceWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const StridedSliceDescriptor & | descriptor ) |
Definition at line 27 of file ClStridedSliceWorkload.cpp.
References ConvertMaskToACLFormat(), TensorInfo::GetNumDimensions(), StridedSliceDescriptor::m_Begin, StridedSliceDescriptor::m_BeginMask, StridedSliceDescriptor::m_DataLayout, StridedSliceDescriptor::m_End, StridedSliceDescriptor::m_EndMask, StridedSliceDescriptor::m_ShrinkAxisMask, StridedSliceDescriptor::m_Stride, numeric_cast(), and SetClStridedSliceData().
Referenced by ClLayerSupport::IsStridedSliceSupported().
| arm_compute::Status ClSubtractionValidate | ( | const TensorInfo & | input0, |
| const TensorInfo & | input1, | ||
| const TensorInfo & | output, | ||
| const ActivationDescriptor * | activationDescriptor ) |
Definition at line 46 of file ClSubtractionWorkload.cpp.
References ConvertActivationDescriptorToAclActivationLayerInfo().
Referenced by ClLayerSupport::IsLayerSupported(), ClLayerSupport::IsSubtractionSupported(), and ClBackend::OptimizeSubgraphView().
|
constexpr |
Definition at line 15 of file ClTensorHandleFactory.hpp.
Referenced by ClTensorHandleFactory::GetIdStatic().
| arm_compute::Status ClTileWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const TileDescriptor & | descriptor ) |
Definition at line 16 of file ClTileWorkload.cpp.
References Boolean, TensorInfo::GetDataType(), and TileDescriptor::m_Multiples.
Referenced by ClLayerSupport::IsTileSupported().
| arm_compute::Status ClTransposeConvolution2dWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const TransposeConvolution2dDescriptor & | descriptor, | ||
| const TensorInfo & | weights, | ||
| const Optional< TensorInfo > & | biases ) |
Definition at line 26 of file ClTransposeConvolution2dWorkload.cpp.
References OptionalBase::has_value(), TransposeConvolution2dDescriptor::m_BiasEnabled, TransposeConvolution2dDescriptor::m_DataLayout, and OptionalReferenceSwitch< IsReference, T >::value().
Referenced by ClLayerSupport::IsTransposeConvolution2dSupported().
| arm_compute::Status ClTransposeWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const TransposeDescriptor & | descriptor ) |
Definition at line 17 of file ClTransposeWorkload.cpp.
References TransposeDescriptor::m_DimMappings.
Referenced by ClLayerSupport::IsTransposeSupported().
| arm_compute::Status ClUnidirectionalSequenceLstmFloatWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | outputStateIn, | ||
| const TensorInfo & | cellStateIn, | ||
| const TensorInfo & | outputStateOut, | ||
| const TensorInfo & | cellStateOut, | ||
| const TensorInfo & | output, | ||
| const UnidirectionalSequenceLstmDescriptor & | descriptor, | ||
| const LstmInputParamsInfo & | paramsInfo ) |
Definition at line 508 of file ClUnidirectionalSequenceLstmFloatWorkload.cpp.
References ConvertLstmActivationFuncToAclLayerInfo(), LstmInputParamsInfo::GetCellBias(), LstmInputParamsInfo::GetCellLayerNormWeights(), LstmInputParamsInfo::GetCellToForgetWeights(), LstmInputParamsInfo::GetCellToInputWeights(), LstmInputParamsInfo::GetCellToOutputWeights(), TensorInfo::GetDataType(), LstmInputParamsInfo::GetForgetGateBias(), LstmInputParamsInfo::GetForgetLayerNormWeights(), LstmInputParamsInfo::GetInputGateBias(), LstmInputParamsInfo::GetInputLayerNormWeights(), LstmInputParamsInfo::GetInputToCellWeights(), LstmInputParamsInfo::GetInputToForgetWeights(), LstmInputParamsInfo::GetInputToInputWeights(), LstmInputParamsInfo::GetInputToOutputWeights(), TensorShape::GetNumDimensions(), LstmInputParamsInfo::GetOutputGateBias(), LstmInputParamsInfo::GetOutputLayerNormWeights(), LstmInputParamsInfo::GetProjectionBias(), LstmInputParamsInfo::GetProjectionWeights(), LstmInputParamsInfo::GetRecurrentToCellWeights(), LstmInputParamsInfo::GetRecurrentToForgetWeights(), LstmInputParamsInfo::GetRecurrentToInputWeights(), LstmInputParamsInfo::GetRecurrentToOutputWeights(), TensorInfo::GetShape(), armnnUtils::GetTensorShape(), LstmDescriptor::m_ActivationFunc, LstmDescriptor::m_CifgEnabled, LstmDescriptor::m_ClippingThresCell, LstmDescriptor::m_ClippingThresProj, LstmDescriptor::m_LayerNormEnabled, LstmDescriptor::m_PeepholeEnabled, LstmInputParamsInfo::m_ProjectionBias, LstmDescriptor::m_ProjectionEnabled, LstmDescriptor::m_TimeMajor, and armnnUtils::Permuted().
Referenced by ClLayerSupport::IsUnidirectionalSequenceLstmSupported().
|
inline |
Definition at line 14 of file NeonBackendOptimizationUtils.hpp.
References TensorInfo::GetNumDimensions(), TensorInfo::GetShape(), and TensorInfo::SetShape().
Referenced by BuildAddMulAddTensorInfoLists().
| MemorySourceFlags Combine | ( | Arg | source, |
| Args... | rest ) |
| MemorySourceFlags Combine | ( | Arg | sourceA, |
| Arg | sourceB ) |
|
inline |
Function to convert ArmNN axis (left to right) to ACL axis (right to left) ranging from [-rank, rank)
Definition at line 246 of file ArmComputeUtils.hpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, and TensorInfo::GetNumDimensions().
Referenced by ClGatherNdWorkload::ClGatherNdWorkload(), ClGatherNdWorkloadValidate(), ClGatherWorkload::ClGatherWorkload(), ClGatherWorkloadValidate(), ClLogSoftmaxWorkload::ClLogSoftmaxWorkload(), ClLogSoftmaxWorkloadValidate(), ClSoftmaxWorkload::ClSoftmaxWorkload(), ClSoftmaxWorkloadValidate(), NeonGatherNdWorkload::NeonGatherNdWorkload(), NeonGatherNdWorkloadValidate(), NeonGatherWorkload::NeonGatherWorkload(), NeonGatherWorkloadValidate(), NeonLogSoftmaxWorkload::NeonLogSoftmaxWorkload(), NeonLogSoftmaxWorkloadValidate(), NeonSoftmaxWorkload::NeonSoftmaxWorkload(), and NeonSoftmaxWorkloadValidate().
|
inline |
Utility function used to setup an arm_compute::Conv3dInfo object from convolution3d descriptor.
Definition at line 261 of file ArmComputeUtils.hpp.
References ConvertActivationDescriptorToAclActivationLayerInfo(), Convolution3dDescriptor::m_DilationX, Convolution3dDescriptor::m_DilationY, Convolution3dDescriptor::m_DilationZ, Convolution3dDescriptor::m_PadBack, Convolution3dDescriptor::m_PadBottom, Convolution3dDescriptor::m_PadFront, Convolution3dDescriptor::m_PadLeft, Convolution3dDescriptor::m_PadRight, Convolution3dDescriptor::m_PadTop, Convolution3dDescriptor::m_StrideX, Convolution3dDescriptor::m_StrideY, and Convolution3dDescriptor::m_StrideZ.
Referenced by ClConvolution3dWorkload::ClConvolution3dWorkload(), ClConvolution3dWorkloadValidate(), NeonConvolution3dWorkload::NeonConvolution3dWorkload(), and NeonConvolution3dWorkloadValidate().
|
inline |
Definition at line 278 of file ArmComputeUtils.hpp.
References ConvertAdditionalInfoToAclActivationLayerInfo(), and QueueDescriptorWithParameters< LayerDescriptor >::m_Parameters.
|
inline |
Function to compute the output tensor shape based on the axes and if keepDims is set.
Definition at line 320 of file ArmComputeUtils.hpp.
References TensorInfo::GetNumDimensions(), and numeric_cast().
Referenced by ChainReduceLayers(), and IsMultiAxesReduceSupported().
|
inline |
Definition at line 227 of file ArmComputeUtils.hpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, TensorInfo::GetNumDimensions(), and SoftmaxDescriptor::m_Axis.
| std::set< unsigned int > ComputeSplitAxis | ( | const armnn::SplitterDescriptor & | desc, |
| const TensorShape & | input ) |
Calculates the axis values for split operation.
| desc | - Splitter Descriptor |
| input | - Input tensor shape |
Definition at line 377 of file WorkloadUtils.cpp.
References ViewsDescriptor::GetAxis(), ViewsDescriptor::GetNumDimensions(), ViewsDescriptor::GetNumViews(), armnnUtils::GetUnsignedAxis(), ViewsDescriptor::GetViewSizes(), and ViewsDescriptor::HasAxis().
Referenced by ClSplitterWorkload::ClSplitterWorkload(), ClUnidirectionalSequenceLstmFloatWorkload::ClUnidirectionalSequenceLstmFloatWorkload(), ConvertSplitToTosaOperator(), ClLayerSupport::IsSplitterSupported(), NeonLayerSupport::IsSplitterSupported(), NeonSplitterWorkload::NeonSplitterWorkload(), NeonUnidirectionalSequenceLstmFloatWorkload::NeonUnidirectionalSequenceLstmFloatWorkload(), and NeonUnidirectionalSequenceLstmWorkload::NeonUnidirectionalSequenceLstmWorkload().
| void Concatenate | ( | const ConcatQueueDescriptor & | data, |
| std::vector< ITensorHandle * > | inputs, | ||
| std::vector< ITensorHandle * > | outputs ) |
Definition at line 14 of file Concatenate.cpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, Decoder< IType >::Get(), TensorInfo::GetNumDimensions(), TensorInfo::GetNumElements(), TensorInfo::GetShape(), armnnUtils::GetTensorInfo(), ConcatQueueDescriptor::ViewOrigin::m_Origin, ConcatQueueDescriptor::m_ViewOrigins, MakeDecoder(), MakeEncoder(), Map, MaxNumOfTensorDimensions, and Encoder< IType >::Set().
| void ConditionalThrow | ( | bool | condition | ) |
Definition at line 174 of file Exceptions.hpp.
| void ConditionalThrow | ( | bool | condition, |
| const std::string & | message ) |
Definition at line 165 of file Exceptions.hpp.
Referenced by armnnUtils::Filesystem::CreateDirectory(), armnnUtils::Filesystem::ReadFileContentsIntoString(), ConstantLayer::ValidateTensorShapesFromInputs(), InputLayer::ValidateTensorShapesFromInputs(), OutputLayer::ValidateTensorShapesFromInputs(), and Layer::VerifyShapeInferenceType().
| void ConditionalThrowIfNotEqual | ( | const std::string & | message, |
| const ComparedType & | leftHandSide, | ||
| const ComparedType & | rightHandSide ) |
ComparedType must support: operator==(const ComparedType&) operator<<(ostream&, const ComparedType&)
Definition at line 189 of file Exceptions.hpp.
Referenced by ConcatLayer::InferOutputShapes(), MergeLayer::InferOutputShapes(), Layer::ValidateAndCopyShape(), ConcatLayer::ValidateTensorShapesFromInputs(), ConstantLayer::ValidateTensorShapesFromInputs(), FillLayer::ValidateTensorShapesFromInputs(), and StackLayer::ValidateTensorShapesFromInputs().
| void ConfigureDetailsObject | ( | JsonChildObject & | detailsObject, |
| std::string | layerDetailsStr ) |
Definition at line 306 of file Profiling.cpp.
References ExecObjectDesc, JsonChildObject::SetAndParseDetails(), and JsonChildObject::SetType().
| void ConfigureLogging | ( | bool | printToStandardOutput, |
| bool | printToDebugOutput, | ||
| LogSeverity | severity ) |
Configures the logging behaviour of the ARMNN library.
printToStandardOutput: Set to true if log messages should be printed to the standard output. printToDebugOutput: Set to true if log messages be printed to a platform-specific debug output (where supported). severity: All log messages that are at this severity level or higher will be printed, others will be ignored.
Definition at line 20 of file Utils.cpp.
References SetAllLoggingSinks(), and SetLogFilter().
Referenced by ConfigureLoggingTest().
|
inline |
Definition at line 44 of file ArmComputeTuningUtils.hpp.
References ARMNN_LOG, Exhaustive, info, None, Normal, and Rapid.
Referenced by ClBackendContext::ClBackendContext().
|
inline |
Checks the Layer's Connections to see if it's connected to a Layer with the provided layerType.
If dimSize is provided will also check if the connecting Tensor has more than that number of dimensions
Definition at line 313 of file SubgraphUtils.hpp.
References InputSlot::GetConnectedOutputSlot(), OutputSlot::GetConnection(), Layer::GetInputSlot(), TensorInfo::GetNumDimensions(), Layer::GetOutputSlot(), InputSlot::GetOwningLayer(), OutputSlot::GetOwningLayer(), InputSlot::GetTensorInfo(), and Layer::GetType().
|
inline |
Checks if the Layer is connected to any Layer that has an NCHW layout.
Definition at line 292 of file SubgraphUtils.hpp.
References InputSlot::GetConnectedOutputSlot(), OutputSlot::GetConnection(), Layer::GetInputSlot(), Layer::GetOutputSlot(), InputSlot::GetOwningLayer(), OutputSlot::GetOwningLayer(), and IsNCHW().
Referenced by ClBackend::OptimizeSubgraphView(), and NeonBackend::OptimizeSubgraphView().
| std::tuple< TensorInfo, unsigned int > Convert1HWOTensorInfoToAcl | ( | const TensorInfo & | weightInfo, |
| const TensorInfo & | inputInfo, | ||
| const DataLayout | dataLayout ) |
Weights for depthwise have a datalayout of [1,H,W,O] = [1,H,W,I*M] This function coverts a TensorInfo from [1,H,W,I*M] to [1,I*M,H,W] (if NCHW) or keeps it at [1,H,W,I*M] (if NHWC) as required by the compute library Returns a tuple of converted weights tensor info and depth multiplier.
Definition at line 177 of file WorkloadUtils.cpp.
References GetDataLayoutName(), TensorInfo::GetShape(), NCHW, NHWC, and armnnUtils::Permuted().
Referenced by ClDepthwiseConvolutionWorkloadValidate(), and NeonDepthwiseConvolutionWorkloadValidate().
| std::tuple< ConstTensor, unsigned int > Convert1HWOTensorToAcl | ( | const ConstTensorHandle * | weightTensor, |
| const TensorInfo & | inputInfo, | ||
| const DataLayout | dataLayout, | ||
| void * | permuteBuffer ) |
Weights for depthwise have a datalayout of [1,H,W,O] = [1,H,W,I*M] This function coverts a ConstCpuTensorHandle from [1,H,W,I*M] to [1,I*M,H,W] (if NCHW) or keeps it at [1,H,W,I*M] (if NHWC) as required by the compute library.
| weightTensor | - ConstTensorHandle of weights tensor |
| inputInfo | - TensorInfo of input tensor |
| dataLayout | - DataLayout of the input tensor |
| permuteBuffer | - Pointer to memory with the size of tensor. Used for the permutation |
Definition at line 146 of file WorkloadUtils.cpp.
References GetDataLayoutName(), TensorInfo::GetShape(), ConstTensorHandle::GetTensorInfo(), NCHW, NHWC, and PermuteTensor().
| std::tuple< ConstTensor, unsigned int > Convert1HWOtoMIHW | ( | const ConstTensorHandle * | weightTensor, |
| const TensorInfo & | inputInfo, | ||
| const DataLayout & | dataLayout, | ||
| void * | permuteBuffer ) |
Converts a (weights) tensor from [1, H, W, I*M] = [1, H, W, O] to [M, I, H, W].
| weightTensor | - ConstTensorHandle of the weight tensor that should be converted |
| inputInfo | - TensorInfo of the corresponding input tensor |
| dataLayout | - DataLayout of the input tensor e.g. NHWC or NCHW |
| permuteBuffer | - Memory location with the same size as the weight tensor to write converted data to |
Definition at line 208 of file WorkloadUtils.cpp.
References DataLayoutIndexed::GetChannelsIndex(), TensorInfo::GetShape(), ConstTensorHandle::GetTensorInfo(), TensorInfo::HasPerAxisQuantization(), PermuteTensor(), and TensorInfo::SetShape().
|
inline |
Definition at line 87 of file ArmComputeUtils.hpp.
References ConvertActivationFunctionToAclActivationFunction(), ActivationDescriptor::m_A, ActivationDescriptor::m_B, and ActivationDescriptor::m_Function.
Referenced by ClActivationWorkload::ClActivationWorkload(), ClActivationWorkloadValidate(), ClAdditionValidate(), ClBatchMatMulValidate(), ClBatchNormalizationValidate(), ClConvolution2dWorkloadValidate(), ClDepthwiseConvolutionWorkloadValidate(), ClDivisionWorkloadValidate(), ClElementwiseBinaryValidate(), ClFloorDivWorkloadValidate(), ClMultiplicationWorkloadValidate(), ClSqrtWorkload::ClSqrtWorkload(), ClSqrtWorkloadValidate(), ClSubtractionValidate(), ComputeConv3DInfo(), ConvertActivationDescriptorToAclActivationLayerInfo(), ConvertAdditionalInfoToAclActivationLayerInfo(), ConvertFullyConnectedDescriptorToAclFullyConnectedLayerInfo(), NeonActivationWorkload::NeonActivationWorkload(), NeonActivationWorkloadValidate(), NeonAdditionWorkloadValidate(), NeonBatchMatMulValidate(), NeonBatchNormalizationValidate(), NeonConvolution2dWorkloadValidate(), NeonDepthwiseConvolutionWorkloadValidate(), NeonDivisionWorkloadValidate(), NeonElementwiseBinaryWorkloadValidate(), NeonFloorDivWorkloadValidate(), NeonFusedWorkloadValidate(), NeonMultiplicationWorkloadValidate(), NeonSqrtWorkload::NeonSqrtWorkload(), NeonSqrtWorkloadValidate(), and NeonSubtractionWorkloadValidate().
|
inline |
Definition at line 94 of file ArmComputeUtils.hpp.
References ConvertActivationDescriptorToAclActivationLayerInfo().
|
inline |
Definition at line 62 of file ArmComputeUtils.hpp.
References Abs, BoundedReLu, Elu, Gelu, HardSwish, LeakyReLu, Linear, ReLu, Sigmoid, SoftReLu, Sqrt, Square, and TanH.
Referenced by ConvertActivationDescriptorToAclActivationLayerInfo().
|
inline |
Definition at line 105 of file ArmComputeUtils.hpp.
References ConvertActivationDescriptorToAclActivationLayerInfo(), and QueueDescriptor::GetAdditionalInformation().
Referenced by ClAdditionWorkload::ClAdditionWorkload(), ClBatchMatMulWorkload::ClBatchMatMulWorkload(), ClBatchNormalizationFloatWorkload::ClBatchNormalizationFloatWorkload(), ClConvolution2dWorkload::ClConvolution2dWorkload(), ClDepthwiseConvolutionWorkload::ClDepthwiseConvolutionWorkload(), ClDivisionWorkload::ClDivisionWorkload(), ClElementwiseBinaryWorkload::ClElementwiseBinaryWorkload(), ClFloorDivWorkload::ClFloorDivWorkload(), ClFullyConnectedWorkload::ClFullyConnectedWorkload(), ClMultiplicationWorkload::ClMultiplicationWorkload(), ClSubtractionWorkload::ClSubtractionWorkload(), ComputeConv3DInfo(), NeonAdditionWorkload::NeonAdditionWorkload(), NeonBatchMatMulWorkload::NeonBatchMatMulWorkload(), NeonBatchNormalizationWorkload::NeonBatchNormalizationWorkload(), NeonConvolution2dWorkload::NeonConvolution2dWorkload(), NeonDepthwiseConvolutionWorkload::NeonDepthwiseConvolutionWorkload(), NeonDivisionWorkload::NeonDivisionWorkload(), NeonElementwiseBinaryWorkload::NeonElementwiseBinaryWorkload(), NeonFloorDivWorkload::NeonFloorDivWorkload(), NeonFullyConnectedWorkload::NeonFullyConnectedWorkload(), NeonFusedWorkload::NeonFusedWorkload(), NeonMultiplicationWorkload::NeonMultiplicationWorkload(), and NeonSubtractionWorkload::NeonSubtractionWorkload().
|
inline |
Definition at line 141 of file ArmComputeUtils.hpp.
References Equal, Greater, GreaterOrEqual, Less, LessOrEqual, ComparisonDescriptor::m_Operation, and NotEqual.
Referenced by ClComparisonWorkload::ClComparisonWorkload(), ClComparisonWorkloadValidate(), NeonComparisonWorkload::NeonComparisonWorkload(), and NeonComparisonWorkloadValidate().
|
inline |
Definition at line 204 of file ArmComputeUtils.hpp.
References FullyConnectedDescriptor::m_TransposeWeightMatrix.
|
inline |
Definition at line 194 of file ArmComputeUtils.hpp.
References ConvertActivationDescriptorToAclActivationLayerInfo(), and FullyConnectedDescriptor::m_TransposeWeightMatrix.
Referenced by ClFullyConnectedWorkload::ClFullyConnectedWorkload(), ClFullyConnectedWorkloadValidate(), NeonFullyConnectedWorkload::NeonFullyConnectedWorkload(), and NeonFullyConnectedWorkloadValidate().
|
constexpr |
Definition at line 206 of file Logging.hpp.
|
inline |
Definition at line 118 of file ArmComputeUtils.hpp.
Referenced by ClLstmFloatWorkload::ClLstmFloatWorkload(), ClLstmFloatWorkloadValidate(), ClUnidirectionalSequenceLstmFloatWorkload::ClUnidirectionalSequenceLstmFloatWorkload(), ClUnidirectionalSequenceLstmFloatWorkloadValidate(), NeonLstmFloatWorkload::NeonLstmFloatWorkload(), NeonLstmFloatWorkloadValidate(), NeonUnidirectionalSequenceLstmFloatWorkload::NeonUnidirectionalSequenceLstmFloatWorkload(), and NeonUnidirectionalSequenceLstmFloatWorkloadValidate().
| int32_t ConvertMaskToACLFormat | ( | int32_t | mask, |
| int32_t | numDim ) |
Definition at line 299 of file WorkloadUtils.cpp.
References numeric_cast().
Referenced by ClStridedSliceWorkload::ClStridedSliceWorkload(), ClStridedSliceWorkloadValidate(), NeonStridedSliceWorkload::NeonStridedSliceWorkload(), and NeonStridedSliceWorkloadValidate().
|
inline |
Definition at line 182 of file ArmComputeUtils.hpp.
References Across, and Within.
Referenced by NeonNormalizationFloatWorkload::NeonNormalizationFloatWorkload().
|
inline |
Definition at line 168 of file ArmComputeUtils.hpp.
|
inline |
Definition at line 295 of file ArmComputeUtils.hpp.
References Constant, Reflect, and Symmetric.
Referenced by ClPadValidate(), ClPadWorkload::ClPadWorkload(), NeonPadWorkload::NeonPadWorkload(), and NeonPadWorkloadValidate().
|
inline |
Definition at line 155 of file ArmComputeUtils.hpp.
|
inline |
Definition at line 306 of file ArmComputeUtils.hpp.
References ReduceDescriptor::m_ReduceOperation, Max, Mean, Min, Prod, and Sum.
Referenced by ClReduceWorkload::ClReduceWorkload(), ClReduceWorkloadValidate(), NeonReduceWorkload::NeonReduceWorkload(), and NeonReduceWorkloadValidate().
|
inline |
Definition at line 213 of file ArmComputeUtils.hpp.
References Bilinear, and NearestNeighbor.
Referenced by ClResizeWorkload::ClResizeWorkload(), ClResizeWorkloadValidate(), NeonResizeWorkload::NeonResizeWorkload(), and NeonResizeWorkloadValidate().
|
inline |
Utility function used for the two cast layer inputs to convert the output layer tensor types.
Definition at line 19 of file ClFloorDivWorkload.cpp.
References Float32, and TensorInfo::SetDataType().
Referenced by ClFloorDivWorkload::ClFloorDivWorkload(), ClFloorDivWorkloadValidate(), NeonFloorDivWorkload::NeonFloorDivWorkload(), and NeonFloorDivWorkloadValidate().
| armnn::ConstTensor ConvertWeightTensorFromArmnnToAcl | ( | const ConstTensorHandle * | weightTensor, |
| DataLayout | dataLayout, | ||
| void * | permuteBuffer ) |
Definition at line 237 of file WorkloadUtils.cpp.
References Float16, Float32, BaseTensor< MemoryType >::GetDataType(), BaseTensor< MemoryType >::GetInfo(), TensorInfo::GetShape(), ConstTensorHandle::GetTensorInfo(), NCHW, NHWC, PermuteTensor(), QAsymmS8, QAsymmU8, QSymmS8, ReorderWeightChannelsForAcl(), and ReshapeWeightsForAcl().
| TensorInfo ConvertWeightTensorInfoFromArmnnToAcl | ( | const TensorInfo & | weightInfo, |
| DataLayout | dataLayout ) |
Definition at line 122 of file WorkloadUtils.cpp.
References NHWC, armnnUtils::Permuted(), and ReshapeWeightsForAcl().
| void Convolve | ( | const TensorShape & | rInputShape, |
| Decoder< float > & | rInputDecoder, | ||
| const TensorShape & | rOutputShape, | ||
| Encoder< float > & | rOutputEncoder, | ||
| const TensorShape & | rFilterShape, | ||
| Decoder< float > & | rFilterDecoder, | ||
| bool | biasEnabled, | ||
| Decoder< float > * | pBiasDecoder, | ||
| DataLayout | dataLayout, | ||
| unsigned int | paddingTop, | ||
| unsigned int | paddingLeft, | ||
| unsigned int | xStride, | ||
| unsigned int | yStride, | ||
| unsigned int | xDilation, | ||
| unsigned int | yDilation, | ||
| bool | depthwise ) |
Definition at line 68 of file ConvImpl.cpp.
References Decoder< IType >::DecodeTensor(), DataLayoutIndexed::GetChannelsIndex(), DataLayoutIndexed::GetDataLayout(), DataLayoutIndexed::GetHeightIndex(), DataLayoutIndexed::GetWidthIndex(), NHWC, and Encoder< IType >::Set().
| void Convolve3d | ( | const TensorShape & | rInputShape, |
| Decoder< float > & | rInputDecoder, | ||
| const TensorShape & | rOutputShape, | ||
| Encoder< float > & | rOutputEncoder, | ||
| const TensorShape & | rFilterShape, | ||
| Decoder< float > & | rFilterDecoder, | ||
| bool | biasEnabled, | ||
| Decoder< float > * | pBiasDecoder, | ||
| DataLayout | dataLayout, | ||
| unsigned int | paddingTop, | ||
| unsigned int | paddingLeft, | ||
| unsigned int | paddingFront, | ||
| unsigned int | xStride, | ||
| unsigned int | yStride, | ||
| unsigned int | zStride, | ||
| unsigned int | xDilation, | ||
| unsigned int | yDilation, | ||
| unsigned int | zDilation ) |
Definition at line 11 of file Conv3dImpl.cpp.
References Decoder< IType >::DecodeTensor(), DataLayoutIndexed::GetChannelsIndex(), DataLayoutIndexed::GetDataLayout(), DataLayoutIndexed::GetDepthIndex(), DataLayoutIndexed::GetHeightIndex(), DataLayoutIndexed::GetWidthIndex(), NDHWC, and Encoder< IType >::Set().
| uint32_t CoordinatesToIndex | ( | TensorShape & | shape, |
| std::vector< uint32_t > & | coordinates ) |
Definition at line 32 of file Tile.cpp.
References TensorShape::GetNumDimensions().
Referenced by Tile().
| void CopyArmComputeClTensorData | ( | arm_compute::CLTensor & | dstTensor, |
| const T * | srcData ) |
Definition at line 64 of file ClWorkloadUtils.hpp.
References ARMNN_SCOPED_PROFILING_EVENT_CL.
Referenced by ClGatherNdWorkload::ClGatherNdWorkload(), ClConstantWorkload::Execute(), and InitializeArmComputeClTensorData().
| void CopyArmComputeTensorData | ( | arm_compute::Tensor & | dstTensor, |
| const T * | srcData ) |
Definition at line 63 of file NeonWorkloadUtils.hpp.
Referenced by InitializeArmComputeTensorData(), and InitializeArmComputeTensorData().
| void CopyTensorContentsGeneric | ( | const ITensorHandle * | srcTensor, |
| ITensorHandle * | dstTensor, | ||
| CopyFunc | copy ) |
Definition at line 46 of file WorkloadUtils.hpp.
References ARMNN_SCOPED_PROFILING_EVENT, TensorShape::GetNumDimensions(), ITensorHandle::GetShape(), ITensorHandle::GetStrides(), ITensorHandle::Map(), MaxNumOfTensorDimensions, Undefined, and ITensorHandle::Unmap().
Referenced by CopyToOutputTensor(), CopyMemGenericWorkload::Execute(), NeonConvertFp16ToFp32Workload::Execute(), and NeonConvertFp32ToFp16Workload::Execute().
| void CopyToOutputTensor | ( | const Tensor & | outputTensor, |
| ITensorHandle * | outputTensorHandle ) |
Definition at line 1360 of file LoadedNetwork.cpp.
References ARMNN_SCOPED_PROFILING_EVENT, CopyTensorContentsGeneric(), BaseTensor< MemoryType >::GetInfo(), BaseTensor< MemoryType >::GetMemoryArea(), and Undefined.
|
inline |
Definition at line 29 of file ArmComputeUtils.hpp.
References TensorInfo::GetShape(), and NCHW.
|
inline |
Definition at line 57 of file ClContextSchema_generated.h.
References ClContextBuilder::add_programs(), and ClContextBuilder::Finish().
Referenced by CreateClContextDirect(), and ClContextSerializer::Serialize().
|
inline |
Definition at line 65 of file ClContextSchema_generated.h.
References CreateClContext().
| OriginsDescriptor CreateDescriptorForConcatenation | ( | TensorShapeIt | first, |
| TensorShapeIt | last, | ||
| unsigned int | concatenationDimension ) |
Convenience template to create an OriginsDescriptor to use when creating a ConcatLayer for performing concatenation of a number of input tensors.
Definition at line 300 of file Descriptors.hpp.
References OriginsDescriptor::SetConcatAxis(), and OriginsDescriptor::SetViewOriginCoord().
|
inline |
Definition at line 118 of file ClContextSchema_generated.h.
References ProgramBuilder::add_binary(), ProgramBuilder::add_name(), and ProgramBuilder::Finish().
Referenced by CreateProgramDirect(), and ClContextSerializer::Serialize().
|
inline |
Definition at line 128 of file ClContextSchema_generated.h.
References CreateProgram().
| BackendsMap CreateSupportedBackends | ( | TensorHandleFactoryRegistry & | handleFactoryRegistry, |
| BackendSettings & | backendSettings ) |
Definition at line 1354 of file Network.cpp.
References BackendRegistryInstance(), and BackendSettings::m_SupportedBackends.
Referenced by IOptimizedNetwork::Optimize.
| void Debug | ( | const TensorInfo & | inputInfo, |
| const T * | inputData, | ||
| LayerGuid | guid, | ||
| const std::string & | layerName, | ||
| unsigned int | slotIndex, | ||
| bool | outputsToFile ) |
Definition at line 97 of file Debug.cpp.
References PrintOutput().
Referenced by DebugLayer::DebugLayer(), and DebugLayer::DebugLayer().
| template void Debug< BFloat16 > | ( | const TensorInfo & | inputInfo, |
| const BFloat16 * | inputData, | ||
| LayerGuid | guid, | ||
| const std::string & | layerName, | ||
| unsigned int | slotIndex, | ||
| bool | outputsToFile ) |
| template void Debug< float > | ( | const TensorInfo & | inputInfo, |
| const float * | inputData, | ||
| LayerGuid | guid, | ||
| const std::string & | layerName, | ||
| unsigned int | slotIndex, | ||
| bool | outputsToFile ) |
| template void Debug< Half > | ( | const TensorInfo & | inputInfo, |
| const Half * | inputData, | ||
| LayerGuid | guid, | ||
| const std::string & | layerName, | ||
| unsigned int | slotIndex, | ||
| bool | outputsToFile ) |
| template void Debug< int16_t > | ( | const TensorInfo & | inputInfo, |
| const int16_t * | inputData, | ||
| LayerGuid | guid, | ||
| const std::string & | layerName, | ||
| unsigned int | slotIndex, | ||
| bool | outputsToFile ) |
| template void Debug< int32_t > | ( | const TensorInfo & | inputInfo, |
| const int32_t * | inputData, | ||
| LayerGuid | guid, | ||
| const std::string & | layerName, | ||
| unsigned int | slotIndex, | ||
| bool | outputsToFile ) |
| template void Debug< int64_t > | ( | const TensorInfo & | inputInfo, |
| const int64_t * | inputData, | ||
| LayerGuid | guid, | ||
| const std::string & | layerName, | ||
| unsigned int | slotIndex, | ||
| bool | outputsToFile ) |
| template void Debug< int8_t > | ( | const TensorInfo & | inputInfo, |
| const int8_t * | inputData, | ||
| LayerGuid | guid, | ||
| const std::string & | layerName, | ||
| unsigned int | slotIndex, | ||
| bool | outputsToFile ) |
| template void Debug< uint8_t > | ( | const TensorInfo & | inputInfo, |
| const uint8_t * | inputData, | ||
| LayerGuid | guid, | ||
| const std::string & | layerName, | ||
| unsigned int | slotIndex, | ||
| bool | outputsToFile ) |
| void DeleteAsType | ( | const void *const | blob | ) |
Definition at line 27 of file TosaRefBackend.cpp.
Referenced by TosaRefBackend::OptimizeSubgraphView().
| void DepthToSpace | ( | const TensorInfo & | inputInfo, |
| const DepthToSpaceDescriptor & | descriptor, | ||
| const void * | inputData, | ||
| void * | outputData, | ||
| unsigned int | dataTypeSize ) |
Definition at line 16 of file DepthToSpace.cpp.
References DataLayoutIndexed::GetChannelsIndex(), DataLayoutIndexed::GetHeightIndex(), TensorShape::GetNumElements(), TensorInfo::GetShape(), DataLayoutIndexed::GetWidthIndex(), SpaceToDepthDescriptor::m_BlockSize, SpaceToDepthDescriptor::m_DataLayout, NCHW, and Permute.
Referenced by DepthToSpaceLayer::DepthToSpaceLayer().
|
inline |
Definition at line 113 of file RefWorkloadUtils.hpp.
References Dequantize, and info.
| std::vector< float > Dequantize | ( | const T * | quant, |
| const TensorInfo & | info ) |
u8 helpers
Definition at line 102 of file RefWorkloadUtils.hpp.
References Dequantize, and info.
| void Dequantize | ( | Decoder< float > & | inputDecoder, |
| Encoder< float > & | outputEncoder, | ||
| const TensorInfo & | inputInfo, | ||
| const TensorInfo & | outputInfo ) |
Definition at line 11 of file Dequantize.cpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, Decoder< IType >::Get(), TensorInfo::GetNumElements(), and Encoder< IType >::Set().
| float Dequantize | ( | QuantizedType | value, |
| float | scale, | ||
| int32_t | offset ) |
Dequantize an 8-bit data type into a floating point data type.
| value | - The value to dequantize. |
| scale | - The scale (must be non-zero). |
| offset | - The offset. |
Definition at line 48 of file TypesUtils.cpp.
References IsQuantizedType(), and numeric_cast().
Referenced by QASymm8Decoder::DecodeTensor(), QASymmS8Decoder::DecodeTensor(), QSymm16Decoder::DecodeTensor(), QSymm8PerAxisDecoder::DecodeTensor(), QSymmS8Decoder::DecodeTensor(), ScaledInt32PerAxisDecoder::DecodeTensor(), and DequantizeLayer::DequantizeLayer().
| void DetectionPostProcess | ( | const TensorInfo & | boxEncodingsInfo, |
| const TensorInfo & | scoresInfo, | ||
| const TensorInfo & | , | ||
| const TensorInfo & | detectionBoxesInfo, | ||
| const TensorInfo & | , | ||
| const TensorInfo & | , | ||
| const TensorInfo & | , | ||
| const DetectionPostProcessDescriptor & | desc, | ||
| Decoder< float > & | boxEncodings, | ||
| Decoder< float > & | scores, | ||
| Decoder< float > & | anchors, | ||
| float * | detectionBoxes, | ||
| float * | detectionClasses, | ||
| float * | detectionScores, | ||
| float * | numDetections ) |
Definition at line 139 of file DetectionPostProcess.cpp.
References AllocateOutputData(), GenerateRangeK(), Decoder< IType >::Get(), TensorInfo::GetNumElements(), TensorInfo::GetShape(), DetectionPostProcessDescriptor::m_DetectionsPerClass, DetectionPostProcessDescriptor::m_MaxClassesPerDetection, DetectionPostProcessDescriptor::m_MaxDetections, DetectionPostProcessDescriptor::m_NmsIouThreshold, DetectionPostProcessDescriptor::m_NmsScoreThreshold, DetectionPostProcessDescriptor::m_NumClasses, DetectionPostProcessDescriptor::m_ScaleH, DetectionPostProcessDescriptor::m_ScaleW, DetectionPostProcessDescriptor::m_ScaleX, DetectionPostProcessDescriptor::m_ScaleY, DetectionPostProcessDescriptor::m_UseRegularNms, NonMaxSuppression(), numeric_cast(), and TopKSort().
Referenced by DetectionPostProcessLayer::DetectionPostProcessLayer().
| void ExecuteFunction | ( | std::vector< ITensorHandle * > | inputs, |
| std::vector< ITensorHandle * > | outputs, | ||
| BinaryOperation | operation ) |
Definition at line 28 of file RefElementwiseBinaryWorkload.cpp.
References Add, CHECK_LOCATION, Div, FloorDiv, GetBinaryOperationAsCString(), TensorInfo::GetShape(), armnnUtils::GetTensorInfo(), MakeDecoder(), MakeEncoder(), Map, Maximum, Minimum, Mul, Power, SqDiff, and Sub.
| void ExtractJsonObjects | ( | unsigned int | inferenceIndex, |
| const Event * | parentEvent, | ||
| JsonChildObject & | parentObject, | ||
| std::map< const Event *, std::vector< const Event * > > | descendantsMap ) |
Definition at line 314 of file Profiling.cpp.
References JsonChildObject::AddChild(), JsonChildObject::AddMeasurement(), ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, Event, ExtractJsonObjects(), JsonChildObject::GetChild(), Event::GetInstruments(), Event::GetMeasurements(), Event::GetProfilingGuid(), OptionalBase::has_value(), JsonChildObject::m_Children, Measurement, JsonChildObject::NumChildren(), JsonChildObject::SetGuid(), JsonChildObject::SetType(), JsonChildObject::SetUnit(), and OptionalReferenceSwitch< IsReference, T >::value().
Referenced by ExtractJsonObjects(), and ProfilerImpl::Print().
| void FakeQuantization | ( | const float * | inputData, |
| float * | outputData, | ||
| uint32_t | numElements, | ||
| float | min, | ||
| float | max ) |
Definition at line 17 of file RefFakeQuantizationFloat32Workload.cpp.
References numeric_cast(), and Quantize.
Referenced by FakeQuantizationLayer::FakeQuantizationLayer().
| bool FalseFunc | ( | Optional< std::string & > | reasonIfUnsupported, |
| Params &&... | params ) |
Definition at line 62 of file LayerSupportCommon.hpp.
References IgnoreUnused().
| bool FalseFuncF16 | ( | Optional< std::string & > | reasonIfUnsupported, |
| Params &&... | params ) |
Definition at line 70 of file LayerSupportCommon.hpp.
References IgnoreUnused(), and SetValueChecked().
Referenced by NeonLayerSupport::IsFloorSupported().
| bool FalseFuncF32 | ( | Optional< std::string & > | reasonIfUnsupported, |
| Params &&... | params ) |
Definition at line 78 of file LayerSupportCommon.hpp.
References IgnoreUnused(), and SetValueChecked().
| bool FalseFuncI32 | ( | Optional< std::string & > | reasonIfUnsupported, |
| Params &&... | params ) |
Definition at line 94 of file LayerSupportCommon.hpp.
References IgnoreUnused(), and SetValueChecked().
Referenced by RefLayerSupport::IsConvertFp16ToFp32Supported(), RefLayerSupport::IsConvertFp32ToFp16Supported(), and NeonLayerSupport::IsFloorSupported().
| bool FalseFuncU8 | ( | Optional< std::string & > | reasonIfUnsupported, |
| Params &&... | params ) |
Definition at line 86 of file LayerSupportCommon.hpp.
References IgnoreUnused(), and SetValueChecked().
Referenced by RefLayerSupport::IsConvertFp16ToFp32Supported(), RefLayerSupport::IsConvertFp32ToFp16Supported(), and NeonLayerSupport::IsFloorSupported().
| bool FalseInputFuncF16 | ( | Optional< std::string & > | reasonIfUnsupported, |
| Params &&... | params ) |
Definition at line 110 of file LayerSupportCommon.hpp.
References IgnoreUnused(), and SetValueChecked().
Referenced by RefLayerSupport::IsConvertFp32ToFp16Supported().
| bool FalseInputFuncF32 | ( | Optional< std::string & > | reasonIfUnsupported, |
| Params &&... | params ) |
Definition at line 102 of file LayerSupportCommon.hpp.
References IgnoreUnused(), and SetValueChecked().
Referenced by RefLayerSupport::IsConvertFp16ToFp32Supported().
| bool FalseOutputFuncF16 | ( | Optional< std::string & > | reasonIfUnsupported, |
| Params &&... | params ) |
Definition at line 126 of file LayerSupportCommon.hpp.
References IgnoreUnused(), and SetValueChecked().
Referenced by RefLayerSupport::IsConvertFp16ToFp32Supported().
| bool FalseOutputFuncF32 | ( | Optional< std::string & > | reasonIfUnsupported, |
| Params &&... | params ) |
Definition at line 118 of file LayerSupportCommon.hpp.
References IgnoreUnused(), and SetValueChecked().
Referenced by RefLayerSupport::IsConvertFp32ToFp16Supported().
| void Fill | ( | Encoder< float > & | output, |
| const TensorShape & | desiredOutputShape, | ||
| const float | value ) |
Creates a tensor and fills it with a scalar value.
Definition at line 13 of file Fill.cpp.
References TensorShape::GetNumElements(), and Encoder< IType >::Set().
Referenced by FillLayer::FillLayer().
| std::vector< Measurement > FindKernelMeasurements | ( | const Event * | event | ) |
Definition at line 62 of file Profiling.cpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, and Event::GetMeasurements().
| Measurement FindMeasurement | ( | const std::string & | name, |
| const Event * | event ) |
Definition at line 43 of file Profiling.cpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, Event::GetMeasurements(), and Measurement::TIME_MS.
Referenced by ProfilerImpl::AnalyzeEventSequenceAndWriteResults(), and ProfilerImpl::CalculateProfilingEventStats().
|
inline |
Definition at line 171 of file ClContextSchema_generated.h.
References ClContextIdentifier().
|
inline |
Definition at line 177 of file ClContextSchema_generated.h.
References ClContextIdentifier().
| LayerType * FoldPadLayer | ( | OptimizationViews & | optimizationViews, |
| LayerType * | baseLayer, | ||
| LayerType * | replacementLayer, | ||
| PadLayer * | padLayer ) |
Definition at line 276 of file SubgraphUtils.hpp.
References OptimizationViews::AddSubstitution().
Referenced by FoldPadLayer2d().
| void FoldPadLayer2d | ( | OptimizationViews & | optimizationViews, |
| LayerT * | baseLayer, | ||
| Descriptor & | descriptor, | ||
| PadLayer * | padLayer ) |
Definition at line 384 of file SubgraphUtils.hpp.
References INetwork::AddConvolution2dLayer(), OptimizationViews::AddDeletedSubgraph(), INetwork::AddDepthwiseConvolution2dLayer(), INetwork::AddPooling2dLayer(), armnn::FoldPadConstraints::alwaysFalse, FoldPadLayer(), OptimizationViews::GetINetwork(), Layer::GetName(), PolymorphicDowncast(), and ReplaceLayer().
Referenced by ClBackend::OptimizeSubgraphView(), and RefBackend::OptimizeSubgraphView().
| void ForEachLayerInput | ( | LayerSelectionInfo::LayerInfoContainer & | layerInfos, |
| LayerSelectionInfo & | layerInfo, | ||
| Delegate | function ) |
Definition at line 277 of file SubgraphViewSelector.cpp.
References Layer::GetInputSlots(), and PolymorphicDowncast().
Referenced by AssignSplitId(), and IsReadyForSplitAssignment().
| void ForEachLayerOutput | ( | LayerSelectionInfo::LayerInfoContainer & | layerInfos, |
| LayerSelectionInfo & | layerInfo, | ||
| Delegate | function ) |
Definition at line 301 of file SubgraphViewSelector.cpp.
References Layer::GetOutputSlots(), and PolymorphicDowncast().
Referenced by SubgraphViewSelector::SelectSubgraphs().
| void FullyConnected | ( | const TensorShape & | rInputShape, |
| Decoder< float > & | rInputDecoder, | ||
| const TensorShape & | rOutputShape, | ||
| Encoder< float > & | rOutputEncoder, | ||
| const TensorShape & | rWeightsShape, | ||
| Decoder< float > & | rWeightDecoder, | ||
| Decoder< float > * | pBiasDecoder, | ||
| const bool | biasEnabled, | ||
| const unsigned int | K, | ||
| const bool | transposeWeights ) |
Performs a matrix multiplication and optionally adds a bias.
Definition at line 13 of file FullyConnected.cpp.
References Decoder< IType >::DecodeTensor(), and Encoder< IType >::Set().
Referenced by FullyConnectedLayer::FullyConnectedLayer().
| LayerType * FuseAdditionLayer | ( | OptimizationViews & | optimizationViews, |
| LayerType * | baseLayer, | ||
| ActivationLayer * | activationLayer, | ||
| ActivationDescriptor & | activationDesc, | ||
| std::string | name ) |
Definition at line 73 of file ArmComputeSubgraphUtils.hpp.
References INetwork::AddAdditionLayer(), ARMNN_NO_DEPRECATE_WARN_BEGIN, ARMNN_NO_DEPRECATE_WARN_END, FuseLayer(), OptimizationViews::GetINetwork(), and PolymorphicDowncast().
Referenced by ClBackend::OptimizeSubgraphView(), and NeonBackend::OptimizeSubgraphView().
| LayerType * FuseBatchNormalizationLayer | ( | OptimizationViews & | optimizationViews, |
| LayerType * | baseLayer, | ||
| ActivationLayer * | activationLayer, | ||
| ActivationDescriptor & | activationDesc, | ||
| std::string | name ) |
Definition at line 178 of file ArmComputeSubgraphUtils.hpp.
References INetwork::AddBatchNormalizationLayer(), FuseLayer(), OptimizationViews::GetINetwork(), and PolymorphicDowncast().
Referenced by ClBackend::OptimizeSubgraphView(), and NeonBackend::OptimizeSubgraphView().
| LayerType * FuseConvolution2dLayer | ( | OptimizationViews & | optimizationViews, |
| LayerType * | baseLayer, | ||
| ActivationLayer * | activationLayer, | ||
| ActivationDescriptor & | activationDesc, | ||
| std::string | name ) |
Definition at line 208 of file ArmComputeSubgraphUtils.hpp.
References INetwork::AddConvolution2dLayer(), FuseLayer(), OptimizationViews::GetINetwork(), and PolymorphicDowncast().
Referenced by ClBackend::OptimizeSubgraphView(), and NeonBackend::OptimizeSubgraphView().
| LayerType * FuseDepthwiseConvolution2dLayer | ( | OptimizationViews & | optimizationViews, |
| LayerType * | baseLayer, | ||
| ActivationLayer * | activationLayer, | ||
| ActivationDescriptor & | activationDesc, | ||
| std::string | name ) |
Definition at line 230 of file ArmComputeSubgraphUtils.hpp.
References INetwork::AddDepthwiseConvolution2dLayer(), FuseLayer(), OptimizationViews::GetINetwork(), and PolymorphicDowncast().
Referenced by ClBackend::OptimizeSubgraphView(), and NeonBackend::OptimizeSubgraphView().
| LayerType * FuseDivisionLayer | ( | OptimizationViews & | optimizationViews, |
| LayerType * | baseLayer, | ||
| ActivationLayer * | activationLayer, | ||
| ActivationDescriptor & | activationDesc, | ||
| std::string | name ) |
Definition at line 115 of file ArmComputeSubgraphUtils.hpp.
References INetwork::AddDivisionLayer(), ARMNN_NO_DEPRECATE_WARN_BEGIN, ARMNN_NO_DEPRECATE_WARN_END, FuseLayer(), OptimizationViews::GetINetwork(), and PolymorphicDowncast().
Referenced by ClBackend::OptimizeSubgraphView(), and NeonBackend::OptimizeSubgraphView().
| LayerType * FuseElementwiseBinaryLayer | ( | OptimizationViews & | optimizationViews, |
| LayerType * | baseLayer, | ||
| ActivationLayer * | activationLayer, | ||
| ActivationDescriptor & | activationDesc, | ||
| BinaryOperation | operation, | ||
| std::string | name ) |
Definition at line 157 of file ArmComputeSubgraphUtils.hpp.
References INetwork::AddElementwiseBinaryLayer(), FuseLayer(), OptimizationViews::GetINetwork(), and PolymorphicDowncast().
Referenced by ClBackend::OptimizeSubgraphView(), and NeonBackend::OptimizeSubgraphView().
| LayerType * FuseFullyConnectedLayer | ( | OptimizationViews & | optimizationViews, |
| LayerType * | baseLayer, | ||
| ActivationLayer * | activationLayer, | ||
| ActivationDescriptor & | activationDesc, | ||
| std::string | name ) |
Definition at line 252 of file ArmComputeSubgraphUtils.hpp.
References INetwork::AddFullyConnectedLayer(), FuseLayer(), OptimizationViews::GetINetwork(), and PolymorphicDowncast().
Referenced by ClBackend::OptimizeSubgraphView(), and NeonBackend::OptimizeSubgraphView().
| LayerType * FuseLayer | ( | OptimizationViews & | optimizationViews, |
| LayerType * | baseLayer, | ||
| LayerType * | replacementLayer, | ||
| ActivationLayer * | activationLayer, | ||
| ActivationDescriptor & | activationDesc ) |
Definition at line 53 of file ArmComputeSubgraphUtils.hpp.
References OptimizationViews::AddSubstitution().
Referenced by FuseAdditionLayer(), FuseBatchNormalizationLayer(), FuseConvolution2dLayer(), FuseDepthwiseConvolution2dLayer(), FuseDivisionLayer(), FuseElementwiseBinaryLayer(), FuseFullyConnectedLayer(), FuseMultiplicationLayer(), and FuseSubtractionLayer().
| LayerType * FuseMultiplicationLayer | ( | OptimizationViews & | optimizationViews, |
| LayerType * | baseLayer, | ||
| ActivationLayer * | activationLayer, | ||
| ActivationDescriptor & | activationDesc, | ||
| std::string | name ) |
Definition at line 136 of file ArmComputeSubgraphUtils.hpp.
References INetwork::AddMultiplicationLayer(), ARMNN_NO_DEPRECATE_WARN_BEGIN, ARMNN_NO_DEPRECATE_WARN_END, FuseLayer(), OptimizationViews::GetINetwork(), and PolymorphicDowncast().
Referenced by ClBackend::OptimizeSubgraphView(), and NeonBackend::OptimizeSubgraphView().
| LayerType * FuseSubtractionLayer | ( | OptimizationViews & | optimizationViews, |
| LayerType * | baseLayer, | ||
| ActivationLayer * | activationLayer, | ||
| ActivationDescriptor & | activationDesc, | ||
| std::string | name ) |
Definition at line 94 of file ArmComputeSubgraphUtils.hpp.
References INetwork::AddSubtractionLayer(), ARMNN_NO_DEPRECATE_WARN_BEGIN, ARMNN_NO_DEPRECATE_WARN_END, FuseLayer(), OptimizationViews::GetINetwork(), and PolymorphicDowncast().
Referenced by ClBackend::OptimizeSubgraphView(), and NeonBackend::OptimizeSubgraphView().
| template void Gather | ( | const TensorInfo & | paramsInfo, |
| const TensorInfo & | indicesInfo, | ||
| const TensorInfo & | outputInfo, | ||
| Decoder< double_t > & | params, | ||
| const int32_t * | indices, | ||
| Encoder< double_t > & | output, | ||
| const int32_t | axis_int ) |
| template void Gather | ( | const TensorInfo & | paramsInfo, |
| const TensorInfo & | indicesInfo, | ||
| const TensorInfo & | outputInfo, | ||
| Decoder< float > & | params, | ||
| const int32_t * | indices, | ||
| Encoder< float > & | output, | ||
| const int32_t | axis_int ) |
| void Gather | ( | const TensorInfo & | paramsInfo, |
| const TensorInfo & | indicesInfo, | ||
| const TensorInfo & | outputInfo, | ||
| Decoder< I > & | params, | ||
| const int32_t * | indices, | ||
| Encoder< O > & | output, | ||
| const int32_t | axis_int ) |
Definition at line 15 of file Gather.cpp.
References Decoder< IType >::Get(), TensorInfo::GetNumDimensions(), TensorInfo::GetNumElements(), TensorInfo::GetShape(), IgnoreUnused(), and Encoder< IType >::Set().
Referenced by GatherLayer::GatherLayer().
| void GatherTensorHandlePairs | ( | const DescriptorType & | descriptor, |
| std::vector< std::pair< SrcTensorHandleType *, DstTensorHandleType * > > & | tensorHandlePairs ) |
Definition at line 214 of file WorkloadUtils.hpp.
References PolymorphicDowncast().
Referenced by CopyMemGenericWorkload::CopyMemGenericWorkload(), NeonConvertFp16ToFp32Workload::NeonConvertFp16ToFp32Workload(), and NeonConvertFp32ToFp16Workload::NeonConvertFp32ToFp16Workload().
| armnn::PermutationVector GeneratePermutationVectorOnLastTwoDimensions | ( | unsigned int | rank | ) |
Generates a permutation vector of size rank that permutes the 2 most right dimensions.
| rank | - Tensor rank, i.e. number of dimensions in the tensors |
Definition at line 357 of file WorkloadUtils.cpp.
| std::vector< unsigned int > GenerateRangeK | ( | unsigned int | k | ) |
Definition at line 16 of file DetectionPostProcess.cpp.
Referenced by DetectionPostProcess(), and NonMaxSuppression().
|
constexpr |
Definition at line 31 of file TypesUtils.hpp.
References Abs, BoundedReLu, Elu, Gelu, HardSwish, LeakyReLu, Linear, ReLu, Sigmoid, SoftReLu, Sqrt, Square, and TanH.
Referenced by StringifyLayerParameters< ActivationDescriptor >::Serialize().
|
constexpr |
Definition at line 52 of file TypesUtils.hpp.
Definition at line 28 of file WorkloadData.cpp.
References BFloat16, Float16, Float32, QAsymmS8, QAsymmU8, QSymmS16, QSymmS8, and Signed32.
Referenced by Convolution2dQueueDescriptor::Validate(), Convolution3dQueueDescriptor::Validate(), DepthwiseConvolution2dQueueDescriptor::Validate(), FullyConnectedQueueDescriptor::Validate(), and TransposeConvolution2dQueueDescriptor::Validate().
|
inline |
Definition at line 13 of file LayerSupportRules.hpp.
References Float16, Float32, QAsymmS8, QAsymmU8, QSymmS16, QSymmS8, Signed32, and OptionalReferenceSwitch< IsReference, T >::value().
Referenced by BiasAndWeightsTypesMatch::BiasAndWeightsTypesMatch().
|
constexpr |
Definition at line 76 of file TypesUtils.hpp.
References Add, Div, FloorDiv, Maximum, Minimum, Mul, Power, SqDiff, and Sub.
Referenced by ExecuteFunction(), and StringifyLayerParameters< ElementwiseBinaryDescriptor >::Serialize().
| Optional< const BackendOptions::BackendOption > GetCapability | ( | const std::string & | backendCapabilityName, |
| const armnn::BackendId & | backend ) |
Returns a BackendCapability if the backend lists the capability The BackendCapability must then be inspected to check whether or not that BackendCapability is supported Otherwise returns an EmptyOptional if the BackendCapability is unlisted.
Definition at line 51 of file BackendHelper.cpp.
References BackendRegistryInstance(), and GetCapability().
| Optional< const BackendOptions::BackendOption > GetCapability | ( | const std::string & | backendCapabilityName, |
| const BackendCapabilities & | capabilities ) |
Returns a BackendCapability if the backend lists the capability The BackendCapability must then be inspected to check whether or not that BackendCapability is supported Otherwise returns an EmptyOptional if the BackendCapability is unlisted.
Definition at line 37 of file BackendHelper.cpp.
References BackendOptions::GetOption(), and BackendOptions::GetOptionCount().
Referenced by AttemptBackendAssignment(), GetCapability(), HasCapability(), HasCapability(), LayerSupportHandle::IsConvolution2dSupported(), LayerSupportHandle::IsDepthwiseConvolutionSupported(), LayerSupportHandle::IsDilatedDepthwiseConvolutionSupported(), and LayerSupportHandle::IsFullyConnectedSupported().
|
inline |
Definition at line 140 of file ClContextSchema_generated.h.
Referenced by ClContextDeserializer::DeserializeFromBinary().
|
constexpr |
Definition at line 62 of file TypesUtils.hpp.
References Equal, Greater, GreaterOrEqual, Less, LessOrEqual, and NotEqual.
Referenced by StringifyLayerParameters< ComparisonDescriptor >::Serialize().
|
constexpr |
Deprecated function that will be removed together with the Compute enum.
Definition at line 36 of file BackendId.hpp.
References CpuAcc, CpuRef, GpuAcc, and TosaRef.
Referenced by BackendId::BackendId(), BackendId::BackendId(), BackendId::IsCpuAcc(), BackendId::IsCpuRef(), BackendId::IsGpuAcc(), BackendId::IsUndefined(), operator<<(), operator<<(), and operator<<().
|
inline |
Definition at line 46 of file ClWorkloadUtils.hpp.
Referenced by ClConvolution2dWorkload::ClConvolution2dWorkload(), and NeonConvolution2dWorkload::NeonConvolution2dWorkload().
|
constexpr |
Definition at line 254 of file TypesUtils.hpp.
References NCDHW, NCHW, NDHWC, and NHWC.
Referenced by Convert1HWOTensorInfoToAcl(), Convert1HWOTensorToAcl(), MakeTensorShape(), StringifyLayerParameters< BatchNormalizationDescriptor >::Serialize(), StringifyLayerParameters< BatchToSpaceNdDescriptor >::Serialize(), StringifyLayerParameters< Convolution2dDescriptor >::Serialize(), StringifyLayerParameters< Convolution3dDescriptor >::Serialize(), StringifyLayerParameters< DepthwiseConvolution2dDescriptor >::Serialize(), StringifyLayerParameters< L2NormalizationDescriptor >::Serialize(), StringifyLayerParameters< NormalizationDescriptor >::Serialize(), StringifyLayerParameters< Pooling2dDescriptor >::Serialize(), StringifyLayerParameters< Pooling3dDescriptor >::Serialize(), StringifyLayerParameters< ResizeDescriptor >::Serialize(), StringifyLayerParameters< SpaceToBatchNdDescriptor >::Serialize(), StringifyLayerParameters< SpaceToDepthDescriptor >::Serialize(), StringifyLayerParameters< StridedSliceDescriptor >::Serialize(), StringifyLayerParameters< TransposeConvolution2dDescriptor >::Serialize(), and PermuteDepthwiseConv2dWeightsImpl::~PermuteDepthwiseConv2dWeightsImpl().
|
constexpr |
Definition at line 234 of file TypesUtils.hpp.
References BFloat16, Boolean, Float16, Float32, QAsymmS8, QAsymmU8, QSymmS16, QSymmS8, Signed32, and Signed64.
Referenced by AttemptBackendAssignment(), RefDebugWorkload< DataType::BFloat16 >::GetName(), armnnUtils::GetPerAxisParams(), ConstantLayer::SerializeLayerParameters(), armnnUtils::ToFloatArray(), VerifyTensorInfoDataType(), ConvertConstDequantisationLayersToConstLayersImpl::~ConvertConstDequantisationLayersToConstLayersImpl(), and TurboConvertConstDequantisationLayersToConstLayersImpl::~TurboConvertConstDequantisationLayersToConstLayersImpl().
|
constexpr |
Definition at line 183 of file TypesUtils.hpp.
References BFloat16, Boolean, Float16, Float32, QAsymmS8, QAsymmU8, QSymmS16, QSymmS8, Signed32, and Signed64.
Referenced by RefTensorHandle::CanBeImported(), TosaRefTensorHandle::CanBeImported(), TensorInfo::GetNumBytes(), GetUnpaddedTensorStrides(), NeonTensorHandle::NeonTensorHandle(), NeonTensorHandle::NeonTensorHandle(), PermuteTensor(), ConvertConstDequantisationLayersToConstLayersImpl::~ConvertConstDequantisationLayersToConstLayersImpl(), ConvertConstPermuteLayersToConstLayers::~ConvertConstPermuteLayersToConstLayers(), and TurboConvertConstDequantisationLayersToConstLayersImpl::~TurboConvertConstDequantisationLayersToConstLayersImpl().
Definition at line 109 of file Profiling.cpp.
Referenced by ProfilerImpl::AnalyzeEventSequenceAndWriteResults().
Definition at line 110 of file Profiling.cpp.
|
inline |
Definition at line 71 of file NeonBackendOptimizationUtils.hpp.
References Layer::GetNameStr().
Referenced by NeonBackend::OptimizeSubgraphView().
|
constexpr |
Definition at line 120 of file TypesUtils.hpp.
References AddMulAdd.
Referenced by StringifyLayerParameters< FusedDescriptor >::Serialize().
| Graph & GetGraphForTesting | ( | IOptimizedNetwork * | optNet | ) |
Definition at line 49 of file TestUtils.cpp.
| LayerSupportHandle GetILayerSupportByBackendId | ( | const armnn::BackendId & | backend | ) |
Convenience function to retrieve the ILayerSupportHandle for a backend.
Definition at line 23 of file BackendHelper.cpp.
References BackendRegistryInstance(), BackendRegistry::GetFactory(), and BackendRegistry::IsBackendRegistered().
| const armnn::ConstTensor GetInputTensor | ( | const LayerBindingId | layerId, |
| const InputTensors & | inputTensors ) |
Definition at line 1376 of file LoadedNetwork.cpp.
| const DataType * GetInputTensorData | ( | unsigned int | idx, |
| const PayloadType & | data ) |
Definition at line 42 of file RefWorkloadUtils.hpp.
References ITensorHandle::Map().
Referenced by GetInputTensorDataBFloat16(), GetInputTensorDataFloat(), GetInputTensorDataHalf(), and Splitter().
| const BFloat16 * GetInputTensorDataBFloat16 | ( | unsigned int | idx, |
| const PayloadType & | data ) |
Definition at line 86 of file RefWorkloadUtils.hpp.
References GetInputTensorData().
| const float * GetInputTensorDataFloat | ( | unsigned int | idx, |
| const PayloadType & | data ) |
Definition at line 62 of file RefWorkloadUtils.hpp.
References GetInputTensorData().
| const Half * GetInputTensorDataHalf | ( | unsigned int | idx, |
| const PayloadType & | data ) |
Definition at line 74 of file RefWorkloadUtils.hpp.
References GetInputTensorData().
Definition at line 1036 of file Network.cpp.
References Float32, InputSlot::GetConnectedOutputSlot(), TensorInfo::GetDataType(), Layer::GetInputSlot(), Layer::GetNumInputSlots(), Layer::GetNumOutputSlots(), Layer::GetOutputSlot(), and OutputSlot::GetTensorInfo().
Referenced by AssignBackends(), and AssignBackendsIConnectable().
| char const * GetLayerTypeAsCString | ( | LayerType | type | ) |
Definition at line 13 of file InternalTypes.cpp.
References LIST_OF_LAYER_TYPE.
Referenced by AttemptBackendAssignment(), Connect(), StrategyBase< DefaultStrategy >::ExecuteStrategy(), Layer::InferOutputShapes(), Graph::InferTensorInfos(), Graph::Print(), ReturnWithError(), Layer::SerializeLayerParameters(), Graph::SerializeToDot(), ElementwiseBaseLayer::ValidateTensorShapesFromInputs(), ElementwiseBinaryLayer::ValidateTensorShapesFromInputs(), ElementwiseUnaryLayer::ValidateTensorShapesFromInputs(), Graph::VerifyConstantLayerSetTensorInfo(), and Layer::VerifyLayerConnections().
|
constexpr |
Definition at line 110 of file TypesUtils.hpp.
References LogicalAnd, and LogicalOr.
|
constexpr |
Definition at line 296 of file TypesUtils.hpp.
References MultiAxisPacking, and SingleAxisPacking.
Referenced by RuntimeImpl::RuntimeImpl().
|
inline |
Definition at line 36 of file MemoryOptimizerStrategyLibrary.hpp.
Referenced by RuntimeImpl::RuntimeImpl().
|
inline |
Definition at line 47 of file MemoryOptimizerStrategyLibrary.hpp.
| ModelOptions & GetModelOptionsForTesting | ( | IOptimizedNetwork * | optNet | ) |
Definition at line 54 of file TestUtils.cpp.
|
constexpr |
Definition at line 266 of file TypesUtils.hpp.
References Across, and Within.
Referenced by StringifyLayerParameters< NormalizationDescriptor >::Serialize().
|
constexpr |
Definition at line 276 of file TypesUtils.hpp.
References LocalBrightness, and LocalContrast.
Referenced by StringifyLayerParameters< NormalizationDescriptor >::Serialize().
| unsigned int GetNumActivations | ( | const TensorInfo & | inputInfo | ) |
Definition at line 16 of file RefFullyConnectedWorkload.cpp.
References TensorInfo::GetNumDimensions(), and TensorInfo::GetShape().
Referenced by RefFullyConnectedWorkload::RefFullyConnectedWorkload().
| unsigned int GetNumberOfCacheFiles | ( | const armnn::BackendId & | backend | ) |
Returns the number of cached files if backend supports caching.
Definition at line 130 of file BackendHelper.cpp.
References BackendRegistryInstance().
| uint32_t GetNumInputs | ( | bool | biasEnabled | ) |
Definition at line 455 of file Descriptors.cpp.
Referenced by Convolution2dLayer::Convolution2dLayer(), Convolution3dLayer::Convolution3dLayer(), DepthwiseConvolution2dLayer::DepthwiseConvolution2dLayer(), FullyConnectedLayer::FullyConnectedLayer(), Convolution2dDescriptor::GetNumInputs(), Convolution3dDescriptor::GetNumInputs(), DepthwiseConvolution2dDescriptor::GetNumInputs(), and FullyConnectedDescriptor::GetNumInputs().
| unsigned int GetOffset | ( | const TensorShape & | shape, |
| unsigned int | b, | ||
| unsigned int | h, | ||
| unsigned int | w, | ||
| unsigned int | c, | ||
| const DataLayoutIndexed & | dataLayout ) |
Definition at line 15 of file SpaceToBatchNd.cpp.
References CHECK_LOCATION, DataLayoutIndexed::GetChannelsIndex(), DataLayoutIndexed::GetDataLayout(), DataLayoutIndexed::GetHeightIndex(), TensorShape::GetNumDimensions(), DataLayoutIndexed::GetWidthIndex(), NCHW, and NHWC.
Referenced by SpaceToBatchNd(), and SpaceToDepth().
|
constexpr |
Definition at line 140 of file TypesUtils.hpp.
References Ceiling, and Floor.
Referenced by StringifyLayerParameters< Pooling2dDescriptor >::Serialize(), and StringifyLayerParameters< Pooling3dDescriptor >::Serialize().
| const armnn::Tensor GetOutputTensor | ( | const LayerBindingId | layerId, |
| const OutputTensors & | outputTensors ) |
Definition at line 1389 of file LoadedNetwork.cpp.
| DataType * GetOutputTensorData | ( | ITensorHandle * | tensorHandle | ) |
Definition at line 56 of file RefWorkloadUtils.hpp.
References ITensorHandle::Map().
| DataType * GetOutputTensorData | ( | unsigned int | idx, |
| const PayloadType & | data ) |
Definition at line 180 of file ClWorkloadUtils.hpp.
References ITensorHandle::Map().
Referenced by ClRankWorkload::Execute(), NeonRankWorkload::Execute(), GetOutputTensorDataBFloat16(), GetOutputTensorDataFloat(), GetOutputTensorDataHalf(), and Splitter().
| BFloat16 * GetOutputTensorDataBFloat16 | ( | unsigned int | idx, |
| const PayloadType & | data ) |
Definition at line 92 of file RefWorkloadUtils.hpp.
References GetOutputTensorData().
| float * GetOutputTensorDataFloat | ( | unsigned int | idx, |
| const PayloadType & | data ) |
Definition at line 68 of file RefWorkloadUtils.hpp.
References GetOutputTensorData().
| Half * GetOutputTensorDataHalf | ( | unsigned int | idx, |
| const PayloadType & | data ) |
Definition at line 80 of file RefWorkloadUtils.hpp.
References GetOutputTensorData().
|
constexpr |
Definition at line 150 of file TypesUtils.hpp.
References Exclude, and IgnoreValue.
Referenced by StringifyLayerParameters< Pooling2dDescriptor >::Serialize(), and StringifyLayerParameters< Pooling3dDescriptor >::Serialize().
|
constexpr |
Definition at line 160 of file TypesUtils.hpp.
References Constant, Reflect, and Symmetric.
Referenced by StringifyLayerParameters< PadDescriptor >::Serialize().
|
constexpr |
Definition at line 129 of file TypesUtils.hpp.
References Average, L2, and Max.
Referenced by StringifyLayerParameters< Pooling2dDescriptor >::Serialize(), and StringifyLayerParameters< Pooling3dDescriptor >::Serialize().
| arm::pipe::IProfilingService & GetProfilingService | ( | armnn::RuntimeImpl * | runtime | ) |
Definition at line 59 of file TestUtils.cpp.
|
constexpr |
Definition at line 171 of file TypesUtils.hpp.
References Max, Mean, Min, Prod, and Sum.
Referenced by ConvertReduceToTosaOperator(), and StringifyLayerParameters< ReduceDescriptor >::Serialize().
|
constexpr |
Definition at line 286 of file TypesUtils.hpp.
References Bilinear, and NearestNeighbor.
Referenced by StringifyLayerParameters< ResizeDescriptor >::Serialize().
|
inline |
Definition at line 144 of file ClContextSchema_generated.h.
|
constexpr |
Definition at line 21 of file TypesUtils.hpp.
References Failure, and Success.
Referenced by operator<<().
|
inline |
float32 helpers
Definition at line 33 of file RefWorkloadUtils.hpp.
References PolymorphicDowncast().
|
inline |
Definition at line 19 of file Timer.hpp.
References GetTimeNow().
Referenced by RuntimeImpl::EnqueueWorkload(), and RuntimeImpl::~RuntimeImpl().
|
inline |
Definition at line 14 of file Timer.hpp.
Referenced by RuntimeImpl::EnqueueWorkload(), GetTimeDuration(), RuntimeImpl::RuntimeImpl(), and RuntimeImpl::~RuntimeImpl().
|
constexpr |
Definition at line 93 of file TypesUtils.hpp.
References Abs, Ceil, Exp, Log, LogicalNot, Neg, Rsqrt, Sin, and Sqrt.
Referenced by StringifyLayerParameters< ElementwiseUnaryDescriptor >::Serialize().
| TensorShape GetUnpaddedTensorStrides | ( | const TensorInfo & | tensorInfo | ) |
Definition at line 15 of file TensorHandle.cpp.
References TensorInfo::GetDataType(), GetDataTypeSize(), TensorShape::GetNumDimensions(), and TensorInfo::GetShape().
Referenced by ConstTensorHandle::GetStrides(), RefTensorHandle::GetStrides(), RefTensorHandleDecorator::GetStrides(), and TosaRefTensorHandle::GetStrides().
| const std::string GetVersion | ( | ) |
Definition at line 81 of file Utils.cpp.
References ARMNN_VERSION.
| bool HasCapability | ( | const BackendOptions::BackendOption & | backendOption, |
| const armnn::BackendId & | backend ) |
Convenience function to check if a given capability matches a capability in a backend.
Definition at line 80 of file BackendHelper.cpp.
References HasMatchingCapability().
| bool HasCapability | ( | const BackendOptions::BackendOption & | capability, |
| const BackendCapabilities & | capabilities ) |
Convenience function to check if a given capability matches a capability in a BackendCapabilities struct.
Definition at line 75 of file BackendHelper.cpp.
References HasMatchingCapability().
| bool HasCapability | ( | const std::string & | name, |
| const armnn::BackendId & | backend ) |
Convenience function to check if a capability exists in a backend.
Definition at line 70 of file BackendHelper.cpp.
References GetCapability().
| bool HasCapability | ( | const std::string & | name, |
| const BackendCapabilities & | capabilities ) |
Convenience function to check if a capability exists in a BackendCapabilites struct.
Definition at line 65 of file BackendHelper.cpp.
References GetCapability().
Referenced by AttemptBackendAssignment().
| bool HasMatchingCapability | ( | const BackendOptions::BackendOption & | backendOption, |
| const armnn::BackendId & | backend ) |
Convenience function to check if a given capability matches a capability in a backend.
Definition at line 117 of file BackendHelper.cpp.
References BackendRegistryInstance(), and HasMatchingCapability().
| bool HasMatchingCapability | ( | const BackendOptions::BackendOption & | capability, |
| const BackendCapabilities & | capabilities ) |
Convenience function to check if a given capability matches a capability in a BackendCapabilities struct.
Definition at line 85 of file BackendHelper.cpp.
References BackendOptions::Var::AsBool(), BackendOptions::Var::AsFloat(), BackendOptions::Var::AsInt(), BackendOptions::Var::AsString(), BackendOptions::Var::AsUnsignedInt(), BackendOptions::BackendOption::GetName(), BackendOptions::GetOption(), BackendOptions::GetOptionCount(), BackendOptions::BackendOption::GetValue(), BackendOptions::Var::IsBool(), BackendOptions::Var::IsFloat(), BackendOptions::Var::IsInt(), BackendOptions::Var::IsString(), and BackendOptions::Var::IsUnsignedInt().
Referenced by CheckFp16Support(), HasCapability(), HasCapability(), HasMatchingCapability(), and RuntimeImpl::RuntimeImpl().
|
inline |
Definition at line 14 of file IgnoreUnused.hpp.
Referenced by ClBackendDefaultAllocator::allocate(), DefaultAllocator::allocate(), ArgMinMax(), CalculateSlotOptionForOutput(), ClTensorHandle::CanBeImported(), ITensorHandle::CanBeImported(), ClConvolution3dWorkload::ClConvolution3dWorkload(), SpaceToBatchNdLayer::Clone(), SpaceToDepthLayer::Clone(), DynamicBackendUtils::CloseHandle(), RefTensorHandleFactory::CreateSubTensorHandle(), RefWorkloadFactory::CreateSubTensorHandle(), TosaRefTensorHandleFactory::CreateSubTensorHandle(), TosaRefWorkloadFactory::CreateSubTensorHandle(), ClWorkloadFactory::CreateTensorHandle(), ClWorkloadFactory::CreateTensorHandle(), ITensorHandleFactory::CreateTensorHandle(), ITensorHandleFactory::CreateTensorHandle(), MockTensorHandleFactory::CreateTensorHandle(), MockTensorHandleFactory::CreateTensorHandle(), RefTensorHandleFactory::CreateTensorHandle(), RefTensorHandleFactory::CreateTensorHandle(), RefWorkloadFactory::CreateTensorHandle(), TosaRefTensorHandleFactory::CreateTensorHandle(), TosaRefTensorHandleFactory::CreateTensorHandle(), TosaRefWorkloadFactory::CreateTensorHandle(), OutputLayer::CreateTensorHandles(), InputLayer::CreateWorkload(), MapLayer::CreateWorkload(), MemCopyLayer::CreateWorkload(), MemImportLayer::CreateWorkload(), MergeLayer::CreateWorkload(), OutputLayer::CreateWorkload(), StandInLayer::CreateWorkload(), UnmapLayer::CreateWorkload(), IBackendInternal::CreateWorkloadFactory(), ITensorHandle::DecorateTensorHandle(), SelectiveQuantizer< armnn::BFloat16, false >::Dequantize(), SelectiveQuantizer< armnn::Half, false >::Dequantize(), SelectiveQuantizer< T, false >::Dequantize(), FakeQuantizationLayer::ExecuteStrategy(), MapLayer::ExecuteStrategy(), MemImportLayer::ExecuteStrategy(), StrategyBase< DefaultStrategy >::ExecuteStrategy(), UnmapLayer::ExecuteStrategy(), SerializerStrategy::ExecuteStrategy(), ExecutionFrame::ExecuteWorkloads(), FalseFunc(), FalseFuncF16(), FalseFuncF32(), FalseFuncI32(), FalseFuncU8(), FalseInputFuncF16(), FalseInputFuncF32(), FalseOutputFuncF16(), FalseOutputFuncF32(), Gather(), ClImportTensorHandleFactory::GetCapabilities(), ITensorHandleFactory::GetCapabilities(), NeonTensorHandleFactory::GetCapabilities(), DynamicBackendUtils::GetEntryPoint(), ClBackendDefaultAllocator::GetMemoryRegionAtOffset(), DefaultAllocator::GetMemoryRegionAtOffset(), ICustomAllocator::GetMemoryRegionAtOffset(), IDeserializer::DeserializerImpl::GetNetworkInputBindingInfo(), IDeserializer::DeserializerImpl::GetNetworkOutputBindingInfo(), IDeserializer::DeserializerImpl::GetNormalizationDescriptor(), IDeserializer::DeserializerImpl::GetPooling2dDescriptor(), IDeserializer::DeserializerImpl::GetPooling3dDescriptor(), DynamicBackendUtils::GetSharedObjects(), ClTensorHandle::Import(), ITensorHandle::Import(), ReshapeLayer::InferOutputShapes(), SliceLayer::InferOutputShapes(), StandInLayer::InferOutputShapes(), NeonLayerSupport::IsActivationSupported(), RefLayerSupport::IsArgMinMaxSupported(), RefLayerSupport::IsBatchMatMulSupported(), RefLayerSupport::IsBatchNormalizationSupported(), RefLayerSupport::IsBatchToSpaceNdSupported(), RefLayerSupport::IsBroadcastToSupported(), RefLayerSupport::IsChannelShuffleSupported(), RefLayerSupport::IsComparisonSupported(), RefLayerSupport::IsConcatSupported(), RefLayerSupport::IsConvolution2dSupported(), RefLayerSupport::IsConvolution3dSupported(), RefLayerSupport::IsDepthToSpaceSupported(), RefLayerSupport::IsDepthwiseConvolutionSupported(), RefLayerSupport::IsDetectionPostProcessSupported(), RefLayerSupport::IsElementwiseUnarySupported(), RefLayerSupport::IsFakeQuantizationSupported(), ClLayerSupport::IsFillSupported(), NeonLayerSupport::IsFillSupported(), RefLayerSupport::IsFillSupported(), NeonLayerSupport::IsFloorSupported(), RefLayerSupport::IsGatherSupported(), RefLayerSupport::IsInstanceNormalizationSupported(), RefLayerSupport::IsL2NormalizationSupported(), ILayerSupport::IsLayerSupported(), TosaRefLayerSupport::IsLayerSupported(), ClLayerSupport::IsLogicalBinarySupported(), RefLayerSupport::IsLogicalBinarySupported(), RefLayerSupport::IsLogSoftmaxSupported(), RefLayerSupport::IsLstmSupported(), RefLayerSupport::IsNormalizationSupported(), RefLayerSupport::IsPadSupported(), RefLayerSupport::IsPermuteSupported(), RefLayerSupport::IsPooling2dSupported(), RefLayerSupport::IsPooling3dSupported(), RefLayerSupport::IsQLstmSupported(), RefLayerSupport::IsRankSupported(), RefLayerSupport::IsReduceSupported(), ClLayerSupport::IsReshapeSupported(), NeonLayerSupport::IsReshapeSupported(), RefLayerSupport::IsReshapeSupported(), RefLayerSupport::IsResizeSupported(), RefLayerSupport::IsScatterNdSupported(), RefLayerSupport::IsShapeSupported(), RefLayerSupport::IsSliceSupported(), RefLayerSupport::IsSoftmaxSupported(), RefLayerSupport::IsSpaceToBatchNdSupported(), RefLayerSupport::IsSpaceToDepthSupported(), ClLayerSupport::IsSplitterSupported(), NeonLayerSupport::IsSplitterSupported(), RefLayerSupport::IsSplitterSupported(), RefLayerSupport::IsStackSupported(), RefLayerSupport::IsStridedSliceSupported(), RefLayerSupport::IsTileSupported(), RefLayerSupport::IsTransposeConvolution2dSupported(), RefLayerSupport::IsTransposeSupported(), RefLayerSupport::IsUnidirectionalSequenceLstmSupported(), Layer::Layer(), ClImportSubTensorHandle::Map(), ClImportTensorHandle::Map(), ClBackend::ClBackendCustomAllocatorMemoryRegion::map(), NeonConvertFp16ToFp32WorkloadValidate(), NeonConvertFp32ToFp16WorkloadValidate(), NeonConvolution3dWorkload::NeonConvolution3dWorkload(), DynamicBackendUtils::OpenHandle(), SelectiveQuantizer< armnn::BFloat16, false >::Quantize(), SelectiveQuantizer< armnn::Half, false >::Quantize(), SelectiveQuantizer< T, false >::Quantize(), BaseWorkload< QueueDescriptor >::ReplaceInputTensorHandle(), BaseWorkload< QueueDescriptor >::ReplaceOutputTensorHandle(), ConvertConstants< Float16ToFloat32, IsFloat32Layer >::Run(), FuseBatchNorm< ConvLayer, ArmnnType, T >::Run(), OptimizeInverseConversionsImpl::Run(), OptimizeInversePermutesImpl< PermuteType >::Run(), SquashEqualSiblingsImpl< Comparable >::Run(), ClImportSubTensorHandle::SetMemoryGroup(), ClImportTensorHandle::SetMemoryGroup(), OpenClTimer::Start(), MemoryManager::StoreMemToAllocate(), TrueFunc(), ClBackend::ClBackendCustomAllocatorMemoryRegion::unmap(), ClBackend::UseCustomMemoryAllocator(), IBackendInternal::UseCustomMemoryAllocator(), ConvertConstPermuteLayersToConstLayers::~ConvertConstPermuteLayersToConstLayers(), Graph::LayerInGraph< InputLayer >::~LayerInGraph(), and Graph::LayerInGraph< OutputLayer >::~LayerInGraph().
| std::vector< uint32_t > IndexToCoordinates | ( | std::vector< uint32_t > & | shape, |
| uint32_t | index ) |
Definition at line 16 of file Tile.cpp.
Referenced by Tile().
|
inline |
Definition at line 124 of file ClWorkloadUtils.hpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, BFloat16, CopyArmComputeClTensorData(), Float16, Float32, ConstTensorHandle::GetConstTensor(), TensorInfo::GetDataType(), ConstTensorHandle::GetTensorInfo(), QAsymmS8, QAsymmU8, QSymmS16, QSymmS8, and Signed32.
Referenced by ClBatchNormalizationFloatWorkload::ClBatchNormalizationFloatWorkload(), ClLstmFloatWorkload::ClLstmFloatWorkload(), ClQLstmWorkload::ClQLstmWorkload(), ClQuantizedLstmWorkload::ClQuantizedLstmWorkload(), ClTransposeConvolution2dWorkload::ClTransposeConvolution2dWorkload(), and ClUnidirectionalSequenceLstmFloatWorkload::ClUnidirectionalSequenceLstmFloatWorkload().
|
inline |
Definition at line 104 of file NeonWorkloadUtils.hpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, BFloat16, CopyArmComputeTensorData(), Float16, Float32, ConstTensorHandle::GetConstTensor(), TensorInfo::GetDataType(), ConstTensorHandle::GetTensorInfo(), QAsymmS8, QAsymmU8, QSymmS16, QSymmS8, and Signed32.
|
inline |
Definition at line 69 of file NeonWorkloadUtils.hpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, BFloat16, CopyArmComputeTensorData(), Float16, Float32, TensorInfo::GetDataType(), ITensorHandle::Map(), QAsymmS8, QAsymmU8, QSymmS16, QSymmS8, and Signed32.
Referenced by NeonConvolution2dWorkload::Execute(), NeonFullyConnectedWorkload::Execute(), NeonBatchNormalizationWorkload::NeonBatchNormalizationWorkload(), NeonDetectionPostProcessWorkload::NeonDetectionPostProcessWorkload(), NeonLstmFloatWorkload::NeonLstmFloatWorkload(), NeonQLstmWorkload::NeonQLstmWorkload(), NeonQuantizedLstmWorkload::NeonQuantizedLstmWorkload(), NeonTransposeConvolution2dWorkload::NeonTransposeConvolution2dWorkload(), NeonUnidirectionalSequenceLstmFloatWorkload::NeonUnidirectionalSequenceLstmFloatWorkload(), and NeonUnidirectionalSequenceLstmWorkload::NeonUnidirectionalSequenceLstmWorkload().
| std::vector< ConvertFp16ToFp32Layer * > InsertConvertFp16ToFp32LayersBefore | ( | Graph & | graph, |
| Layer & | layer, | ||
| bool | expectCorrectInputType ) |
Definition at line 40 of file NetworkUtils.cpp.
References Layer::BeginInputSlots(), Layer::EndInputSlots(), Float16, Float32, InputSlot::GetConnectedOutputSlot(), TensorInfo::GetDataType(), Layer::GetInputSlot(), Layer::GetName(), Layer::GetNumInputSlots(), Layer::GetOutputSlot(), OutputSlot::GetTensorInfo(), Graph::InsertNewLayer(), TensorInfo::SetDataType(), and OutputSlot::SetTensorInfo().
Referenced by AttemptBackendAssignment(), and ConvertFp32NetworkToFp16Impl::Run().
| std::vector< ConvertFp32ToFp16Layer * > InsertConvertFp32ToFp16LayersAfter | ( | Graph & | graph, |
| Layer & | layer ) |
Definition at line 79 of file NetworkUtils.cpp.
References Float16, Float32, InputSlot::GetConnectedOutputSlot(), TensorInfo::GetDataType(), Layer::GetInputSlot(), Layer::GetName(), Layer::GetNumOutputSlots(), Layer::GetOutputSlot(), OutputSlot::GetTensorInfo(), Graph::InsertNewLayer(), TensorInfo::SetDataType(), and OutputSlot::SetTensorInfo().
Referenced by AttemptBackendAssignment(), and ConvertFp32NetworkToFp16Impl::Run().
| std::vector< DebugLayer * > InsertDebugLayerAfter | ( | Graph & | graph, |
| Layer & | layer, | ||
| bool | toFile ) |
Definition at line 112 of file NetworkUtils.cpp.
References Layer::BeginOutputSlots(), CpuRef, Layer::EndOutputSlots(), InputSlot::GetConnectedOutputSlot(), Layer::GetInputSlot(), Layer::GetNameStr(), Layer::GetNumOutputSlots(), Layer::GetOutputSlot(), OutputSlot::GetTensorInfo(), Graph::InsertNewLayer(), Layer::SetBackendId(), and OutputSlot::SetTensorInfo().
Referenced by AddDebugImpl::Run(), and AddDebugToFileImpl::Run().
| void InstanceNorm | ( | const InstanceNormalizationQueueDescriptor & | data, |
| const TensorInfo & | inputInfo, | ||
| Decoder< float > & | inputDecoder, | ||
| Encoder< float > & | outputEncoder ) |
Definition at line 18 of file InstanceNorm.cpp.
References Decoder< IType >::Get(), DataLayoutIndexed::GetChannelsIndex(), DataLayoutIndexed::GetHeightIndex(), DataLayoutIndexed::GetIndex(), TensorInfo::GetShape(), DataLayoutIndexed::GetWidthIndex(), InstanceNormalizationDescriptor::m_Beta, InstanceNormalizationDescriptor::m_DataLayout, InstanceNormalizationDescriptor::m_Eps, InstanceNormalizationDescriptor::m_Gamma, QueueDescriptorWithParameters< LayerDescriptor >::m_Parameters, and Encoder< IType >::Set().
| float IntersectionOverUnion | ( | const float * | boxI, |
| const float * | boxJ ) |
Definition at line 29 of file DetectionPostProcess.cpp.
Referenced by NonMaxSuppression().
|
inline |
Utility function used to check if a vector of tensors are Signed32.
Definition at line 25 of file NeonFloorDivWorkload.cpp.
References Signed32.
Referenced by NeonFloorDivWorkloadValidate().
| bool IsBFloat16 | ( | const WorkloadInfo & | info | ) |
Definition at line 56 of file RefWorkloadFactory.cpp.
References info, and IsDataType().
Referenced by RefWorkloadFactory::CreateWorkload().
| bool IsBoolean | ( | const WorkloadInfo & | info | ) |
Definition at line 80 of file RefWorkloadFactory.cpp.
References info, and IsDataType().
Referenced by RefWorkloadFactory::CreateWorkload().
| bool IsDataType | ( | const WorkloadInfo & | info | ) |
Definition at line 33 of file RefWorkloadFactory.cpp.
References info.
Referenced by IsBFloat16(), IsBoolean(), IsFloat16(), IsQAsymmS8(), IsQAsymmU8(), IsQSymmS16(), IsQSymmS8(), IsSigned32(), and IsSigned64().
| bool IsFloat16 | ( | const WorkloadInfo & | info | ) |
Definition at line 60 of file RefWorkloadFactory.cpp.
References info, and IsDataType().
Referenced by RefWorkloadFactory::CreateWorkload().
| bool IsLayerOptimizable | ( | const armnn::Layer & | layer | ) |
Definition at line 99 of file MockBackend.cpp.
References IsLayerOptimizable().
| bool IsLayerOptimizable | ( | const armnn::Layer * | layer | ) |
Definition at line 88 of file MockBackend.cpp.
References ARMNN_ASSERT, Layer::GetName(), and IsLayerOptimizable().
Referenced by IsLayerOptimizable(), and IsLayerOptimizable().
| bool IsLayerSequence | ( | Layer & | currentLayer, |
| TYPE | first, | ||
| TYPE | second, | ||
| TYPE | third, | ||
| Layer * | layerList[4], | ||
| bool | handleValidActivates, | ||
| const std::vector< ActivationFunction > & | validActivates ) |
Definition at line 452 of file SubgraphUtils.hpp.
References Activation, IOutputSlot::GetConnection(), TensorInfo::GetDataType(), IConnectableLayer::GetInputSlot(), IConnectableLayer::GetOutputSlot(), IInputSlot::GetTensorInfo(), IsSequenceLayerType(), and PolymorphicDowncast().
Referenced by NeonBackend::OptimizeSubgraphView().
| bool IsLayerSupported | ( | const armnn::Layer & | layer | ) |
Definition at line 83 of file MockBackend.cpp.
References IsLayerSupported().
| bool IsLayerSupported | ( | const armnn::Layer * | layer | ) |
Definition at line 62 of file MockBackend.cpp.
References Addition, ARMNN_ASSERT, Constant, Convolution2d, ElementwiseBinary, Layer::GetType(), Input, IsLayerSupported(), and Output.
Referenced by IsLayerSupported(), and IsLayerSupported().
| bool IsLayerTypeSupported | ( | const LayerType & | type, |
| const std::vector< TensorInfo > & | infos, | ||
| const BaseDescriptor & | descriptor, | ||
| const Optional< LstmInputParamsInfo > & | lstmParamsInfo, | ||
| const Optional< QuantizedLstmInputParamsInfo > & | quantizedLstmParamsInfo, | ||
| Optional< std::string & > | reasonIfUnsupported, | ||
| const NeonLayerSupport & | support ) |
Definition at line 185 of file NeonLayerSupport.cpp.
References Activation, Add, Addition, ArgMinMax, BatchMatMul, BatchNormalization, BatchToSpaceNd, Cast, ChannelShuffle, Comparison, Concat, Constant, ConvertFp16ToFp32, ConvertFp32ToFp16, Convolution2d, Convolution3d, DepthToSpace, DepthwiseConvolution2d, Dequantize, DetectionPostProcess, Div, Division, ElementwiseBinary, ElementwiseUnary, Fill, Floor, FloorDiv, FORWARD_WORKLOAD_VALIDATE_FUNC, FullyConnected, Fused, Gather, GatherNd, Input, InstanceNormalization, NeonLayerSupport::IsActivationSupported(), NeonLayerSupport::IsAdditionSupported(), NeonLayerSupport::IsArgMinMaxSupported(), NeonLayerSupport::IsBatchMatMulSupported(), NeonLayerSupport::IsBatchNormalizationSupported(), NeonLayerSupport::IsBatchToSpaceNdSupported(), NeonLayerSupport::IsCastSupported(), NeonLayerSupport::IsChannelShuffleSupported(), NeonLayerSupport::IsComparisonSupported(), NeonLayerSupport::IsConcatSupported(), NeonLayerSupport::IsConstantSupported(), NeonLayerSupport::IsConvertFp16ToFp32Supported(), NeonLayerSupport::IsConvertFp32ToFp16Supported(), NeonLayerSupport::IsConvolution2dSupported(), NeonLayerSupport::IsConvolution3dSupported(), NeonLayerSupport::IsDepthToSpaceSupported(), NeonLayerSupport::IsDepthwiseConvolutionSupported(), NeonLayerSupport::IsDequantizeSupported(), LayerSupportBase::IsDetectionPostProcessSupported(), NeonLayerSupport::IsDivisionSupported(), NeonLayerSupport::IsElementwiseUnarySupported(), NeonLayerSupport::IsFillSupported(), NeonLayerSupport::IsFloorSupported(), NeonLayerSupport::IsFullyConnectedSupported(), NeonLayerSupport::IsFusedSupported(), NeonLayerSupport::IsGatherNdSupported(), NeonLayerSupport::IsGatherSupported(), NeonLayerSupport::IsInputSupported(), NeonLayerSupport::IsInstanceNormalizationSupported(), NeonLayerSupport::IsL2NormalizationSupported(), NeonLayerSupport::IsLogicalBinarySupported(), NeonLayerSupport::IsLogSoftmaxSupported(), NeonLayerSupport::IsLstmSupported(), NeonLayerSupport::IsMaximumSupported(), NeonLayerSupport::IsMeanSupported(), LayerSupportBase::IsMemCopySupported(), LayerSupportBase::IsMemImportSupported(), LayerSupportBase::IsMergeSupported(), NeonLayerSupport::IsMinimumSupported(), NeonLayerSupport::IsMultiplicationSupported(), NeonLayerSupport::IsNormalizationSupported(), NeonLayerSupport::IsOutputSupported(), NeonLayerSupport::IsPadSupported(), NeonLayerSupport::IsPermuteSupported(), NeonLayerSupport::IsPooling2dSupported(), NeonLayerSupport::IsPooling3dSupported(), NeonLayerSupport::IsPreluSupported(), NeonLayerSupport::IsQLstmSupported(), NeonLayerSupport::IsQuantizedLstmSupported(), NeonLayerSupport::IsQuantizeSupported(), NeonLayerSupport::IsReduceSupported(), NeonLayerSupport::IsReshapeSupported(), NeonLayerSupport::IsResizeSupported(), NeonLayerSupport::IsReverseV2Supported(), LayerSupportBase::IsShapeSupported(), NeonLayerSupport::IsSliceSupported(), NeonLayerSupport::IsSoftmaxSupported(), NeonLayerSupport::IsSpaceToBatchNdSupported(), NeonLayerSupport::IsSpaceToDepthSupported(), NeonLayerSupport::IsSplitterSupported(), NeonLayerSupport::IsStackSupported(), NeonLayerSupport::IsStridedSliceSupported(), NeonLayerSupport::IsSubtractionSupported(), NeonLayerSupport::IsTileSupported(), NeonLayerSupport::IsTransposeConvolution2dSupported(), NeonLayerSupport::IsTransposeSupported(), NeonLayerSupport::IsUnidirectionalSequenceLstmSupported(), L2Normalization, LogicalBinary, LogSoftmax, Lstm, Map, Maximum, Mean, MemCopy, MemImport, Merge, Minimum, Mul, Multiplication, NeonAdditionWorkloadValidate(), NeonDivisionWorkloadValidate(), NeonElementwiseBinaryWorkloadValidate(), NeonFloorDivWorkloadValidate(), NeonMaximumWorkloadValidate(), NeonMinimumWorkloadValidate(), NeonMultiplicationWorkloadValidate(), NeonSubtractionWorkloadValidate(), Normalization, numeric_cast(), Output, Pad, Permute, PolymorphicDowncast(), Pooling2d, Pooling3d, Power, Prelu, QLstm, Quantize, QuantizedLstm, Rank, Reduce, Reshape, Resize, ReverseV2, Shape, Slice, Softmax, SpaceToBatchNd, SpaceToDepth, Splitter, SqDiff, Stack, StridedSlice, Sub, Subtraction, Tile, Transpose, TransposeConvolution2d, UnidirectionalSequenceLstm, Unmap, and OptionalReferenceSwitch< IsReference, T >::value().
Referenced by NeonLayerSupport::IsLayerSupported().
|
inline |
Function to check if layer with multiple axes is supported on each backend.
Definition at line 378 of file ArmComputeUtils.hpp.
References ComputeReductionTensorShape(), TensorInfo::GetQuantizationOffset(), TensorInfo::GetQuantizationScale(), TensorInfo::GetQuantizationScales(), TensorInfo::GetShape(), TensorInfo::HasMultipleQuantizationScales(), ReduceDescriptor::m_KeepDims, ReduceDescriptor::m_vAxis, TensorInfo::SetQuantizationOffset(), TensorInfo::SetQuantizationScale(), and TensorInfo::SetQuantizationScales().
Referenced by ClReduceWorkloadValidate(), and NeonReduceWorkloadValidate().
|
inline |
Definition at line 236 of file SubgraphUtils.hpp.
References Layer::ExecuteStrategy().
Referenced by ConnectedToLayerWithNCHW().
|
constexpr |
Definition at line 22 of file RefWorkloadFactory.hpp.
|
constexpr |
Definition at line 20 of file RefWorkloadFactory.hpp.
|
constexpr |
Definition at line 24 of file RefWorkloadFactory.hpp.
|
constexpr |
Definition at line 18 of file RefWorkloadFactory.hpp.
| bool IsQAsymmS8 | ( | const WorkloadInfo & | info | ) |
Definition at line 72 of file RefWorkloadFactory.cpp.
References info, and IsDataType().
Referenced by RefWorkloadFactory::CreateWorkload().
| bool IsQAsymmU8 | ( | const WorkloadInfo & | info | ) |
Definition at line 76 of file RefWorkloadFactory.cpp.
References info, and IsDataType().
Referenced by RefWorkloadFactory::CreateWorkload().
| bool IsQSymmS16 | ( | const WorkloadInfo & | info | ) |
Definition at line 64 of file RefWorkloadFactory.cpp.
References info, and IsDataType().
Referenced by RefWorkloadFactory::CreateWorkload().
| bool IsQSymmS8 | ( | const WorkloadInfo & | info | ) |
Definition at line 68 of file RefWorkloadFactory.cpp.
References info, and IsDataType().
Referenced by RefWorkloadFactory::CreateWorkload().
|
constexpr |
Definition at line 317 of file TypesUtils.hpp.
References QAsymmS8, QAsymmU8, and QSymmS8.
Referenced by ConvertPReluToTosaOperator(), ConvertSoftmaxToTosaOperator(), RefLayerSupport::IsConvolution2dSupported(), RefLayerSupport::IsConvolution3dSupported(), RefLayerSupport::IsDepthwiseConvolutionSupported(), IsQuantizedType(), and RefLayerSupport::IsTransposeConvolution2dSupported().
|
constexpr |
Definition at line 312 of file TypesUtils.hpp.
Referenced by ClMultiplicationWorkload::ClMultiplicationWorkload(), ClMultiplicationWorkloadValidate(), RefWorkloadFactory::CreateWorkload(), Dequantize(), TensorInfo::IsQuantized(), NeonMultiplicationWorkload::NeonMultiplicationWorkload(), NeonMultiplicationWorkloadValidate(), Quantize(), and QuantizeQueueDescriptor::Validate().
|
constexpr |
Definition at line 324 of file TypesUtils.hpp.
References IsQuantized8BitType(), and QSymmS16.
| bool IsReadyForSplitAssignment | ( | LayerSelectionInfo::LayerInfoContainer & | layerInfos, |
| LayerSelectionInfo & | layerInfo ) |
Definition at line 387 of file SubgraphViewSelector.cpp.
References ForEachLayerInput().
Referenced by SubgraphViewSelector::SelectSubgraphs().
|
inline |
Definition at line 444 of file SubgraphUtils.hpp.
References ElementwiseBinary, Layer::GetType(), and PolymorphicDowncast().
Definition at line 439 of file SubgraphUtils.hpp.
References Layer::GetType().
Referenced by BuildAddMulAddTensorInfoLists(), and IsLayerSequence().
| bool IsSigned32 | ( | const WorkloadInfo & | info | ) |
Definition at line 52 of file RefWorkloadFactory.cpp.
References info, and IsDataType().
Referenced by RefWorkloadFactory::CreateWorkload().
| bool IsSigned64 | ( | const WorkloadInfo & | info | ) |
Definition at line 48 of file RefWorkloadFactory.cpp.
References info, and IsDataType().
Referenced by RefWorkloadFactory::CreateWorkload().
| bool IsSupportedForDataTypeGeneric | ( | Optional< std::string & > | reasonIfUnsupported, |
| DataType | dataType, | ||
| Float16Func | float16FuncPtr, | ||
| Float32Func | float32FuncPtr, | ||
| Uint8Func | uint8FuncPtr, | ||
| Int32Func | int32FuncPtr, | ||
| BooleanFunc | booleanFuncPtr, | ||
| Params &&... | params ) |
Definition at line 27 of file LayerSupportCommon.hpp.
References Boolean, Float16, Float32, QAsymmU8, and Signed32.
Referenced by RefLayerSupport::IsConvertFp16ToFp32Supported(), RefLayerSupport::IsConvertFp32ToFp16Supported(), and NeonLayerSupport::IsFloorSupported().
| bool IsTfLiteTurboModel | ( | const Graph & | optGraph | ) |
Definition at line 1976 of file Network.cpp.
References Graph::begin(), Constant, Dequantize, Graph::end(), Float16, and Float32.
Referenced by IOptimizedNetwork::Optimize.
|
inline |
Utility function used to check if statuses are returning 'OK'.
Definition at line 42 of file ClFloorDivWorkload.cpp.
Referenced by ClFloorDivWorkloadValidate(), and NeonFloorDivWorkloadValidate().
|
constexpr |
Definition at line 115 of file LayersFwd.hpp.
|
constexpr |
Definition at line 116 of file LayersFwd.hpp.
|
constexpr |
Definition at line 117 of file LayersFwd.hpp.
|
constexpr |
Definition at line 118 of file LayersFwd.hpp.
|
constexpr |
Definition at line 119 of file LayersFwd.hpp.
|
constexpr |
Definition at line 120 of file LayersFwd.hpp.
|
constexpr |
Definition at line 121 of file LayersFwd.hpp.
Definition at line 122 of file LayersFwd.hpp.
|
constexpr |
Definition at line 123 of file LayersFwd.hpp.
|
constexpr |
Definition at line 124 of file LayersFwd.hpp.
|
constexpr |
Definition at line 125 of file LayersFwd.hpp.
|
constexpr |
Definition at line 126 of file LayersFwd.hpp.
|
constexpr |
Definition at line 127 of file LayersFwd.hpp.
|
constexpr |
Definition at line 128 of file LayersFwd.hpp.
|
constexpr |
Definition at line 129 of file LayersFwd.hpp.
|
constexpr |
Definition at line 130 of file LayersFwd.hpp.
|
constexpr |
Definition at line 131 of file LayersFwd.hpp.
|
constexpr |
Definition at line 132 of file LayersFwd.hpp.
|
constexpr |
Definition at line 133 of file LayersFwd.hpp.
|
constexpr |
Definition at line 134 of file LayersFwd.hpp.
|
constexpr |
Definition at line 135 of file LayersFwd.hpp.
|
constexpr |
Definition at line 136 of file LayersFwd.hpp.
|
constexpr |
Definition at line 137 of file LayersFwd.hpp.
|
constexpr |
Definition at line 138 of file LayersFwd.hpp.
|
constexpr |
Definition at line 139 of file LayersFwd.hpp.
Definition at line 140 of file LayersFwd.hpp.
|
constexpr |
Definition at line 141 of file LayersFwd.hpp.
|
constexpr |
Definition at line 142 of file LayersFwd.hpp.
|
constexpr |
Definition at line 143 of file LayersFwd.hpp.
|
constexpr |
Definition at line 144 of file LayersFwd.hpp.
|
constexpr |
Definition at line 145 of file LayersFwd.hpp.
|
constexpr |
Definition at line 146 of file LayersFwd.hpp.
|
constexpr |
Definition at line 147 of file LayersFwd.hpp.
|
constexpr |
Definition at line 148 of file LayersFwd.hpp.
|
constexpr |
Definition at line 149 of file LayersFwd.hpp.
|
constexpr |
Definition at line 150 of file LayersFwd.hpp.
Definition at line 151 of file LayersFwd.hpp.
Definition at line 152 of file LayersFwd.hpp.
|
constexpr |
Definition at line 153 of file LayersFwd.hpp.
Definition at line 154 of file LayersFwd.hpp.
|
constexpr |
Definition at line 155 of file LayersFwd.hpp.
|
constexpr |
Definition at line 156 of file LayersFwd.hpp.
|
constexpr |
Definition at line 157 of file LayersFwd.hpp.
|
constexpr |
Definition at line 158 of file LayersFwd.hpp.
|
constexpr |
Definition at line 159 of file LayersFwd.hpp.
|
constexpr |
Definition at line 160 of file LayersFwd.hpp.
|
constexpr |
Definition at line 161 of file LayersFwd.hpp.
Definition at line 162 of file LayersFwd.hpp.
|
constexpr |
Definition at line 163 of file LayersFwd.hpp.
|
constexpr |
Definition at line 164 of file LayersFwd.hpp.
|
constexpr |
Definition at line 165 of file LayersFwd.hpp.
|
constexpr |
Definition at line 166 of file LayersFwd.hpp.
|
constexpr |
Definition at line 167 of file LayersFwd.hpp.
|
constexpr |
Definition at line 169 of file LayersFwd.hpp.
|
constexpr |
Definition at line 170 of file LayersFwd.hpp.
|
constexpr |
Definition at line 168 of file LayersFwd.hpp.
Definition at line 171 of file LayersFwd.hpp.
|
constexpr |
Definition at line 172 of file LayersFwd.hpp.
|
constexpr |
Definition at line 173 of file LayersFwd.hpp.
|
constexpr |
Definition at line 174 of file LayersFwd.hpp.
|
constexpr |
Definition at line 175 of file LayersFwd.hpp.
|
constexpr |
Definition at line 176 of file LayersFwd.hpp.
|
constexpr |
Definition at line 177 of file LayersFwd.hpp.
|
constexpr |
Definition at line 178 of file LayersFwd.hpp.
|
constexpr |
Definition at line 179 of file LayersFwd.hpp.
|
constexpr |
Definition at line 180 of file LayersFwd.hpp.
|
constexpr |
Definition at line 181 of file LayersFwd.hpp.
|
constexpr |
Definition at line 182 of file LayersFwd.hpp.
|
constexpr |
Definition at line 183 of file LayersFwd.hpp.
|
constexpr |
Definition at line 184 of file LayersFwd.hpp.
|
constexpr |
Definition at line 185 of file LayersFwd.hpp.
|
constexpr |
Definition at line 186 of file LayersFwd.hpp.
|
constexpr |
Definition at line 187 of file LayersFwd.hpp.
|
constexpr |
Definition at line 188 of file LayersFwd.hpp.
|
constexpr |
Definition at line 190 of file LayersFwd.hpp.
|
constexpr |
Definition at line 189 of file LayersFwd.hpp.
|
constexpr |
Definition at line 191 of file LayersFwd.hpp.
|
constexpr |
Definition at line 192 of file LayersFwd.hpp.
|
inline |
Definition at line 22 of file Logging.hpp.
References Debug, Error, Fatal, Info, Trace, and Warning.
Referenced by ScopedRecord::ScopedRecord().
| void LogSoftmax | ( | Decoder< float > & | input, |
| Encoder< float > & | output, | ||
| const TensorInfo & | inputInfo, | ||
| const LogSoftmaxDescriptor & | descriptor ) |
Definition at line 27 of file LogSoftmax.cpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, Decoder< IType >::Get(), TensorInfo::GetNumDimensions(), TensorShape::GetNumDimensions(), armnnUtils::GetNumElementsBetween(), TensorInfo::GetShape(), SoftmaxDescriptor::m_Axis, SoftmaxDescriptor::m_Beta, numeric_cast(), and Encoder< IType >::Set().
Referenced by LogSoftmaxLayer::LogSoftmaxLayer().
| void LstmImpl | ( | const LstmDescriptor & | descriptor, |
| const TensorInfo & | inputInfo, | ||
| const TensorInfo & | outputInfo, | ||
| const TensorShape & | inputToOutputWeightsShape, | ||
| const TensorShape & | recurrentToOutputWeightsShape, | ||
| std::unique_ptr< Decoder< float > > & | inputData, | ||
| std::unique_ptr< Decoder< float > > & | outputStateIn, | ||
| std::unique_ptr< Decoder< float > > & | cellStateIn, | ||
| std::unique_ptr< Encoder< float > > & | outputStateOut, | ||
| std::unique_ptr< Encoder< float > > & | cellStateOut, | ||
| std::unique_ptr< Encoder< float > > & | output, | ||
| std::unique_ptr< Decoder< float > > & | cellStateOutDecoder, | ||
| std::unique_ptr< Decoder< float > > & | outputDecoder, | ||
| std::unique_ptr< Decoder< float > > & | inputToInputWeightsTensor, | ||
| std::unique_ptr< Decoder< float > > & | inputToForgetWeightsTensor, | ||
| std::unique_ptr< Decoder< float > > & | inputToCellWeightsTensor, | ||
| std::unique_ptr< Decoder< float > > & | inputToOutputWeightsTensor, | ||
| std::unique_ptr< Decoder< float > > & | recurrentToInputWeightsTensor, | ||
| std::unique_ptr< Decoder< float > > & | recurrentToForgetWeightsTensor, | ||
| std::unique_ptr< Decoder< float > > & | recurrentToCellWeightsTensor, | ||
| std::unique_ptr< Decoder< float > > & | recurrentToOutputWeightsTensor, | ||
| std::unique_ptr< Decoder< float > > & | cellToInputWeightsTensor, | ||
| std::unique_ptr< Decoder< float > > & | cellToForgetWeightsTensor, | ||
| std::unique_ptr< Decoder< float > > & | cellToOutputWeightsTensor, | ||
| std::unique_ptr< Decoder< float > > & | inputGateBiasTensor, | ||
| std::unique_ptr< Decoder< float > > & | forgetGateBiasTensor, | ||
| std::unique_ptr< Decoder< float > > & | cellBiasTensor, | ||
| std::unique_ptr< Decoder< float > > & | outputGateBiasTensor, | ||
| std::unique_ptr< Decoder< float > > & | projectionWeightsTensor, | ||
| std::unique_ptr< Decoder< float > > & | projectionBiasTensor, | ||
| std::unique_ptr< Decoder< float > > & | inputLayerNormWeights, | ||
| std::unique_ptr< Decoder< float > > & | forgetLayerNormWeights, | ||
| std::unique_ptr< Decoder< float > > & | cellLayerNormWeights, | ||
| std::unique_ptr< Decoder< float > > & | outputLayerNormWeights, | ||
| std::unique_ptr< Encoder< float > > & | inputGateScratch, | ||
| std::unique_ptr< Encoder< float > > & | cellScratch, | ||
| std::unique_ptr< Encoder< float > > & | forgetGateScratch, | ||
| std::unique_ptr< Encoder< float > > & | outputGateScratch, | ||
| std::unique_ptr< Decoder< float > > & | inputGateScratchDecoder, | ||
| std::unique_ptr< Decoder< float > > & | cellScratchDecoder, | ||
| std::unique_ptr< Decoder< float > > & | forgetGateScratchDecoder, | ||
| std::unique_ptr< Decoder< float > > & | outputGateScratchDecoder, | ||
| float | layerNormEpsilon ) |
Definition at line 13 of file Lstm.cpp.
References Activation, ClipVector(), CopyVector(), TensorInfo::GetDataType(), TensorInfo::GetShape(), LstmDescriptor::m_ActivationFunc, LstmDescriptor::m_CifgEnabled, LstmDescriptor::m_ClippingThresCell, LstmDescriptor::m_ClippingThresProj, LstmDescriptor::m_LayerNormEnabled, LstmDescriptor::m_PeepholeEnabled, LstmDescriptor::m_ProjectionEnabled, MatrixBatchVectorMultiplyAccumulate(), MeanStddevNormalization(), SetActivationParameters(), Sigmoid, Sub1Vector(), VectorBatchVectorAdd(), VectorBatchVectorAssign(), VectorBatchVectorCwiseProduct(), VectorBatchVectorCwiseProductAccumulate(), VectorVectorCwiseProduct(), VectorVectorCwiseProductAccumulate(), and ZeroVector().
|
inline |
Definition at line 64 of file Decoders.hpp.
References Boolean, Float16, Float32, info, QAsymmS8, QAsymmU8, QSymmS16, QSymmS8, and Signed32.
|
inline |
Definition at line 129 of file Decoders.hpp.
|
inline |
Definition at line 147 of file Decoders.hpp.
|
inline |
Definition at line 165 of file Decoders.hpp.
|
inline |
References info.
Referenced by Concatenate(), ExecuteFunction(), MirrorPad(), Pad(), RefTransposeConvolution2dWorkload::RefTransposeConvolution2dWorkload(), and Split().
|
inline |
Definition at line 19 of file Encoders.hpp.
References Float16, Float32, armnnUtils::GetPerAxisParams(), info, QAsymmS8, QAsymmU8, QSymmS16, QSymmS8, and Signed32.
|
inline |
Definition at line 94 of file Encoders.hpp.
|
inline |
Definition at line 112 of file Encoders.hpp.
|
inline |
Definition at line 130 of file Encoders.hpp.
|
inline |
References info.
Referenced by Concatenate(), ConvertToDataType(), ConvertToDataType(), ExecuteFunction(), MirrorPad(), Pad(), and Split().
| arm_compute::DetectionPostProcessLayerInfo MakeInfo | ( | const DetectionPostProcessDescriptor & | descriptor | ) |
Definition at line 17 of file NeonDetectionPostProcessWorkload.cpp.
References DetectionPostProcessDescriptor::m_DetectionsPerClass, DetectionPostProcessDescriptor::m_MaxClassesPerDetection, DetectionPostProcessDescriptor::m_MaxDetections, DetectionPostProcessDescriptor::m_NmsIouThreshold, DetectionPostProcessDescriptor::m_NmsScoreThreshold, DetectionPostProcessDescriptor::m_NumClasses, and DetectionPostProcessDescriptor::m_UseRegularNms.
Referenced by NeonDetectionPostProcessValidate(), and NeonDetectionPostProcessWorkload::NeonDetectionPostProcessWorkload().
| Optimizer::Optimizations MakeOptimizations | ( | Args &&... | args | ) |
Definition at line 43 of file Optimizer.hpp.
References Append().
Referenced by ApplyBackendOptimizations(), and IOptimizedNetwork::Optimize.
| Optional< T > MakeOptional | ( | Args &&... | args | ) |
Utility template that constructs an object of type T in-place and wraps it inside an Optional<T> object.
Definition at line 305 of file Optional.hpp.
References CONSTRUCT_IN_PLACE.
Referenced by TensorInfo::SetQuantizationOffset(), TensorInfo::TensorInfo(), TensorInfo::TensorInfo(), Convolution2dQueueDescriptor::Validate(), Convolution3dQueueDescriptor::Validate(), DepthwiseConvolution2dQueueDescriptor::Validate(), and TransposeConvolution2dQueueDescriptor::Validate().
|
constexpr |
Definition at line 90 of file TransformIterator.hpp.
| void MirrorPad | ( | const TensorInfo & | inputInfo, |
| const TensorInfo & | outputInfo, | ||
| const ITensorHandle * | inputHandle, | ||
| ITensorHandle * | outputHandle, | ||
| const PadQueueDescriptor & | data ) |
Definition at line 59 of file MirrorPad.cpp.
References Decoder< IType >::Get(), TensorShape::GetNumDimensions(), TensorInfo::GetNumElements(), TensorInfo::GetShape(), PadDescriptor::m_PaddingMode, PadDescriptor::m_PadList, QueueDescriptorWithParameters< LayerDescriptor >::m_Parameters, MakeDecoder(), MakeEncoder(), ITensorHandle::Map(), Reflect, Encoder< IType >::Set(), and Symmetric.
|
constexpr |
Definition at line 14 of file MockTensorHandleFactory.hpp.
Referenced by MockTensorHandleFactory::GetIdStatic().
| arm_compute::Status NeonAbsWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output ) |
Definition at line 17 of file NeonAbsWorkload.cpp.
Referenced by NeonLayerSupport::IsElementwiseUnarySupported().
| arm_compute::Status NeonActivationWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const ActivationDescriptor & | descriptor ) |
Definition at line 17 of file NeonActivationWorkload.cpp.
References ConvertActivationDescriptorToAclActivationLayerInfo().
Referenced by NeonLayerSupport::IsActivationSupported().
| arm_compute::Status NeonAdditionWorkloadValidate | ( | const TensorInfo & | input0, |
| const TensorInfo & | input1, | ||
| const TensorInfo & | output, | ||
| const ActivationDescriptor * | activationDescriptor ) |
Definition at line 20 of file NeonAdditionWorkload.cpp.
References ConvertActivationDescriptorToAclActivationLayerInfo().
Referenced by NeonLayerSupport::IsAdditionSupported(), IsLayerTypeSupported(), and NeonBackend::OptimizeSubgraphView().
| arm_compute::Status NeonArgMinMaxWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const ArgMinMaxDescriptor & | descriptor ) |
Definition at line 31 of file NeonArgMinMaxWorkload.cpp.
References TensorInfo::GetNumDimensions(), armnnUtils::GetUnsignedAxis(), ArgMinMaxDescriptor::m_Axis, ArgMinMaxDescriptor::m_Function, Max, and numeric_cast().
Referenced by NeonLayerSupport::IsArgMinMaxSupported().
|
constexpr |
Definition at line 10 of file NeonBackendId.hpp.
Referenced by NeonBackend::GetIdStatic().
| arm_compute::Status NeonBatchMatMulValidate | ( | const TensorInfo & | inputInfoX, |
| const TensorInfo & | inputInfoY, | ||
| const TensorInfo & | outputInfo, | ||
| const BatchMatMulDescriptor & | descriptor, | ||
| const bool | isFastMathEnabled, | ||
| const ActivationDescriptor * | activationDescriptor ) |
Definition at line 19 of file NeonBatchMatMulWorkload.cpp.
References ConvertActivationDescriptorToAclActivationLayerInfo(), BatchMatMulDescriptor::m_AdjointX, BatchMatMulDescriptor::m_AdjointY, BatchMatMulDescriptor::m_DataLayoutX, BatchMatMulDescriptor::m_DataLayoutY, BatchMatMulDescriptor::m_TransposeX, BatchMatMulDescriptor::m_TransposeY, and NCHW.
Referenced by NeonLayerSupport::IsBatchMatMulSupported().
| arm_compute::Status NeonBatchNormalizationValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const TensorInfo & | mean, | ||
| const TensorInfo & | var, | ||
| const TensorInfo & | beta, | ||
| const TensorInfo & | gamma, | ||
| const BatchNormalizationDescriptor & | descriptor, | ||
| const ActivationDescriptor * | activationDescriptor ) |
Definition at line 24 of file NeonBatchNormalizationWorkload.cpp.
References ConvertActivationDescriptorToAclActivationLayerInfo(), BatchNormalizationDescriptor::m_DataLayout, and BatchNormalizationDescriptor::m_Eps.
Referenced by NeonLayerSupport::IsBatchNormalizationSupported(), and NeonBackend::OptimizeSubgraphView().
| arm_compute::Status NeonBatchToSpaceNdWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const BatchToSpaceNdDescriptor & | descriptor ) |
Definition at line 15 of file NeonBatchToSpaceNdWorkload.cpp.
References CHECK_LOCATION, TensorInfo::GetNumDimensions(), BatchToSpaceNdDescriptor::m_BlockShape, BatchToSpaceNdDescriptor::m_DataLayout, NCHW, NHWC, and numeric_cast().
Referenced by NeonLayerSupport::IsBatchToSpaceNdSupported().
| arm_compute::Status NeonCastValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output ) |
Definition at line 19 of file NeonCastWorkload.cpp.
References Boolean, and TensorInfo::GetDataType().
Referenced by NeonLayerSupport::IsCastSupported().
| arm_compute::Status NeonChannelShuffleValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const ChannelShuffleDescriptor & | descriptor ) |
Definition at line 17 of file NeonChannelShuffleWorkload.cpp.
References TensorInfo::GetNumDimensions(), ChannelShuffleDescriptor::m_Axis, ChannelShuffleDescriptor::m_NumGroups, NCHW, and NHWC.
Referenced by NeonLayerSupport::IsChannelShuffleSupported().
| arm_compute::Status NeonComparisonWorkloadValidate | ( | const TensorInfo & | input0, |
| const TensorInfo & | input1, | ||
| const TensorInfo & | output, | ||
| const ComparisonDescriptor & | descriptor ) |
Definition at line 16 of file NeonComparisonWorkload.cpp.
References ConvertComparisonOperationToAcl().
Referenced by NeonLayerSupport::IsComparisonSupported().
| arm_compute::Status NeonConcatWorkloadValidate | ( | const std::vector< const TensorInfo * > & | inputs, |
| const TensorInfo & | output, | ||
| const OriginsDescriptor & | descriptor ) |
Definition at line 27 of file NeonConcatWorkload.cpp.
References NCHW.
Referenced by NeonLayerSupport::IsConcatSupported().
| arm_compute::Status NeonConstantWorkloadValidate | ( | const TensorInfo & | output | ) |
Definition at line 20 of file NeonConstantWorkload.cpp.
Referenced by NeonLayerSupport::IsConstantSupported().
| arm_compute::Status NeonConvertFp16ToFp32WorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output ) |
Definition at line 19 of file NeonConvertFp16ToFp32Workload.cpp.
References IgnoreUnused().
Referenced by NeonLayerSupport::IsConvertFp16ToFp32Supported().
| arm_compute::Status NeonConvertFp32ToFp16WorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output ) |
Definition at line 21 of file NeonConvertFp32ToFp16Workload.cpp.
References IgnoreUnused().
Referenced by NeonLayerSupport::IsConvertFp32ToFp16Supported().
| arm_compute::Status NeonConvolution2dWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const Convolution2dDescriptor & | descriptor, | ||
| const TensorInfo & | weights, | ||
| const Optional< TensorInfo > & | biases, | ||
| bool | isFastMathEnabled, | ||
| const ActivationDescriptor * | activationDescriptor ) |
Definition at line 24 of file NeonConvolution2dWorkload.cpp.
References ConvertActivationDescriptorToAclActivationLayerInfo(), OptionalBase::has_value(), TensorInfo::IsConstant(), Convolution2dDescriptor::m_BiasEnabled, Convolution2dDescriptor::m_DataLayout, Convolution2dDescriptor::m_DilationX, Convolution2dDescriptor::m_DilationY, and OptionalReferenceSwitch< IsReference, T >::value().
Referenced by NeonLayerSupport::IsConvolution2dSupported(), and NeonBackend::OptimizeSubgraphView().
| arm_compute::Status NeonConvolution3dWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const Convolution3dDescriptor & | descriptor, | ||
| const TensorInfo & | weights, | ||
| const Optional< TensorInfo > & | biases, | ||
| bool | isFastMathEnabled, | ||
| const ActivationDescriptor * | activationDescriptor ) |
Definition at line 24 of file NeonConvolution3dWorkload.cpp.
References ComputeConv3DInfo(), OptionalBase::has_value(), Convolution3dDescriptor::m_BiasEnabled, Convolution3dDescriptor::m_DataLayout, and OptionalReferenceSwitch< IsReference, T >::value().
Referenced by NeonLayerSupport::IsConvolution3dSupported().
| arm_compute::Status NeonDepthToSpaceWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const DepthToSpaceDescriptor & | descriptor ) |
Definition at line 19 of file NeonDepthToSpaceWorkload.cpp.
References SpaceToDepthDescriptor::m_BlockSize, SpaceToDepthDescriptor::m_DataLayout, and numeric_cast().
Referenced by NeonLayerSupport::IsDepthToSpaceSupported().
| arm_compute::Status NeonDepthwiseConvolutionWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const DepthwiseConvolution2dDescriptor & | descriptor, | ||
| const TensorInfo & | weights, | ||
| const Optional< TensorInfo > & | biases, | ||
| const ActivationDescriptor * | activationDescriptor ) |
Definition at line 29 of file NeonDepthwiseConvolutionWorkload.cpp.
References Convert1HWOTensorInfoToAcl(), ConvertActivationDescriptorToAclActivationLayerInfo(), OptionalBase::has_value(), TensorInfo::IsConstant(), DepthwiseConvolution2dDescriptor::m_BiasEnabled, DepthwiseConvolution2dDescriptor::m_DataLayout, DepthwiseConvolution2dDescriptor::m_DilationX, DepthwiseConvolution2dDescriptor::m_DilationY, and OptionalReferenceSwitch< IsReference, T >::value().
Referenced by NeonLayerSupport::IsDepthwiseConvolutionSupported(), NeonLayerSupport::IsDilatedDepthwiseConvolutionSupported(), and NeonBackend::OptimizeSubgraphView().
| arm_compute::Status NeonDequantizeWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output ) |
Definition at line 22 of file NeonDequantizeWorkload.cpp.
Referenced by NeonLayerSupport::IsDequantizeSupported().
| bool NeonDetected | ( | ) |
Definition at line 39 of file Utils.cpp.
| arm_compute::Status NeonDetectionPostProcessValidate | ( | const TensorInfo & | boxEncodings, |
| const TensorInfo & | scores, | ||
| const TensorInfo & | anchors, | ||
| const TensorInfo & | detectionBoxes, | ||
| const TensorInfo & | detectionClasses, | ||
| const TensorInfo & | detectionScores, | ||
| const TensorInfo & | numDetections, | ||
| const DetectionPostProcessDescriptor & | descriptor ) |
Definition at line 32 of file NeonDetectionPostProcessWorkload.cpp.
References info, and MakeInfo().
| arm_compute::Status NeonDivisionWorkloadValidate | ( | const TensorInfo & | input0, |
| const TensorInfo & | input1, | ||
| const TensorInfo & | output, | ||
| const ActivationDescriptor * | activationDescriptor ) |
Definition at line 18 of file NeonDivisionWorkload.cpp.
References ConvertActivationDescriptorToAclActivationLayerInfo().
Referenced by NeonLayerSupport::IsDivisionSupported(), IsLayerTypeSupported(), and NeonBackend::OptimizeSubgraphView().
| arm_compute::Status NeonElementwiseBinaryWorkloadValidate | ( | const TensorInfo & | input0, |
| const TensorInfo & | input1, | ||
| const TensorInfo & | output, | ||
| const ElementwiseBinaryDescriptor & | descriptor, | ||
| const ActivationDescriptor * | activationDescriptor ) |
Definition at line 20 of file NeonElementwiseBinaryWorkload.cpp.
References CHECK_LOCATION, ConvertActivationDescriptorToAclActivationLayerInfo(), ElementwiseBinaryDescriptor::m_Operation, Power, and SqDiff.
Referenced by IsLayerTypeSupported().
| arm_compute::Status NeonExpWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output ) |
Definition at line 17 of file NeonExpWorkload.cpp.
Referenced by NeonLayerSupport::IsElementwiseUnarySupported().
| arm_compute::Status NeonFloorDivWorkloadValidate | ( | const TensorInfo & | input0Info, |
| const TensorInfo & | input1Info, | ||
| const TensorInfo & | outputInfo, | ||
| const ActivationDescriptor * | activationDescriptor ) |
Validation for the Floor Div Workload.
Definition at line 52 of file NeonFloorDivWorkload.cpp.
References ConvertActivationDescriptorToAclActivationLayerInfo(), ConvertTensorToFloat32(), IsAllTensorsSigned32(), and IsValidationPassing().
Referenced by IsLayerTypeSupported().
| arm_compute::Status NeonFullyConnectedWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const TensorInfo & | weights, | ||
| const Optional< TensorInfo > & | biases, | ||
| const FullyConnectedDescriptor & | descriptor, | ||
| const ActivationDescriptor * | activationDescriptor ) |
Definition at line 24 of file NeonFullyConnectedWorkload.cpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, ConvertFullyConnectedDescriptorToAclFullyConnectedLayerInfo(), OptionalBase::has_value(), TensorInfo::IsConstant(), FullyConnectedDescriptor::m_BiasEnabled, and OptionalReferenceSwitch< IsReference, T >::value().
Referenced by NeonLayerSupport::IsFullyConnectedSupported(), and NeonBackend::OptimizeSubgraphView().
| arm_compute::Status NeonFusedWorkloadValidate | ( | const std::vector< std::reference_wrapper< TensorInfo > > & | inputInfos, |
| const std::vector< std::reference_wrapper< TensorInfo > > & | outputInfos, | ||
| const FusedDescriptor & | fusedDescriptor, | ||
| const ActivationDescriptor * | activationDescriptor ) |
Definition at line 22 of file NeonFusedWorkload.cpp.
References AddMulAdd, ConvertActivationDescriptorToAclActivationLayerInfo(), and FusedDescriptor::m_FusedKernelType.
Referenced by NeonLayerSupport::IsFusedSupported(), and NeonBackend::OptimizeSubgraphView().
| arm_compute::Status NeonGatherNdWorkloadValidate | ( | const TensorInfo & | paramsInfo, |
| const TensorInfo & | indicesInfo, | ||
| const TensorInfo & | outputInfo ) |
Validate Mul
Validate ReduceSum
Validate Gather
Validate Reshape
Return OK if all the layers are valid
Definition at line 14 of file NeonGatherNdWorkload.cpp.
References CalculateGatherNdKeyIndices(), ComputeAclAxis(), TensorInfo::GetNumDimensions(), and TensorInfo::SetShape().
Referenced by NeonLayerSupport::IsGatherNdSupported().
| arm_compute::Status NeonGatherWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | indices, | ||
| const TensorInfo & | output, | ||
| const GatherDescriptor & | descriptor ) |
Definition at line 13 of file NeonGatherWorkload.cpp.
References ComputeAclAxis(), and GatherDescriptor::m_Axis.
Referenced by NeonLayerSupport::IsGatherSupported().
| arm_compute::Status NeonInstanceNormalizationWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const InstanceNormalizationDescriptor & | descriptor ) |
Definition at line 19 of file NeonInstanceNormalizationWorkload.cpp.
References InstanceNormalizationDescriptor::m_Beta, InstanceNormalizationDescriptor::m_DataLayout, InstanceNormalizationDescriptor::m_Eps, and InstanceNormalizationDescriptor::m_Gamma.
Referenced by NeonLayerSupport::IsInstanceNormalizationSupported().
| arm_compute::Status NeonL2NormalizationWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const L2NormalizationDescriptor & | descriptor ) |
Definition at line 19 of file NeonL2NormalizationFloatWorkload.cpp.
References L2NormalizationDescriptor::m_DataLayout, L2NormalizationDescriptor::m_Eps, and NCHW.
Referenced by NeonLayerSupport::IsL2NormalizationSupported().
| arm_compute::Status NeonLogicalAndWorkloadValidate | ( | const TensorInfo & | input0, |
| const TensorInfo & | input1, | ||
| const TensorInfo & | output ) |
Definition at line 18 of file NeonLogicalAndWorkload.cpp.
Referenced by NeonLayerSupport::IsLogicalBinarySupported().
| arm_compute::Status NeonLogicalNotWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output ) |
Definition at line 19 of file NeonLogicalNotWorkload.cpp.
Referenced by NeonLayerSupport::IsElementwiseUnarySupported().
| arm_compute::Status NeonLogicalOrWorkloadValidate | ( | const TensorInfo & | input0, |
| const TensorInfo & | input1, | ||
| const TensorInfo & | output ) |
Definition at line 18 of file NeonLogicalOrWorkload.cpp.
Referenced by NeonLayerSupport::IsLogicalBinarySupported().
| arm_compute::Status NeonLogSoftmaxWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const LogSoftmaxDescriptor & | descriptor ) |
Definition at line 19 of file NeonLogSoftmaxWorkload.cpp.
References ComputeAclAxis(), SoftmaxDescriptor::m_Axis, and SoftmaxDescriptor::m_Beta.
Referenced by NeonLayerSupport::IsLogSoftmaxSupported().
| arm_compute::Status NeonLogWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output ) |
Definition at line 17 of file NeonLogWorkload.cpp.
Referenced by NeonLayerSupport::IsElementwiseUnarySupported().
| arm_compute::Status NeonLstmFloatWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | outputStateIn, | ||
| const TensorInfo & | cellStateIn, | ||
| const TensorInfo & | scratchBuffer, | ||
| const TensorInfo & | outputStateOut, | ||
| const TensorInfo & | cellStateOut, | ||
| const TensorInfo & | output, | ||
| const LstmDescriptor & | descriptor, | ||
| const LstmInputParamsInfo & | paramsInfo ) |
Definition at line 253 of file NeonLstmFloatWorkload.cpp.
References ConvertLstmActivationFuncToAclLayerInfo(), LstmInputParamsInfo::GetCellBias(), LstmInputParamsInfo::GetCellLayerNormWeights(), LstmInputParamsInfo::GetCellToForgetWeights(), LstmInputParamsInfo::GetCellToInputWeights(), LstmInputParamsInfo::GetCellToOutputWeights(), LstmInputParamsInfo::GetForgetGateBias(), LstmInputParamsInfo::GetForgetLayerNormWeights(), LstmInputParamsInfo::GetInputGateBias(), LstmInputParamsInfo::GetInputLayerNormWeights(), LstmInputParamsInfo::GetInputToCellWeights(), LstmInputParamsInfo::GetInputToForgetWeights(), LstmInputParamsInfo::GetInputToInputWeights(), LstmInputParamsInfo::GetInputToOutputWeights(), LstmInputParamsInfo::GetOutputGateBias(), LstmInputParamsInfo::GetOutputLayerNormWeights(), LstmInputParamsInfo::GetProjectionBias(), LstmInputParamsInfo::GetProjectionWeights(), LstmInputParamsInfo::GetRecurrentToCellWeights(), LstmInputParamsInfo::GetRecurrentToForgetWeights(), LstmInputParamsInfo::GetRecurrentToInputWeights(), LstmInputParamsInfo::GetRecurrentToOutputWeights(), LstmDescriptor::m_ActivationFunc, LstmDescriptor::m_CifgEnabled, LstmDescriptor::m_ClippingThresCell, LstmDescriptor::m_ClippingThresProj, LstmDescriptor::m_LayerNormEnabled, LstmDescriptor::m_PeepholeEnabled, LstmInputParamsInfo::m_ProjectionBias, and LstmDescriptor::m_ProjectionEnabled.
Referenced by NeonLayerSupport::IsLstmSupported().
| arm_compute::Status NeonMaximumWorkloadValidate | ( | const TensorInfo & | input0, |
| const TensorInfo & | input1, | ||
| const TensorInfo & | output ) |
Definition at line 14 of file NeonMaximumWorkload.cpp.
Referenced by IsLayerTypeSupported(), and NeonLayerSupport::IsMaximumSupported().
| arm_compute::Status NeonMeanWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const MeanDescriptor & | descriptor ) |
Definition at line 18 of file NeonMeanWorkload.cpp.
References TensorInfo::GetNumDimensions(), MeanDescriptor::m_Axis, and MeanDescriptor::m_KeepDims.
Referenced by NeonLayerSupport::IsMeanSupported().
| arm_compute::Status NeonMinimumWorkloadValidate | ( | const TensorInfo & | input0, |
| const TensorInfo & | input1, | ||
| const TensorInfo & | output ) |
Validate function for validating the inputs and output.
| [in] | input0 | The input0 value to be validated. |
| [in] | input1 | The input1 value to be validated. |
| [in] | output | The output value to be validated. |
Definition at line 15 of file NeonMinimumWorkload.cpp.
Referenced by IsLayerTypeSupported(), and NeonLayerSupport::IsMinimumSupported().
| arm_compute::Status NeonMultiplicationWorkloadValidate | ( | const TensorInfo & | input0, |
| const TensorInfo & | input1, | ||
| const TensorInfo & | output, | ||
| const ActivationDescriptor * | activationDescriptor ) |
Definition at line 19 of file NeonMultiplicationWorkload.cpp.
References ConvertActivationDescriptorToAclActivationLayerInfo(), TensorInfo::GetDataType(), and IsQuantizedType().
Referenced by IsLayerTypeSupported(), NeonLayerSupport::IsMultiplicationSupported(), and NeonBackend::OptimizeSubgraphView().
| arm_compute::Status NeonNegWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output ) |
Definition at line 17 of file NeonNegWorkload.cpp.
Referenced by NeonLayerSupport::IsElementwiseUnarySupported().
| arm_compute::Status NeonNormalizationWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const NormalizationDescriptor & | descriptor ) |
Definition at line 49 of file NeonNormalizationFloatWorkload.cpp.
References NormalizationDescriptor::m_DataLayout.
Referenced by NeonLayerSupport::IsNormalizationSupported().
| arm_compute::Status NeonPadWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const PadDescriptor & | descriptor ) |
Definition at line 59 of file NeonPadWorkload.cpp.
References ConvertPaddingModeToAcl(), PadDescriptor::m_PaddingMode, PadDescriptor::m_PadList, and PadDescriptor::m_PadValue.
Referenced by NeonLayerSupport::IsPadSupported().
| arm_compute::Status NeonPermuteWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const PermuteDescriptor & | descriptor ) |
Definition at line 15 of file NeonPermuteWorkload.cpp.
References PermuteDescriptor::m_DimMappings.
Referenced by NeonLayerSupport::IsPermuteSupported().
| arm_compute::Status NeonPooling2dWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const Pooling2dDescriptor & | descriptor ) |
Definition at line 22 of file NeonPooling2dWorkload.cpp.
References Pooling2dDescriptor::m_DataLayout.
Referenced by NeonLayerSupport::IsPooling2dSupported().
| arm_compute::Status NeonPooling3dWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const Pooling3dDescriptor & | descriptor ) |
Definition at line 15 of file NeonPooling3dWorkload.cpp.
References Pooling3dDescriptor::m_DataLayout.
Referenced by NeonLayerSupport::IsPooling3dSupported().
| arm_compute::Status NeonPreluWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | alpha, | ||
| const TensorInfo & | output ) |
Definition at line 17 of file NeonPreluWorkload.cpp.
Referenced by NeonLayerSupport::IsPreluSupported().
| arm_compute::Status NeonQLstmWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | cellStateIn, | ||
| const TensorInfo & | outputStateIn, | ||
| const TensorInfo & | cellStateOut, | ||
| const TensorInfo & | outputStateOut, | ||
| const TensorInfo & | output, | ||
| const QLstmDescriptor & | descriptor, | ||
| const LstmInputParamsInfo & | paramsInfo ) |
Definition at line 243 of file NeonQLstmWorkload.cpp.
References LstmInputParamsInfo::GetCellBias(), LstmInputParamsInfo::GetCellLayerNormWeights(), LstmInputParamsInfo::GetCellToForgetWeights(), LstmInputParamsInfo::GetCellToInputWeights(), LstmInputParamsInfo::GetCellToOutputWeights(), LstmInputParamsInfo::GetForgetGateBias(), LstmInputParamsInfo::GetForgetLayerNormWeights(), LstmInputParamsInfo::GetInputGateBias(), LstmInputParamsInfo::GetInputLayerNormWeights(), LstmInputParamsInfo::GetInputToCellWeights(), LstmInputParamsInfo::GetInputToForgetWeights(), LstmInputParamsInfo::GetInputToInputWeights(), LstmInputParamsInfo::GetInputToOutputWeights(), LstmInputParamsInfo::GetOutputGateBias(), LstmInputParamsInfo::GetOutputLayerNormWeights(), LstmInputParamsInfo::GetProjectionBias(), LstmInputParamsInfo::GetProjectionWeights(), LstmInputParamsInfo::GetRecurrentToCellWeights(), LstmInputParamsInfo::GetRecurrentToForgetWeights(), LstmInputParamsInfo::GetRecurrentToInputWeights(), LstmInputParamsInfo::GetRecurrentToOutputWeights(), QLstmDescriptor::m_CellClip, QLstmDescriptor::m_CellIntermediateScale, LstmInputParamsInfo::m_CellToInputWeights, QLstmDescriptor::m_CifgEnabled, QLstmDescriptor::m_ForgetIntermediateScale, QLstmDescriptor::m_HiddenStateScale, QLstmDescriptor::m_HiddenStateZeroPoint, QLstmDescriptor::m_InputIntermediateScale, LstmInputParamsInfo::m_InputLayerNormWeights, QLstmDescriptor::m_LayerNormEnabled, QLstmDescriptor::m_OutputIntermediateScale, QLstmDescriptor::m_PeepholeEnabled, LstmInputParamsInfo::m_ProjectionBias, QLstmDescriptor::m_ProjectionClip, and QLstmDescriptor::m_ProjectionEnabled.
Referenced by NeonLayerSupport::IsQLstmSupported().
| arm_compute::Status NeonQuantizedLstmWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | cellStateIn, | ||
| const TensorInfo & | outputStateIn, | ||
| const TensorInfo & | cellStateOut, | ||
| const TensorInfo & | outputStateOut, | ||
| const QuantizedLstmInputParamsInfo & | paramsInfo ) |
Definition at line 131 of file NeonQuantizedLstmWorkload.cpp.
References QuantizedLstmInputParamsInfo::GetCellBias(), QuantizedLstmInputParamsInfo::GetForgetGateBias(), QuantizedLstmInputParamsInfo::GetInputGateBias(), QuantizedLstmInputParamsInfo::GetInputToCellWeights(), QuantizedLstmInputParamsInfo::GetInputToForgetWeights(), QuantizedLstmInputParamsInfo::GetInputToInputWeights(), QuantizedLstmInputParamsInfo::GetInputToOutputWeights(), QuantizedLstmInputParamsInfo::GetOutputGateBias(), QuantizedLstmInputParamsInfo::GetRecurrentToCellWeights(), QuantizedLstmInputParamsInfo::GetRecurrentToForgetWeights(), QuantizedLstmInputParamsInfo::GetRecurrentToInputWeights(), and QuantizedLstmInputParamsInfo::GetRecurrentToOutputWeights().
Referenced by NeonLayerSupport::IsQuantizedLstmSupported().
| arm_compute::Status NeonQuantizeWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output ) |
Definition at line 18 of file NeonQuantizeWorkload.cpp.
Referenced by NeonLayerSupport::IsQuantizeSupported().
| arm_compute::Status NeonReduceWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const ReduceDescriptor & | descriptor ) |
Definition at line 19 of file NeonReduceWorkload.cpp.
References ConvertReductionOperationToAcl(), TensorInfo::GetNumDimensions(), IsMultiAxesReduceSupported(), ReduceDescriptor::m_KeepDims, ReduceDescriptor::m_vAxis, and NeonReduceWorkloadValidate().
Referenced by NeonLayerSupport::IsReduceSupported(), and NeonReduceWorkloadValidate().
| arm_compute::Status NeonReshapeWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output ) |
Definition at line 17 of file NeonReshapeWorkload.cpp.
Referenced by NeonLayerSupport::IsReshapeSupported().
| arm_compute::Status NeonResizeWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const ResizeDescriptor & | descriptor ) |
Definition at line 22 of file NeonResizeWorkload.cpp.
References ConvertResizeMethodToAclInterpolationPolicy(), ResizeDescriptor::m_AlignCorners, ResizeDescriptor::m_DataLayout, ResizeDescriptor::m_HalfPixelCenters, and ResizeDescriptor::m_Method.
Referenced by NeonLayerSupport::IsResizeSupported().
| arm_compute::Status NeonReverseV2WorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | axis, | ||
| const TensorInfo & | output ) |
Definition at line 14 of file NeonReverseV2Workload.cpp.
Referenced by NeonLayerSupport::IsReverseV2Supported().
| arm_compute::Status NeonRsqrtWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output ) |
Definition at line 18 of file NeonRsqrtWorkload.cpp.
Referenced by NeonLayerSupport::IsElementwiseUnarySupported().
| arm_compute::Status NeonSinWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output ) |
Definition at line 17 of file NeonSinWorkload.cpp.
Referenced by NeonLayerSupport::IsElementwiseUnarySupported().
| arm_compute::Status NeonSliceWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const SliceDescriptor & | descriptor ) |
Definition at line 21 of file NeonSliceWorkload.cpp.
References SliceDescriptor::m_Begin, SliceDescriptor::m_Size, and SetNeonSliceData().
Referenced by NeonLayerSupport::IsSliceSupported().
| arm_compute::Status NeonSoftmaxWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const SoftmaxDescriptor & | descriptor ) |
Definition at line 19 of file NeonSoftmaxWorkload.cpp.
References ComputeAclAxis(), SoftmaxDescriptor::m_Axis, and SoftmaxDescriptor::m_Beta.
Referenced by NeonLayerSupport::IsSoftmaxSupported().
| arm_compute::Status NeonSpaceToBatchNdWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const SpaceToBatchNdDescriptor & | descriptor ) |
Definition at line 15 of file NeonSpaceToBatchNdWorkload.cpp.
References CHECK_LOCATION, TensorInfo::GetNumDimensions(), SpaceToBatchNdDescriptor::m_BlockShape, SpaceToBatchNdDescriptor::m_DataLayout, SpaceToBatchNdDescriptor::m_PadList, NCHW, NHWC, and numeric_cast().
Referenced by NeonLayerSupport::IsSpaceToBatchNdSupported().
| arm_compute::Status NeonSpaceToDepthWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const SpaceToDepthDescriptor & | descriptor ) |
Definition at line 19 of file NeonSpaceToDepthWorkload.cpp.
References SpaceToDepthDescriptor::m_BlockSize, SpaceToDepthDescriptor::m_DataLayout, and numeric_cast().
Referenced by NeonLayerSupport::IsSpaceToDepthSupported().
| arm_compute::Status NeonSplitterWorkloadValidate | ( | const TensorInfo & | input, |
| const std::vector< std::reference_wrapper< TensorInfo > > & | outputs, | ||
| unsigned int | splitAxis ) |
Definition at line 33 of file NeonSplitterWorkload.cpp.
References TensorInfo::GetNumDimensions().
Referenced by NeonLayerSupport::IsSplitterSupported().
| arm_compute::Status NeonSqrtWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output ) |
Definition at line 18 of file NeonSqrtWorkload.cpp.
References ConvertActivationDescriptorToAclActivationLayerInfo(), ActivationDescriptor::m_Function, and Sqrt.
Referenced by NeonLayerSupport::IsElementwiseUnarySupported().
| arm_compute::Status NeonStackWorkloadValidate | ( | const std::vector< const TensorInfo * > & | inputs, |
| const TensorInfo & | output, | ||
| const StackDescriptor & | descriptor ) |
Definition at line 27 of file NeonStackWorkload.cpp.
References TensorShape::GetNumDimensions(), StackDescriptor::m_Axis, StackDescriptor::m_InputShape, and NCHW.
Referenced by NeonLayerSupport::IsStackSupported().
| arm_compute::Status NeonStridedSliceWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const StridedSliceDescriptor & | descriptor ) |
Definition at line 19 of file NeonStridedSliceWorkload.cpp.
References ConvertMaskToACLFormat(), TensorInfo::GetNumDimensions(), StridedSliceDescriptor::m_Begin, StridedSliceDescriptor::m_BeginMask, StridedSliceDescriptor::m_DataLayout, StridedSliceDescriptor::m_End, StridedSliceDescriptor::m_EndMask, StridedSliceDescriptor::m_ShrinkAxisMask, StridedSliceDescriptor::m_Stride, numeric_cast(), and SetNeonStridedSliceData().
Referenced by NeonLayerSupport::IsStridedSliceSupported().
| arm_compute::Status NeonSubtractionWorkloadValidate | ( | const TensorInfo & | input0, |
| const TensorInfo & | input1, | ||
| const TensorInfo & | output, | ||
| const ActivationDescriptor * | activationDescriptor ) |
Definition at line 22 of file NeonSubtractionWorkload.cpp.
References ConvertActivationDescriptorToAclActivationLayerInfo().
Referenced by IsLayerTypeSupported(), NeonLayerSupport::IsSubtractionSupported(), and NeonBackend::OptimizeSubgraphView().
|
constexpr |
Definition at line 14 of file NeonTensorHandleFactory.hpp.
Referenced by NeonTensorHandleFactory::GetIdStatic().
| arm_compute::Status NeonTileWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const TileDescriptor & | descriptor ) |
Definition at line 14 of file NeonTileWorkload.cpp.
References Boolean, TensorInfo::GetDataType(), and TileDescriptor::m_Multiples.
Referenced by NeonLayerSupport::IsTileSupported().
| arm_compute::Status NeonTransposeConvolution2dWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const TransposeConvolution2dDescriptor & | descriptor, | ||
| const TensorInfo & | weights, | ||
| const Optional< TensorInfo > & | biases ) |
Definition at line 25 of file NeonTransposeConvolution2dWorkload.cpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, OptionalBase::has_value(), TransposeConvolution2dDescriptor::m_BiasEnabled, TransposeConvolution2dDescriptor::m_DataLayout, and OptionalReferenceSwitch< IsReference, T >::value().
Referenced by NeonLayerSupport::IsTransposeConvolution2dSupported().
| arm_compute::Status NeonTransposeWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | output, | ||
| const TransposeDescriptor & | descriptor ) |
Definition at line 15 of file NeonTransposeWorkload.cpp.
References TransposeDescriptor::m_DimMappings.
Referenced by NeonLayerSupport::IsTransposeSupported().
| arm_compute::Status NeonUnidirectionalSequenceLstmFloatWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | outputStateIn, | ||
| const TensorInfo & | cellStateIn, | ||
| const TensorInfo & | outputStateOut, | ||
| const TensorInfo & | cellStateOut, | ||
| const TensorInfo & | output, | ||
| const UnidirectionalSequenceLstmDescriptor & | descriptor, | ||
| const LstmInputParamsInfo & | paramsInfo ) |
Definition at line 510 of file NeonUnidirectionalSequenceLstmFloatWorkload.cpp.
References ConvertLstmActivationFuncToAclLayerInfo(), LstmInputParamsInfo::GetCellBias(), LstmInputParamsInfo::GetCellLayerNormWeights(), LstmInputParamsInfo::GetCellToForgetWeights(), LstmInputParamsInfo::GetCellToInputWeights(), LstmInputParamsInfo::GetCellToOutputWeights(), TensorInfo::GetDataType(), LstmInputParamsInfo::GetForgetGateBias(), LstmInputParamsInfo::GetForgetLayerNormWeights(), LstmInputParamsInfo::GetInputGateBias(), LstmInputParamsInfo::GetInputLayerNormWeights(), LstmInputParamsInfo::GetInputToCellWeights(), LstmInputParamsInfo::GetInputToForgetWeights(), LstmInputParamsInfo::GetInputToInputWeights(), LstmInputParamsInfo::GetInputToOutputWeights(), TensorShape::GetNumDimensions(), LstmInputParamsInfo::GetOutputGateBias(), LstmInputParamsInfo::GetOutputLayerNormWeights(), LstmInputParamsInfo::GetProjectionBias(), LstmInputParamsInfo::GetProjectionWeights(), LstmInputParamsInfo::GetRecurrentToCellWeights(), LstmInputParamsInfo::GetRecurrentToForgetWeights(), LstmInputParamsInfo::GetRecurrentToInputWeights(), LstmInputParamsInfo::GetRecurrentToOutputWeights(), TensorInfo::GetShape(), armnnUtils::GetTensorShape(), LstmDescriptor::m_ActivationFunc, LstmDescriptor::m_CifgEnabled, LstmDescriptor::m_ClippingThresCell, LstmDescriptor::m_ClippingThresProj, LstmDescriptor::m_LayerNormEnabled, LstmDescriptor::m_PeepholeEnabled, LstmInputParamsInfo::m_ProjectionBias, LstmDescriptor::m_ProjectionEnabled, LstmDescriptor::m_TimeMajor, and armnnUtils::Permuted().
Referenced by NeonLayerSupport::IsUnidirectionalSequenceLstmSupported().
| arm_compute::Status NeonUnidirectionalSequenceLstmWorkloadValidate | ( | const TensorInfo & | input, |
| const TensorInfo & | outputStateIn, | ||
| const TensorInfo & | cellStateIn, | ||
| const TensorInfo & | outputStateOut, | ||
| const TensorInfo & | cellStateOut, | ||
| const TensorInfo & | output, | ||
| const UnidirectionalSequenceLstmDescriptor & | descriptor, | ||
| const LstmInputParamsInfo & | paramsInfo ) |
Definition at line 491 of file NeonUnidirectionalSequenceLstmWorkload.cpp.
References LstmInputParamsInfo::GetCellBias(), LstmInputParamsInfo::GetCellLayerNormWeights(), LstmInputParamsInfo::GetCellToForgetWeights(), LstmInputParamsInfo::GetCellToInputWeights(), LstmInputParamsInfo::GetCellToOutputWeights(), TensorInfo::GetDataType(), LstmInputParamsInfo::GetForgetGateBias(), LstmInputParamsInfo::GetForgetLayerNormWeights(), LstmInputParamsInfo::GetInputGateBias(), LstmInputParamsInfo::GetInputLayerNormWeights(), LstmInputParamsInfo::GetInputToCellWeights(), LstmInputParamsInfo::GetInputToForgetWeights(), LstmInputParamsInfo::GetInputToInputWeights(), LstmInputParamsInfo::GetInputToOutputWeights(), TensorShape::GetNumDimensions(), LstmInputParamsInfo::GetOutputGateBias(), LstmInputParamsInfo::GetOutputLayerNormWeights(), LstmInputParamsInfo::GetProjectionBias(), LstmInputParamsInfo::GetProjectionWeights(), LstmInputParamsInfo::GetRecurrentToCellWeights(), LstmInputParamsInfo::GetRecurrentToForgetWeights(), LstmInputParamsInfo::GetRecurrentToInputWeights(), LstmInputParamsInfo::GetRecurrentToOutputWeights(), TensorInfo::GetShape(), armnnUtils::GetTensorShape(), LstmDescriptor::m_CellIntermediateScale, LstmDescriptor::m_CifgEnabled, LstmDescriptor::m_ClippingThresCell, LstmDescriptor::m_ClippingThresProj, LstmDescriptor::m_ForgetIntermediateScale, LstmDescriptor::m_HiddenStateScale, LstmDescriptor::m_HiddenStateZeroPoint, LstmDescriptor::m_InputIntermediateScale, LstmDescriptor::m_LayerNormEnabled, LstmDescriptor::m_OutputIntermediateScale, LstmDescriptor::m_PeepholeEnabled, LstmInputParamsInfo::m_ProjectionBias, LstmDescriptor::m_ProjectionEnabled, LstmDescriptor::m_TimeMajor, and armnnUtils::Permuted().
Referenced by NeonLayerSupport::IsUnidirectionalSequenceLstmSupported().
| bool NextIndex | ( | const unsigned int | numDims, |
| const armnn::TensorShape & | dims, | ||
| std::vector< unsigned int > & | current ) |
Definition at line 19 of file Reduce.cpp.
Referenced by Reduce().
| std::vector< unsigned int > NonMaxSuppression | ( | unsigned int | numBoxes, |
| const std::vector< float > & | boxCorners, | ||
| const std::vector< float > & | scores, | ||
| float | nmsScoreThreshold, | ||
| unsigned int | maxDetection, | ||
| float | nmsIouThreshold ) |
Definition at line 48 of file DetectionPostProcess.cpp.
References GenerateRangeK(), IntersectionOverUnion(), numeric_cast(), and TopKSort().
Referenced by DetectionPostProcess().
| std::enable_if_t< std::is_unsigned< Source >::value &&std::is_unsigned< Dest >::value, Dest > numeric_cast | ( | Source | source | ) |
Definition at line 35 of file NumericCast.hpp.
References ARMNN_NUMERIC_CAST_CHECK.
Referenced by Graph::AddCompatibilityLayers(), AllocateOutputData(), ArgMinMax(), ClArgMinMaxWorkload::ClArgMinMaxWorkload(), ClArgMinMaxWorkloadValidate(), ClBatchToSpaceNdWorkload::ClBatchToSpaceNdWorkload(), ClBatchToSpaceNdWorkloadValidate(), ClDepthToSpaceWorkload::ClDepthToSpaceWorkload(), ClDepthToSpaceWorkloadValidate(), ClLstmFloatWorkload::ClLstmFloatWorkload(), ClSpaceToBatchNdWorkload::ClSpaceToBatchNdWorkload(), ClSpaceToBatchNdWorkloadValidate(), ClSpaceToDepthWorkload::ClSpaceToDepthWorkload(), ClSpaceToDepthWorkloadValidate(), ClStridedSliceWorkload::ClStridedSliceWorkload(), ClStridedSliceWorkloadValidate(), ComputeReductionTensorShape(), armnnTfLiteParser::ComputeWrappedIndex(), OutputSlot::Connect(), ConvertMaskToACLFormat(), ClImportTensorHandleFactory::CreateSubTensorHandle(), ClTensorHandleFactory::CreateSubTensorHandle(), ClWorkloadFactory::CreateSubTensorHandle(), NeonTensorHandleFactory::CreateSubTensorHandle(), NeonWorkloadFactory::CreateSubTensorHandle(), Dequantize(), DetectionPostProcess(), armnnUtils::ExpandDims(), FakeQuantization(), OutputSlot::GetNumConnections(), SubgraphView::GetNumInputSlots(), SubgraphView::GetNumOutputSlots(), StridedSliceDescriptor::GetStartForAxis(), StridedSliceDescriptor::GetStopForAxis(), armnnUtils::GetUnsignedAxis(), MeanLayer::InferOutputShapes(), PreluLayer::InferOutputShapes(), ReduceLayer::InferOutputShapes(), SliceLayer::InferOutputShapes(), StridedSliceLayer::InferOutputShapes(), InputSlot::Insert(), IsLayerTypeSupported(), RefLayerSupport::IsMeanSupported(), LogSoftmax(), NeonArgMinMaxWorkload::NeonArgMinMaxWorkload(), NeonArgMinMaxWorkloadValidate(), NeonBatchToSpaceNdWorkload::NeonBatchToSpaceNdWorkload(), NeonBatchToSpaceNdWorkloadValidate(), NeonDepthToSpaceWorkload::NeonDepthToSpaceWorkload(), NeonDepthToSpaceWorkloadValidate(), NeonLstmFloatWorkload::NeonLstmFloatWorkload(), NeonSpaceToBatchNdWorkload::NeonSpaceToBatchNdWorkload(), NeonSpaceToBatchNdWorkloadValidate(), NeonSpaceToDepthWorkload::NeonSpaceToDepthWorkload(), NeonSpaceToDepthWorkloadValidate(), NeonStridedSliceWorkload::NeonStridedSliceWorkload(), NeonStridedSliceWorkloadValidate(), NonMaxSuppression(), power< T >::operator()(), squaredDifference< T >::operator()(), IDeserializer::DeserializerImpl::OutputShapeOfReshape(), TfLiteParserImpl::OutputShapeOfReshape(), PermutationVector::PermutationVector(), Pooling2d(), Pooling3d(), armnnUtils::QuantizedVector(), Reduce(), ClContextSerializer::SaveSerializedToStream(), ISerializer::SerializerImpl::SaveSerializedToStream(), StridedSlice(), TensorShape::TensorShape(), and MeanQueueDescriptor::Validate().
| std::enable_if_t< std::is_signed< Source >::value &&std::is_integral< Source >::value &&std::is_signed< Dest >::value &&std::is_integral< Dest >::value, Dest > numeric_cast | ( | Source | source | ) |
Definition at line 58 of file NumericCast.hpp.
References ARMNN_NUMERIC_CAST_CHECK.
| std::enable_if_t< std::is_floating_point< Source >::value &&std::is_floating_point< Dest >::value, Dest > numeric_cast | ( | Source | source | ) |
Definition at line 83 of file NumericCast.hpp.
References ARMNN_NUMERIC_CAST_CHECK.
| std::enable_if_t< std::is_floating_point< Source >::value &&std::is_signed< Dest >::value &&std::is_integral< Dest >::value, Dest > numeric_cast | ( | Source | source | ) |
Definition at line 109 of file NumericCast.hpp.
References ARMNN_NUMERIC_CAST_CHECK.
| std::enable_if_t< std::is_signed< Source >::value &&std::is_integral< Source >::value &&std::is_floating_point< Dest >::value, Dest > numeric_cast | ( | Source | source | ) |
Definition at line 135 of file NumericCast.hpp.
References ARMNN_NUMERIC_CAST_CHECK.
| std::enable_if_t< std::is_signed< Dest >::value &&std::is_integral< Dest >::value &&std::is_unsigned< Source >::value, Dest > numeric_cast | ( | Source | sValue | ) |
Definition at line 165 of file NumericCast.hpp.
References ARMNN_NUMERIC_CAST_CHECK.
| std::enable_if_t< std::is_floating_point< Dest >::value &&std::is_unsigned< Source >::value, Dest > numeric_cast | ( | Source | sValue | ) |
Definition at line 184 of file NumericCast.hpp.
References ARMNN_NUMERIC_CAST_CHECK.
| std::enable_if_t< std::is_unsigned< Dest >::value &&std::is_signed< Source >::value &&std::is_integral< Source >::value, Dest > numeric_cast | ( | Source | sValue | ) |
Definition at line 206 of file NumericCast.hpp.
References ARMNN_NUMERIC_CAST_CHECK.
| std::enable_if_t< std::is_unsigned< Dest >::value &&std::is_floating_point< Source >::value, Dest > numeric_cast | ( | Source | sValue | ) |
Definition at line 230 of file NumericCast.hpp.
References ARMNN_NUMERIC_CAST_CHECK.
| unsigned int Offset | ( | const TensorShape & | shape, |
| unsigned int | batch, | ||
| unsigned int | height, | ||
| unsigned int | width, | ||
| unsigned int | channels, | ||
| const DataLayoutIndexed & | dataLayout ) |
Definition at line 15 of file BatchToSpaceNd.cpp.
References CHECK_LOCATION, DataLayoutIndexed::GetChannelsIndex(), DataLayoutIndexed::GetDataLayout(), DataLayoutIndexed::GetHeightIndex(), TensorShape::GetNumDimensions(), DataLayoutIndexed::GetWidthIndex(), NCHW, and NHWC.
Referenced by BatchToSpaceNd().
|
inline |
Definition at line 336 of file TypesUtils.hpp.
References TensorShape::GetDimensionality(), TensorShape::GetDimensionSpecificity(), TensorShape::GetNumDimensions(), and NotSpecified.
|
inline |
Definition at line 179 of file BackendId.hpp.
|
inline |
Definition at line 68 of file IBackendInternal.hpp.
References BackendVersion::m_Major, and BackendVersion::m_Minor.
|
inline |
Definition at line 122 of file BFloat16.hpp.
References BFloat16::ToFloat32(), and BFloat16::Val().
|
inline |
Deprecated function that will be removed together with the Compute enum.
Definition at line 72 of file BackendId.hpp.
References GetComputeDeviceAsCString().
|
inline |
Deprecated function that will be removed together with the Compute enum.
Definition at line 61 of file BackendId.hpp.
References GetComputeDeviceAsCString().
|
inline |
Deprecated function that will be removed together with the Compute enum.
Definition at line 50 of file BackendId.hpp.
References GetComputeDeviceAsCString().
| std::ostream & operator<< | ( | std::ostream & | os, |
| const TContainer< BackendId, TContainerTemplateArgs... > & | ids ) |
Definition at line 186 of file BackendId.hpp.
|
inline |
Definition at line 329 of file TypesUtils.hpp.
References GetStatusAsCString().
| IOptimizedNetworkPtr Optimize | ( | const Graph & | inGraph, |
| const std::vector< BackendId > & | backendPreferences, | ||
| const IDeviceSpec & | deviceSpec, | ||
| const OptimizerOptions & | options, | ||
| Optional< std::vector< std::string > & > | messages = EmptyOptional() ) |
Accept legacy OptimizerOptions.
Definition at line 2013 of file Network.cpp.
References Optimize().
| IOptimizedNetworkPtr Optimize | ( | const Graph & | inGraph, |
| const std::vector< BackendId > & | backendPreferences, | ||
| const IDeviceSpec & | deviceSpec, | ||
| const OptimizerOptionsOpaque & | options, | ||
| Optional< std::vector< std::string > & > | messages = EmptyOptional() ) |
Create an optimized version of the network.
| inGraph | Graph to be optimized. |
| backendPreferences | The choice of the backend ordered by user preferences. |
| deviceSpec | DeviceSpec object as queried from the runtime. See IRuntime::GetDeviceSpec() |
| messages | If there are failures or warnings a string describing same will be added to the vector |
| options | OptimizerOptions object with optimizer configuration options |
Definition at line 2026 of file Network.cpp.
| IOptimizedNetworkPtr Optimize | ( | const INetwork & | network, |
| const std::vector< BackendId > & | backendPreferences, | ||
| const IDeviceSpec & | deviceSpec, | ||
| const OptimizerOptions & | options, | ||
| Optional< std::vector< std::string > & > | messages = EmptyOptional() ) |
Accept legacy OptimizerOptions.
Definition at line 2274 of file Network.cpp.
| IOptimizedNetworkPtr Optimize | ( | const INetwork & | network, |
| const std::vector< BackendId > & | backendPreferences, | ||
| const IDeviceSpec & | deviceSpec, | ||
| const OptimizerOptionsOpaque & | options = OptimizerOptionsOpaque(), | ||
| Optional< std::vector< std::string > & > | messages = EmptyOptional() ) |
Create an optimized version of the network.
| network | INetwork description of the network to be optimized. |
| backendPreferences | The choice of the backend ordered by user preferences. |
| deviceSpec | DeviceSpec object as queried from the runtime. See IRuntime::GetDeviceSpec() |
| messages | If there are failures or warnings a string describing same will be added to the vector |
| options | OptimizerOptions object with optimizer configuration options |
Definition at line 2287 of file Network.cpp.
Referenced by INetwork::Optimize, INetwork::Optimize, Optimize(), ParserPrototxtFixture< TParser >::Setup(), ParserPrototxtFixture< TParser >::Setup(), ParserPrototxtFixture< TParser >::Setup(), and ParserPrototxtFixture< TParser >::SetupOptimizedNetwork().
| void Pad | ( | const TensorInfo & | inputInfo, |
| const TensorInfo & | outputInfo, | ||
| const ITensorHandle * | inputHandle, | ||
| ITensorHandle * | outputHandle, | ||
| const PadQueueDescriptor & | data ) |
Definition at line 39 of file Pad.cpp.
References Decoder< IType >::Get(), TensorInfo::GetDataType(), TensorShape::GetNumDimensions(), TensorInfo::GetNumElements(), TensorInfo::GetShape(), TensorInfo::IsQuantized(), PadDescriptor::m_PadList, PadDescriptor::m_PadValue, QueueDescriptorWithParameters< LayerDescriptor >::m_Parameters, MakeDecoder(), MakeEncoder(), ITensorHandle::Map(), and Encoder< IType >::Set().
Referenced by PadLayer::PadLayer().
|
inline |
Definition at line 312 of file BackendOptions.hpp.
References BackendOptions::Var::AsBool(), and BackendOptions::Var::IsBool().
|
constexpr |
Deprecated function that will be removed together with the Compute enum.
Definition at line 214 of file TypesUtils.hpp.
References CpuAcc, CpuRef, GpuAcc, StrEqual(), and Undefined.
|
inline |
Definition at line 330 of file BackendOptions.hpp.
References BackendOptions::Var::AsInt(), and BackendOptions::Var::IsInt().
| void ParseOptions | ( | const std::vector< BackendOptions > & | options, |
| BackendId | backend, | ||
| F | f ) |
Definition at line 297 of file BackendOptions.hpp.
References BackendOptions::BackendOption::GetName(), and BackendOptions::BackendOption::GetValue().
Referenced by CheckFastMathSupport(), ClBackendContext::ClBackendContext(), and RuntimeImpl::RuntimeImpl().
|
inline |
Definition at line 321 of file BackendOptions.hpp.
References BackendOptions::Var::AsString(), and BackendOptions::Var::IsString().
Referenced by ClBackendContext::ClBackendContext().
|
inline |
Definition at line 26 of file ArmComputeTuningUtils.hpp.
References ARMNN_LOG, BackendOptions::Var::AsInt(), Exhaustive, BackendOptions::Var::IsInt(), None, and warning.
Referenced by ClBackendContext::ClBackendContext().
| armnn::ConstTensor PermuteTensor | ( | const ConstTensorHandle * | tensor, |
| const PermutationVector & | permutationVector, | ||
| void * | permuteBuffer ) |
Definition at line 19 of file WorkloadUtils.cpp.
References ConstTensorHandle::GetConstTensor(), TensorInfo::GetDataType(), GetDataTypeSize(), TensorInfo::GetNumBytes(), TensorInfo::GetShape(), PermutationVector::GetSize(), ConstTensorHandle::GetTensorInfo(), Permute, armnnUtils::Permuted(), and TensorInfo::SetConstant().
Referenced by Convert1HWOTensorToAcl(), Convert1HWOtoMIHW(), and ConvertWeightTensorFromArmnnToAcl().
| DestType PolymorphicDowncast | ( | SourceType * | value | ) |
Polymorphic downcast for build in pointers only.
Usage: Child* pChild = PolymorphicDowncast<Child*>(pBase);
| DestType | Pointer type to the target object (Child pointer type) |
| SourceType | Pointer type to the source object (Base pointer type) |
| value | Pointer to the source object |
Definition at line 74 of file PolymorphicDowncast.hpp.
References ARMNN_POLYMORPHIC_CAST_CHECK.
Referenced by ApplyBackendOptimizations(), AssignBackends(), AssignBackends(), AssignBackendsIConnectable(), AttemptBackendAssignment(), BuildAddMulAddTensorInfoLists(), ChainReduceLayers(), ClAbsWorkload::ClAbsWorkload(), ClBackendContext::ClBackendContext(), ClBatchMatMulWorkload::ClBatchMatMulWorkload(), ClCastWorkload::ClCastWorkload(), ClChannelShuffleWorkload::ClChannelShuffleWorkload(), ClDepthwiseConvolutionWorkload::ClDepthwiseConvolutionWorkload(), ClExpWorkload::ClExpWorkload(), ClFloorDivWorkload::ClFloorDivWorkload(), ClFullyConnectedWorkload::ClFullyConnectedWorkload(), ClLogicalAndWorkload::ClLogicalAndWorkload(), ClLogicalNotWorkload::ClLogicalNotWorkload(), ClLogicalOrWorkload::ClLogicalOrWorkload(), ClLogWorkload::ClLogWorkload(), ClNegWorkload::ClNegWorkload(), ClRsqrtWorkload::ClRsqrtWorkload(), ClSinWorkload::ClSinWorkload(), ClSliceWorkload::ClSliceWorkload(), ClSqrtWorkload::ClSqrtWorkload(), OutputSlot::Connect(), ConvertConstantToTosaOperator(), ConvertTransposeConv2dToTosaOperator(), ClImportTensorHandleFactory::CreateSubTensorHandle(), ClTensorHandleFactory::CreateSubTensorHandle(), ClWorkloadFactory::CreateSubTensorHandle(), NeonTensorHandleFactory::CreateSubTensorHandle(), NeonWorkloadFactory::CreateSubTensorHandle(), ClWorkloadFactory::CreateWorkload(), NeonWorkloadFactory::CreateWorkload(), RefWorkloadFactory::CreateWorkload(), TosaRefWorkloadFactory::CreateWorkload(), SampleDynamicWorkloadFactory::CreateWorkload(), OutputSlot::Disconnect(), ClRankWorkload::Execute(), NeonConstantWorkload::Execute(), NeonRankWorkload::Execute(), armnn::optimizations::pad_fold::FoldPadIntoLayer2dImpl(), FoldPadLayer2d(), ForEachLayerInput(), ForEachLayerOutput(), FuseAdditionLayer(), FuseBatchNormalizationLayer(), FuseConvolution2dLayer(), FuseDepthwiseConvolution2dLayer(), FuseDivisionLayer(), FuseElementwiseBinaryLayer(), FuseFullyConnectedLayer(), FuseMultiplicationLayer(), FuseSubtractionLayer(), GatherTensorHandlePairs(), NeonTensorHandleFactory::GetCapabilities(), GetTensorInfo(), GetTosaMapping(), SubgraphView::GetWorkingCopy(), IsConnected(), PermuteLayer::IsEqual(), ReshapeLayer::IsEqual(), TransposeLayer::IsEqual(), PermuteLayer::IsInverse(), TransposeLayer::IsInverse(), IsLayerSequence(), ClLayerSupport::IsLayerSupported(), IWorkloadFactory::IsLayerSupported(), IWorkloadFactory::IsLayerSupported(), LayerSupportBase::IsLayerSupported(), RefLayerSupport::IsLayerSupported(), TosaRefLayerSupport::IsLayerSupported(), IsLayerTypeSupported(), IsSequenceLayerType(), NeonAbsWorkload::NeonAbsWorkload(), NeonActivationWorkload::NeonActivationWorkload(), NeonAdditionWorkload::NeonAdditionWorkload(), NeonArgMinMaxWorkload::NeonArgMinMaxWorkload(), NeonBatchMatMulWorkload::NeonBatchMatMulWorkload(), NeonBatchNormalizationWorkload::NeonBatchNormalizationWorkload(), NeonCastWorkload::NeonCastWorkload(), NeonChannelShuffleWorkload::NeonChannelShuffleWorkload(), NeonComparisonWorkload::NeonComparisonWorkload(), NeonConvertFp16ToFp32Workload::NeonConvertFp16ToFp32Workload(), NeonConvertFp32ToFp16Workload::NeonConvertFp32ToFp16Workload(), NeonConvolution2dWorkload::NeonConvolution2dWorkload(), NeonConvolution3dWorkload::NeonConvolution3dWorkload(), NeonDepthwiseConvolutionWorkload::NeonDepthwiseConvolutionWorkload(), NeonDequantizeWorkload::NeonDequantizeWorkload(), NeonDetectionPostProcessWorkload::NeonDetectionPostProcessWorkload(), NeonDivisionWorkload::NeonDivisionWorkload(), NeonElementwiseBinaryWorkload::NeonElementwiseBinaryWorkload(), NeonExpWorkload::NeonExpWorkload(), NeonFloorDivWorkload::NeonFloorDivWorkload(), NeonFloorFloatWorkload::NeonFloorFloatWorkload(), NeonFullyConnectedWorkload::NeonFullyConnectedWorkload(), NeonGatherNdWorkload::NeonGatherNdWorkload(), NeonGatherWorkload::NeonGatherWorkload(), NeonL2NormalizationFloatWorkload::NeonL2NormalizationFloatWorkload(), NeonLogicalAndWorkload::NeonLogicalAndWorkload(), NeonLogicalNotWorkload::NeonLogicalNotWorkload(), NeonLogicalOrWorkload::NeonLogicalOrWorkload(), NeonLogSoftmaxWorkload::NeonLogSoftmaxWorkload(), NeonLogWorkload::NeonLogWorkload(), NeonMaximumWorkload::NeonMaximumWorkload(), NeonMinimumWorkload::NeonMinimumWorkload(), NeonMultiplicationWorkload::NeonMultiplicationWorkload(), NeonNegWorkload::NeonNegWorkload(), NeonNormalizationFloatWorkload::NeonNormalizationFloatWorkload(), NeonPooling2dWorkload::NeonPooling2dWorkload(), NeonPreluWorkload::NeonPreluWorkload(), NeonReshapeWorkload::NeonReshapeWorkload(), NeonResizeWorkload::NeonResizeWorkload(), NeonReverseV2Workload::NeonReverseV2Workload(), NeonRsqrtWorkload::NeonRsqrtWorkload(), NeonSinWorkload::NeonSinWorkload(), NeonSliceWorkload::NeonSliceWorkload(), NeonSoftmaxWorkload::NeonSoftmaxWorkload(), NeonSpaceToDepthWorkload::NeonSpaceToDepthWorkload(), NeonSplitterWorkload::NeonSplitterWorkload(), NeonSqrtWorkload::NeonSqrtWorkload(), NeonStridedSliceWorkload::NeonStridedSliceWorkload(), NeonSubtractionWorkload::NeonSubtractionWorkload(), NeonTransposeConvolution2dWorkload::NeonTransposeConvolution2dWorkload(), ClBackend::OptimizeSubgraphView(), NeonBackend::OptimizeSubgraphView(), RefBackend::OptimizeSubgraphView(), TosaRefBackend::OptimizeSubgraphView(), Graph::PtrCast(), AddBroadcastReshapeLayerImpl::Run(), ConvertConstDequantisationLayersToConstLayersImpl::Run(), ConvertConstPermuteLayersToConstLayers::Run(), FuseBatchNorm< ConvLayer, ArmnnType, T >::Run(), MaxMinIntoBoundedReluImpl::Run(), MovePermuteUpImpl::Run(), MoveTransposeUpImpl::Run(), OptimizeInversePermutesImpl< PermuteType >::Run(), FoldPadIntoConvolution2dImpl::Run(), FoldPadIntoDepthwiseConvolution2dImpl::Run(), PermuteDepthwiseConv2dWeightsImpl::Run(), SquashEqualSiblingsImpl< Comparable >::Run(), TurboConvertConstDequantisationLayersToConstLayersImpl::Run(), OptimizeForTypeImpl< BaseType, Wrapped >::Run(), SubgraphViewSelector::SelectSubgraphs(), SubgraphView::SubgraphView(), SubgraphView::SubgraphView(), Graph::SubstituteSubgraph(), SubgraphView::SubstituteSubgraph(), ConvertConstDequantisationLayersToConstLayersImpl::~ConvertConstDequantisationLayersToConstLayersImpl(), MaxMinIntoBoundedReluImpl::~MaxMinIntoBoundedReluImpl(), MovePermuteUpImpl::~MovePermuteUpImpl(), MoveTransposeUpImpl::~MoveTransposeUpImpl(), and TurboConvertConstDequantisationLayersToConstLayersImpl::~TurboConvertConstDequantisationLayersToConstLayersImpl().
| auto PolymorphicPointerDowncast | ( | const SourceType & | value | ) |
Polymorphic downcast for shared pointers and build in pointers.
Usage: auto pChild = PolymorphicPointerDowncast<Child>(pBase)
| DestType | Type of the target object (Child type) |
| SourceType | Pointer type to the source object (Base (shared) pointer type) |
| value | Pointer to the source object |
Definition at line 93 of file PolymorphicDowncast.hpp.
References ARMNN_POLYMORPHIC_CAST_CHECK, armnn::utility::DynamicPointerCast(), and armnn::utility::StaticPointerCast().
Referenced by ClConcatWorkload::ClConcatWorkload(), ClDepthToSpaceWorkload::ClDepthToSpaceWorkload(), ClDequantizeWorkload::ClDequantizeWorkload(), ClSpaceToBatchNdWorkload::ClSpaceToBatchNdWorkload(), ClSplitterWorkload::ClSplitterWorkload(), ClStackWorkload::ClStackWorkload(), ClBackend::CreateWorkloadFactory(), ClBackend::CreateWorkloadFactory(), ClBackend::CreateWorkloadFactory(), ClBackend::CreateWorkloadFactory(), ClBackend::CreateWorkloadFactory(), NeonBackend::CreateWorkloadFactory(), NeonBackend::CreateWorkloadFactory(), NeonBackend::CreateWorkloadFactory(), NeonBackend::CreateWorkloadFactory(), RefBackend::CreateWorkloadFactory(), RefBackend::CreateWorkloadFactory(), TosaRefBackend::CreateWorkloadFactory(), TosaRefBackend::CreateWorkloadFactory(), NeonBatchToSpaceNdWorkload::NeonBatchToSpaceNdWorkload(), NeonConcatWorkload::NeonConcatWorkload(), NeonDepthToSpaceWorkload::NeonDepthToSpaceWorkload(), NeonQuantizeWorkload::NeonQuantizeWorkload(), NeonSpaceToBatchNdWorkload::NeonSpaceToBatchNdWorkload(), NeonSplitterWorkload::NeonSplitterWorkload(), NeonStackWorkload::NeonStackWorkload(), ClTensorHandle::SetMemoryGroup(), NeonTensorHandle::SetMemoryGroup(), and NeonTimer::Start().
| void Pooling2d | ( | Decoder< float > & | rInputDecoder, |
| Encoder< float > & | rOutputEncoder, | ||
| const TensorInfo & | inputInfo, | ||
| const TensorInfo & | outputInfo, | ||
| const Pooling2dDescriptor & | params ) |
Computes the Pooling2d operation.
Definition at line 142 of file Pooling2d.cpp.
References Decoder< IType >::DecodeTensor(), Exclude, DataLayoutIndexed::GetChannelsIndex(), DataLayoutIndexed::GetDataLayout(), DataLayoutIndexed::GetHeightIndex(), TensorInfo::GetShape(), DataLayoutIndexed::GetWidthIndex(), IgnoreValue, Pooling2dDescriptor::m_DataLayout, Pooling2dDescriptor::m_PadBottom, Pooling2dDescriptor::m_PaddingMethod, Pooling2dDescriptor::m_PadLeft, Pooling2dDescriptor::m_PadRight, Pooling2dDescriptor::m_PadTop, Pooling2dDescriptor::m_PoolHeight, Pooling2dDescriptor::m_PoolType, Pooling2dDescriptor::m_PoolWidth, Pooling2dDescriptor::m_StrideX, Pooling2dDescriptor::m_StrideY, NHWC, numeric_cast(), and Encoder< IType >::Set().
Referenced by Pooling2dLayer::Pooling2dLayer().
| void Pooling3d | ( | Decoder< float > & | rInputDecoder, |
| Encoder< float > & | rOutputEncoder, | ||
| const TensorInfo & | inputInfo, | ||
| const TensorInfo & | outputInfo, | ||
| const Pooling3dDescriptor & | params ) |
Computes the Pooling3d operation.
Definition at line 172 of file Pooling3d.cpp.
References Decoder< IType >::DecodeTensor(), Exclude, DataLayoutIndexed::GetChannelsIndex(), DataLayoutIndexed::GetDepthIndex(), DataLayoutIndexed::GetHeightIndex(), TensorInfo::GetShape(), DataLayoutIndexed::GetWidthIndex(), IgnoreValue, Pooling3dDescriptor::m_DataLayout, Pooling3dDescriptor::m_PadBack, Pooling3dDescriptor::m_PadBottom, Pooling3dDescriptor::m_PaddingMethod, Pooling3dDescriptor::m_PadFront, Pooling3dDescriptor::m_PadLeft, Pooling3dDescriptor::m_PadRight, Pooling3dDescriptor::m_PadTop, Pooling3dDescriptor::m_PoolDepth, Pooling3dDescriptor::m_PoolHeight, Pooling3dDescriptor::m_PoolType, Pooling3dDescriptor::m_PoolWidth, Pooling3dDescriptor::m_StrideX, Pooling3dDescriptor::m_StrideY, Pooling3dDescriptor::m_StrideZ, numeric_cast(), and Encoder< IType >::Set().
Referenced by Pooling3dLayer::Pooling3dLayer().
| void PreluImpl | ( | const TensorInfo & | inputInfo, |
| const TensorInfo & | alphaInfo, | ||
| const TensorInfo & | outputInfo, | ||
| Decoder< float > & | inputData, | ||
| Decoder< float > & | alphaData, | ||
| Encoder< float > & | outputData ) |
Definition at line 13 of file PreluImpl.cpp.
References TensorInfo::GetShape(), and BroadcastLoop::Unroll().
| void PrintOutput | ( | const TensorInfo & | inputInfo, |
| const T * | inputData, | ||
| LayerGuid | guid, | ||
| const std::string & | layerName, | ||
| unsigned int | slotIndex, | ||
| std::ostream & | os ) |
Definition at line 23 of file Debug.cpp.
References TensorInfo::GetNumDimensions(), TensorInfo::GetNumElements(), and TensorInfo::GetShape().
Referenced by Debug().
|
inline |
< Profiler used
Definition at line 180 of file Profiling.hpp.
| template int64_t Quantize< int64_t > | ( | float | value, |
| float | scale, | ||
| int32_t | offset ) |
Quantize a floating point data type into an 8-bit data type.
Explicit specialization of Quantize for int64_t.
Explicit specialization of Quantize for int32_t.
Explicit specialization of Quantize for int16_t.
Explicit specialization of Quantize for uint8_t.
Explicit specialization of Quantize for int8_t.
| value | - The value to quantize. |
| scale | - The scale (must be non-zero). |
| offset | - The offset. |
Definition at line 30 of file TypesUtils.cpp.
References IsQuantizedType().
Referenced by QuantizeLayer::QuantizeLayer().
|
inline |
Definition at line 121 of file RefWorkloadUtils.hpp.
| void Reduce | ( | const TensorInfo & | inputInfo, |
| const TensorInfo & | outputInfo, | ||
| Decoder< float > & | input, | ||
| Encoder< float > & | output, | ||
| const std::vector< uint32_t > | axis, | ||
| const ReduceOperation | reduceOperation ) |
Definition at line 70 of file Reduce.cpp.
References Decoder< IType >::Get(), TensorInfo::GetNumDimensions(), TensorInfo::GetNumElements(), TensorInfo::GetShape(), Max, Mean, Min, NextIndex(), numeric_cast(), Prod, ReducedOutputOffset(), Encoder< IType >::Set(), and Sum.
Referenced by ReduceLayer::ReduceLayer().
| unsigned int ReducedOutputOffset | ( | const unsigned int | numDims, |
| const armnn::TensorShape & | dims, | ||
| std::vector< unsigned int > & | index, | ||
| const unsigned int | numAxis, | ||
| const std::vector< unsigned int > & | axis ) |
Definition at line 40 of file Reduce.cpp.
Referenced by Reduce().
|
constexpr |
Definition at line 10 of file RefBackendId.hpp.
Referenced by RefBackend::GetIdStatic().
|
constexpr |
Definition at line 15 of file RefTensorHandleFactory.hpp.
Referenced by RefTensorHandleFactory::GetIdStatic().
|
inline |
Definition at line 335 of file SubgraphUtils.hpp.
References OptimizationViews::AddDeletedSubgraph(), Constant, InputSlot::GetConnectedOutputSlot(), OutputSlot::GetConnection(), Layer::GetGuid(), Layer::GetInputSlot(), Layer::GetOutputSlot(), InputSlot::GetOwningLayer(), OutputSlot::GetOwningLayer(), LayerWithParameters< Parameters >::GetParameters(), InputSlot::GetSlotIndex(), OutputSlot::GetTensorInfo(), Layer::GetType(), Input, ReshapeDescriptor::m_TargetShape, Output, TensorInfo::SetShape(), and InputSlot::SetTensorInfo().
Referenced by ClBackend::OptimizeSubgraphView(), NeonBackend::OptimizeSubgraphView(), and RefBackend::OptimizeSubgraphView().
| ConstTensor ReorderWeightChannelsForAcl | ( | const ConstTensor & | weightHandle, |
| DataLayout | dataLayout, | ||
| void * | permuteBuffer ) |
Definition at line 74 of file WorkloadUtils.cpp.
References BaseTensor< MemoryType >::GetInfo(), TensorInfo::GetNumBytes(), BaseTensor< MemoryType >::GetShape(), NCHW, and NHWC.
Referenced by ConvertWeightTensorFromArmnnToAcl().
| LayerType * ReplaceLayer | ( | OptimizationViews & | optimizationViews, |
| LayerType * | baseLayer, | ||
| LayerType * | replacementLayer ) |
Definition at line 257 of file SubgraphUtils.hpp.
References OptimizationViews::AddSubstitution().
Referenced by FoldPadLayer2d().
| void ReplaceLayers | ( | OptimizationViews & | optimizationViews, |
| LayerType * | baseLayer, | ||
| std::vector< IConnectableLayer * > & | layers ) |
Definition at line 339 of file ArmComputeSubgraphUtils.hpp.
References OptimizationViews::AddSubstitution().
Referenced by ClBackend::OptimizeSubgraphView(), and NeonBackend::OptimizeSubgraphView().
| void ReplaceMultipleLayers | ( | OptimizationViews & | optimizationViews, |
| std::vector< IConnectableLayer * > & | originalLayers, | ||
| LayerType * | baseLayer, | ||
| const std::vector< SlotList > | inputLayersSlotLists, | ||
| const std::vector< SlotList > | outputLayersSlotLists ) |
Definition at line 357 of file ArmComputeSubgraphUtils.hpp.
References OptimizationViews::AddSubstitution().
Referenced by NeonBackend::OptimizeSubgraphView().
| void ReportError | ( | const std::string & | errorMessage, |
| Optional< std::vector< std::string > & > | errorMessages ) |
Definition at line 762 of file Network.cpp.
References ARMNN_LOG, and warning.
Referenced by AssignBackends(), AssignBackends(), CheckScaleSetOnQuantizedType(), IOptimizedNetwork::Optimize, and ReturnWithError().
|
inline |
Definition at line 243 of file SubgraphUtils.hpp.
References OptimizationViews::AddUntouchedSubgraph().
Referenced by ClBackend::OptimizeSubgraphView(), NeonBackend::OptimizeSubgraphView(), and RefBackend::OptimizeSubgraphView().
| void ReportWarning | ( | const std::string & | warningMessage, |
| Optional< std::vector< std::string > & > | warningMessages ) |
Definition at line 774 of file Network.cpp.
References ARMNN_LOG, and warning.
Referenced by ApplyBackendOptimizations(), and AttemptBackendAssignment().
| bool RequiresCopy | ( | ITensorHandleFactory::FactoryId | src, |
| ITensorHandleFactory::FactoryId | dst, | ||
| TensorHandleFactoryRegistry & | registry ) |
Definition at line 1510 of file Network.cpp.
References ITensorHandleFactory::GetExportFlags(), TensorHandleFactoryRegistry::GetFactory(), and ITensorHandleFactory::GetImportFlags().
Referenced by CalculateSlotOption().
| void ReshapeWeightsForAcl | ( | TensorInfo & | weightInfo, |
| DataLayout | dataLayout ) |
Definition at line 48 of file WorkloadUtils.cpp.
References TensorInfo::GetShape(), NCHW, NHWC, and TensorInfo::SetShape().
Referenced by ConvertWeightTensorFromArmnnToAcl(), and ConvertWeightTensorInfoFromArmnnToAcl().
| void Resize | ( | Decoder< float > & | in, |
| const TensorInfo & | inputInfo, | ||
| Encoder< float > & | out, | ||
| const TensorInfo & | outputInfo, | ||
| DataLayoutIndexed | dataLayout, | ||
| ResizeMethod | resizeMethod, | ||
| bool | alignCorners, | ||
| bool | halfPixelCenters ) |
Definition at line 65 of file Resize.cpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, Bilinear, Decoder< IType >::Get(), DataLayoutIndexed::GetChannelsIndex(), DataLayoutIndexed::GetHeightIndex(), DataLayoutIndexed::GetIndex(), TensorInfo::GetShape(), DataLayoutIndexed::GetWidthIndex(), NearestNeighbor, roundf(), and Encoder< IType >::Set().
Referenced by ResizeLayer::ResizeLayer().
| OptimizationResult ReturnWithError | ( | OptimizationResult | res, |
| const Layer * | layer, | ||
| const BackendSettings & | backendSettings, | ||
| Optional< std::vector< std::string > & > | errMessages ) |
Definition at line 787 of file Network.cpp.
References GetLayerTypeAsCString(), Layer::GetType(), OptimizationResult::m_Error, BackendSettings::m_PreferredBackends, and ReportError().
Referenced by AssignBackendsIConnectable(), and AttemptBackendAssignment().
| unsigned int ReverseGetFlatIdx | ( | const std::vector< unsigned int > & | idxList, |
| unsigned int | inputRank, | ||
| std::vector< unsigned int > & | elementNumInner ) |
Definition at line 34 of file ReverseV2Impl.cpp.
Referenced by ReverseRelocateIdx().
| std::vector< unsigned int > ReverseGetMultIdx | ( | const unsigned int | idx, |
| unsigned int | inputRank, | ||
| std::vector< unsigned int > & | elementNumInner ) |
Definition at line 16 of file ReverseV2Impl.cpp.
Referenced by ReverseRelocateIdx().
| unsigned int ReverseRelocateIdx | ( | unsigned int | idx, |
| unsigned int | inputRank, | ||
| std::vector< bool > & | axisFlag, | ||
| std::vector< unsigned int > & | dimSize, | ||
| std::vector< unsigned int > & | elementNumInner ) |
Definition at line 49 of file ReverseV2Impl.cpp.
References ReverseGetFlatIdx(), and ReverseGetMultIdx().
Referenced by ReverseV2().
| void ReverseV2 | ( | const TensorInfo & | inputInfo, |
| const TensorInfo & | axisInfo, | ||
| Decoder< float > & | inputDecoder, | ||
| Decoder< int > & | axisDecoder, | ||
| Encoder< float > & | outputEncoder ) |
Definition at line 78 of file ReverseV2Impl.cpp.
References Decoder< IType >::Get(), TensorInfo::GetNumDimensions(), TensorInfo::GetNumElements(), TensorInfo::GetShape(), ReverseRelocateIdx(), and Encoder< IType >::Set().
Referenced by ReverseV2Layer::ReverseV2Layer().
| bool RevertConstantWeightsToFP32 | ( | Layer * | layer | ) |
|
inline |
|
inline |
Definition at line 167 of file ClWorkloadUtils.hpp.
References error, and WrapClError().
Referenced by ClAbsWorkload::Execute(), ClActivationWorkload::Execute(), ClAdditionWorkload::Execute(), ClArgMinMaxWorkload::Execute(), ClBatchMatMulWorkload::Execute(), ClBatchNormalizationFloatWorkload::Execute(), ClBatchToSpaceNdWorkload::Execute(), ClCastWorkload::Execute(), ClChannelShuffleWorkload::Execute(), ClComparisonWorkload::Execute(), ClConvertFp16ToFp32Workload::Execute(), ClConvertFp32ToFp16Workload::Execute(), ClConvolution2dWorkload::Execute(), ClConvolution3dWorkload::Execute(), ClDepthToSpaceWorkload::Execute(), ClDepthwiseConvolutionWorkload::Execute(), ClDivisionWorkload::Execute(), ClExpWorkload::Execute(), ClFillWorkload::Execute(), ClFloorDivWorkload::Execute(), ClFloorFloatWorkload::Execute(), ClFullyConnectedWorkload::Execute(), ClGatherNdWorkload::Execute(), ClGatherWorkload::Execute(), ClInstanceNormalizationWorkload::Execute(), ClL2NormalizationFloatWorkload::Execute(), ClLogSoftmaxWorkload::Execute(), ClLogWorkload::Execute(), ClLstmFloatWorkload::Execute(), ClMaximumWorkload::Execute(), ClMinimumWorkload::Execute(), ClMultiplicationWorkload::Execute(), ClNegWorkload::Execute(), ClNormalizationFloatWorkload::Execute(), ClPadWorkload::Execute(), ClPermuteWorkload::Execute(), ClPooling2dWorkload::Execute(), ClPooling3dWorkload::Execute(), ClPreluWorkload::Execute(), ClQuantizedLstmWorkload::Execute(), ClQuantizeWorkload::Execute(), ClReshapeWorkload::Execute(), ClResizeWorkload::Execute(), ClRsqrtWorkload::Execute(), ClScatterNdWorkload::Execute(), ClSinWorkload::Execute(), ClSliceWorkload::Execute(), ClSoftmaxWorkload::Execute(), ClSpaceToBatchNdWorkload::Execute(), ClSpaceToDepthWorkload::Execute(), ClSqrtWorkload::Execute(), ClStridedSliceWorkload::Execute(), ClSubtractionWorkload::Execute(), ClTransposeConvolution2dWorkload::Execute(), and ClTransposeWorkload::Execute().
| void ScatterNd | ( | const TensorInfo & | indicesInfo, |
| const TensorInfo & | updatesInfo, | ||
| const TensorInfo & | shapeInfo, | ||
| Decoder< int > & | indices, | ||
| Decoder< float > & | updates, | ||
| Decoder< int > & | shape, | ||
| Encoder< float > & | output, | ||
| ScatterNdDescriptor | descriptor ) |
Definition at line 181 of file ScatterNd.cpp.
References Decoder< IType >::DecodeTensor(), Decoder< IType >::Get(), TensorInfo::GetNumDimensions(), TensorInfo::GetNumElements(), TensorInfo::GetShape(), ScatterNdDescriptor::m_AxisEnabled, ScatterNdDescriptor::m_Function, ScatterOperation(), Encoder< IType >::Set(), and Update.
| void ScatterNd | ( | const TensorInfo & | inputInfo, |
| const TensorInfo & | indicesInfo, | ||
| const TensorInfo & | updatesInfo, | ||
| Decoder< float > & | input, | ||
| Decoder< int > & | indices, | ||
| Decoder< float > & | updates, | ||
| Encoder< float > & | output, | ||
| ScatterNdDescriptor | descriptor ) |
Definition at line 41 of file ScatterNd.cpp.
References Decoder< IType >::Get(), TensorInfo::GetNumDimensions(), TensorInfo::GetNumElements(), TensorInfo::GetShape(), ScatterNdDescriptor::m_AxisEnabled, ScatterNdDescriptor::m_Function, ScatterOperation(), Encoder< IType >::Set(), and Update.
Referenced by ScatterNdLayer::ScatterNdLayer().
| float ScatterOperation | ( | ScatterNdFunction | operation, |
| float | input, | ||
| float | update ) |
Definition at line 18 of file ScatterNd.cpp.
References Add, Max, Min, Mul, Sub, and Update.
Referenced by ScatterNd(), and ScatterNd().
| OptimizationResult SelectTensorHandleStrategy | ( | Graph & | optGraph, |
| BackendsMap & | backends, | ||
| TensorHandleFactoryRegistry & | registry, | ||
| bool | importEnabled, | ||
| bool | exportEnabled, | ||
| Optional< std::vector< std::string > & > | errMessages ) |
Definition at line 1877 of file Network.cpp.
References ARMNN_SCOPED_PROFILING_EVENT, CalculateEdgeStrategy(), CalculateSlotOption(), CalculateSlotOptionForInput(), CalculateSlotOptionForOutput(), Graph::ForEachLayer(), Layer::GetBackendId(), OutputSlot::GetConnections(), Layer::GetNumOutputSlots(), Layer::GetOutputSlot(), Layer::GetType(), Input, ITensorHandleFactory::LegacyFactoryId, OptimizationResult::m_Error, Output, OutputSlot::SetEdgeStrategy(), OutputSlot::SetTensorHandleFactory(), and Undefined.
Referenced by IOptimizedNetwork::Optimize.
| void SetAllLoggingSinks | ( | bool | standardOut, |
| bool | debugOut, | ||
| bool | coloured ) |
Definition at line 191 of file Logging.cpp.
References SetLoggingSinks().
Referenced by ConfigureLogging(), and TEST_SUITE().
|
inline |
Definition at line 100 of file ClWorkloadUtils.hpp.
Referenced by ClSliceWorkload::ClSliceWorkload(), and ClSliceWorkloadValidate().
|
inline |
Definition at line 79 of file ClWorkloadUtils.hpp.
Referenced by ClStridedSliceWorkload::ClStridedSliceWorkload(), and ClStridedSliceWorkloadValidate().
| void SetLogFilter | ( | LogSeverity | level | ) |
Definition at line 73 of file Logging.cpp.
References ARMNN_FALLTHROUGH, Debug, SimpleLogger< Level >::Enable(), Error, Fatal, SimpleLogger< Level >::Get(), Info, Trace, and Warning.
Referenced by ConfigureLogging(), and TEST_SUITE().
|
inline |
Definition at line 167 of file Logging.cpp.
References SimpleLogger< Level >::AddSink(), SimpleLogger< Level >::Get(), and SimpleLogger< Level >::RemoveAllSinks().
Referenced by SetAllLoggingSinks().
|
inline |
Definition at line 160 of file NeonWorkloadUtils.hpp.
Referenced by NeonSliceWorkload::NeonSliceWorkload(), and NeonSliceWorkloadValidate().
|
inline |
Definition at line 138 of file NeonWorkloadUtils.hpp.
Referenced by NeonStridedSliceWorkload::NeonStridedSliceWorkload(), and NeonStridedSliceWorkloadValidate().
| void SetValueChecked | ( | Optional< T & > | optionalRef, |
| V && | val ) |
Definition at line 17 of file LayerSupportCommon.hpp.
References OptionalReferenceSwitch< IsReference, T >::value().
Referenced by FalseFuncF16(), FalseFuncF32(), FalseFuncI32(), FalseFuncU8(), FalseInputFuncF16(), FalseInputFuncF32(), FalseOutputFuncF16(), FalseOutputFuncF32(), ClLayerSupport::IsConcatSupported(), NeonLayerSupport::IsConcatSupported(), ClLayerSupport::IsSplitterSupported(), and NeonLayerSupport::IsSplitterSupported().
| void Slice | ( | const TensorInfo & | inputInfo, |
| const SliceDescriptor & | descriptor, | ||
| const void * | inputData, | ||
| void * | outputData, | ||
| unsigned int | dataTypeSize ) |
Definition at line 14 of file Slice.cpp.
References TensorShape::GetNumDimensions(), TensorInfo::GetShape(), SliceDescriptor::m_Begin, and SliceDescriptor::m_Size.
Referenced by SliceLayer::SliceLayer().
| void Softmax | ( | Decoder< float > & | in, |
| Encoder< float > & | out, | ||
| const TensorInfo & | inputTensorInfo, | ||
| float | beta, | ||
| int | axis ) |
Computes the softmax function on some inputs, into outputs, with a shape given by tensorInfo.
Definition at line 17 of file Softmax.cpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, Decoder< IType >::Get(), TensorInfo::GetNumDimensions(), TensorShape::GetNumDimensions(), armnnUtils::GetNumElementsBetween(), TensorInfo::GetShape(), and Encoder< IType >::Set().
Referenced by SoftmaxLayer::SoftmaxLayer().
| void SpaceToBatchNd | ( | const TensorInfo & | inputInfo, |
| const TensorInfo & | outputInfo, | ||
| const SpaceToBatchNdDescriptor & | params, | ||
| Decoder< float > & | inputData, | ||
| Encoder< float > & | outputData ) |
Definition at line 48 of file SpaceToBatchNd.cpp.
References CHECK_LOCATION, Decoder< IType >::Get(), DataLayoutIndexed::GetChannelsIndex(), DataLayoutIndexed::GetHeightIndex(), TensorInfo::GetNumDimensions(), GetOffset(), TensorInfo::GetShape(), DataLayoutIndexed::GetWidthIndex(), SpaceToBatchNdDescriptor::m_BlockShape, SpaceToBatchNdDescriptor::m_DataLayout, SpaceToBatchNdDescriptor::m_PadList, NCHW, and Encoder< IType >::Set().
Referenced by SpaceToBatchNdLayer::SpaceToBatchNdLayer().
| void SpaceToDepth | ( | const TensorInfo & | inputInfo, |
| const TensorInfo & | outputInfo, | ||
| const SpaceToDepthDescriptor & | params, | ||
| Decoder< float > & | inputData, | ||
| Encoder< float > & | outputData ) |
Definition at line 36 of file SpaceToDepth.cpp.
References Decoder< IType >::Get(), DataLayoutIndexed::GetChannelsIndex(), DataLayoutIndexed::GetHeightIndex(), GetOffset(), TensorInfo::GetShape(), DataLayoutIndexed::GetWidthIndex(), SpaceToDepthDescriptor::m_BlockSize, SpaceToDepthDescriptor::m_DataLayout, and Encoder< IType >::Set().
Referenced by SpaceToDepthLayer::SpaceToDepthLayer().
| void Split | ( | const SplitterQueueDescriptor & | data, |
| std::vector< ITensorHandle * > | inputs, | ||
| std::vector< ITensorHandle * > | outputs ) |
Definition at line 20 of file Splitter.cpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, Decoder< IType >::Get(), TensorInfo::GetNumDimensions(), TensorInfo::GetNumElements(), TensorInfo::GetShape(), armnnUtils::GetTensorInfo(), SplitterQueueDescriptor::ViewOrigin::m_Origin, SplitterQueueDescriptor::m_ViewOrigins, MakeDecoder(), MakeEncoder(), Map, MaxNumOfTensorDimensions, and Encoder< IType >::Set().
| void Splitter | ( | const SplitterQueueDescriptor & | data, |
| std::vector< ITensorHandle * > | inputs, | ||
| std::vector< ITensorHandle * > | outputs ) |
Definition at line 17 of file Splitter.hpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, GetInputTensorData(), TensorInfo::GetNumDimensions(), TensorInfo::GetNumElements(), GetOutputTensorData(), TensorInfo::GetShape(), armnnUtils::GetTensorInfo(), SplitterQueueDescriptor::ViewOrigin::m_Origin, SplitterQueueDescriptor::m_ViewOrigins, and MaxNumOfTensorDimensions.
Referenced by SplitterLayer::SplitterLayer().
| void Stack | ( | const StackQueueDescriptor & | data, |
| std::vector< std::unique_ptr< Decoder< float > > > & | inputs, | ||
| Encoder< float > & | output, | ||
| const TensorInfo & | inputInfo, | ||
| const TensorInfo & | outputInfo ) |
Definition at line 12 of file Stack.cpp.
References TensorInfo::GetNumDimensions(), TensorInfo::GetNumElements(), TensorInfo::GetShape(), StackDescriptor::m_Axis, QueueDescriptor::m_Inputs, StackDescriptor::m_NumInputs, QueueDescriptorWithParameters< LayerDescriptor >::m_Parameters, and Encoder< IType >::Set().
Referenced by StackLayer::StackLayer().
|
constexpr |
Definition at line 202 of file TypesUtils.hpp.
Referenced by ParseComputeDevice().
| void StridedSlice | ( | const TensorInfo & | inputInfo, |
| const StridedSliceDescriptor & | params, | ||
| const void * | inputData, | ||
| void * | outputData, | ||
| unsigned int | dataTypeSize ) |
Definition at line 86 of file StridedSlice.cpp.
References TensorInfo::GetShape(), StridedSliceDescriptor::GetStartForAxis(), StridedSliceDescriptor::GetStopForAxis(), StridedSliceDescriptor::m_EllipsisMask, StridedSliceDescriptor::m_NewAxisMask, StridedSliceDescriptor::m_Stride, and numeric_cast().
Referenced by StridedSliceLayer::StridedSliceLayer().
|
inline |
Definition at line 43 of file Logging.hpp.
| void swap | ( | OriginsDescriptor & | first, |
| OriginsDescriptor & | second ) |
Definition at line 357 of file Descriptors.cpp.
Referenced by OriginsDescriptor::swap, and ViewsDescriptor::swap.
| void swap | ( | ViewsDescriptor & | first, |
| ViewsDescriptor & | second ) |
| template void Tile | ( | const TileDescriptor & | params, |
| const TensorInfo & | inputInfo, | ||
| Decoder< double_t > & | inputDecoder, | ||
| Encoder< double_t > & | outputEncoder ) |
| template void Tile | ( | const TileDescriptor & | params, |
| const TensorInfo & | inputInfo, | ||
| Decoder< float > & | inputDecoder, | ||
| Encoder< float > & | outputEncoder ) |
| void Tile | ( | const TileDescriptor & | params, |
| const TensorInfo & | inputInfo, | ||
| Decoder< I > & | inputDecoder, | ||
| Encoder< O > & | outputEncoder ) |
Definition at line 46 of file Tile.cpp.
References CoordinatesToIndex(), Decoder< IType >::DecodeTensor(), Decoder< IType >::Get(), TensorInfo::GetNumDimensions(), TensorInfo::GetNumElements(), TensorInfo::GetShape(), IndexToCoordinates(), TileDescriptor::m_Multiples, and Encoder< IType >::Set().
Referenced by TileLayer::TileLayer().
| void TopKSort | ( | unsigned int | k, |
| unsigned int * | indices, | ||
| const float * | values, | ||
| unsigned int | numElement ) |
Definition at line 23 of file DetectionPostProcess.cpp.
Referenced by DetectionPostProcess(), and NonMaxSuppression().
|
constexpr |
Definition at line 10 of file TosaRefBackendId.hpp.
Referenced by TosaRefBackend::GetIdStatic().
| bool TosaRefPreCompiledWorkloadValidate | ( | std::string * | ) |
Definition at line 166 of file TosaRefPreCompiledWorkload.cpp.
|
constexpr |
Definition at line 15 of file TosaRefTensorHandleFactory.hpp.
Referenced by TosaRefTensorHandleFactory::GetIdStatic().
| void TransposeConvolution2dImpl | ( | const TransposeConvolution2dDescriptor & | descriptor, |
| const TensorShape & | inputShape, | ||
| Decoder< float > & | inputDecoder, | ||
| const TensorShape & | outputShape, | ||
| Encoder< float > & | outputEncoder, | ||
| const TensorShape & | weightsShape, | ||
| Decoder< float > & | weightsDecoder, | ||
| Decoder< float > * | biasesDecoder ) |
Definition at line 15 of file TransposeConvolution2d.cpp.
References Decoder< IType >::DecodeTensor(), Decoder< IType >::Get(), DataLayoutIndexed::GetChannelsIndex(), DataLayoutIndexed::GetHeightIndex(), DataLayoutIndexed::GetIndex(), TensorShape::GetNumElements(), DataLayoutIndexed::GetWidthIndex(), TransposeConvolution2dDescriptor::m_BiasEnabled, TransposeConvolution2dDescriptor::m_DataLayout, TransposeConvolution2dDescriptor::m_PadLeft, TransposeConvolution2dDescriptor::m_PadTop, TransposeConvolution2dDescriptor::m_StrideX, TransposeConvolution2dDescriptor::m_StrideY, NHWC, and Encoder< IType >::Set().
| bool TrueFunc | ( | Optional< std::string & > | reasonIfUnsupported, |
| Params &&... | params ) |
Definition at line 54 of file LayerSupportCommon.hpp.
References IgnoreUnused().
Referenced by RefLayerSupport::IsConvertFp16ToFp32Supported(), RefLayerSupport::IsConvertFp32ToFp16Supported(), and NeonLayerSupport::IsFloorSupported().
| void ValidateSourcesMatchOptimizedNetwork | ( | std::vector< BackendOptions > | optimizedOptions, |
| const INetworkProperties & | networkProperties ) |
This function performs a sanity check to ensure that the combination of input and output memory source matches the values for importEnabled and exportEnabled that were specified during optimization.
During optimization the tensor handle factories are chosen based on whether import and export are enabled. If the user then specifies something incompatible here it can lead to problems.
| optimizedOptions | |
| networkProperties |
Definition at line 98 of file LoadedNetwork.cpp.
|
inline |
Definition at line 157 of file ClContextSchema_generated.h.
References ClContextIdentifier().
|
inline |
Definition at line 162 of file ClContextSchema_generated.h.
References ClContextIdentifier().
|
inline |
Definition at line 383 of file TypesUtils.hpp.
References GetDataTypeName(), and info.
|
inline |
Definition at line 159 of file ClWorkloadUtils.hpp.
References Exception::what().
Referenced by RunClFunction().
| const BackendCapabilities cpuAccCapabilities("CpuAcc", { {"NonConstWeights", true}, {"ProtectedContentAllocation", false}, {"ConstantTensorsAsInputs", true}, {"PreImportIOTensors", false}, {"ExternallyManagedMemory", true}, {"MultiAxisPacking", false}, {"SingleAxisPacking", true}, {"HasFp16", arm_compute::CPUInfo::get().has_fp16()}, {"AllOrNothing", false} }) | ( | "CpuAcc" | , |
| { {"NonConstWeights", true}, {"ProtectedContentAllocation", false}, {"ConstantTensorsAsInputs", true}, {"PreImportIOTensors", false}, {"ExternallyManagedMemory", true}, {"MultiAxisPacking", false}, {"SingleAxisPacking", true}, {"HasFp16", arm_compute::CPUInfo::get().has_fp16()}, {"AllOrNothing", false} } | ) |
Referenced by NeonBackend::GetCapabilities().
| const BackendCapabilities cpuRefCapabilities("CpuRef", { {"NonConstWeights", true}, {"ProtectedContentAllocation", false}, {"ConstantTensorsAsInputs", true}, {"PreImportIOTensors", true}, {"ExternallyManagedMemory", true}, {"MultiAxisPacking", false}, {"SingleAxisPacking", true}, {"HasFp16", true}, {"AllOrNothing", false} }) | ( | "CpuRef" | , |
| { {"NonConstWeights", true}, {"ProtectedContentAllocation", false}, {"ConstantTensorsAsInputs", true}, {"PreImportIOTensors", true}, {"ExternallyManagedMemory", true}, {"MultiAxisPacking", false}, {"SingleAxisPacking", true}, {"HasFp16", true}, {"AllOrNothing", false} } | ) |
Referenced by RefBackend::GetCapabilities().
|
constexpr |
|
constexpr |
Definition at line 37 of file Profiling.cpp.
Referenced by ProfilerImpl::AnalyzeEventsAndWriteResults().
|
constexpr |
Definition at line 29 of file Profiling.cpp.
Referenced by ProfilerImpl::ProfilerImpl().
|
constexpr |
Definition at line 32 of file Profiling.cpp.
Referenced by ProfilerImpl::AnalyzeEventSequenceAndWriteResults().
|
constexpr |
Definition at line 41 of file Profiling.cpp.
Referenced by ProfilerImpl::~ProfilerImpl().
|
constexpr |
The lowest performance data capture interval we support is 10 miliseconds.
Definition at line 34 of file Types.hpp.
Referenced by IRuntime::CreationOptions::ExternalProfilingOptions::ExternalProfilingOptions().
|
constexpr |
Definition at line 31 of file Types.hpp.
Referenced by BaseTensor< const void * >::BaseTensor(), Concatenate(), CopyTensorContentsGeneric(), PermutationVector::PermutationVector(), armnnUtils::Permuted(), Split(), Splitter(), TosaTensorNumDimensionsWithinBounds::TosaTensorNumDimensionsWithinBounds(), armnnDeserializer::ToTensorInfo(), and armnnUtils::TransposeTensorShape().
| const std::set<armnn::BackendCapability> oldCpuRefCapabilities |
Definition at line 25 of file RefBackend.hpp.
| const std::set<armnn::LayerType> paddingRequiredLayers |
Definition at line 16 of file NeonTensorHandleFactory.hpp.
Referenced by NeonTensorHandleFactory::GetCapabilities().
| thread_local IProfiler* tl_Profiler = nullptr |
Definition at line 600 of file Profiling.cpp.
Referenced by ProfilerManager::GetProfiler(), and ProfilerManager::RegisterProfiler().
|
constexpr |
Definition at line 22 of file SingleAxisPriorityList.cpp.