24.08
|
Copyright (c) 2021 ARM Limited and Contributors. More...
Namespaces | |
experimental | |
optimizations | |
profiling | |
stringUtils | |
timelinedecoder | |
utility | |
Functions | |
LayerSupportHandle | GetILayerSupportByBackendId (const armnn::BackendId &backend) |
Convenience function to retrieve the ILayerSupportHandle for a backend. More... | |
bool | HasCapability (const std::string &name, const BackendCapabilities &capabilities) |
Convenience function to check if a capability exists in a BackendCapabilites struct. More... | |
bool | HasCapability (const std::string &name, const armnn::BackendId &backend) |
Convenience function to check if a capability exists in a backend. More... | |
bool | HasCapability (const BackendOptions::BackendOption &capability, const BackendCapabilities &capabilities) |
Convenience function to check if a given capability matches a capability in a BackendCapabilities struct. More... | |
bool | HasCapability (const BackendOptions::BackendOption &backendOption, const armnn::BackendId &backend) |
Convenience function to check if a given capability matches a capability in a backend. More... | |
bool | HasMatchingCapability (const BackendOptions::BackendOption &capability, const BackendCapabilities &capabilities) |
Convenience function to check if a given capability matches a capability in a BackendCapabilities struct. More... | |
bool | HasMatchingCapability (const BackendOptions::BackendOption &backendOption, const armnn::BackendId &backend) |
Convenience function to check if a given capability matches a capability in a backend. More... | |
Optional< const BackendOptions::BackendOption > | GetCapability (const std::string &backendCapabilityName, const BackendCapabilities &capabilities) |
Returns a BackendCapability if the backend lists the capability The BackendCapability must then be inspected to check whether or not that BackendCapability is supported Otherwise returns an EmptyOptional if the BackendCapability is unlisted. More... | |
Optional< const BackendOptions::BackendOption > | GetCapability (const std::string &backendCapabilityName, const armnn::BackendId &backend) |
Returns a BackendCapability if the backend lists the capability The BackendCapability must then be inspected to check whether or not that BackendCapability is supported Otherwise returns an EmptyOptional if the BackendCapability is unlisted. More... | |
unsigned int | GetNumberOfCacheFiles (const armnn::BackendId &backend) |
Returns the number of cached files if backend supports caching. More... | |
constexpr char const * | GetComputeDeviceAsCString (Compute compute) |
Deprecated function that will be removed together with the Compute enum. More... | |
std::ostream & | operator<< (std::ostream &os, const std::vector< Compute > &compute) |
Deprecated function that will be removed together with the Compute enum. More... | |
std::ostream & | operator<< (std::ostream &os, const std::set< Compute > &compute) |
Deprecated function that will be removed together with the Compute enum. More... | |
std::ostream & | operator<< (std::ostream &os, const Compute &compute) |
Deprecated function that will be removed together with the Compute enum. More... | |
std::ostream & | operator<< (std::ostream &os, const BackendId &id) |
template<template< typename... > class TContainer, typename... TContainerTemplateArgs> | |
std::ostream & | operator<< (std::ostream &os, const TContainer< BackendId, TContainerTemplateArgs... > &ids) |
template<typename F > | |
void | ParseOptions (const std::vector< BackendOptions > &options, BackendId backend, F f) |
bool | ParseBooleanBackendOption (const armnn::BackendOptions::Var &value, bool defaultValue) |
std::string | ParseStringBackendOption (const armnn::BackendOptions::Var &value, std::string defaultValue) |
int | ParseIntBackendOption (const armnn::BackendOptions::Var &value, int defaultValue) |
BackendRegistry & | BackendRegistryInstance () |
std::ostream & | operator<< (std::ostream &os, const BackendVersion &backendVersion) |
TensorShape | GetUnpaddedTensorStrides (const TensorInfo &tensorInfo) |
DataType | GetBiasDataType (DataType inputDataType) |
template<typename TensorShapeIt > | |
OriginsDescriptor | CreateDescriptorForConcatenation (TensorShapeIt first, TensorShapeIt last, unsigned int concatenationDimension) |
Convenience template to create an OriginsDescriptor to use when creating a ConcatLayer for performing concatenation of a number of input tensors. More... | |
template<typename ExceptionType > | |
void | ConditionalThrow (bool condition, const std::string &message) |
template<typename ExceptionType > | |
void | ConditionalThrow (bool condition) |
template<typename ExceptionType , typename ComparedType > | |
void | ConditionalThrowIfNotEqual (const std::string &message, const ComparedType &leftHandSide, const ComparedType &rightHandSide) |
ComparedType must support: operator==(const ComparedType&) operator<<(ostream&, const ComparedType&) More... | |
IOptimizedNetworkPtr | Optimize (const INetwork &network, const std::vector< BackendId > &backendPreferences, const IDeviceSpec &deviceSpec, const OptimizerOptionsOpaque &options=OptimizerOptionsOpaque(), Optional< std::vector< std::string > & > messages=EmptyOptional()) |
Create an optimized version of the network. More... | |
IOptimizedNetworkPtr | Optimize (const Graph &inGraph, const std::vector< BackendId > &backendPreferences, const IDeviceSpec &deviceSpec, const OptimizerOptionsOpaque &options, Optional< std::vector< std::string > & > messages=EmptyOptional()) |
Create an optimized version of the network. More... | |
IOptimizedNetworkPtr | Optimize (const Graph &inGraph, const std::vector< BackendId > &backendPreferences, const IDeviceSpec &deviceSpec, const OptimizerOptions &options, Optional< std::vector< std::string > & > messages=EmptyOptional()) |
Accept legacy OptimizerOptions. More... | |
IOptimizedNetworkPtr | Optimize (const INetwork &network, const std::vector< BackendId > &backendPreferences, const IDeviceSpec &deviceSpec, const OptimizerOptions &options, Optional< std::vector< std::string > & > messages=EmptyOptional()) |
Accept legacy OptimizerOptions. More... | |
std::string | LevelToString (LogSeverity level) |
LogSeverity | StringToLogLevel (std::string level) |
void | SetLogFilter (LogSeverity level) |
void | SetAllLoggingSinks (bool standardOut, bool debugOut, bool coloured) |
constexpr LogSeverity | ConvertLogSeverity (BoostLogSeverityMapping severity) |
template<typename Arg , typename std::enable_if< IsMemorySource< Arg >::value >::type * = nullptr> | |
MemorySourceFlags | Combine (Arg sourceA, Arg sourceB) |
template<typename Arg , typename ... Args, typename std::enable_if< IsMemorySource< Arg >::value >::type * = nullptr> | |
MemorySourceFlags | Combine (Arg source, Args... rest) |
bool | CheckFlag (MemorySourceFlags flags, MemorySource source) |
template<typename T , class... Args> | |
Optional< T > | MakeOptional (Args &&... args) |
Utility template that constructs an object of type T in-place and wraps it inside an Optional<T> object. More... | |
const char * | GetLayerTypeAsCString (LayerType type) |
constexpr char const * | GetStatusAsCString (Status status) |
constexpr char const * | GetActivationFunctionAsCString (ActivationFunction activation) |
constexpr char const * | GetArgMinMaxFunctionAsCString (ArgMinMaxFunction function) |
constexpr char const * | GetComparisonOperationAsCString (ComparisonOperation operation) |
constexpr char const * | GetBinaryOperationAsCString (BinaryOperation operation) |
constexpr char const * | GetUnaryOperationAsCString (UnaryOperation operation) |
constexpr char const * | GetLogicalBinaryOperationAsCString (LogicalBinaryOperation operation) |
constexpr char const * | GetFusedTypeAsCString (FusedKernelType type) |
constexpr char const * | GetPoolingAlgorithmAsCString (PoolingAlgorithm pooling) |
constexpr char const * | GetOutputShapeRoundingAsCString (OutputShapeRounding rounding) |
constexpr char const * | GetPaddingMethodAsCString (PaddingMethod method) |
constexpr char const * | GetPaddingModeAsCString (PaddingMode mode) |
constexpr char const * | GetReduceOperationAsCString (ReduceOperation reduce_operation) |
constexpr unsigned int | GetDataTypeSize (DataType dataType) |
template<unsigned N> | |
constexpr bool | StrEqual (const char *strA, const char(&strB)[N]) |
constexpr armnn::Compute | ParseComputeDevice (const char *str) |
Deprecated function that will be removed together with the Compute enum. More... | |
constexpr const char * | GetDataTypeName (DataType dataType) |
constexpr const char * | GetDataLayoutName (DataLayout dataLayout) |
constexpr const char * | GetNormalizationAlgorithmChannelAsCString (NormalizationAlgorithmChannel channel) |
constexpr const char * | GetNormalizationAlgorithmMethodAsCString (NormalizationAlgorithmMethod method) |
constexpr const char * | GetResizeMethodAsCString (ResizeMethod method) |
constexpr const char * | GetMemBlockStrategyTypeName (MemBlockStrategyType memBlockStrategyType) |
template<typename T > | |
constexpr bool | IsQuantizedType () |
constexpr bool | IsQuantized8BitType (DataType dataType) |
constexpr bool | IsQuantizedType (DataType dataType) |
std::ostream & | operator<< (std::ostream &os, Status stat) |
std::ostream & | operator<< (std::ostream &os, const armnn::TensorShape &shape) |
template<typename QuantizedType > | |
QuantizedType | Quantize (float value, float scale, int32_t offset) |
Quantize a floating point data type into an 8-bit data type. More... | |
template<typename QuantizedType > | |
float | Dequantize (QuantizedType value, float scale, int32_t offset) |
Dequantize an 8-bit data type into a floating point data type. More... | |
void | VerifyTensorInfoDataType (const armnn::TensorInfo &info, armnn::DataType dataType) |
template<typename ... Ts> | |
void | IgnoreUnused (Ts &&...) |
template<typename Dest , typename Source > | |
std::enable_if_t< std::is_unsigned< Source >::value &&std::is_unsigned< Dest >::value, Dest > | numeric_cast (Source source) |
template<typename Dest , typename Source > | |
std::enable_if_t< std::is_signed< Source >::value &&std::is_integral< Source >::value &&std::is_signed< Dest >::value &&std::is_integral< Dest >::value, Dest > | numeric_cast (Source source) |
template<typename Dest , typename Source > | |
std::enable_if_t< std::is_floating_point< Source >::value &&std::is_floating_point< Dest >::value, Dest > | numeric_cast (Source source) |
template<typename Dest , typename Source > | |
std::enable_if_t< std::is_floating_point< Source >::value &&std::is_signed< Dest >::value &&std::is_integral< Dest >::value, Dest > | numeric_cast (Source source) |
template<typename Dest , typename Source > | |
std::enable_if_t< std::is_signed< Source >::value &&std::is_integral< Source >::value &&std::is_floating_point< Dest >::value, Dest > | numeric_cast (Source source) |
template<typename Dest , typename Source > | |
std::enable_if_t< std::is_signed< Dest >::value &&std::is_integral< Dest >::value &&std::is_unsigned< Source >::value, Dest > | numeric_cast (Source sValue) |
template<typename Dest , typename Source > | |
std::enable_if_t< std::is_floating_point< Dest >::value &&std::is_unsigned< Source >::value, Dest > | numeric_cast (Source sValue) |
template<typename Dest , typename Source > | |
std::enable_if_t< std::is_unsigned< Dest >::value &&std::is_signed< Source >::value &&std::is_integral< Source >::value, Dest > | numeric_cast (Source sValue) |
template<typename Dest , typename Source > | |
std::enable_if_t< std::is_unsigned< Dest >::value &&std::is_floating_point< Source >::value, Dest > | numeric_cast (Source sValue) |
template<typename DestType , typename SourceType > | |
DestType | PolymorphicDowncast (SourceType *value) |
Polymorphic downcast for build in pointers only. More... | |
template<typename DestType , typename SourceType > | |
auto | PolymorphicPointerDowncast (const SourceType &value) |
Polymorphic downcast for shared pointers and build in pointers. More... | |
std::chrono::high_resolution_clock::time_point | GetTimeNow () |
std::chrono::duration< double, std::milli > | GetTimeDuration (std::chrono::high_resolution_clock::time_point start_time) |
template<typename Function , typename Iterator > | |
constexpr TransformIterator< Function, Iterator > | MakeTransformIterator (Iterator i, Function f) |
void | ConfigureLogging (bool printToStandardOutput, bool printToDebugOutput, LogSeverity severity) |
Configures the logging behaviour of the ARMNN library. More... | |
bool | NeonDetected () |
const std::string | GetVersion () |
float | roundf (float value) |
void | swap (OriginsDescriptor &first, OriginsDescriptor &second) |
void | swap (ViewsDescriptor &first, ViewsDescriptor &second) |
uint32_t | GetNumInputs (bool biasEnabled) |
void | AssertNumberOfInputSlots (Layer &layer) |
template<typename T > | |
constexpr LayerType | LayerEnumOf (const T *=nullptr) |
template<> | |
constexpr LayerType | LayerEnumOf (const ActivationLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const AdditionLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const ArgMinMaxLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const BatchMatMulLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const BatchNormalizationLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const BatchToSpaceNdLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const BroadcastToLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const CastLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const ChannelShuffleLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const ComparisonLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const ConcatLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const ConstantLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const ConvertFp16ToFp32Layer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const ConvertFp32ToFp16Layer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const Convolution2dLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const Convolution3dLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const DebugLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const DepthToSpaceLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const DepthwiseConvolution2dLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const DequantizeLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const DetectionPostProcessLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const DivisionLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const ElementwiseBinaryLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const ElementwiseUnaryLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const FakeQuantizationLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const FillLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const FloorLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const FullyConnectedLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const FusedLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const GatherLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const GatherNdLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const InputLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const InstanceNormalizationLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const L2NormalizationLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const LogicalBinaryLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const LogSoftmaxLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const LstmLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const MapLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const MaximumLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const MeanLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const MemCopyLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const MemImportLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const MergeLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const MinimumLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const MultiplicationLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const NormalizationLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const OutputLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const PadLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const PermuteLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const Pooling2dLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const Pooling3dLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const PreCompiledLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const PreluLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const QuantizeLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const QLstmLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const QuantizedLstmLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const RankLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const ReduceLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const ReshapeLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const ResizeLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const ReverseV2Layer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const ScatterNdLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const ShapeLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const SliceLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const SoftmaxLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const SpaceToBatchNdLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const SpaceToDepthLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const SplitterLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const StackLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const StandInLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const StridedSliceLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const SubtractionLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const SwitchLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const TileLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const TransposeLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const TransposeConvolution2dLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const UnidirectionalSequenceLstmLayer *) |
template<> | |
constexpr LayerType | LayerEnumOf (const UnmapLayer *) |
template<typename T , typename V > | |
void | SetValueChecked (Optional< T & > optionalRef, V &&val) |
template<typename Float16Func , typename Float32Func , typename Uint8Func , typename Int32Func , typename BooleanFunc , typename ... Params> | |
bool | IsSupportedForDataTypeGeneric (Optional< std::string & > reasonIfUnsupported, DataType dataType, Float16Func float16FuncPtr, Float32Func float32FuncPtr, Uint8Func uint8FuncPtr, Int32Func int32FuncPtr, BooleanFunc booleanFuncPtr, Params &&... params) |
template<typename ... Params> | |
bool | TrueFunc (Optional< std::string & > reasonIfUnsupported, Params &&... params) |
template<typename ... Params> | |
bool | FalseFunc (Optional< std::string & > reasonIfUnsupported, Params &&... params) |
template<typename ... Params> | |
bool | FalseFuncF16 (Optional< std::string & > reasonIfUnsupported, Params &&... params) |
template<typename ... Params> | |
bool | FalseFuncF32 (Optional< std::string & > reasonIfUnsupported, Params &&... params) |
template<typename ... Params> | |
bool | FalseFuncU8 (Optional< std::string & > reasonIfUnsupported, Params &&... params) |
template<typename ... Params> | |
bool | FalseFuncI32 (Optional< std::string & > reasonIfUnsupported, Params &&... params) |
template<typename ... Params> | |
bool | FalseInputFuncF32 (Optional< std::string & > reasonIfUnsupported, Params &&... params) |
template<typename ... Params> | |
bool | FalseInputFuncF16 (Optional< std::string & > reasonIfUnsupported, Params &&... params) |
template<typename ... Params> | |
bool | FalseOutputFuncF32 (Optional< std::string & > reasonIfUnsupported, Params &&... params) |
template<typename ... Params> | |
bool | FalseOutputFuncF16 (Optional< std::string & > reasonIfUnsupported, Params &&... params) |
void | ValidateSourcesMatchOptimizedNetwork (std::vector< BackendOptions > optimizedOptions, const INetworkProperties &networkProperties) |
This function performs a sanity check to ensure that the combination of input and output memory source matches the values for importEnabled and exportEnabled that were specified during optimization. More... | |
void | CopyToOutputTensor (const Tensor &outputTensor, ITensorHandle *outputTensorHandle) |
const armnn::ConstTensor | GetInputTensor (const LayerBindingId layerId, const InputTensors &inputTensors) |
const armnn::Tensor | GetOutputTensor (const LayerBindingId layerId, const OutputTensors &outputTensors) |
template<LogSeverity Level> | |
void | SetLoggingSinks (bool standardOut, bool debugOut, bool coloured) |
void | ReportError (const std::string &errorMessage, Optional< std::vector< std::string > & > errorMessages) |
void | ReportWarning (const std::string &warningMessage, Optional< std::vector< std::string > & > warningMessages) |
OptimizationResult | ReturnWithError (OptimizationResult res, const Layer *layer, const BackendSettings &backendSettings, Optional< std::vector< std::string > & > errMessages) |
bool | CheckScaleSetOnQuantizedType (Layer *layer, Optional< std::vector< std::string > & > errMessages) |
OptimizationResult | AttemptBackendAssignment (BackendSettings &backendSettings, Graph &graph, Layer *layer, BackendId backend, DataType dataTypeIn, DataType dataTypeOut, const std::vector< BackendId > &availablePreferredBackends, std::string &reasonIfUnsupported, Optional< std::vector< std::string > & > errMessages) |
std::vector< DataType > | GetLayerInOutDatatype (const Layer *layer) |
bool | CheckFp16Support (BackendsMap &backends, const std::vector< BackendId > &availablePreferredBackends) |
void | AssignBackendsIConnectable (OptimizedNetworkImpl *optNetObjPtr, IConnectableLayer *it, Optional< std::vector< std::string > & > errMessages, OptimizationResult &result, BackendSettings &backendSettings, std::vector< BackendId > &availablePreferredBackends) |
OptimizationResult | AssignBackends (OptimizedNetworkImpl *optNetObjPtr, BackendSettings &backendSettings, Graph::Iterator &firstLayer, Graph::Iterator &lastLayer, Optional< std::vector< std::string > & > errMessages) |
OptimizationResult | AssignBackends (OptimizedNetworkImpl *optNetObjPtr, BackendSettings &backendSettings, SubgraphView::IConnectableLayerIterator &firstLayer, SubgraphView::IConnectableLayerIterator &lastLayer, Optional< std::vector< std::string > & > errMessages) |
OptimizationResult | AssignBackends (OptimizedNetworkImpl *optNetObjPtr, BackendSettings &backendSettings, SubgraphView &subgraph, Optional< std::vector< std::string > & > errMessages) |
BackendsMap | CreateSupportedBackends (TensorHandleFactoryRegistry &handleFactoryRegistry, BackendSettings &backendSettings) |
OptimizationResult | ApplyBackendOptimizations (OptimizedNetworkImpl *optNetObjPtr, BackendSettings &backendSettings, BackendsMap &backends, const ModelOptions &modelOptions, Optional< std::vector< std::string > & > errMessages) |
bool | RequiresCopy (ITensorHandleFactory::FactoryId src, ITensorHandleFactory::FactoryId dst, TensorHandleFactoryRegistry ®istry) |
ITensorHandleFactory::FactoryId | CalculateSlotOptionForInput (BackendsMap &backends, OutputSlot &slot, TensorHandleFactoryRegistry ®istry, bool importEnabled) |
ITensorHandleFactory::FactoryId | CalculateSlotOptionForOutput (BackendsMap &backends, OutputSlot &slot, TensorHandleFactoryRegistry ®istry) |
ITensorHandleFactory::FactoryId | CalculateSlotOption (BackendsMap &backends, OutputSlot &outputSlot, TensorHandleFactoryRegistry ®istry, bool exportEnabled) |
EdgeStrategy | CalculateEdgeStrategy (BackendsMap &backends, ITensorHandleFactory::FactoryId srcFactoryId, const Layer &layer, const Layer &connectedLayer, TensorHandleFactoryRegistry ®istry, bool importEnabled) |
OptimizationResult | SelectTensorHandleStrategy (Graph &optGraph, BackendsMap &backends, TensorHandleFactoryRegistry ®istry, bool importEnabled, bool exportEnabled, Optional< std::vector< std::string > & > errMessages) |
std::vector< ConvertFp16ToFp32Layer * > | InsertConvertFp16ToFp32LayersBefore (Graph &graph, Layer &layer, bool expectCorrectInputType) |
std::vector< ConvertFp32ToFp16Layer * > | InsertConvertFp32ToFp16LayersAfter (Graph &graph, Layer &layer) |
std::vector< DebugLayer * > | InsertDebugLayerAfter (Graph &graph, Layer &layer, bool toFile) |
bool | RevertConstantWeightsToFP32 (Layer *layer) |
template<typename T > | |
void | Append (Optimizer::Optimizations &optimizations, T &&optimization) |
template<typename Front , typename... Others> | |
void | Append (Optimizer::Optimizations &optimizations, Front &&front, Others &&... others) |
template<typename... Args> | |
Optimizer::Optimizations | MakeOptimizations (Args &&... args) |
Measurement | FindMeasurement (const std::string &name, const Event *event) |
std::vector< Measurement > | FindKernelMeasurements (const Event *event) |
const Event * | GetEventPtr (const Event *ptr) |
const Event * | GetEventPtr (const std::unique_ptr< Event > &ptr) |
int | CalcLevel (const Event *eventPtr) |
void | ConfigureDetailsObject (JsonChildObject &detailsObject, std::string layerDetailsStr) |
void | ExtractJsonObjects (unsigned int inferenceIndex, const Event *parentEvent, JsonChildObject &parentObject, std::map< const Event *, std::vector< const Event * >> descendantsMap) |
template<typename DescriptorType > | |
void | ProfilingUpdateDescriptions (const std::string &name, const DescriptorType &desc, const WorkloadInfo &infos, const arm::pipe::ProfilingGuid guid) |
template<typename Delegate > | |
void | ForEachLayerInput (LayerSelectionInfo::LayerInfoContainer &layerInfos, LayerSelectionInfo &layerInfo, Delegate function) |
template<typename Delegate > | |
void | ForEachLayerOutput (LayerSelectionInfo::LayerInfoContainer &layerInfos, LayerSelectionInfo &layerInfo, Delegate function) |
void | AssignSplitId (LayerSelectionInfo::LayerInfoContainer &layerInfos, LayerSelectionInfo &layerInfo) |
bool | IsReadyForSplitAssignment (LayerSelectionInfo::LayerInfoContainer &layerInfos, LayerSelectionInfo &layerInfo) |
bool | IsLayerSupported (const armnn::Layer *layer) |
bool | IsLayerSupported (const armnn::Layer &layer) |
bool | IsLayerOptimizable (const armnn::Layer *layer) |
bool | IsLayerOptimizable (const armnn::Layer &layer) |
constexpr const char * | MockTensorHandleFactoryId () |
Graph & | GetGraphForTesting (IOptimizedNetwork *optNet) |
ModelOptions & | GetModelOptionsForTesting (IOptimizedNetwork *optNet) |
arm::pipe::IProfilingService & | GetProfilingService (armnn::RuntimeImpl *runtime) |
std::ostream & | operator<< (std::ostream &os, const BFloat16 &b) |
template<typename LayerType > | |
LayerType * | FuseLayer (OptimizationViews &optimizationViews, LayerType *baseLayer, LayerType *replacementLayer, ActivationLayer *activationLayer, ActivationDescriptor &activationDesc) |
template<typename LayerType > | |
LayerType * | FuseAdditionLayer (OptimizationViews &optimizationViews, LayerType *baseLayer, ActivationLayer *activationLayer, ActivationDescriptor &activationDesc, std::string name) |
template<typename LayerType > | |
LayerType * | FuseSubtractionLayer (OptimizationViews &optimizationViews, LayerType *baseLayer, ActivationLayer *activationLayer, ActivationDescriptor &activationDesc, std::string name) |
template<typename LayerType > | |
LayerType * | FuseDivisionLayer (OptimizationViews &optimizationViews, LayerType *baseLayer, ActivationLayer *activationLayer, ActivationDescriptor &activationDesc, std::string name) |
template<typename LayerType > | |
LayerType * | FuseMultiplicationLayer (OptimizationViews &optimizationViews, LayerType *baseLayer, ActivationLayer *activationLayer, ActivationDescriptor &activationDesc, std::string name) |
template<typename LayerType > | |
LayerType * | FuseElementwiseBinaryLayer (OptimizationViews &optimizationViews, LayerType *baseLayer, ActivationLayer *activationLayer, ActivationDescriptor &activationDesc, BinaryOperation operation, std::string name) |
template<typename LayerType > | |
LayerType * | FuseBatchNormalizationLayer (OptimizationViews &optimizationViews, LayerType *baseLayer, ActivationLayer *activationLayer, ActivationDescriptor &activationDesc, std::string name) |
template<typename LayerType > | |
LayerType * | FuseConvolution2dLayer (OptimizationViews &optimizationViews, LayerType *baseLayer, ActivationLayer *activationLayer, ActivationDescriptor &activationDesc, std::string name) |
template<typename LayerType > | |
LayerType * | FuseDepthwiseConvolution2dLayer (OptimizationViews &optimizationViews, LayerType *baseLayer, ActivationLayer *activationLayer, ActivationDescriptor &activationDesc, std::string name) |
template<typename LayerType > | |
LayerType * | FuseFullyConnectedLayer (OptimizationViews &optimizationViews, LayerType *baseLayer, ActivationLayer *activationLayer, ActivationDescriptor &activationDesc, std::string name) |
template<typename LayerType > | |
std::vector< IConnectableLayer * > | ChainReduceLayers (OptimizationViews &optimizationViews, LayerType *baseLayer, ReduceDescriptor &desc) |
template<typename LayerType > | |
void | ReplaceLayers (OptimizationViews &optimizationViews, LayerType *baseLayer, std::vector< IConnectableLayer * > &layers) |
template<typename LayerType > | |
void | ReplaceMultipleLayers (OptimizationViews &optimizationViews, std::vector< IConnectableLayer * > &originalLayers, LayerType *baseLayer, const std::vector< SlotList > inputLayersSlotLists, const std::vector< SlotList > outputLayersSlotLists) |
TuningLevel | ParseTuningLevel (const BackendOptions::Var &value, TuningLevel defaultValue) |
void | ConfigureTuner (arm_compute::CLTuner &tuner, TuningLevel level) |
arm_compute::NormalizationLayerInfo | CreateAclNormalizationLayerInfoForL2Normalization (const armnn::TensorInfo &tensorInfo, armnn::DataLayout dataLayout) |
arm_compute::ActivationLayerInfo::ActivationFunction | ConvertActivationFunctionToAclActivationFunction (ActivationFunction armnnFunction) |
arm_compute::ActivationLayerInfo | ConvertActivationDescriptorToAclActivationLayerInfo (const ActivationDescriptor &actDesc) |
arm_compute::ActivationLayerInfo | ConvertActivationDescriptorToAclActivationLayerInfo (const ActivationDescriptor *activationDescPtr) |
arm_compute::ActivationLayerInfo | ConvertAdditionalInfoToAclActivationLayerInfo (const QueueDescriptor &queueDescriptor) |
arm_compute::ActivationLayerInfo | ConvertLstmActivationFuncToAclLayerInfo (uint32_t activationFunction) |
arm_compute::ComparisonOperation | ConvertComparisonOperationToAcl (const ComparisonDescriptor &descriptor) |
arm_compute::PoolingType | ConvertPoolingAlgorithmToAclPoolingType (PoolingAlgorithm poolingAlgorithm) |
arm_compute::DimensionRoundingType | ConvertOutputShapeRoundingToAclDimensionRoundingType (OutputShapeRounding rounding) |
arm_compute::NormType | ConvertNormalizationAlgorithmChannelToAclNormType (NormalizationAlgorithmChannel channelType) |
arm_compute::FullyConnectedLayerInfo | ConvertFullyConnectedDescriptorToAclFullyConnectedLayerInfo (const FullyConnectedDescriptor &fullyConnectedDesc, const ActivationDescriptor *activationDesc) |
arm_compute::FullyConnectedLayerInfo | ConvertFullyConnectedDescriptorToAclFullyConnectedLayerInfo (const FullyConnectedDescriptor &fullyConnectedDesc, arm_compute::ActivationLayerInfo activationLayerInfo) |
arm_compute::InterpolationPolicy | ConvertResizeMethodToAclInterpolationPolicy (ResizeMethod resizeMethod) |
template<typename T > | |
T | ComputeSoftmaxAclAxis (const SoftmaxDescriptor &softmaxDesc, const armnn::TensorInfo &tensor) |
int | ComputeAclAxis (const int &armnnAxis, const armnn::TensorInfo &tensor) |
Function to convert ArmNN axis (left to right) to ACL axis (right to left) ranging from [-rank, rank) More... | |
arm_compute::Conv3dInfo | ComputeConv3DInfo (const armnn::Convolution3dDescriptor descriptor, bool isFastMathEnabled, const ActivationDescriptor *activationDescriptor) |
Utility function used to setup an arm_compute::Conv3dInfo object from convolution3d descriptor. More... | |
arm_compute::Conv3dInfo | ComputeConv3DInfo (const armnn::Convolution3dQueueDescriptor queueDescriptor, bool isFastMathEnabled) |
arm_compute::PaddingMode | ConvertPaddingModeToAcl (const PaddingMode &paddingMode) |
arm_compute::ReductionOperation | ConvertReductionOperationToAcl (const ReduceDescriptor &descriptor) |
const TensorInfo | ComputeReductionTensorShape (const armnn::TensorInfo &input, const std::vector< uint32_t > &vAxis, const bool keepDims) |
Function to compute the output tensor shape based on the axes and if keepDims is set. More... | |
armnn::Optional< armnn::DataType > | GetBiasTypeFromWeightsType (armnn::Optional< armnn::DataType > weightsType) |
template<typename F > | |
bool | CheckSupportRule (F rule, Optional< std::string & > reasonIfUnsupported, const char *reason) |
template<typename T > | |
bool | AllTypesAreEqualImpl (T) |
template<typename T , typename... Rest> | |
bool | AllTypesAreEqualImpl (T t1, T t2, Rest... rest) |
std::unique_ptr< IMemoryOptimizerStrategy > | GetMemoryOptimizerStrategy (const std::string &strategyName) |
const std::vector< std::string > | GetMemoryOptimizerStrategyNames () |
bool | IsNCHW (armnn::Layer &layer) |
void | ReportUntouchedLayers (OptimizationViews &optimizationViews, std::map< LayerGuid, Layer * > untouched) |
template<typename LayerType > | |
LayerType * | FoldPadLayer (OptimizationViews &optimizationViews, LayerType *baseLayer, LayerType *replacementLayer, PadLayer *padLayer) |
bool | ConnectedToLayerWithNCHW (Layer *baseLayer) |
Checks if the Layer is connected to any Layer that has an NCHW layout. More... | |
bool | ConnectedToLayerType (Layer *baseLayer, LayerType layerType, unsigned int dimSize=0) |
Checks the Layer's Connections to see if it's connected to a Layer with the provided layerType. More... | |
void | RemoveReshapeLayer (ReshapeLayer *baseLayer, std::map< LayerGuid, Layer * > &untouched, OptimizationViews &optimizationViews) |
template<typename LayerType > | |
LayerType * | FoldPadIntoAveragePool2d (OptimizationViews &optimizationViews, Pooling2dLayer *baseLayer, Pooling2dDescriptor &poolDescriptor, PadLayer *padLayer) |
bool | IsSequenceLayerType (Layer &layer, LayerType type) |
bool | IsSequenceLayerType (Layer &layer, BinaryOperation type) |
template<typename TYPE > | |
bool | IsLayerSequence (Layer ¤tLayer, TYPE first, TYPE second, TYPE third, Layer *layerList[4], bool handleValidActivates, const std::vector< ActivationFunction > &validActivates) |
armnn::ConstTensor | PermuteTensor (const ConstTensorHandle *tensor, const PermutationVector &permutationVector, void *permuteBuffer) |
void | ReshapeWeightsForAcl (TensorInfo &weightInfo, DataLayout dataLayout) |
template<typename DataType > | |
ConstTensor | ReorderWeightChannelsForAcl (const ConstTensor &weightHandle, DataLayout dataLayout, void *permuteBuffer) |
TensorInfo | ConvertWeightTensorInfoFromArmnnToAcl (const TensorInfo &weightInfo, DataLayout dataLayout) |
std::tuple< ConstTensor, unsigned int > | Convert1HWOTensorToAcl (const ConstTensorHandle *weightTensor, const TensorInfo &inputInfo, const DataLayout dataLayout, void *permuteBuffer) |
Weights for depthwise have a datalayout of [1,H,W,O] = [1,H,W,I*M] This function coverts a ConstCpuTensorHandle from [1,H,W,I*M] to [1,I*M,H,W] (if NCHW) or keeps it at [1,H,W,I*M] (if NHWC) as required by the compute library. More... | |
std::tuple< TensorInfo, unsigned int > | Convert1HWOTensorInfoToAcl (const TensorInfo &weightInfo, const TensorInfo &inputInfo, const DataLayout dataLayout) |
Weights for depthwise have a datalayout of [1,H,W,O] = [1,H,W,I*M] This function coverts a TensorInfo from [1,H,W,I*M] to [1,I*M,H,W] (if NCHW) or keeps it at [1,H,W,I*M] (if NHWC) as required by the compute library Returns a tuple of converted weights tensor info and depth multiplier. More... | |
std::tuple< ConstTensor, unsigned int > | Convert1HWOtoMIHW (const ConstTensorHandle *weightTensor, const TensorInfo &inputInfo, const DataLayout &dataLayout, void *permuteBuffer) |
Converts a (weights) tensor from [1, H, W, I*M] = [1, H, W, O] to [M, I, H, W]. More... | |
armnn::ConstTensor | ConvertWeightTensorFromArmnnToAcl (const ConstTensorHandle *weightTensor, DataLayout dataLayout, void *permuteBuffer) |
int32_t | ConvertMaskToACLFormat (int32_t mask, int32_t numDim) |
std::map< std::string, unsigned int > | CalculateGatherNdKeyIndices (TensorInfo inputInfo0, TensorInfo inputInfo1) |
Calculates the key index values needed for GatherNd: N, ND, K, W, C (N is always 1) More... | |
armnn::PermutationVector | GeneratePermutationVectorOnLastTwoDimensions (unsigned int rank) |
Generates a permutation vector of size rank that permutes the 2 most right dimensions. More... | |
std::set< unsigned int > | ComputeSplitAxis (const armnn::SplitterDescriptor &desc, const TensorShape &input) |
Calculates the axis values for split operation. More... | |
template<typename CopyFunc > | |
void | CopyTensorContentsGeneric (const ITensorHandle *srcTensor, ITensorHandle *dstTensor, CopyFunc copy) |
template<typename SrcTensorHandleType , typename DstTensorHandleType , typename DescriptorType > | |
void | GatherTensorHandlePairs (const DescriptorType &descriptor, std::vector< std::pair< SrcTensorHandleType *, DstTensorHandleType * >> &tensorHandlePairs) |
constexpr const char * | ClBackendId () |
flatbuffers::Offset< ClContext > | CreateClContext (flatbuffers::FlatBufferBuilder &_fbb, flatbuffers::Offset< flatbuffers::Vector< flatbuffers::Offset< armnn::Program >>> programs=0) |
flatbuffers::Offset< ClContext > | CreateClContextDirect (flatbuffers::FlatBufferBuilder &_fbb, const std::vector< flatbuffers::Offset< armnn::Program >> *programs=nullptr) |
flatbuffers::Offset< Program > | CreateProgram (flatbuffers::FlatBufferBuilder &_fbb, flatbuffers::Offset< flatbuffers::String > name=0, flatbuffers::Offset< flatbuffers::Vector< uint8_t >> binary=0) |
flatbuffers::Offset< Program > | CreateProgramDirect (flatbuffers::FlatBufferBuilder &_fbb, const char *name=nullptr, const std::vector< uint8_t > *binary=nullptr) |
const armnn::ClContext * | GetClContext (const void *buf) |
const armnn::ClContext * | GetSizePrefixedClContext (const void *buf) |
const char * | ClContextIdentifier () |
bool | ClContextBufferHasIdentifier (const void *buf) |
bool | VerifyClContextBuffer (flatbuffers::Verifier &verifier) |
bool | VerifySizePrefixedClContextBuffer (flatbuffers::Verifier &verifier) |
const char * | ClContextExtension () |
void | FinishClContextBuffer (flatbuffers::FlatBufferBuilder &fbb, flatbuffers::Offset< armnn::ClContext > root) |
void | FinishSizePrefixedClContextBuffer (flatbuffers::FlatBufferBuilder &fbb, flatbuffers::Offset< armnn::ClContext > root) |
constexpr const char * | ClImportTensorHandleFactoryId () |
constexpr const char * | ClTensorHandleFactoryId () |
arm_compute::Status | ClAbsWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
arm_compute::Status | ClActivationWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const ActivationDescriptor &descriptor) |
arm_compute::Status | ClAdditionValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ActivationDescriptor *activationDescriptor) |
arm_compute::Status | ClArgMinMaxWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const ArgMinMaxDescriptor &descriptor) |
arm_compute::Status | ClBatchMatMulValidate (const TensorInfo &inputInfoX, const TensorInfo &inputInfoY, const TensorInfo &outputInfo, const BatchMatMulDescriptor &descriptor, const ActivationDescriptor *activationDescriptor) |
arm_compute::Status | ClBatchNormalizationValidate (const TensorInfo &input, const TensorInfo &output, const TensorInfo &mean, const TensorInfo &var, const TensorInfo &beta, const TensorInfo &gamma, const BatchNormalizationDescriptor &descriptor, const ActivationDescriptor *activationDescriptor) |
arm_compute::Status | ClBatchToSpaceNdWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const BatchToSpaceNdDescriptor &descriptor) |
arm_compute::Status | ClCastValidate (const TensorInfo &input, const TensorInfo &output) |
arm_compute::Status | ClChannelShuffleValidate (const TensorInfo &input, const TensorInfo &output, const ChannelShuffleDescriptor &descriptor) |
arm_compute::Status | ClComparisonWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ComparisonDescriptor &descriptor) |
arm_compute::Status | ClConcatWorkloadValidate (const std::vector< const TensorInfo * > &inputs, const TensorInfo &output, const OriginsDescriptor &descriptor) |
arm_compute::Status | ClConstantWorkloadValidate (const TensorInfo &output) |
arm_compute::Status | ClConvertFp16ToFp32WorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
arm_compute::Status | ClConvertFp32ToFp16WorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
arm_compute::Status | ClConvolution2dWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const Convolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, bool isFastMathEnabled, const ActivationDescriptor *activationDescriptor) |
arm_compute::Status | ClConvolution3dWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const Convolution3dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, bool isFastMathEnabled, const ActivationDescriptor *activationDescriptor) |
arm_compute::Status | ClDepthToSpaceWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const DepthToSpaceDescriptor &descriptor) |
arm_compute::Status | ClDepthwiseConvolutionWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const DepthwiseConvolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, const ActivationDescriptor *activationDescriptor) |
arm_compute::Status | ClDequantizeWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
arm_compute::Status | ClDivisionWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ActivationDescriptor *activationDescriptor) |
arm_compute::Status | ClElementwiseBinaryValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ElementwiseBinaryDescriptor &descriptor, const ActivationDescriptor *activationDescriptor) |
arm_compute::Status | ClExpWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
arm_compute::Status | ClFloorWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
arm_compute::Status | ClFullyConnectedWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const TensorInfo &weights, const Optional< TensorInfo > &biases, const FullyConnectedDescriptor &descriptor, const ActivationDescriptor *activationDescriptor) |
arm_compute::Status | ClGatherNdWorkloadValidate (const TensorInfo ¶msInfo, const TensorInfo &indicesInfo, const TensorInfo &outputInfo) |
arm_compute::Status | ClGatherWorkloadValidate (const TensorInfo &input, const TensorInfo &indices, const TensorInfo &output, const GatherDescriptor &descriptor) |
arm_compute::Status | ClInstanceNormalizationWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const InstanceNormalizationDescriptor &descriptor) |
arm_compute::Status | ClL2NormalizationWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const L2NormalizationDescriptor &descriptor) |
arm_compute::Status | ClLogicalAndWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output) |
arm_compute::Status | ClLogicalNotWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
arm_compute::Status | ClLogicalOrWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output) |
arm_compute::Status | ClLogSoftmaxWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const LogSoftmaxDescriptor &descriptor) |
arm_compute::Status | ClLogWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
arm_compute::Status | ClLstmFloatWorkloadValidate (const TensorInfo &input, const TensorInfo &outputStateIn, const TensorInfo &cellStateIn, const TensorInfo &scratchBuffer, const TensorInfo &outputStateOut, const TensorInfo &cellStateOut, const TensorInfo &output, const LstmDescriptor &descriptor, const LstmInputParamsInfo ¶msInfo) |
arm_compute::Status | ClMaximumWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output) |
arm_compute::Status | ClMeanValidate (const TensorInfo &input, const TensorInfo &output, const MeanDescriptor &descriptor) |
arm_compute::Status | ClMinimumWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output) |
arm_compute::Status | ClMultiplicationWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ActivationDescriptor *activationDescriptor) |
arm_compute::Status | ClNegWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
arm_compute::Status | ClNormalizationWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const NormalizationDescriptor &descriptor) |
arm_compute::Status | ClPadValidate (const TensorInfo &input, const TensorInfo &output, const PadDescriptor &descriptor) |
arm_compute::Status | ClPermuteWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const PermuteDescriptor &descriptor) |
arm_compute::Status | ClPooling2dWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const Pooling2dDescriptor &descriptor) |
arm_compute::Status | ClPooling3dWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const Pooling3dDescriptor &descriptor) |
arm_compute::Status | ClPreluWorkloadValidate (const TensorInfo &input, const TensorInfo &alpha, const TensorInfo &output) |
arm_compute::Status | ClQLstmWorkloadValidate (const TensorInfo &input, const TensorInfo &cellStateIn, const TensorInfo &outputStateIn, const TensorInfo &cellStateOut, const TensorInfo &outputStateOut, const TensorInfo &output, const QLstmDescriptor &descriptor, const LstmInputParamsInfo ¶msInfo) |
arm_compute::Status | ClQuantizedLstmWorkloadValidate (const TensorInfo &input, const TensorInfo &previousCellStateIn, const TensorInfo &previousOutputIn, const TensorInfo &cellStateOut, const TensorInfo &output, const QuantizedLstmInputParamsInfo ¶msInfo) |
arm_compute::Status | ClQuantizeWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
arm_compute::Status | ClReduceWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const ReduceDescriptor &descriptor) |
arm_compute::Status | ClReshapeWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
arm_compute::Status | ClResizeWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const ResizeDescriptor &descriptor) |
arm_compute::Status | ClReverseV2WorkloadValidate (const TensorInfo &input, const TensorInfo &axis, const TensorInfo &output) |
arm_compute::Status | ClRsqrtWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
arm_compute::Status | ClScatterNdWorkloadValidate (const TensorInfo &inputInfo, const TensorInfo &indicesInfo, const TensorInfo &updatesInfo, const TensorInfo &outputInfo, const ScatterNdDescriptor &descriptor) |
arm_compute::Status | ClSinWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
arm_compute::Status | ClSliceWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const SliceDescriptor &descriptor) |
arm_compute::Status | ClSoftmaxWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const SoftmaxDescriptor &descriptor) |
arm_compute::Status | ClSpaceToBatchNdWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const SpaceToBatchNdDescriptor &descriptor) |
arm_compute::Status | ClSpaceToDepthWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const SpaceToDepthDescriptor &descriptor) |
arm_compute::Status | ClSplitterWorkloadValidate (const TensorInfo &input, const std::vector< std::reference_wrapper< TensorInfo >> &outputs, unsigned int splitAxis) |
arm_compute::Status | ClSqrtWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
arm_compute::Status | ClStackWorkloadValidate (const std::vector< const TensorInfo * > &inputs, const TensorInfo &output, const StackDescriptor &descriptor) |
arm_compute::Status | ClStridedSliceWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const StridedSliceDescriptor &descriptor) |
arm_compute::Status | ClSubtractionValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ActivationDescriptor *activationDescriptor) |
arm_compute::Status | ClTileWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const TileDescriptor &descriptor) |
arm_compute::Status | ClTransposeConvolution2dWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const TransposeConvolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases) |
arm_compute::Status | ClTransposeWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const TransposeDescriptor &descriptor) |
arm_compute::Status | ClUnidirectionalSequenceLstmFloatWorkloadValidate (const TensorInfo &input, const TensorInfo &outputStateIn, const TensorInfo &cellStateIn, const TensorInfo &outputStateOut, const TensorInfo &cellStateOut, const TensorInfo &output, const UnidirectionalSequenceLstmDescriptor &descriptor, const LstmInputParamsInfo ¶msInfo) |
std::string | GetConvolutionMethodString (arm_compute::ConvolutionMethod &convolutionMethod) |
template<typename T > | |
void | CopyArmComputeClTensorData (arm_compute::CLTensor &dstTensor, const T *srcData) |
auto | SetClStridedSliceData (const std::vector< int > &m_begin, const std::vector< int > &m_end, const std::vector< int > &m_stride) |
auto | SetClSliceData (const std::vector< unsigned int > &m_begin, const std::vector< unsigned int > &m_size) |
void | InitializeArmComputeClTensorData (arm_compute::CLTensor &clTensor, const ConstTensorHandle *handle) |
RuntimeException | WrapClError (const cl::Error &clError, const CheckLocation &location) |
void | RunClFunction (arm_compute::IFunction &function, const CheckLocation &location) |
template<typename DataType , typename PayloadType > | |
DataType * | GetOutputTensorData (unsigned int idx, const PayloadType &data) |
template<typename T > | |
void | DeleteAsType (const void *const blob) |
SubgraphView::InputSlots | CreateInputsFrom (Layer *layer) |
SubgraphView::OutputSlots | CreateOutputsFrom (Layer *layer) |
SubgraphView::SubgraphViewPtr | CreateSubgraphViewFrom (SubgraphView::InputSlots &&inputs, SubgraphView::OutputSlots &&outputs, SubgraphView::Layers &&layers) |
constexpr const char * | GpuFsaBackendId () |
template<typename ... Args> | |
bool | IsGpuFsaBackendSupported (Optional< std::string & > reasonIfUnsupported, Args... args) |
constexpr const char * | GpuFsaTensorHandleFactoryId () |
template<DataType ArmnnType> | |
bool | IsDataType (const WorkloadInfo &info) |
arm_compute::Status | GpuFsaActivationValidate (const TensorInfo &input, const ActivationDescriptor &descriptor) |
void | GpuFsaActivationCreateOp (GpuFsaPreCompiledBlob *blob, const TensorInfo &input, const ActivationDescriptor &descriptor) |
arm_compute::Status | GpuFsaBatchMatMulValidate (const TensorInfo &input0, const TensorInfo &input1, const BatchMatMulDescriptor &descriptor) |
void | GpuFsaBatchMatMulCreateOp (GpuFsaPreCompiledBlob *blob, const TensorInfo &input0, const TensorInfo &input1, const BatchMatMulDescriptor &descriptor) |
arm_compute::Status | GpuFsaCastValidate (const TensorInfo &input, const TensorInfo &output) |
void | GpuFsaCastCreateOp (GpuFsaPreCompiledBlob *blob, const TensorInfo &input, const TensorInfo &output) |
arm_compute::Status | GpuFsaConvolution2dValidate (const TensorInfo &input, const Convolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases) |
void | GpuFsaConvolution2dCreateOp (GpuFsaPreCompiledBlob *blob, const TensorInfo &input, const Convolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases) |
arm_compute::Status | GpuFsaDepthwiseConvolution2dValidate (const TensorInfo &input, const DepthwiseConvolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases) |
void | GpuFsaDepthwiseConvolution2dCreateOp (GpuFsaPreCompiledBlob *blob, const TensorInfo &input, const DepthwiseConvolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases) |
arm_compute::Status | GpuFsaElementwiseBinaryValidate (const TensorInfo &input0, const TensorInfo &input1, const ElementwiseBinaryDescriptor &descriptor) |
void | GpuFsaElementwiseBinaryCreateOp (GpuFsaPreCompiledBlob *blob, const TensorInfo &input0, const TensorInfo &input1, const ElementwiseBinaryDescriptor &descriptor) |
arm_compute::Status | GpuFsaPooling2dValidate (const TensorInfo &input, const Pooling2dDescriptor &descriptor) |
void | GpuFsaPooling2dCreateOp (GpuFsaPreCompiledBlob *blob, const TensorInfo &input, const Pooling2dDescriptor &descriptor) |
arm_compute::Status | GpuFsaReshapeValidate (const TensorInfo &input, const ReshapeDescriptor &descriptor) |
void | GpuFsaReshapeCreateOp (GpuFsaPreCompiledBlob *blob, const TensorInfo &input, const ReshapeDescriptor &descriptor) |
arm_compute::Status | GpuFsaResizeValidate (const TensorInfo &input, const ResizeDescriptor &descriptor) |
void | GpuFsaResizeCreateOp (GpuFsaPreCompiledBlob *blob, const TensorInfo &input, const ResizeDescriptor &descriptor) |
arm_compute::Status | GpuFsaSoftmaxValidate (const TensorInfo &input, const TensorInfo &output, const SoftmaxDescriptor &descriptor) |
void | GpuFsaSoftmaxCreateOp (GpuFsaPreCompiledBlob *blob, const TensorInfo &input, const TensorInfo &output, const SoftmaxDescriptor &descriptor) |
arm_compute::Status | GpuFsaConstantWorkloadValidate (const TensorInfo &output) |
bool | GpuFsaPreCompiledWorkloadValidate (std::string *reasonIfUnsupported) |
constexpr const char * | NeonBackendId () |
bool | CollapseLeadingUnitDimensions (const TensorInfo &in, TensorInfo &out) |
template<typename SlotListType > | |
void | BuildAddMulAddSlotLists (bool handleReLu, bool multipleOutputs, std::vector< SlotListType > &inputLayersSlotLists, std::vector< SlotListType > &outputLayersSlotLists) |
void | GetFusedName (Layer *layerList[4], std::string &fusedName) |
template<typename Type > | |
bool | BuildAddMulAddTensorInfoLists (Type *layerList[4], unsigned int &numInputs, unsigned int &numOutputs, std::vector< TensorInfo > &inputInfos, std::vector< TensorInfo > &outputInfos, const ActivationDescriptor *&activationDescriptor, bool &fuseReLu) |
bool | IsLayerTypeSupported (const LayerType &type, const std::vector< TensorInfo > &infos, const BaseDescriptor &descriptor, const Optional< LstmInputParamsInfo > &lstmParamsInfo, const Optional< QuantizedLstmInputParamsInfo > &quantizedLstmParamsInfo, Optional< std::string & > reasonIfUnsupported, const NeonLayerSupport &support) |
constexpr const char * | NeonTensorHandleFactoryId () |
arm_compute::Status | NeonAbsWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
arm_compute::Status | NeonActivationWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const ActivationDescriptor &descriptor) |
arm_compute::Status | NeonAdditionWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ActivationDescriptor *activationDescriptor) |
arm_compute::Status | NeonArgMinMaxWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const ArgMinMaxDescriptor &descriptor) |
arm_compute::Status | NeonBatchMatMulValidate (const TensorInfo &inputInfoX, const TensorInfo &inputInfoY, const TensorInfo &outputInfo, const BatchMatMulDescriptor &descriptor, const bool isFastMathEnabled, const ActivationDescriptor *activationDescriptor) |
arm_compute::Status | NeonBatchNormalizationValidate (const TensorInfo &input, const TensorInfo &output, const TensorInfo &mean, const TensorInfo &var, const TensorInfo &beta, const TensorInfo &gamma, const BatchNormalizationDescriptor &descriptor, const ActivationDescriptor *activationDescriptor) |
arm_compute::Status | NeonBatchToSpaceNdWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const BatchToSpaceNdDescriptor &descriptor) |
arm_compute::Status | NeonCastValidate (const TensorInfo &input, const TensorInfo &output) |
arm_compute::Status | NeonChannelShuffleValidate (const TensorInfo &input, const TensorInfo &output, const ChannelShuffleDescriptor &descriptor) |
arm_compute::Status | NeonComparisonWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ComparisonDescriptor &descriptor) |
arm_compute::Status | NeonConcatWorkloadValidate (const std::vector< const TensorInfo * > &inputs, const TensorInfo &output, const OriginsDescriptor &descriptor) |
arm_compute::Status | NeonConstantWorkloadValidate (const TensorInfo &output) |
arm_compute::Status | NeonConvertFp16ToFp32WorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
arm_compute::Status | NeonConvertFp32ToFp16WorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
arm_compute::Status | NeonConvolution2dWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const Convolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, bool isFastMathEnabled, const ActivationDescriptor *activationDescriptor) |
arm_compute::Status | NeonConvolution3dWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const Convolution3dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, bool isFastMathEnabled, const ActivationDescriptor *activationDescriptor) |
arm_compute::Status | NeonDepthToSpaceWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const DepthToSpaceDescriptor &descriptor) |
arm_compute::Status | NeonDepthwiseConvolutionWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const DepthwiseConvolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, const ActivationDescriptor *activationDescriptor) |
arm_compute::Status | NeonDequantizeWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
arm_compute::DetectionPostProcessLayerInfo | MakeInfo (const DetectionPostProcessDescriptor &descriptor) |
arm_compute::Status | NeonDetectionPostProcessValidate (const TensorInfo &boxEncodings, const TensorInfo &scores, const TensorInfo &anchors, const TensorInfo &detectionBoxes, const TensorInfo &detectionClasses, const TensorInfo &detectionScores, const TensorInfo &numDetections, const DetectionPostProcessDescriptor &descriptor) |
arm_compute::Status | NeonDivisionWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ActivationDescriptor *activationDescriptor) |
arm_compute::Status | NeonElementwiseBinaryWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ElementwiseBinaryDescriptor &descriptor, const ActivationDescriptor *activationDescriptor) |
arm_compute::Status | NeonExpWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
arm_compute::Status | NeonFullyConnectedWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const TensorInfo &weights, const Optional< TensorInfo > &biases, const FullyConnectedDescriptor &descriptor, const ActivationDescriptor *activationDescriptor) |
arm_compute::Status | NeonFusedWorkloadValidate (const std::vector< std::reference_wrapper< TensorInfo >> &inputInfos, const std::vector< std::reference_wrapper< TensorInfo >> &outputInfos, const FusedDescriptor &fusedDescriptor, const ActivationDescriptor *activationDescriptor) |
arm_compute::Status | NeonGatherNdWorkloadValidate (const TensorInfo ¶msInfo, const TensorInfo &indicesInfo, const TensorInfo &outputInfo) |
arm_compute::Status | NeonGatherWorkloadValidate (const TensorInfo &input, const TensorInfo &indices, const TensorInfo &output, const GatherDescriptor &descriptor) |
arm_compute::Status | NeonInstanceNormalizationWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const InstanceNormalizationDescriptor &descriptor) |
arm_compute::Status | NeonL2NormalizationWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const L2NormalizationDescriptor &descriptor) |
arm_compute::Status | NeonLogicalAndWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output) |
arm_compute::Status | NeonLogicalNotWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
arm_compute::Status | NeonLogicalOrWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output) |
arm_compute::Status | NeonLogSoftmaxWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const LogSoftmaxDescriptor &descriptor) |
arm_compute::Status | NeonLogWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
arm_compute::Status | NeonLstmFloatWorkloadValidate (const TensorInfo &input, const TensorInfo &outputStateIn, const TensorInfo &cellStateIn, const TensorInfo &scratchBuffer, const TensorInfo &outputStateOut, const TensorInfo &cellStateOut, const TensorInfo &output, const LstmDescriptor &descriptor, const LstmInputParamsInfo ¶msInfo) |
arm_compute::Status | NeonMaximumWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output) |
arm_compute::Status | NeonMeanWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const MeanDescriptor &descriptor) |
arm_compute::Status | NeonMinimumWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output) |
Validate function for validating the inputs and output. More... | |
arm_compute::Status | NeonMultiplicationWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ActivationDescriptor *activationDescriptor) |
arm_compute::Status | NeonNegWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
arm_compute::Status | NeonNormalizationWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const NormalizationDescriptor &descriptor) |
arm_compute::Status | NeonPadWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const PadDescriptor &descriptor) |
arm_compute::Status | NeonPermuteWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const PermuteDescriptor &descriptor) |
arm_compute::Status | NeonPooling2dWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const Pooling2dDescriptor &descriptor) |
arm_compute::Status | NeonPooling3dWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const Pooling3dDescriptor &descriptor) |
arm_compute::Status | NeonPreluWorkloadValidate (const TensorInfo &input, const TensorInfo &alpha, const TensorInfo &output) |
arm_compute::Status | NeonQLstmWorkloadValidate (const TensorInfo &input, const TensorInfo &cellStateIn, const TensorInfo &outputStateIn, const TensorInfo &cellStateOut, const TensorInfo &outputStateOut, const TensorInfo &output, const QLstmDescriptor &descriptor, const LstmInputParamsInfo ¶msInfo) |
arm_compute::Status | NeonQuantizedLstmWorkloadValidate (const TensorInfo &input, const TensorInfo &cellStateIn, const TensorInfo &outputStateIn, const TensorInfo &cellStateOut, const TensorInfo &outputStateOut, const QuantizedLstmInputParamsInfo ¶msInfo) |
arm_compute::Status | NeonQuantizeWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
arm_compute::Status | NeonReduceWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const ReduceDescriptor &descriptor) |
arm_compute::Status | NeonReshapeWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
arm_compute::Status | NeonResizeWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const ResizeDescriptor &descriptor) |
arm_compute::Status | NeonReverseV2WorkloadValidate (const TensorInfo &input, const TensorInfo &axis, const TensorInfo &output) |
arm_compute::Status | NeonRsqrtWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
arm_compute::Status | NeonSinWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
arm_compute::Status | NeonSliceWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const SliceDescriptor &descriptor) |
arm_compute::Status | NeonSoftmaxWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const SoftmaxDescriptor &descriptor) |
arm_compute::Status | NeonSpaceToBatchNdWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const SpaceToBatchNdDescriptor &descriptor) |
arm_compute::Status | NeonSpaceToDepthWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const SpaceToDepthDescriptor &descriptor) |
arm_compute::Status | NeonSplitterWorkloadValidate (const TensorInfo &input, const std::vector< std::reference_wrapper< TensorInfo >> &outputs, unsigned int splitAxis) |
arm_compute::Status | NeonSqrtWorkloadValidate (const TensorInfo &input, const TensorInfo &output) |
arm_compute::Status | NeonStackWorkloadValidate (const std::vector< const TensorInfo * > &inputs, const TensorInfo &output, const StackDescriptor &descriptor) |
arm_compute::Status | NeonStridedSliceWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const StridedSliceDescriptor &descriptor) |
arm_compute::Status | NeonSubtractionWorkloadValidate (const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ActivationDescriptor *activationDescriptor) |
arm_compute::Status | NeonTileWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const TileDescriptor &descriptor) |
arm_compute::Status | NeonTransposeConvolution2dWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const TransposeConvolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases) |
arm_compute::Status | NeonTransposeWorkloadValidate (const TensorInfo &input, const TensorInfo &output, const TransposeDescriptor &descriptor) |
arm_compute::Status | NeonUnidirectionalSequenceLstmFloatWorkloadValidate (const TensorInfo &input, const TensorInfo &outputStateIn, const TensorInfo &cellStateIn, const TensorInfo &outputStateOut, const TensorInfo &cellStateOut, const TensorInfo &output, const UnidirectionalSequenceLstmDescriptor &descriptor, const LstmInputParamsInfo ¶msInfo) |
arm_compute::Status | NeonUnidirectionalSequenceLstmWorkloadValidate (const TensorInfo &input, const TensorInfo &outputStateIn, const TensorInfo &cellStateIn, const TensorInfo &outputStateOut, const TensorInfo &cellStateOut, const TensorInfo &output, const UnidirectionalSequenceLstmDescriptor &descriptor, const LstmInputParamsInfo ¶msInfo) |
template<typename T > | |
void | CopyArmComputeTensorData (arm_compute::Tensor &dstTensor, const T *srcData) |
void | InitializeArmComputeTensorData (arm_compute::Tensor &tensor, TensorInfo tensorInfo, const ITensorHandle *handle) |
void | InitializeArmComputeTensorData (arm_compute::Tensor &tensor, const ConstTensorHandle *handle) |
auto | SetNeonStridedSliceData (const std::vector< int > &m_begin, const std::vector< int > &m_end, const std::vector< int > &m_stride) |
auto | SetNeonSliceData (const std::vector< unsigned int > &m_begin, const std::vector< unsigned int > &m_size) |
constexpr const char * | RefBackendId () |
constexpr const char * | RefTensorHandleFactoryId () |
bool | IsSigned64 (const WorkloadInfo &info) |
bool | IsSigned32 (const WorkloadInfo &info) |
bool | IsBFloat16 (const WorkloadInfo &info) |
bool | IsFloat16 (const WorkloadInfo &info) |
bool | IsQSymmS16 (const WorkloadInfo &info) |
bool | IsQSymmS8 (const WorkloadInfo &info) |
bool | IsQAsymmS8 (const WorkloadInfo &info) |
bool | IsQAsymmU8 (const WorkloadInfo &info) |
bool | IsBoolean (const WorkloadInfo &info) |
template<typename QueueDescriptorType > | |
constexpr bool | IsOperationQueueDescriptor (const QueueDescriptorType &) |
template<> | |
constexpr bool | IsOperationQueueDescriptor (const MemCopyQueueDescriptor &) |
template<> | |
constexpr bool | IsOperationQueueDescriptor (const ConstantQueueDescriptor &) |
template<> | |
constexpr bool | IsOperationQueueDescriptor (const PermuteQueueDescriptor &) |
float | Activation (float in, ActivationFunction function, float a, float b) |
void | Activation (Decoder< float > &in, Encoder< float > &out, const TensorInfo &tensorInfo, ActivationFunction function, float a, float b) |
template<typename OUT > | |
void | ArgMinMax (Decoder< float > &in, OUT *out, const TensorInfo &inputTensorInfo, const TensorInfo &outputTensorInfo, ArgMinMaxFunction function, int axis) |
template void | ArgMinMax (Decoder< float > &in, int32_t *out, const TensorInfo &inputTensorInfo, const TensorInfo &outputTensorInfo, ArgMinMaxFunction function, int axis) |
template void | ArgMinMax (Decoder< float > &in, int64_t *out, const TensorInfo &inputTensorInfo, const TensorInfo &outputTensorInfo, ArgMinMaxFunction function, int axis) |
void | BatchNormImpl (const BatchNormalizationQueueDescriptor &data, Decoder< float > &meanDecoder, Decoder< float > &varianceDecoder, Decoder< float > &betaDecoder, Decoder< float > &gammaDecoder, Decoder< float > &inputDecoder, Encoder< float > &outputEncoder) |
unsigned int | Offset (const TensorShape &shape, unsigned int batch, unsigned int height, unsigned int width, unsigned int channels, const DataLayoutIndexed &dataLayout) |
void | BatchToSpaceNd (const TensorInfo &inputInfo, const TensorInfo &outputInfo, const BatchToSpaceNdDescriptor ¶ms, Decoder< float > &inputData, Encoder< float > &outputData) |
void | Concatenate (const ConcatQueueDescriptor &data, std::vector< ITensorHandle * > inputs, std::vector< ITensorHandle * > outputs) |
void | Convolve3d (const TensorShape &rInputShape, Decoder< float > &rInputDecoder, const TensorShape &rOutputShape, Encoder< float > &rOutputEncoder, const TensorShape &rFilterShape, Decoder< float > &rFilterDecoder, bool biasEnabled, Decoder< float > *pBiasDecoder, DataLayout dataLayout, unsigned int paddingTop, unsigned int paddingLeft, unsigned int paddingFront, unsigned int xStride, unsigned int yStride, unsigned int zStride, unsigned int xDilation, unsigned int yDilation, unsigned int zDilation) |
void | Convolve (const TensorShape &rInputShape, Decoder< float > &rInputDecoder, const TensorShape &rOutputShape, Encoder< float > &rOutputEncoder, const TensorShape &rFilterShape, Decoder< float > &rFilterDecoder, bool biasEnabled, Decoder< float > *pBiasDecoder, DataLayout dataLayout, unsigned int paddingTop, unsigned int paddingLeft, unsigned int xStride, unsigned int yStride, unsigned int xDilation, unsigned int yDilation, bool depthwise) |
template<typename T > | |
void | PrintOutput (const TensorInfo &inputInfo, const T *inputData, LayerGuid guid, const std::string &layerName, unsigned int slotIndex, std::ostream &os) |
template<typename T > | |
void | Debug (const TensorInfo &inputInfo, const T *inputData, LayerGuid guid, const std::string &layerName, unsigned int slotIndex, bool outputsToFile) |
template void | Debug< BFloat16 > (const TensorInfo &inputInfo, const BFloat16 *inputData, LayerGuid guid, const std::string &layerName, unsigned int slotIndex, bool outputsToFile) |
template void | Debug< Half > (const TensorInfo &inputInfo, const Half *inputData, LayerGuid guid, const std::string &layerName, unsigned int slotIndex, bool outputsToFile) |
template void | Debug< float > (const TensorInfo &inputInfo, const float *inputData, LayerGuid guid, const std::string &layerName, unsigned int slotIndex, bool outputsToFile) |
template void | Debug< uint8_t > (const TensorInfo &inputInfo, const uint8_t *inputData, LayerGuid guid, const std::string &layerName, unsigned int slotIndex, bool outputsToFile) |
template void | Debug< int8_t > (const TensorInfo &inputInfo, const int8_t *inputData, LayerGuid guid, const std::string &layerName, unsigned int slotIndex, bool outputsToFile) |
template void | Debug< int16_t > (const TensorInfo &inputInfo, const int16_t *inputData, LayerGuid guid, const std::string &layerName, unsigned int slotIndex, bool outputsToFile) |
template void | Debug< int32_t > (const TensorInfo &inputInfo, const int32_t *inputData, LayerGuid guid, const std::string &layerName, unsigned int slotIndex, bool outputsToFile) |
template void | Debug< int64_t > (const TensorInfo &inputInfo, const int64_t *inputData, LayerGuid guid, const std::string &layerName, unsigned int slotIndex, bool outputsToFile) |
template<typename T > | |
std::unique_ptr< Decoder< T > > | MakeDecoder (const TensorInfo &info, const void *data=nullptr) |
template<> | |
std::unique_ptr< Decoder< float > > | MakeDecoder (const TensorInfo &info, const void *data) |
void | DepthToSpace (const TensorInfo &inputInfo, const DepthToSpaceDescriptor &descriptor, const void *inputData, void *outputData, unsigned int dataTypeSize) |
void | Dequantize (Decoder< float > &inputDecoder, Encoder< float > &outputEncoder, const TensorInfo &inputInfo, const TensorInfo &outputInfo) |
std::vector< unsigned int > | GenerateRangeK (unsigned int k) |
void | TopKSort (unsigned int k, unsigned int *indices, const float *values, unsigned int numElement) |
float | IntersectionOverUnion (const float *boxI, const float *boxJ) |
std::vector< unsigned int > | NonMaxSuppression (unsigned int numBoxes, const std::vector< float > &boxCorners, const std::vector< float > &scores, float nmsScoreThreshold, unsigned int maxDetection, float nmsIouThreshold) |
void | AllocateOutputData (unsigned int numOutput, unsigned int numSelected, const std::vector< float > &boxCorners, const std::vector< unsigned int > &outputIndices, const std::vector< unsigned int > &selectedBoxes, const std::vector< unsigned int > &selectedClasses, const std::vector< float > &selectedScores, float *detectionBoxes, float *detectionScores, float *detectionClasses, float *numDetections) |
void | DetectionPostProcess (const TensorInfo &boxEncodingsInfo, const TensorInfo &scoresInfo, const TensorInfo &, const TensorInfo &detectionBoxesInfo, const TensorInfo &, const TensorInfo &, const TensorInfo &, const DetectionPostProcessDescriptor &desc, Decoder< float > &boxEncodings, Decoder< float > &scores, Decoder< float > &anchors, float *detectionBoxes, float *detectionClasses, float *detectionScores, float *numDetections) |
template<typename T > | |
std::unique_ptr< Encoder< T > > | MakeEncoder (const TensorInfo &info, void *data=nullptr) |
template<> | |
std::unique_ptr< Encoder< float > > | MakeEncoder (const TensorInfo &info, void *data) |
void | Fill (Encoder< float > &output, const TensorShape &desiredOutputShape, const float value) |
Creates a tensor and fills it with a scalar value. More... | |
void | FullyConnected (const TensorShape &rInputShape, Decoder< float > &rInputDecoder, const TensorShape &rOutputShape, Encoder< float > &rOutputEncoder, const TensorShape &rWeightsShape, Decoder< float > &rWeightDecoder, Decoder< float > *rBiasDecoder, bool biasEnabled, unsigned int K, bool transposeWeights) |
Performs a matrix multiplication and optionally adds a bias. More... | |
void | Gather (const TensorInfo ¶msInfo, const TensorInfo &indicesInfo, const TensorInfo &outputInfo, Decoder< float > ¶ms, const int32_t *indices, Encoder< float > &output, const int32_t axis_int) |
void | InstanceNorm (const InstanceNormalizationQueueDescriptor &data, const TensorInfo &inputInfo, Decoder< float > &inputDecoder, Encoder< float > &outputEncoder) |
void | LogSoftmax (Decoder< float > &input, Encoder< float > &output, const TensorInfo &inputInfo, const LogSoftmaxDescriptor &descriptor) |
void | LstmImpl (const LstmDescriptor &descriptor, const TensorInfo &inputInfo, const TensorInfo &outputInfo, const TensorShape &inputToOutputWeightsShape, const TensorShape &recurrentToOutputWeightsShape, std::unique_ptr< Decoder< float >> &inputData, std::unique_ptr< Decoder< float >> &outputStateIn, std::unique_ptr< Decoder< float >> &cellStateIn, std::unique_ptr< Encoder< float >> &outputStateOut, std::unique_ptr< Encoder< float >> &cellStateOut, std::unique_ptr< Encoder< float >> &output, std::unique_ptr< Decoder< float >> &cellStateOutDecoder, std::unique_ptr< Decoder< float >> &outputDecoder, std::unique_ptr< Decoder< float >> &inputToInputWeightsTensor, std::unique_ptr< Decoder< float >> &inputToForgetWeightsTensor, std::unique_ptr< Decoder< float >> &inputToCellWeightsTensor, std::unique_ptr< Decoder< float >> &inputToOutputWeightsTensor, std::unique_ptr< Decoder< float >> &recurrentToInputWeightsTensor, std::unique_ptr< Decoder< float >> &recurrentToForgetWeightsTensor, std::unique_ptr< Decoder< float >> &recurrentToCellWeightsTensor, std::unique_ptr< Decoder< float >> &recurrentToOutputWeightsTensor, std::unique_ptr< Decoder< float >> &cellToInputWeightsTensor, std::unique_ptr< Decoder< float >> &cellToForgetWeightsTensor, std::unique_ptr< Decoder< float >> &cellToOutputWeightsTensor, std::unique_ptr< Decoder< float >> &inputGateBiasTensor, std::unique_ptr< Decoder< float >> &forgetGateBiasTensor, std::unique_ptr< Decoder< float >> &cellBiasTensor, std::unique_ptr< Decoder< float >> &outputGateBiasTensor, std::unique_ptr< Decoder< float >> &projectionWeightsTensor, std::unique_ptr< Decoder< float >> &projectionBiasTensor, std::unique_ptr< Decoder< float >> &inputLayerNormWeights, std::unique_ptr< Decoder< float >> &forgetLayerNormWeights, std::unique_ptr< Decoder< float >> &cellLayerNormWeights, std::unique_ptr< Decoder< float >> &outputLayerNormWeights, std::unique_ptr< Encoder< float >> &inputGateScratch, std::unique_ptr< Encoder< float >> &cellScratch, std::unique_ptr< Encoder< float >> &forgetGateScratch, std::unique_ptr< Encoder< float >> &outputGateScratch, std::unique_ptr< Decoder< float >> &inputGateScratchDecoder, std::unique_ptr< Decoder< float >> &cellScratchDecoder, std::unique_ptr< Decoder< float >> &forgetGateScratchDecoder, std::unique_ptr< Decoder< float >> &outputGateScratchDecoder, float layerNormEpsilon) |
void | MirrorPad (const TensorInfo &inputInfo, const TensorInfo &outputInfo, const ITensorHandle *inputHandle, ITensorHandle *outputHandle, const PadQueueDescriptor &data) |
void | Pad (const TensorInfo &inputInfo, const TensorInfo &outputInfo, const ITensorHandle *inputHandle, ITensorHandle *outputHandle, const PadQueueDescriptor &data) |
void | Pooling2d (Decoder< float > &rInputDecoder, Encoder< float > &rOutputEncoder, const TensorInfo &inputInfo, const TensorInfo &outputInfo, const Pooling2dDescriptor ¶ms) |
Computes the Pooling2d operation. More... | |
void | Pooling3d (Decoder< float > &rInputDecoder, Encoder< float > &rOutputEncoder, const TensorInfo &inputInfo, const TensorInfo &outputInfo, const Pooling3dDescriptor ¶ms) |
Computes the Pooling3d operation. More... | |
void | PreluImpl (const TensorInfo &inputInfo, const TensorInfo &alphaInfo, const TensorInfo &outputInfo, Decoder< float > &inputData, Decoder< float > &alphaData, Encoder< float > &outputData) |
bool | NextIndex (const unsigned int numDims, const armnn::TensorShape &dims, std::vector< unsigned int > ¤t) |
unsigned int | ReducedOutputOffset (const unsigned int numDims, const armnn::TensorShape &dims, std::vector< unsigned int > &index, const unsigned int numAxis, const std::vector< unsigned int > &axis) |
void | Reduce (const TensorInfo &inputInfo, const TensorInfo &outputInfo, Decoder< float > &input, Encoder< float > &output, const std::vector< uint32_t > axis, const ReduceOperation reduceOperation) |
template<typename DataType > | |
void | ExecuteFunction (std::vector< ITensorHandle * > inputs, std::vector< ITensorHandle * > outputs, BinaryOperation operation, const std::string &layerName="") |
void | FakeQuantization (const float *inputData, float *outputData, uint32_t numElements, float min, float max) |
unsigned int | GetNumActivations (const TensorInfo &inputInfo) |
template<typename TensorHandleType = RefTensorHandle> | |
const TensorInfo & | GetTensorInfo (const ITensorHandle *tensorHandle) |
float32 helpers More... | |
template<typename DataType , typename PayloadType > | |
const DataType * | GetInputTensorData (unsigned int idx, const PayloadType &data) |
template<typename DataType > | |
DataType * | GetOutputTensorData (ITensorHandle *tensorHandle) |
template<typename PayloadType > | |
const float * | GetInputTensorDataFloat (unsigned int idx, const PayloadType &data) |
template<typename PayloadType > | |
float * | GetOutputTensorDataFloat (unsigned int idx, const PayloadType &data) |
template<typename PayloadType > | |
const Half * | GetInputTensorDataHalf (unsigned int idx, const PayloadType &data) |
template<typename PayloadType > | |
Half * | GetOutputTensorDataHalf (unsigned int idx, const PayloadType &data) |
template<typename PayloadType > | |
const BFloat16 * | GetInputTensorDataBFloat16 (unsigned int idx, const PayloadType &data) |
template<typename PayloadType > | |
BFloat16 * | GetOutputTensorDataBFloat16 (unsigned int idx, const PayloadType &data) |
template<typename T > | |
std::vector< float > | Dequantize (const T *quant, const TensorInfo &info) |
u8 helpers More... | |
template<typename T > | |
void | Dequantize (const T *inputData, float *outputData, const TensorInfo &info) |
void | Quantize (uint8_t *quant, const float *dequant, const TensorInfo &info) |
void | Resize (Decoder< float > &in, const TensorInfo &inputInfo, Encoder< float > &out, const TensorInfo &outputInfo, DataLayoutIndexed dataLayout, ResizeMethod resizeMethod, bool alignCorners, bool halfPixelCenters) |
std::vector< unsigned int > | ReverseGetMultIdx (const unsigned int idx, unsigned int inputRank, std::vector< unsigned int > &elementNumInner) |
unsigned int | ReverseGetFlatIdx (const std::vector< unsigned int > &idxList, unsigned int inputRank, std::vector< unsigned int > &elementNumInner) |
unsigned int | ReverseRelocateIdx (unsigned int idx, unsigned int inputRank, std::vector< bool > &axisFlag, std::vector< unsigned int > &dimSize, std::vector< unsigned int > &elementNumInner) |
void | ReverseV2 (const TensorInfo &inputInfo, const TensorInfo &axisInfo, Decoder< float > &inputDecoder, Decoder< int > &axisDecoder, Encoder< float > &outputEncoder) |
float | ScatterOperation (ScatterNdFunction operation, float input, float update) |
void | ScatterNd (const TensorInfo &inputInfo, const TensorInfo &indicesInfo, const TensorInfo &updatesInfo, Decoder< float > &input, Decoder< int > &indices, Decoder< float > &updates, Encoder< float > &output, ScatterNdDescriptor descriptor) |
void | ScatterNd (const TensorInfo &indicesInfo, const TensorInfo &updatesInfo, const TensorInfo &shapeInfo, Decoder< int > &indices, Decoder< float > &updates, Decoder< int > &shape, Encoder< float > &output, ScatterNdDescriptor descriptor) |
void | Slice (const TensorInfo &inputInfo, const SliceDescriptor &descriptor, const void *inputData, void *outputData, unsigned int dataTypeSize) |
void | Softmax (Decoder< float > &in, Encoder< float > &out, const TensorInfo &inputTensorInfo, float beta, int axis) |
Computes the softmax function on some inputs, into outputs, with a shape given by tensorInfo. More... | |
unsigned int | GetOffset (const TensorShape &shape, unsigned int b, unsigned int h, unsigned int w, unsigned int c, const DataLayoutIndexed &dataLayout) |
void | SpaceToBatchNd (const TensorInfo &inputInfo, const TensorInfo &outputInfo, const SpaceToBatchNdDescriptor ¶ms, Decoder< float > &inputData, Encoder< float > &outputData) |
void | SpaceToDepth (const TensorInfo &inputInfo, const TensorInfo &outputInfo, const SpaceToDepthDescriptor ¶ms, Decoder< float > &inputData, Encoder< float > &outputData) |
void | Split (const SplitterQueueDescriptor &data, std::vector< ITensorHandle * > inputs, std::vector< ITensorHandle * > outputs) |
template<typename DataType > | |
void | Splitter (const SplitterQueueDescriptor &data, std::vector< ITensorHandle * > inputs, std::vector< ITensorHandle * > outputs) |
void | Stack (const StackQueueDescriptor &data, std::vector< std::unique_ptr< Decoder< float >>> &inputs, Encoder< float > &output, const TensorInfo &inputInfo, const TensorInfo &outputInfo) |
void | StridedSlice (const TensorInfo &inputInfo, const StridedSliceDescriptor ¶ms, const void *inputData, void *outputData, unsigned int dataTypeSize) |
std::vector< uint32_t > | IndexToCoordinates (std::vector< uint32_t > &shape, uint32_t index) |
uint32_t | CoordinatesToIndex (TensorShape &shape, std::vector< uint32_t > &coordinates) |
void | Tile (const TileDescriptor ¶ms, const TensorInfo &inputInfo, Decoder< float > &inputDecoder, Encoder< float > &outputEncoder) |
void | TransposeConvolution2dImpl (const TransposeConvolution2dDescriptor &descriptor, const TensorShape &inputShape, Decoder< float > &inputDecoder, const TensorShape &outputShape, Encoder< float > &outputEncoder, const TensorShape &weightsShape, Decoder< float > &weightsDecoder, Decoder< float > *biasesDecoder) |
constexpr const char * | TosaRefBackendId () |
constexpr const char * | TosaRefTensorHandleFactoryId () |
bool | TosaRefPreCompiledWorkloadValidate (std::string *) |
Variables | |
constexpr unsigned int | MaxNumOfTensorDimensions = 5U |
constexpr unsigned int | LOWEST_CAPTURE_PERIOD = 10000u |
The lowest performance data capture interval we support is 10 miliseconds. More... | |
constexpr unsigned int | EXPIRE_RATE = 3U |
Variable to control expire rate of priority queue. More... | |
constexpr std::size_t | g_ProfilingEventCountHint = 1024 |
constexpr bool | g_WriteProfilingEventSequence = true |
constexpr bool | g_AggregateProfilingEventsByInference = true |
constexpr bool | g_WriteReportToStdOutOnProfilerDestruction = false |
thread_local IProfiler * | tl_Profiler = nullptr |
constexpr size_t | wordSize = sizeof(size_t) * 8 |
const BackendCapabilities | gpuFsaCapabilities ("GpuFsa", { {"NonConstWeights", false}, {"AsyncExecution", false}, {"ProtectedContentAllocation", false}, {"ConstantTensorsAsInputs", true}, {"PreImportIOTensors", false}, {"ExternallyManagedMemory", false}, {"MultiAxisPacking", false}, {"SingleAxisPacking", false} }) |
const BackendCapabilities | cpuAccCapabilities ("CpuAcc", { {"NonConstWeights", true}, {"AsyncExecution", false}, {"ProtectedContentAllocation", false}, {"ConstantTensorsAsInputs", true}, {"PreImportIOTensors", false}, {"ExternallyManagedMemory", true}, {"MultiAxisPacking", false}, {"SingleAxisPacking", true}, {"HasFp16", arm_compute::CPUInfo::get().has_fp16()} }) |
const std::set< armnn::LayerType > | paddingRequiredLayers |
const BackendCapabilities | cpuRefCapabilities ("CpuRef", { {"NonConstWeights", true}, {"AsyncExecution", true}, {"ProtectedContentAllocation", false}, {"ConstantTensorsAsInputs", true}, {"PreImportIOTensors", true}, {"ExternallyManagedMemory", true}, {"MultiAxisPacking", false}, {"SingleAxisPacking", true}, {"HasFp16", true} }) |
const std::set< armnn::BackendCapability > | oldCpuRefCapabilities |
Copyright (c) 2021 ARM Limited and Contributors.
Optional is a drop in replacement for std::optional until we migrate to c++-17.
Copyright (c) 2022-2023 ARM Limited and Contributors.
Copyright (c) 2021-2023 ARM Limited and Contributors.
Copyright (c) 2022-2024 Arm Ltd and Contributors.
Copyright (c) 2022 ARM Limited and Contributors.
Create pages for each tool so they appear nicely in the doxygen tree-view.
Copyright (c) 2021-2024 ARM Limited and Contributors.
All rights reserved.
SPDX-License-Identifier: MIT
Subpages are not listed there. Also we can overwrite the page name this way.
Subpages are not listed there.
Note: The parser, serializer and deserializer pages are created in 01_parsers.dox or 02_deserializer_serializer.dox
Only a subset of the optional features are implemented that we intend to use in ArmNN. There are two distinct implementations here:
1, for normal constructable/destructable types and reference types 2, for reference types The std::optional features we support are:
using ACLMemManagerOnDemand = std::shared_ptr<arm_compute::MemoryManagerOnDemand> |
Definition at line 22 of file NeonFullyConnectedWorkload.cpp.
using AdditionalInfoObjectPtr = std::shared_ptr<void> |
using BackendCapabilities = BackendOptions |
Definition at line 19 of file BackendOptions.hpp.
using BackendIdSet = std::unordered_set<BackendId> |
Definition at line 193 of file BackendId.hpp.
using BackendIdVector = std::vector<BackendId> |
Definition at line 192 of file BackendId.hpp.
using BackendsMap = std::map<BackendId, std::unique_ptr<class IBackendInternal> > |
Definition at line 285 of file Network.hpp.
using BaseFloat32ComparisonWorkload = MultiTypedWorkload<QueueDescriptor, armnn::DataType::Float32, armnn::DataType::Boolean> |
Definition at line 234 of file Workload.hpp.
using BaseUint8ComparisonWorkload = MultiTypedWorkload<QueueDescriptor, armnn::DataType::QAsymmU8, armnn::DataType::Boolean> |
Definition at line 239 of file Workload.hpp.
using BFloat16ToFloat32Workload = MultiTypedWorkload<QueueDescriptor, armnn::DataType::BFloat16, armnn::DataType::Float32> |
Definition at line 244 of file Workload.hpp.
using BindingPointInfo = std::pair<armnn::LayerBindingId, armnn::TensorInfo> |
Definition at line 276 of file Tensor.hpp.
Definition at line 229 of file Workload.hpp.
typedef std::function< void(const void *)> CompiledBlobDeleter |
Definition at line 342 of file INetwork.hpp.
typedef std::unique_ptr< void, CompiledBlobDeleter > CompiledBlobPtr |
Definition at line 343 of file INetwork.hpp.
using ConcatDescriptor = OriginsDescriptor |
Definition at line 60 of file DescriptorsFwd.hpp.
using Coordinates = std::array<unsigned int, MaxNumOfTensorDimensions> |
Definition at line 15 of file InternalTypes.hpp.
using CopyAndImportFactoryPairs = std::map<ITensorHandleFactory::FactoryId, ITensorHandleFactory::FactoryId> |
Definition at line 20 of file TensorHandleFactoryRegistry.hpp.
using DebugCallbackFunction = std::function<void(LayerGuid guid, unsigned int slotIndex, ITensorHandle* tensorHandle)> |
Define the type of callback for the Debug layer to call.
guid | - guid of layer connected to the input of the Debug layer |
slotIndex | - index of the output slot connected to the input of the Debug layer |
tensorHandle | - TensorHandle for the input tensor to the Debug layer |
A DepthToSpaceDescriptor for the DepthToSpaceLayer.
Definition at line 1099 of file Descriptors.hpp.
using Dimensions = std::array<unsigned int, MaxNumOfTensorDimensions> |
Definition at line 16 of file InternalTypes.hpp.
using DynamicBackendPtr = std::unique_ptr<DynamicBackend> |
Definition at line 54 of file DynamicBackend.hpp.
Definition at line 12 of file MockTensorHandleFactory.cpp.
using Float16ToFloat32Workload = MultiTypedWorkload<QueueDescriptor, armnn::DataType::Float16, armnn::DataType::Float32> |
Definition at line 254 of file Workload.hpp.
using Float32ToBFloat16Workload = MultiTypedWorkload<QueueDescriptor, armnn::DataType::Float32, armnn::DataType::BFloat16> |
Definition at line 249 of file Workload.hpp.
using Float32ToFloat16Workload = MultiTypedWorkload<QueueDescriptor, armnn::DataType::Float32, armnn::DataType::Float16> |
Definition at line 259 of file Workload.hpp.
Definition at line 220 of file Workload.hpp.
using FloatWorkload = TypedWorkload<QueueDescriptor, armnn::DataType::Float16, armnn::DataType::Float32> |
Definition at line 217 of file Workload.hpp.
using HighResolutionClock = std::chrono::high_resolution_clock::time_point |
using IBackendContextUniquePtr = std::unique_ptr<IBackendContext> |
Definition at line 34 of file IBackendContext.hpp.
typedef std::unique_ptr< IBackendInternal > IBackendInternalUniquePtr |
Definition at line 32 of file BackendRegistry.hpp.
using IBackendSharedPtr = std::shared_ptr<IBackend> |
using IBackendUniquePtr = std::unique_ptr<IBackend, void(*)(IBackend* backend)> |
using IGpuAccTunedParametersPtr = std::shared_ptr<IGpuAccTunedParameters> |
The following API is replaced by the backend options API.
Definition at line 301 of file IRuntime.hpp.
using IInitialiseProfilingService = arm::pipe::IInitialiseProfilingService |
Definition at line 28 of file Runtime.hpp.
using ILayerSupportSharedPtr = std::shared_ptr<ILayerSupport> |
Definition at line 40 of file ILayerSupport.hpp.
using IMemoryManagerUniquePtr = std::unique_ptr<IMemoryManager> |
Definition at line 24 of file IMemoryManager.hpp.
using ImportedInputId = unsigned int |
using ImportedOutputId = unsigned int |
using INetworkPtr = std::unique_ptr<INetwork, void(*)(INetwork* network)> |
Definition at line 339 of file INetwork.hpp.
using InferenceTimingPair = std::pair<HighResolutionClock, HighResolutionClock> |
Definition at line 91 of file WorkloadData.hpp.
using InputTensors = std::vector<std::pair<LayerBindingId, class ConstTensor> > |
Definition at line 394 of file Tensor.hpp.
Definition at line 226 of file Workload.hpp.
using IOptimizedNetworkPtr = std::unique_ptr<IOptimizedNetwork, void(*)(IOptimizedNetwork* network)> |
Definition at line 340 of file INetwork.hpp.
using IReportStructure = arm::pipe::IReportStructure |
Definition at line 27 of file Runtime.hpp.
using IRuntimePtr = std::unique_ptr<IRuntime, void(*)(IRuntime* runtime)> |
Definition at line 41 of file IRuntime.hpp.
using LayerBindingId = int |
using LayerPriority = unsigned int |
using LayerTypeOf = typename LayerTypeOfImpl<Type>::Type |
Definition at line 95 of file LayersFwd.hpp.
using LoadedNetworks = std::unordered_map<NetworkId, std::unique_ptr<LoadedNetwork> > |
Definition at line 26 of file Runtime.hpp.
A LogSoftmaxDescriptor for the LogSoftmaxLayer.
Definition at line 196 of file Descriptors.hpp.
using MemoryOptimizerStrategiesMapRef = std::unordered_map<BackendId, std::shared_ptr<IMemoryOptimizerStrategy> > |
Definition at line 33 of file BackendRegistry.hpp.
using MemorySourceFlags = unsigned int |
Definition at line 15 of file MemorySources.hpp.
using MergerDescriptor = OriginsDescriptor |
MergerDescriptor is deprecated, use ConcatDescriptor instead.
Definition at line 64 of file DescriptorsFwd.hpp.
Definition at line 149 of file WorkloadData.hpp.
using ModelOptions = std::vector<BackendOptions> |
Definition at line 18 of file BackendOptions.hpp.
typedef int NetworkId |
Definition at line 35 of file IRuntime.hpp.
using NetworkImplPtr = std::unique_ptr<NetworkImpl, void (*)(NetworkImpl* network)> |
Definition at line 29 of file Network.hpp.
using NetworkOptions = std::vector<BackendOptions> |
Definition at line 16 of file BackendOptions.hpp.
Definition at line 92 of file WorkloadData.hpp.
using OutputTensors = std::vector<std::pair<LayerBindingId, class Tensor> > |
Definition at line 395 of file Tensor.hpp.
using ParameterStringifyFunction = std::function<void(const std::string& name, const std::string& value)> |
Definition at line 14 of file SerializeLayerParameters.hpp.
using PreCompiledObjectDeleter = std::function<void(const void*)> |
Definition at line 19 of file PreCompiledLayer.hpp.
using PreCompiledObjectPtr = std::unique_ptr<void, PreCompiledObjectDeleter> |
Definition at line 20 of file PreCompiledLayer.hpp.
using RefAdditionWorkload = RefElementwiseWorkload<std::plus<DataType>, AdditionQueueDescriptor, StringMapping::RefAdditionWorkload_Execute> |
Definition at line 40 of file RefElementwiseWorkload.hpp.
Definition at line 42 of file RefDebugWorkload.hpp.
Definition at line 51 of file RefDebugWorkload.hpp.
Definition at line 43 of file RefDebugWorkload.hpp.
Definition at line 44 of file RefDebugWorkload.hpp.
Definition at line 46 of file RefDebugWorkload.hpp.
Definition at line 45 of file RefDebugWorkload.hpp.
Definition at line 47 of file RefDebugWorkload.hpp.
Definition at line 48 of file RefDebugWorkload.hpp.
Definition at line 49 of file RefDebugWorkload.hpp.
Definition at line 50 of file RefDebugWorkload.hpp.
using RefDivisionWorkload = RefElementwiseWorkload<std::divides<DataType>, DivisionQueueDescriptor, StringMapping::RefDivisionWorkload_Execute> |
Definition at line 58 of file RefElementwiseWorkload.hpp.
using RefMaximumWorkload = RefElementwiseWorkload<armnn::maximum<DataType>, MaximumQueueDescriptor, StringMapping::RefMaximumWorkload_Execute> |
Definition at line 64 of file RefElementwiseWorkload.hpp.
using RefMinimumWorkload = RefElementwiseWorkload<armnn::minimum<DataType>, MinimumQueueDescriptor, StringMapping::RefMinimumWorkload_Execute> |
Definition at line 70 of file RefElementwiseWorkload.hpp.
using RefMultiplicationWorkload = RefElementwiseWorkload<std::multiplies<DataType>, MultiplicationQueueDescriptor, StringMapping::RefMultiplicationWorkload_Execute> |
Definition at line 52 of file RefElementwiseWorkload.hpp.
Definition at line 27 of file RefPermuteWorkload.hpp.
Definition at line 28 of file RefPermuteWorkload.hpp.
Definition at line 29 of file RefPermuteWorkload.hpp.
Definition at line 31 of file RefPermuteWorkload.hpp.
Definition at line 30 of file RefPermuteWorkload.hpp.
Definition at line 32 of file RefPermuteWorkload.hpp.
using RefSubtractionWorkload = RefElementwiseWorkload<std::minus<DataType>, SubtractionQueueDescriptor, StringMapping::RefSubtractionWorkload_Execute> |
Definition at line 46 of file RefElementwiseWorkload.hpp.
Definition at line 27 of file RefTransposeWorkload.hpp.
Definition at line 28 of file RefTransposeWorkload.hpp.
Definition at line 29 of file RefTransposeWorkload.hpp.
Definition at line 31 of file RefTransposeWorkload.hpp.
Definition at line 30 of file RefTransposeWorkload.hpp.
Definition at line 32 of file RefTransposeWorkload.hpp.
using ResolveType = typename ResolveTypeImpl<DT>::Type |
Definition at line 79 of file ResolveType.hpp.
using SplitterDescriptor = ViewsDescriptor |
Definition at line 65 of file DescriptorsFwd.hpp.
using TensorInfos = std::vector<TensorInfo> |
Definition at line 152 of file BackendHelper.cpp.
using Uint8ToFloat32Workload = MultiTypedWorkload<QueueDescriptor, armnn::DataType::QAsymmU8, armnn::DataType::Float32> |
Definition at line 264 of file Workload.hpp.
Definition at line 223 of file Workload.hpp.
Definition at line 1169 of file Descriptors.hpp.
using WorkloadQueue = std::vector< std::unique_ptr<IWorkload> > |
Definition at line 13 of file ExecutionFrame.hpp.
|
strong |
|
strong |
|
strong |
|
strong |
|
strong |
|
strong |
Capability class to calculate in the GetCapabilities function so that only the capability in the scope can be choose to calculate.
Enumerator | |
---|---|
PaddingRequired | |
FallbackImportDisabled | |
CapabilityClassMax |
Definition at line 24 of file ITensorHandleFactory.hpp.
|
strong |
|
strong |
The Compute enum is now deprecated and it is now being replaced by BackendId.
Enumerator | |
---|---|
Undefined | |
CpuRef | CPU Execution: Reference C++ kernels. |
CpuAcc | CPU Execution: NEON: ArmCompute. |
GpuAcc | GPU Execution: OpenCL: ArmCompute. |
Definition at line 21 of file BackendId.hpp.
|
strong |
|
strong |
|
strong |
|
strong |
Definition at line 104 of file ITensorHandleFactory.hpp.
|
strong |
|
strong |
Enumerator | |
---|---|
LayerAdded | |
LayerErased |
Definition at line 12 of file IGraphObservable.hpp.
|
strong |
Enumerator | |
---|---|
Measurement | |
Event | |
ExecObjectDesc |
Definition at line 20 of file JsonPrinter.hpp.
|
strong |
When adding a new layer, adapt also the LastLayer enum value in the enum class LayerType below.
Definition at line 491 of file Types.hpp.
|
strong |
|
strong |
|
strong |
|
strong |
|
strong |
|
strong |
|
strong |
|
strong |
The padding method modifies the output of pooling layers.
In both supported methods, the values are ignored (they are not even zeroes, which would make a difference for max pooling a tensor with negative values). The difference between IgnoreValue and Exclude is that the former counts the padding fields in the divisor of Average and L2 pooling, while Exclude does not.
Enumerator | |
---|---|
IgnoreValue | The padding fields count, but are ignored. |
Exclude | The padding fields don't count and are ignored. |
Definition at line 188 of file Types.hpp.
|
strong |
|
strong |
|
strong |
|
strong |
|
strong |
|
strong |
|
strong |
|
strong |
The ShapeInferenceMethod modify how the output shapes are treated.
When ValidateOnly is selected, the output shapes are inferred from the input parameters of the layer and any mismatch is reported. When InferAndValidate is selected 2 actions are performed: (1)infer output shape from inputs and (2)validate the shapes as in ValidateOnly. This option has been added to work with tensors which rank or dimension sizes are not specified explicitly, however this information can be calculated from the inputs.
Enumerator | |
---|---|
ValidateOnly | Validate all output shapes. |
InferAndValidate | Infer missing output shapes and validate all output shapes. |
Definition at line 235 of file Types.hpp.
|
strong |
|
strong |
Enumerator | |
---|---|
None | |
Rapid | |
Normal | |
Exhaustive |
Definition at line 18 of file ArmComputeTuningUtils.hpp.
|
strong |
void Activation | ( | Decoder< float > & | in, |
Encoder< float > & | out, | ||
const TensorInfo & | tensorInfo, | ||
ActivationFunction | function, | ||
float | a, | ||
float | b | ||
) |
Definition at line 102 of file Activation.cpp.
References Activation(), Decoder< IType >::Get(), TensorInfo::GetNumElements(), and Encoder< IType >::Set().
float Activation | ( | float | in, |
ActivationFunction | function, | ||
float | a, | ||
float | b | ||
) |
Definition at line 13 of file Activation.cpp.
References Abs, BoundedReLu, Elu, Gelu, HardSwish, LeakyReLu, Linear, ReLu, Sigmoid, SoftReLu, Sqrt, Square, and TanH.
Referenced by Activation(), and LstmImpl().
void armnn::AllocateOutputData | ( | unsigned int | numOutput, |
unsigned int | numSelected, | ||
const std::vector< float > & | boxCorners, | ||
const std::vector< unsigned int > & | outputIndices, | ||
const std::vector< unsigned int > & | selectedBoxes, | ||
const std::vector< unsigned int > & | selectedClasses, | ||
const std::vector< float > & | selectedScores, | ||
float * | detectionBoxes, | ||
float * | detectionScores, | ||
float * | detectionClasses, | ||
float * | numDetections | ||
) |
bool armnn::AllTypesAreEqualImpl | ( | T | t1, |
T | t2, | ||
Rest... | rest | ||
) |
bool armnn::AllTypesAreEqualImpl | ( | T | ) |
Definition at line 58 of file LayerSupportRules.hpp.
Referenced by AllTypesAreEqualImpl(), and TypesAreEqual::TypesAreEqual().
void armnn::Append | ( | Optimizer::Optimizations & | optimizations, |
Front && | front, | ||
Others &&... | others | ||
) |
void armnn::Append | ( | Optimizer::Optimizations & | optimizations, |
T && | optimization | ||
) |
Definition at line 30 of file Optimizer.hpp.
Referenced by Append(), and MakeOptimizations().
OptimizationResult armnn::ApplyBackendOptimizations | ( | OptimizedNetworkImpl * | optNetObjPtr, |
BackendSettings & | backendSettings, | ||
BackendsMap & | backends, | ||
const ModelOptions & | modelOptions, | ||
Optional< std::vector< std::string > & > | errMessages | ||
) |
Definition at line 1320 of file Network.cpp.
References ARMNN_SCOPED_PROFILING_EVENT, AssignBackends(), OutputSlot::Connect(), CpuAcc, OutputSlot::Disconnect(), Graph::EraseLayer(), Layer::GetBackendId(), OutputSlot::GetConnection(), OptimizationViews::GetDeletedSubgraphs(), OptimizationViews::GetFailedSubgraphs(), OptimizedNetworkImpl::GetGraph(), SubgraphView::GetIConnectableLayers(), Layer::GetInputSlot(), OutputSlot::GetNumConnections(), Layer::GetNumInputSlots(), Layer::GetOutputSlot(), OptimizationViews::GetSubstitutions(), Layer::GetType(), GpuAcc, Input, OptimizationResult::m_Error, BackendSettings::m_IgnoredBackends, BackendSettings::m_SelectedBackends, MakeOptimizations(), Output, Optimizer::Pass(), ReportWarning(), SubgraphViewSelector::SelectSubgraphs(), Graph::SubstituteSubgraph(), Undefined, and OptimizationViews::Validate().
Referenced by Optimize().
template void armnn::ArgMinMax | ( | Decoder< float > & | in, |
int32_t * | out, | ||
const TensorInfo & | inputTensorInfo, | ||
const TensorInfo & | outputTensorInfo, | ||
ArgMinMaxFunction | function, | ||
int | axis | ||
) |
template void armnn::ArgMinMax | ( | Decoder< float > & | in, |
int64_t * | out, | ||
const TensorInfo & | inputTensorInfo, | ||
const TensorInfo & | outputTensorInfo, | ||
ArgMinMaxFunction | function, | ||
int | axis | ||
) |
void ArgMinMax | ( | Decoder< float > & | in, |
OUT * | out, | ||
const TensorInfo & | inputTensorInfo, | ||
const TensorInfo & | outputTensorInfo, | ||
ArgMinMaxFunction | function, | ||
int | axis | ||
) |
Definition at line 17 of file ArgMinMax.cpp.
References Decoder< IType >::Get(), TensorInfo::GetNumDimensions(), armnnUtils::GetNumElementsBetween(), TensorInfo::GetShape(), armnnUtils::GetUnsignedAxis(), IgnoreUnused(), Max, and Min.
void armnn::AssertNumberOfInputSlots | ( | Layer & | layer | ) |
Definition at line 28 of file Layer.cpp.
References Convolution2d, DepthwiseConvolution2d, FullyConnected, Layer::GetNumInputSlots(), and Layer::GetType().
Referenced by InputSlot::Insert().
OptimizationResult AssignBackends | ( | OptimizedNetworkImpl * | optNetObjPtr, |
BackendSettings & | backendSettings, | ||
Graph::Iterator & | firstLayer, | ||
Graph::Iterator & | lastLayer, | ||
Optional< std::vector< std::string > & > | errMessages | ||
) |
Definition at line 1179 of file Network.cpp.
References ARMNN_SCOPED_PROFILING_EVENT, AssignBackendsIConnectable(), Float16, BackendSettings::GetAvailablePreferredBackends(), GetLayerInOutDatatype(), Input, OptimizationResult::m_Error, ReportError(), and Undefined.
Referenced by ApplyBackendOptimizations(), and AssignBackends().
OptimizationResult armnn::AssignBackends | ( | OptimizedNetworkImpl * | optNetObjPtr, |
BackendSettings & | backendSettings, | ||
SubgraphView & | subgraph, | ||
Optional< std::vector< std::string > & > | errMessages | ||
) |
Definition at line 1288 of file Network.cpp.
References AssignBackends(), SubgraphView::begin(), and SubgraphView::end().
Referenced by Optimize().
OptimizationResult AssignBackends | ( | OptimizedNetworkImpl * | optNetObjPtr, |
BackendSettings & | backendSettings, | ||
SubgraphView::IConnectableLayerIterator & | firstLayer, | ||
SubgraphView::IConnectableLayerIterator & | lastLayer, | ||
Optional< std::vector< std::string > & > | errMessages | ||
) |
Definition at line 1244 of file Network.cpp.
References ARMNN_SCOPED_PROFILING_EVENT, AssignBackendsIConnectable(), BackendSettings::GetAvailablePreferredBackends(), Input, OptimizationResult::m_Error, ReportError(), and Undefined.
void armnn::AssignBackendsIConnectable | ( | OptimizedNetworkImpl * | optNetObjPtr, |
IConnectableLayer * | it, | ||
Optional< std::vector< std::string > & > | errMessages, | ||
OptimizationResult & | result, | ||
BackendSettings & | backendSettings, | ||
std::vector< BackendId > & | availablePreferredBackends | ||
) |
Definition at line 1073 of file Network.cpp.
References AttemptBackendAssignment(), CheckScaleSetOnQuantizedType(), Constant, CpuRef, OptimizedNetworkImpl::GetGraph(), GetLayerInOutDatatype(), Input, BackendSettings::IsBackendSupported(), BackendSettings::IsCpuRefUsed(), OptimizationResult::IsError(), OptimizationResult::IsOk(), OptimizationResult::m_Error, BackendSettings::m_SelectedBackends, MemCopy, Permute, and ReturnWithError().
Referenced by AssignBackends().
void armnn::AssignSplitId | ( | LayerSelectionInfo::LayerInfoContainer & | layerInfos, |
LayerSelectionInfo & | layerInfo | ||
) |
Definition at line 322 of file SubgraphViewSelector.cpp.
References ForEachLayerInput().
Referenced by SubgraphViewSelector::SelectSubgraphs().
OptimizationResult armnn::AttemptBackendAssignment | ( | BackendSettings & | backendSettings, |
Graph & | graph, | ||
Layer * | layer, | ||
BackendId | backend, | ||
DataType | dataTypeIn, | ||
DataType | dataTypeOut, | ||
const std::vector< BackendId > & | availablePreferredBackends, | ||
std::string & | reasonIfUnsupported, | ||
Optional< std::vector< std::string > & > | errMessages | ||
) |
Definition at line 844 of file Network.cpp.
References Constant, FloatingPointConverter::ConvertFloat16To32(), ConvertFp16ToFp32, ConvertFp32ToFp16, Float16, Float32, BackendId::Get(), Layer::GetBackendId(), GetDataTypeName(), Layer::GetInputSlots(), GetLayerTypeAsCString(), Layer::GetOutputSlot(), Layer::GetType(), info, InsertConvertFp16ToFp32LayersBefore(), InsertConvertFp32ToFp16LayersAfter(), IWorkloadFactory::IsLayerSupported(), ConstantLayer::m_LayerOutput, ReportWarning(), ReturnWithError(), Layer::SetBackendId(), TensorInfo::SetDataType(), and OutputSlot::SetTensorInfo().
Referenced by AssignBackendsIConnectable().
BackendRegistry & BackendRegistryInstance | ( | ) |
Definition at line 15 of file BackendRegistry.cpp.
Referenced by CreateBackendObject(), CreateSupportedBackends(), DynamicBackendUtils::DeregisterDynamicBackends(), GetCapability(), GetILayerSupportByBackendId(), GetNumberOfCacheFiles(), HasMatchingCapability(), ArmNNProfilingServiceInitialiser::InitialiseProfilingService(), DynamicBackendUtils::RegisterDynamicBackends(), RuntimeImpl::RuntimeImpl(), and RuntimeImpl::~RuntimeImpl().
void BatchNormImpl | ( | const BatchNormalizationQueueDescriptor & | data, |
Decoder< float > & | meanDecoder, | ||
Decoder< float > & | varianceDecoder, | ||
Decoder< float > & | betaDecoder, | ||
Decoder< float > & | gammaDecoder, | ||
Decoder< float > & | inputDecoder, | ||
Encoder< float > & | outputEncoder | ||
) |
Definition at line 18 of file BatchNormImpl.cpp.
References Decoder< IType >::Get(), DataLayoutIndexed::GetChannelsIndex(), DataLayoutIndexed::GetHeightIndex(), DataLayoutIndexed::GetIndex(), TensorInfo::GetShape(), GetTensorInfo(), DataLayoutIndexed::GetWidthIndex(), BatchNormalizationDescriptor::m_DataLayout, BatchNormalizationDescriptor::m_Eps, QueueDescriptor::m_Inputs, QueueDescriptorWithParameters< LayerDescriptor >::m_Parameters, and Encoder< IType >::Set().
void BatchToSpaceNd | ( | const TensorInfo & | inputInfo, |
const TensorInfo & | outputInfo, | ||
const BatchToSpaceNdDescriptor & | params, | ||
Decoder< float > & | inputData, | ||
Encoder< float > & | outputData | ||
) |
Definition at line 50 of file BatchToSpaceNd.cpp.
References BatchToSpaceNd(), CHECK_LOCATION, Decoder< IType >::Get(), DataLayoutIndexed::GetChannelsIndex(), DataLayoutIndexed::GetHeightIndex(), TensorInfo::GetNumDimensions(), TensorInfo::GetShape(), DataLayoutIndexed::GetWidthIndex(), BatchToSpaceNdDescriptor::m_BlockShape, BatchToSpaceNdDescriptor::m_Crops, BatchToSpaceNdDescriptor::m_DataLayout, Offset(), and Encoder< IType >::Set().
Referenced by BatchToSpaceNd(), and BatchToSpaceNdLayer::BatchToSpaceNdLayer().
void armnn::BuildAddMulAddSlotLists | ( | bool | handleReLu, |
bool | multipleOutputs, | ||
std::vector< SlotListType > & | inputLayersSlotLists, | ||
std::vector< SlotListType > & | outputLayersSlotLists | ||
) |
Definition at line 36 of file NeonBackendOptimizationUtils.hpp.
bool armnn::BuildAddMulAddTensorInfoLists | ( | Type * | layerList[4], |
unsigned int & | numInputs, | ||
unsigned int & | numOutputs, | ||
std::vector< TensorInfo > & | inputInfos, | ||
std::vector< TensorInfo > & | outputInfos, | ||
const ActivationDescriptor *& | activationDescriptor, | ||
bool & | fuseReLu | ||
) |
Definition at line 87 of file NeonBackendOptimizationUtils.hpp.
References Add, ARMNN_THROW_INVALIDARG_IF_FALSE, BoundedReLu, CollapseLeadingUnitDimensions(), Constant, InputSlot::GetConnectedOutputSlot(), OutputSlot::GetConnection(), Layer::GetInputSlot(), OutputSlot::GetNumConnections(), TensorInfo::GetNumDimensions(), Layer::GetOutputSlot(), InputSlot::GetOwningLayer(), OutputSlot::GetOwningLayer(), InputSlot::GetTensorInfo(), OutputSlot::GetTensorInfo(), Layer::GetType(), IsSequenceLayerType(), ActivationDescriptor::m_Function, ConstantLayer::m_LayerOutput, Mul, ReLu, and OutputSlot::SetTensorInfo().
int armnn::CalcLevel | ( | const Event * | eventPtr | ) |
Definition at line 257 of file Profiling.cpp.
References Event::GetParentEvent().
Referenced by ProfilerImpl::AnalyzeEventsAndWriteResults(), and ProfilerImpl::PopulateParent().
EdgeStrategy armnn::CalculateEdgeStrategy | ( | BackendsMap & | backends, |
ITensorHandleFactory::FactoryId | srcFactoryId, | ||
const Layer & | layer, | ||
const Layer & | connectedLayer, | ||
TensorHandleFactoryRegistry & | registry, | ||
bool | importEnabled | ||
) |
Definition at line 1729 of file Network.cpp.
References CopyToTarget, DirectCompatibility, ExportToTarget, FallbackImportDisabled, Layer::GetBackendId(), ITensorHandleFactory::GetCapabilities(), ITensorHandleFactory::GetExportFlags(), TensorHandleFactoryRegistry::GetFactory(), ITensorHandleFactory::GetImportFlags(), Layer::GetType(), ITensorHandleFactory::LegacyFactoryId, Output, PaddingRequired, ITensorHandleFactory::SupportsMapUnmap(), and Undefined.
std::map< std::string, unsigned int > CalculateGatherNdKeyIndices | ( | TensorInfo | inputInfo0, |
TensorInfo | inputInfo1 | ||
) |
Calculates the key index values needed for GatherNd: N, ND, K, W, C (N is always 1)
inputInfo0 | - TensorInfo of the corresponding input tensor: params |
inputInfo1 | - TensorInfo of the corresponding input tensor: indices |
Definition at line 313 of file WorkloadUtils.cpp.
References TensorInfo::GetNumDimensions(), and TensorInfo::GetShape().
Referenced by ClGatherNdWorkload::ClGatherNdWorkload(), ClGatherNdWorkloadValidate(), NeonGatherNdWorkload::NeonGatherNdWorkload(), and NeonGatherNdWorkloadValidate().
ITensorHandleFactory::FactoryId armnn::CalculateSlotOption | ( | BackendsMap & | backends, |
OutputSlot & | outputSlot, | ||
TensorHandleFactoryRegistry & | registry, | ||
bool | exportEnabled | ||
) |
Definition at line 1576 of file Network.cpp.
References FallbackImportDisabled, Layer::GetBackendId(), ITensorHandleFactory::GetCapabilities(), OutputSlot::GetConnections(), ITensorHandleFactory::GetExportFlags(), TensorHandleFactoryRegistry::GetFactory(), IBackendInternal::GetHandleFactoryPreferences(), Layer::GetInputSlots(), OutputSlot::GetOwningLayer(), Layer::GetType(), ITensorHandleFactory::LegacyFactoryId, Output, RequiresCopy(), and ITensorHandleFactory::SupportsMapUnmap().
ITensorHandleFactory::FactoryId armnn::CalculateSlotOptionForInput | ( | BackendsMap & | backends, |
OutputSlot & | slot, | ||
TensorHandleFactoryRegistry & | registry, | ||
bool | importEnabled | ||
) |
Definition at line 1474 of file Network.cpp.
References Layer::GetBackendId(), OutputSlot::GetConnections(), TensorHandleFactoryRegistry::GetFactory(), ITensorHandleFactory::GetImportFlags(), OutputSlot::GetOwningLayer(), Layer::GetType(), Input, ITensorHandleFactory::LegacyFactoryId, and ITensorHandleFactory::SupportsMapUnmap().
ITensorHandleFactory::FactoryId armnn::CalculateSlotOptionForOutput | ( | BackendsMap & | backends, |
OutputSlot & | slot, | ||
TensorHandleFactoryRegistry & | registry | ||
) |
Definition at line 1566 of file Network.cpp.
References ITensorHandleFactory::DeferredFactoryId, and IgnoreUnused().
std::vector<IConnectableLayer*> armnn::ChainReduceLayers | ( | OptimizationViews & | optimizationViews, |
LayerType * | baseLayer, | ||
ReduceDescriptor & | desc | ||
) |
Definition at line 278 of file ArmComputeSubgraphUtils.hpp.
References INetwork::AddReduceLayer(), ComputeReductionTensorShape(), OptimizationViews::GetINetwork(), Layer::GetInputSlot(), Layer::GetOutputSlot(), ReduceDescriptor::m_KeepDims, ReduceDescriptor::m_vAxis, and OutputSlot::SetTensorInfo().
|
inline |
Definition at line 41 of file MemorySources.hpp.
Referenced by LoadedNetwork::ImportInputs(), and LoadedNetwork::ImportOutputs().
bool armnn::CheckFp16Support | ( | BackendsMap & | backends, |
const std::vector< BackendId > & | availablePreferredBackends | ||
) |
Definition at line 1026 of file Network.cpp.
References ARMNN_LOG, debug, HasMatchingCapability(), and warning.
Referenced by Optimize().
bool armnn::CheckScaleSetOnQuantizedType | ( | Layer * | layer, |
Optional< std::vector< std::string > & > | errMessages | ||
) |
Definition at line 801 of file Network.cpp.
References ARMNN_LOG, Layer::GetNumOutputSlots(), Layer::GetOutputSlot(), OutputSlot::GetTensorInfo(), Layer::GetType(), info, QAsymmS8, QAsymmU8, QSymmS16, QSymmS8, ReportError(), OutputSlot::SetTensorInfo(), Softmax, and warning.
Referenced by AssignBackendsIConnectable().
bool armnn::CheckSupportRule | ( | F | rule, |
Optional< std::string & > | reasonIfUnsupported, | ||
const char * | reason | ||
) |
Definition at line 37 of file LayerSupportRules.hpp.
References OptionalReferenceSwitch< std::is_reference< T >::value, T >::value().
Referenced by RefLayerSupport::IsActivationSupported(), RefLayerSupport::IsAdditionSupported(), RefLayerSupport::IsArgMinMaxSupported(), RefLayerSupport::IsBatchMatMulSupported(), RefLayerSupport::IsBatchNormalizationSupported(), RefLayerSupport::IsBatchToSpaceNdSupported(), RefLayerSupport::IsBroadcastToSupported(), RefLayerSupport::IsCastSupported(), RefLayerSupport::IsChannelShuffleSupported(), RefLayerSupport::IsComparisonSupported(), RefLayerSupport::IsConcatSupported(), RefLayerSupport::IsConstantSupported(), RefLayerSupport::IsConvolution2dSupported(), RefLayerSupport::IsConvolution3dSupported(), RefLayerSupport::IsDebugSupported(), RefLayerSupport::IsDepthToSpaceSupported(), RefLayerSupport::IsDepthwiseConvolutionSupported(), RefLayerSupport::IsDequantizeSupported(), RefLayerSupport::IsDetectionPostProcessSupported(), RefLayerSupport::IsDivisionSupported(), RefLayerSupport::IsElementwiseUnarySupported(), RefLayerSupport::IsFakeQuantizationSupported(), RefLayerSupport::IsFillSupported(), RefLayerSupport::IsFloorSupported(), RefLayerSupport::IsFullyConnectedSupported(), RefLayerSupport::IsGatherNdSupported(), RefLayerSupport::IsGatherSupported(), RefLayerSupport::IsInstanceNormalizationSupported(), RefLayerSupport::IsL2NormalizationSupported(), RefLayerSupport::IsLayerSupported(), RefLayerSupport::IsLogicalBinarySupported(), RefLayerSupport::IsLogSoftmaxSupported(), RefLayerSupport::IsLstmSupported(), RefLayerSupport::IsMaximumSupported(), RefLayerSupport::IsMeanSupported(), RefLayerSupport::IsMemCopySupported(), RefLayerSupport::IsMinimumSupported(), RefLayerSupport::IsMultiplicationSupported(), RefLayerSupport::IsNormalizationSupported(), RefLayerSupport::IsPadSupported(), RefLayerSupport::IsPermuteSupported(), RefLayerSupport::IsPooling2dSupported(), RefLayerSupport::IsPooling3dSupported(), RefLayerSupport::IsPreluSupported(), RefLayerSupport::IsQuantizeSupported(), RefLayerSupport::IsRankSupported(), RefLayerSupport::IsReduceSupported(), RefLayerSupport::IsReshapeSupported(), RefLayerSupport::IsResizeSupported(), RefLayerSupport::IsReverseV2Supported(), RefLayerSupport::IsScatterNdSupported(), RefLayerSupport::IsShapeSupported(), RefLayerSupport::IsSliceSupported(), RefLayerSupport::IsSoftmaxSupported(), RefLayerSupport::IsSpaceToBatchNdSupported(), RefLayerSupport::IsSpaceToDepthSupported(), RefLayerSupport::IsSplitterSupported(), RefLayerSupport::IsStackSupported(), RefLayerSupport::IsStridedSliceSupported(), RefLayerSupport::IsSubtractionSupported(), RefLayerSupport::IsTileSupported(), RefLayerSupport::IsTransposeConvolution2dSupported(), RefLayerSupport::IsTransposeSupported(), and RefLayerSupport::IsUnidirectionalSequenceLstmSupported().
arm_compute::Status ClAbsWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 19 of file ClAbsWorkload.cpp.
Referenced by ClLayerSupport::IsElementwiseUnarySupported().
arm_compute::Status ClActivationWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const ActivationDescriptor & | descriptor | ||
) |
Definition at line 17 of file ClActivationWorkload.cpp.
Referenced by ClLayerSupport::IsActivationSupported().
arm_compute::Status ClAdditionValidate | ( | const TensorInfo & | input0, |
const TensorInfo & | input1, | ||
const TensorInfo & | output, | ||
const ActivationDescriptor * | activationDescriptor | ||
) |
Definition at line 45 of file ClAdditionWorkload.cpp.
Referenced by ClLayerSupport::IsAdditionSupported(), ClLayerSupport::IsLayerSupported(), and ClBackend::OptimizeSubgraphView().
arm_compute::Status ClArgMinMaxWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const ArgMinMaxDescriptor & | descriptor | ||
) |
Definition at line 31 of file ClArgMinMaxWorkload.cpp.
Referenced by ClLayerSupport::IsArgMinMaxSupported().
|
constexpr |
Definition at line 10 of file ClBackendId.hpp.
Referenced by ClBackend::GetIdStatic().
arm_compute::Status ClBatchMatMulValidate | ( | const TensorInfo & | inputInfoX, |
const TensorInfo & | inputInfoY, | ||
const TensorInfo & | outputInfo, | ||
const BatchMatMulDescriptor & | descriptor, | ||
const ActivationDescriptor * | activationDescriptor | ||
) |
Definition at line 24 of file ClBatchMatMulWorkload.cpp.
References BatchMatMulDescriptor::m_AdjointX, BatchMatMulDescriptor::m_AdjointY, BatchMatMulDescriptor::m_DataLayoutX, BatchMatMulDescriptor::m_DataLayoutY, and NCHW.
Referenced by ClLayerSupport::IsBatchMatMulSupported().
arm_compute::Status ClBatchNormalizationValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const TensorInfo & | mean, | ||
const TensorInfo & | var, | ||
const TensorInfo & | beta, | ||
const TensorInfo & | gamma, | ||
const BatchNormalizationDescriptor & | descriptor, | ||
const ActivationDescriptor * | activationDescriptor | ||
) |
Definition at line 19 of file ClBatchNormalizationFloatWorkload.cpp.
Referenced by ClLayerSupport::IsBatchNormalizationSupported(), and ClBackend::OptimizeSubgraphView().
arm_compute::Status ClBatchToSpaceNdWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const BatchToSpaceNdDescriptor & | descriptor | ||
) |
Definition at line 17 of file ClBatchToSpaceNdWorkload.cpp.
Referenced by ClLayerSupport::IsBatchToSpaceNdSupported().
arm_compute::Status ClCastValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 20 of file ClCastWorkload.cpp.
Referenced by ClLayerSupport::IsCastSupported().
arm_compute::Status ClChannelShuffleValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const ChannelShuffleDescriptor & | descriptor | ||
) |
Definition at line 20 of file ClChannelShuffleWorkload.cpp.
Referenced by ClLayerSupport::IsChannelShuffleSupported().
arm_compute::Status ClComparisonWorkloadValidate | ( | const TensorInfo & | input0, |
const TensorInfo & | input1, | ||
const TensorInfo & | output, | ||
const ComparisonDescriptor & | descriptor | ||
) |
Definition at line 24 of file ClComparisonWorkload.cpp.
Referenced by ClLayerSupport::IsComparisonSupported().
arm_compute::Status ClConcatWorkloadValidate | ( | const std::vector< const TensorInfo * > & | inputs, |
const TensorInfo & | output, | ||
const OriginsDescriptor & | descriptor | ||
) |
Definition at line 27 of file ClConcatWorkload.cpp.
Referenced by ClLayerSupport::IsConcatSupported().
arm_compute::Status ClConstantWorkloadValidate | ( | const TensorInfo & | output | ) |
Definition at line 18 of file ClConstantWorkload.cpp.
Referenced by ClLayerSupport::IsConstantSupported().
|
inline |
Definition at line 152 of file ClContextSchema_generated.h.
References ClContextIdentifier().
|
inline |
Definition at line 167 of file ClContextSchema_generated.h.
|
inline |
Definition at line 148 of file ClContextSchema_generated.h.
Referenced by ClContextBufferHasIdentifier(), FinishClContextBuffer(), FinishSizePrefixedClContextBuffer(), VerifyClContextBuffer(), and VerifySizePrefixedClContextBuffer().
arm_compute::Status ClConvertFp16ToFp32WorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 44 of file ClConvertFp16ToFp32Workload.cpp.
References Float16, Float32, and TensorInfo::GetDataType().
Referenced by ClLayerSupport::IsConvertFp16ToFp32Supported().
arm_compute::Status ClConvertFp32ToFp16WorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 44 of file ClConvertFp32ToFp16Workload.cpp.
References Float16, Float32, and TensorInfo::GetDataType().
Referenced by ClLayerSupport::IsConvertFp32ToFp16Supported().
arm_compute::Status ClConvolution2dWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const Convolution2dDescriptor & | descriptor, | ||
const TensorInfo & | weights, | ||
const Optional< TensorInfo > & | biases, | ||
bool | isFastMathEnabled, | ||
const ActivationDescriptor * | activationDescriptor | ||
) |
Definition at line 23 of file ClConvolution2dWorkload.cpp.
Referenced by ClLayerSupport::IsConvolution2dSupported(), and ClBackend::OptimizeSubgraphView().
arm_compute::Status ClConvolution3dWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const Convolution3dDescriptor & | descriptor, | ||
const TensorInfo & | weights, | ||
const Optional< TensorInfo > & | biases, | ||
bool | isFastMathEnabled, | ||
const ActivationDescriptor * | activationDescriptor | ||
) |
Definition at line 23 of file ClConvolution3dWorkload.cpp.
Referenced by ClLayerSupport::IsConvolution3dSupported().
arm_compute::Status ClDepthToSpaceWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const DepthToSpaceDescriptor & | descriptor | ||
) |
Definition at line 22 of file ClDepthToSpaceWorkload.cpp.
References SpaceToDepthDescriptor::m_DataLayout.
Referenced by ClLayerSupport::IsDepthToSpaceSupported().
arm_compute::Status ClDepthwiseConvolutionWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const DepthwiseConvolution2dDescriptor & | descriptor, | ||
const TensorInfo & | weights, | ||
const Optional< TensorInfo > & | biases, | ||
const ActivationDescriptor * | activationDescriptor | ||
) |
Definition at line 26 of file ClDepthwiseConvolutionWorkload.cpp.
Referenced by ClLayerSupport::IsDepthwiseConvolutionSupported(), ClLayerSupport::IsDilatedDepthwiseConvolutionSupported(), and ClBackend::OptimizeSubgraphView().
arm_compute::Status ClDequantizeWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 22 of file ClDequantizeWorkload.cpp.
Referenced by ClLayerSupport::IsDequantizeSupported().
arm_compute::Status ClDivisionWorkloadValidate | ( | const TensorInfo & | input0, |
const TensorInfo & | input1, | ||
const TensorInfo & | output, | ||
const ActivationDescriptor * | activationDescriptor | ||
) |
Definition at line 18 of file ClDivisionWorkload.cpp.
Referenced by ClLayerSupport::IsDivisionSupported(), ClLayerSupport::IsLayerSupported(), and ClBackend::OptimizeSubgraphView().
arm_compute::Status ClElementwiseBinaryValidate | ( | const TensorInfo & | input0, |
const TensorInfo & | input1, | ||
const TensorInfo & | output, | ||
const ElementwiseBinaryDescriptor & | descriptor, | ||
const ActivationDescriptor * | activationDescriptor | ||
) |
Definition at line 64 of file ClElementwiseBinaryWorkload.cpp.
Referenced by ClLayerSupport::IsLayerSupported().
arm_compute::Status ClExpWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 18 of file ClExpWorkload.cpp.
Referenced by ClLayerSupport::IsElementwiseUnarySupported().
arm_compute::Status ClFloorWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 14 of file ClFloorFloatWorkload.cpp.
Referenced by ClLayerSupport::IsFloorSupported().
arm_compute::Status ClFullyConnectedWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const TensorInfo & | weights, | ||
const Optional< TensorInfo > & | biases, | ||
const FullyConnectedDescriptor & | descriptor, | ||
const ActivationDescriptor * | activationDescriptor | ||
) |
Definition at line 19 of file ClFullyConnectedWorkload.cpp.
Referenced by ClLayerSupport::IsFullyConnectedSupported(), and ClBackend::OptimizeSubgraphView().
arm_compute::Status ClGatherNdWorkloadValidate | ( | const TensorInfo & | paramsInfo, |
const TensorInfo & | indicesInfo, | ||
const TensorInfo & | outputInfo | ||
) |
Validate Mul
Validate ReduceSum
Validate Gather
Validate Reshape
Return OK if all the layers are valid
Definition at line 16 of file ClGatherNdWorkload.cpp.
References CalculateGatherNdKeyIndices(), and TensorInfo::SetShape().
Referenced by ClLayerSupport::IsGatherNdSupported().
arm_compute::Status ClGatherWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | indices, | ||
const TensorInfo & | output, | ||
const GatherDescriptor & | descriptor | ||
) |
Definition at line 15 of file ClGatherWorkload.cpp.
Referenced by ClLayerSupport::IsGatherSupported().
|
constexpr |
Definition at line 14 of file ClImportTensorHandleFactory.hpp.
Referenced by ClImportTensorHandleFactory::GetIdStatic().
arm_compute::Status ClInstanceNormalizationWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const InstanceNormalizationDescriptor & | descriptor | ||
) |
Definition at line 18 of file ClInstanceNormalizationWorkload.cpp.
Referenced by ClLayerSupport::IsInstanceNormalizationSupported().
arm_compute::Status ClL2NormalizationWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const L2NormalizationDescriptor & | descriptor | ||
) |
Definition at line 17 of file ClL2NormalizationFloatWorkload.cpp.
Referenced by ClLayerSupport::IsL2NormalizationSupported().
arm_compute::Status ClLogicalAndWorkloadValidate | ( | const TensorInfo & | input0, |
const TensorInfo & | input1, | ||
const TensorInfo & | output | ||
) |
Definition at line 20 of file ClLogicalAndWorkload.cpp.
Referenced by ClLayerSupport::IsLogicalBinarySupported().
arm_compute::Status ClLogicalNotWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 20 of file ClLogicalNotWorkload.cpp.
Referenced by ClLayerSupport::IsElementwiseUnarySupported().
arm_compute::Status ClLogicalOrWorkloadValidate | ( | const TensorInfo & | input0, |
const TensorInfo & | input1, | ||
const TensorInfo & | output | ||
) |
Definition at line 20 of file ClLogicalOrWorkload.cpp.
Referenced by ClLayerSupport::IsLogicalBinarySupported().
arm_compute::Status ClLogSoftmaxWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const LogSoftmaxDescriptor & | descriptor | ||
) |
Definition at line 17 of file ClLogSoftmaxWorkload.cpp.
Referenced by ClLayerSupport::IsLogSoftmaxSupported().
arm_compute::Status ClLogWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 18 of file ClLogWorkload.cpp.
Referenced by ClLayerSupport::IsElementwiseUnarySupported().
arm_compute::Status ClLstmFloatWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | outputStateIn, | ||
const TensorInfo & | cellStateIn, | ||
const TensorInfo & | scratchBuffer, | ||
const TensorInfo & | outputStateOut, | ||
const TensorInfo & | cellStateOut, | ||
const TensorInfo & | output, | ||
const LstmDescriptor & | descriptor, | ||
const LstmInputParamsInfo & | paramsInfo | ||
) |
Definition at line 244 of file ClLstmFloatWorkload.cpp.
Referenced by ClLayerSupport::IsLstmSupported().
arm_compute::Status ClMaximumWorkloadValidate | ( | const TensorInfo & | input0, |
const TensorInfo & | input1, | ||
const TensorInfo & | output | ||
) |
Definition at line 24 of file ClMaximumWorkload.cpp.
Referenced by ClLayerSupport::IsLayerSupported(), and ClLayerSupport::IsMaximumSupported().
arm_compute::Status ClMeanValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const MeanDescriptor & | descriptor | ||
) |
Definition at line 17 of file ClMeanWorkload.cpp.
Referenced by ClLayerSupport::IsMeanSupported().
arm_compute::Status ClMinimumWorkloadValidate | ( | const TensorInfo & | input0, |
const TensorInfo & | input1, | ||
const TensorInfo & | output | ||
) |
Definition at line 24 of file ClMinimumWorkload.cpp.
Referenced by ClLayerSupport::IsLayerSupported(), and ClLayerSupport::IsMinimumSupported().
arm_compute::Status ClMultiplicationWorkloadValidate | ( | const TensorInfo & | input0, |
const TensorInfo & | input1, | ||
const TensorInfo & | output, | ||
const ActivationDescriptor * | activationDescriptor | ||
) |
Definition at line 18 of file ClMultiplicationWorkload.cpp.
Referenced by ClLayerSupport::IsLayerSupported(), ClLayerSupport::IsMultiplicationSupported(), and ClBackend::OptimizeSubgraphView().
arm_compute::Status ClNegWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 18 of file ClNegWorkload.cpp.
Referenced by ClLayerSupport::IsElementwiseUnarySupported().
arm_compute::Status ClNormalizationWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const NormalizationDescriptor & | descriptor | ||
) |
Definition at line 19 of file ClNormalizationFloatWorkload.cpp.
Referenced by ClLayerSupport::IsNormalizationSupported().
arm_compute::Status ClPadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const PadDescriptor & | descriptor | ||
) |
Definition at line 62 of file ClPadWorkload.cpp.
Referenced by ClLayerSupport::IsPadSupported().
arm_compute::Status ClPermuteWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const PermuteDescriptor & | descriptor | ||
) |
Definition at line 17 of file ClPermuteWorkload.cpp.
Referenced by ClLayerSupport::IsPermuteSupported().
arm_compute::Status ClPooling2dWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const Pooling2dDescriptor & | descriptor | ||
) |
Definition at line 18 of file ClPooling2dWorkload.cpp.
Referenced by ClLayerSupport::IsPooling2dSupported().
arm_compute::Status ClPooling3dWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const Pooling3dDescriptor & | descriptor | ||
) |
Definition at line 18 of file ClPooling3dWorkload.cpp.
Referenced by ClLayerSupport::IsPooling3dSupported().
arm_compute::Status ClPreluWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | alpha, | ||
const TensorInfo & | output | ||
) |
Definition at line 16 of file ClPreluWorkload.cpp.
Referenced by ClLayerSupport::IsPreluSupported().
arm_compute::Status ClQLstmWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | cellStateIn, | ||
const TensorInfo & | outputStateIn, | ||
const TensorInfo & | cellStateOut, | ||
const TensorInfo & | outputStateOut, | ||
const TensorInfo & | output, | ||
const QLstmDescriptor & | descriptor, | ||
const LstmInputParamsInfo & | paramsInfo | ||
) |
Definition at line 247 of file ClQLstmWorkload.cpp.
Referenced by ClLayerSupport::IsQLstmSupported().
arm_compute::Status ClQuantizedLstmWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | previousCellStateIn, | ||
const TensorInfo & | previousOutputIn, | ||
const TensorInfo & | cellStateOut, | ||
const TensorInfo & | output, | ||
const QuantizedLstmInputParamsInfo & | paramsInfo | ||
) |
Definition at line 18 of file ClQuantizedLstmWorkload.cpp.
Referenced by ClLayerSupport::IsQuantizedLstmSupported().
arm_compute::Status ClQuantizeWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 22 of file ClQuantizeWorkload.cpp.
Referenced by ClLayerSupport::IsQuantizeSupported().
arm_compute::Status ClReduceWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const ReduceDescriptor & | descriptor | ||
) |
Definition at line 18 of file ClReduceWorkload.cpp.
References ReduceDescriptor::m_vAxis.
Referenced by ClLayerSupport::IsReduceSupported().
arm_compute::Status ClReshapeWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 15 of file ClReshapeWorkload.cpp.
Referenced by ClLayerSupport::IsReshapeSupported().
arm_compute::Status ClResizeWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const ResizeDescriptor & | descriptor | ||
) |
Definition at line 22 of file ClResizeWorkload.cpp.
Referenced by ClLayerSupport::IsResizeSupported().
arm_compute::Status ClReverseV2WorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | axis, | ||
const TensorInfo & | output | ||
) |
Definition at line 16 of file ClReverseV2Workload.cpp.
Referenced by ClLayerSupport::IsReverseV2Supported().
arm_compute::Status ClRsqrtWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 18 of file ClRsqrtWorkload.cpp.
Referenced by ClLayerSupport::IsElementwiseUnarySupported().
arm_compute::Status ClScatterNdWorkloadValidate | ( | const TensorInfo & | inputInfo, |
const TensorInfo & | indicesInfo, | ||
const TensorInfo & | updatesInfo, | ||
const TensorInfo & | outputInfo, | ||
const ScatterNdDescriptor & | descriptor | ||
) |
Definition at line 20 of file ClScatterNdWorkload.cpp.
Referenced by ClLayerSupport::IsScatterNdSupported().
arm_compute::Status ClSinWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 18 of file ClSinWorkload.cpp.
Referenced by ClLayerSupport::IsElementwiseUnarySupported().
arm_compute::Status ClSliceWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const SliceDescriptor & | descriptor | ||
) |
Definition at line 18 of file ClSliceWorkload.cpp.
Referenced by ClLayerSupport::IsSliceSupported().
arm_compute::Status ClSoftmaxWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const SoftmaxDescriptor & | descriptor | ||
) |
Definition at line 17 of file ClSoftmaxWorkload.cpp.
Referenced by ClLayerSupport::IsSoftmaxSupported().
arm_compute::Status ClSpaceToBatchNdWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const SpaceToBatchNdDescriptor & | descriptor | ||
) |
Definition at line 16 of file ClSpaceToBatchNdWorkload.cpp.
Referenced by ClLayerSupport::IsSpaceToBatchNdSupported().
arm_compute::Status ClSpaceToDepthWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const SpaceToDepthDescriptor & | descriptor | ||
) |
Definition at line 54 of file ClSpaceToDepthWorkload.cpp.
References SpaceToDepthDescriptor::m_DataLayout.
Referenced by ClLayerSupport::IsSpaceToDepthSupported().
arm_compute::Status ClSplitterWorkloadValidate | ( | const TensorInfo & | input, |
const std::vector< std::reference_wrapper< TensorInfo >> & | outputs, | ||
unsigned int | splitAxis | ||
) |
Definition at line 32 of file ClSplitterWorkload.cpp.
Referenced by ClLayerSupport::IsSplitterSupported().
arm_compute::Status ClSqrtWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 19 of file ClSqrtWorkload.cpp.
Referenced by ClLayerSupport::IsElementwiseUnarySupported().
arm_compute::Status ClStackWorkloadValidate | ( | const std::vector< const TensorInfo * > & | inputs, |
const TensorInfo & | output, | ||
const StackDescriptor & | descriptor | ||
) |
Definition at line 29 of file ClStackWorkload.cpp.
Referenced by ClLayerSupport::IsStackSupported().
arm_compute::Status ClStridedSliceWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const StridedSliceDescriptor & | descriptor | ||
) |
Definition at line 27 of file ClStridedSliceWorkload.cpp.
Referenced by ClLayerSupport::IsStridedSliceSupported().
arm_compute::Status ClSubtractionValidate | ( | const TensorInfo & | input0, |
const TensorInfo & | input1, | ||
const TensorInfo & | output, | ||
const ActivationDescriptor * | activationDescriptor | ||
) |
Definition at line 46 of file ClSubtractionWorkload.cpp.
Referenced by ClLayerSupport::IsLayerSupported(), ClLayerSupport::IsSubtractionSupported(), and ClBackend::OptimizeSubgraphView().
|
constexpr |
Definition at line 15 of file ClTensorHandleFactory.hpp.
Referenced by ClTensorHandleFactory::GetIdStatic().
arm_compute::Status ClTileWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const TileDescriptor & | descriptor | ||
) |
arm_compute::Status ClTransposeConvolution2dWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const TransposeConvolution2dDescriptor & | descriptor, | ||
const TensorInfo & | weights, | ||
const Optional< TensorInfo > & | biases | ||
) |
Definition at line 26 of file ClTransposeConvolution2dWorkload.cpp.
Referenced by ClLayerSupport::IsTransposeConvolution2dSupported().
arm_compute::Status ClTransposeWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const TransposeDescriptor & | descriptor | ||
) |
Definition at line 17 of file ClTransposeWorkload.cpp.
Referenced by ClLayerSupport::IsTransposeSupported().
arm_compute::Status ClUnidirectionalSequenceLstmFloatWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | outputStateIn, | ||
const TensorInfo & | cellStateIn, | ||
const TensorInfo & | outputStateOut, | ||
const TensorInfo & | cellStateOut, | ||
const TensorInfo & | output, | ||
const UnidirectionalSequenceLstmDescriptor & | descriptor, | ||
const LstmInputParamsInfo & | paramsInfo | ||
) |
Definition at line 508 of file ClUnidirectionalSequenceLstmFloatWorkload.cpp.
References TensorShape::GetNumDimensions(), TensorInfo::GetShape(), and LstmDescriptor::m_TimeMajor.
Referenced by ClLayerSupport::IsUnidirectionalSequenceLstmSupported().
|
inline |
Definition at line 14 of file NeonBackendOptimizationUtils.hpp.
References TensorInfo::GetNumDimensions(), TensorInfo::GetShape(), and TensorInfo::SetShape().
Referenced by BuildAddMulAddTensorInfoLists().
MemorySourceFlags armnn::Combine | ( | Arg | source, |
Args... | rest | ||
) |
Definition at line 36 of file MemorySources.hpp.
References Combine().
MemorySourceFlags armnn::Combine | ( | Arg | sourceA, |
Arg | sourceB | ||
) |
|
inline |
Function to convert ArmNN axis (left to right) to ACL axis (right to left) ranging from [-rank, rank)
Definition at line 246 of file ArmComputeUtils.hpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, and TensorInfo::GetNumDimensions().
Referenced by ClGatherWorkload::ClGatherWorkload(), ClLogSoftmaxWorkload::ClLogSoftmaxWorkload(), ClSoftmaxWorkload::ClSoftmaxWorkload(), NeonGatherWorkload::NeonGatherWorkload(), NeonLogSoftmaxWorkload::NeonLogSoftmaxWorkload(), and NeonSoftmaxWorkload::NeonSoftmaxWorkload().
|
inline |
Utility function used to setup an arm_compute::Conv3dInfo object from convolution3d descriptor.
Definition at line 261 of file ArmComputeUtils.hpp.
References ConvertActivationDescriptorToAclActivationLayerInfo(), Convolution3dDescriptor::m_DilationX, Convolution3dDescriptor::m_DilationY, Convolution3dDescriptor::m_DilationZ, Convolution3dDescriptor::m_PadBack, Convolution3dDescriptor::m_PadBottom, Convolution3dDescriptor::m_PadFront, Convolution3dDescriptor::m_PadLeft, Convolution3dDescriptor::m_PadRight, Convolution3dDescriptor::m_PadTop, Convolution3dDescriptor::m_StrideX, Convolution3dDescriptor::m_StrideY, and Convolution3dDescriptor::m_StrideZ.
|
inline |
Definition at line 278 of file ArmComputeUtils.hpp.
References ConvertAdditionalInfoToAclActivationLayerInfo(), QueueDescriptorWithParameters< LayerDescriptor >::m_Parameters, and Convolution3dDescriptor::m_StrideX.
|
inline |
Function to compute the output tensor shape based on the axes and if keepDims is set.
Definition at line 320 of file ArmComputeUtils.hpp.
References TensorInfo::GetNumDimensions().
Referenced by ChainReduceLayers().
|
inline |
Definition at line 227 of file ArmComputeUtils.hpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, TensorInfo::GetNumDimensions(), and SoftmaxDescriptor::m_Axis.
std::set< unsigned int > ComputeSplitAxis | ( | const armnn::SplitterDescriptor & | desc, |
const TensorShape & | input | ||
) |
Calculates the axis values for split operation.
desc | - Splitter Descriptor |
input | - Input tensor shape |
Definition at line 377 of file WorkloadUtils.cpp.
References ViewsDescriptor::GetAxis(), ViewsDescriptor::GetNumDimensions(), ViewsDescriptor::GetNumViews(), armnnUtils::GetUnsignedAxis(), ViewsDescriptor::GetViewSizes(), and ViewsDescriptor::HasAxis().
Referenced by ClSplitterWorkload::ClSplitterWorkload(), NeonLayerSupport::IsSplitterSupported(), ClLayerSupport::IsSplitterSupported(), and NeonSplitterWorkload::NeonSplitterWorkload().
void Concatenate | ( | const ConcatQueueDescriptor & | data, |
std::vector< ITensorHandle * > | inputs, | ||
std::vector< ITensorHandle * > | outputs | ||
) |
Definition at line 14 of file Concatenate.cpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, Decoder< IType >::Get(), TensorInfo::GetNumDimensions(), TensorInfo::GetNumElements(), TensorInfo::GetShape(), GetTensorInfo(), ConcatQueueDescriptor::ViewOrigin::m_Origin, ConcatQueueDescriptor::m_ViewOrigins, Map, MaxNumOfTensorDimensions, and Encoder< IType >::Set().
void armnn::ConditionalThrow | ( | bool | condition | ) |
Definition at line 174 of file Exceptions.hpp.
void armnn::ConditionalThrow | ( | bool | condition, |
const std::string & | message | ||
) |
Definition at line 165 of file Exceptions.hpp.
void armnn::ConditionalThrowIfNotEqual | ( | const std::string & | message, |
const ComparedType & | leftHandSide, | ||
const ComparedType & | rightHandSide | ||
) |
ComparedType must support: operator==(const ComparedType&) operator<<(ostream&, const ComparedType&)
Definition at line 189 of file Exceptions.hpp.
void armnn::ConfigureDetailsObject | ( | JsonChildObject & | detailsObject, |
std::string | layerDetailsStr | ||
) |
Definition at line 306 of file Profiling.cpp.
References ExecObjectDesc, JsonChildObject::SetAndParseDetails(), and JsonChildObject::SetType().
void ConfigureLogging | ( | bool | printToStandardOutput, |
bool | printToDebugOutput, | ||
LogSeverity | severity | ||
) |
Configures the logging behaviour of the ARMNN library.
printToStandardOutput: Set to true if log messages should be printed to the standard output. printToDebugOutput: Set to true if log messages be printed to a platform-specific debug output (where supported). severity: All log messages that are at this severity level or higher will be printed, others will be ignored.
Definition at line 20 of file Utils.cpp.
References SetAllLoggingSinks(), and SetLogFilter().
Referenced by ArmnnDevice::ArmnnDevice(), ConfigureLoggingTest(), and main().
|
inline |
Definition at line 44 of file ArmComputeTuningUtils.hpp.
References ARMNN_LOG, Exhaustive, info, None, Normal, and Rapid.
Referenced by ClBackendContext::ClBackendContext(), and GpuFsaBackendContext::GpuFsaBackendContext().
|
inline |
Checks the Layer's Connections to see if it's connected to a Layer with the provided layerType.
If dimSize is provided will also check if the connecting Tensor has more than that number of dimensions
Definition at line 271 of file SubgraphUtils.hpp.
References InputSlot::GetConnectedOutputSlot(), OutputSlot::GetConnection(), Layer::GetInputSlot(), TensorInfo::GetNumDimensions(), Layer::GetOutputSlot(), InputSlot::GetOwningLayer(), OutputSlot::GetOwningLayer(), InputSlot::GetTensorInfo(), and Layer::GetType().
|
inline |
Checks if the Layer is connected to any Layer that has an NCHW layout.
Definition at line 250 of file SubgraphUtils.hpp.
References InputSlot::GetConnectedOutputSlot(), OutputSlot::GetConnection(), Layer::GetInputSlot(), Layer::GetOutputSlot(), InputSlot::GetOwningLayer(), OutputSlot::GetOwningLayer(), and IsNCHW().
Referenced by NeonBackend::OptimizeSubgraphView(), and ClBackend::OptimizeSubgraphView().
std::tuple< TensorInfo, unsigned int > Convert1HWOTensorInfoToAcl | ( | const TensorInfo & | weightInfo, |
const TensorInfo & | inputInfo, | ||
const DataLayout | dataLayout | ||
) |
Weights for depthwise have a datalayout of [1,H,W,O] = [1,H,W,I*M] This function coverts a TensorInfo from [1,H,W,I*M] to [1,I*M,H,W] (if NCHW) or keeps it at [1,H,W,I*M] (if NHWC) as required by the compute library Returns a tuple of converted weights tensor info and depth multiplier.
Definition at line 177 of file WorkloadUtils.cpp.
References GetDataLayoutName(), TensorInfo::GetShape(), NCHW, NHWC, and armnnUtils::Permuted().
std::tuple< ConstTensor, unsigned int > Convert1HWOTensorToAcl | ( | const ConstTensorHandle * | weightTensor, |
const TensorInfo & | inputInfo, | ||
const DataLayout | dataLayout, | ||
void * | permuteBuffer | ||
) |
Weights for depthwise have a datalayout of [1,H,W,O] = [1,H,W,I*M] This function coverts a ConstCpuTensorHandle from [1,H,W,I*M] to [1,I*M,H,W] (if NCHW) or keeps it at [1,H,W,I*M] (if NHWC) as required by the compute library.
weightTensor | - ConstTensorHandle of weights tensor |
inputInfo | - TensorInfo of input tensor |
dataLayout | - DataLayout of the input tensor |
permuteBuffer | - Pointer to memory with the size of tensor. Used for the permutation |
Definition at line 146 of file WorkloadUtils.cpp.
References GetDataLayoutName(), TensorInfo::GetShape(), ConstTensorHandle::GetTensorInfo(), NCHW, NHWC, and PermuteTensor().
std::tuple< ConstTensor, unsigned int > Convert1HWOtoMIHW | ( | const ConstTensorHandle * | weightTensor, |
const TensorInfo & | inputInfo, | ||
const DataLayout & | dataLayout, | ||
void * | permuteBuffer | ||
) |
Converts a (weights) tensor from [1, H, W, I*M] = [1, H, W, O] to [M, I, H, W].
weightTensor | - ConstTensorHandle of the weight tensor that should be converted |
inputInfo | - TensorInfo of the corresponding input tensor |
dataLayout | - DataLayout of the input tensor e.g. NHWC or NCHW |
permuteBuffer | - Memory location with the same size as the weight tensor to write converted data to |
Definition at line 208 of file WorkloadUtils.cpp.
References DataLayoutIndexed::GetChannelsIndex(), TensorInfo::GetShape(), ConstTensorHandle::GetTensorInfo(), TensorInfo::HasPerAxisQuantization(), PermuteTensor(), and TensorInfo::SetShape().
|
inline |
Definition at line 87 of file ArmComputeUtils.hpp.
References ConvertActivationFunctionToAclActivationFunction(), ActivationDescriptor::m_A, ActivationDescriptor::m_B, and ActivationDescriptor::m_Function.
Referenced by ClActivationWorkload::ClActivationWorkload(), ClSqrtWorkload::ClSqrtWorkload(), ComputeConv3DInfo(), ConvertActivationDescriptorToAclActivationLayerInfo(), ConvertAdditionalInfoToAclActivationLayerInfo(), ConvertFullyConnectedDescriptorToAclFullyConnectedLayerInfo(), NeonActivationWorkload::NeonActivationWorkload(), and NeonSqrtWorkload::NeonSqrtWorkload().
|
inline |
Definition at line 94 of file ArmComputeUtils.hpp.
References ConvertActivationDescriptorToAclActivationLayerInfo().
|
inline |
|
inline |
Definition at line 105 of file ArmComputeUtils.hpp.
References ConvertActivationDescriptorToAclActivationLayerInfo(), and QueueDescriptor::GetAdditionalInformation().
Referenced by ClAdditionWorkload::ClAdditionWorkload(), ClBatchMatMulWorkload::ClBatchMatMulWorkload(), ClDivisionWorkload::ClDivisionWorkload(), ClElementwiseBinaryWorkload::ClElementwiseBinaryWorkload(), ClFullyConnectedWorkload::ClFullyConnectedWorkload(), ClMultiplicationWorkload::ClMultiplicationWorkload(), ClSubtractionWorkload::ClSubtractionWorkload(), ComputeConv3DInfo(), NeonAdditionWorkload::NeonAdditionWorkload(), NeonBatchMatMulWorkload::NeonBatchMatMulWorkload(), NeonDivisionWorkload::NeonDivisionWorkload(), NeonElementwiseBinaryWorkload::NeonElementwiseBinaryWorkload(), NeonFusedWorkload::NeonFusedWorkload(), NeonMultiplicationWorkload::NeonMultiplicationWorkload(), and NeonSubtractionWorkload::NeonSubtractionWorkload().
|
inline |
Definition at line 141 of file ArmComputeUtils.hpp.
References Equal, Greater, GreaterOrEqual, Less, LessOrEqual, ComparisonDescriptor::m_Operation, and NotEqual.
Referenced by ClComparisonWorkload::ClComparisonWorkload(), and NeonComparisonWorkload::NeonComparisonWorkload().
|
inline |
Definition at line 204 of file ArmComputeUtils.hpp.
References FullyConnectedDescriptor::m_TransposeWeightMatrix.
|
inline |
Definition at line 194 of file ArmComputeUtils.hpp.
References ConvertActivationDescriptorToAclActivationLayerInfo(), and FullyConnectedDescriptor::m_TransposeWeightMatrix.
Referenced by ClFullyConnectedWorkload::ClFullyConnectedWorkload().
|
constexpr |
Definition at line 206 of file Logging.hpp.
|
inline |
Definition at line 118 of file ArmComputeUtils.hpp.
int32_t ConvertMaskToACLFormat | ( | int32_t | mask, |
int32_t | numDim | ||
) |
Definition at line 299 of file WorkloadUtils.cpp.
Referenced by ClStridedSliceWorkload::ClStridedSliceWorkload(), and NeonStridedSliceWorkload::NeonStridedSliceWorkload().
|
inline |
Definition at line 182 of file ArmComputeUtils.hpp.
|
inline |
Definition at line 168 of file ArmComputeUtils.hpp.
|
inline |
Definition at line 295 of file ArmComputeUtils.hpp.
|
inline |
Definition at line 155 of file ArmComputeUtils.hpp.
References Average, L2, and Max.
Referenced by CreatePool2dAttributes().
|
inline |
Definition at line 306 of file ArmComputeUtils.hpp.
References ReduceDescriptor::m_ReduceOperation, Max, Mean, Min, Prod, and Sum.
|
inline |
armnn::ConstTensor ConvertWeightTensorFromArmnnToAcl | ( | const ConstTensorHandle * | weightTensor, |
DataLayout | dataLayout, | ||
void * | permuteBuffer | ||
) |
Definition at line 237 of file WorkloadUtils.cpp.
References Float16, Float32, BaseTensor< MemoryType >::GetDataType(), BaseTensor< MemoryType >::GetInfo(), TensorInfo::GetShape(), ConstTensorHandle::GetTensorInfo(), NCHW, NHWC, PermuteTensor(), QAsymmS8, QAsymmU8, QSymmS8, and ReshapeWeightsForAcl().
TensorInfo ConvertWeightTensorInfoFromArmnnToAcl | ( | const TensorInfo & | weightInfo, |
DataLayout | dataLayout | ||
) |
Definition at line 122 of file WorkloadUtils.cpp.
References NHWC, armnnUtils::Permuted(), and ReshapeWeightsForAcl().
void Convolve | ( | const TensorShape & | rInputShape, |
Decoder< float > & | rInputDecoder, | ||
const TensorShape & | rOutputShape, | ||
Encoder< float > & | rOutputEncoder, | ||
const TensorShape & | rFilterShape, | ||
Decoder< float > & | rFilterDecoder, | ||
bool | biasEnabled, | ||
Decoder< float > * | pBiasDecoder, | ||
DataLayout | dataLayout, | ||
unsigned int | paddingTop, | ||
unsigned int | paddingLeft, | ||
unsigned int | xStride, | ||
unsigned int | yStride, | ||
unsigned int | xDilation, | ||
unsigned int | yDilation, | ||
bool | depthwise | ||
) |
Definition at line 68 of file ConvImpl.cpp.
References Decoder< IType >::DecodeTensor(), DataLayoutIndexed::GetChannelsIndex(), DataLayoutIndexed::GetDataLayout(), DataLayoutIndexed::GetHeightIndex(), DataLayoutIndexed::GetWidthIndex(), NHWC, and Encoder< IType >::Set().
void Convolve3d | ( | const TensorShape & | rInputShape, |
Decoder< float > & | rInputDecoder, | ||
const TensorShape & | rOutputShape, | ||
Encoder< float > & | rOutputEncoder, | ||
const TensorShape & | rFilterShape, | ||
Decoder< float > & | rFilterDecoder, | ||
bool | biasEnabled, | ||
Decoder< float > * | pBiasDecoder, | ||
DataLayout | dataLayout, | ||
unsigned int | paddingTop, | ||
unsigned int | paddingLeft, | ||
unsigned int | paddingFront, | ||
unsigned int | xStride, | ||
unsigned int | yStride, | ||
unsigned int | zStride, | ||
unsigned int | xDilation, | ||
unsigned int | yDilation, | ||
unsigned int | zDilation | ||
) |
Definition at line 11 of file Conv3dImpl.cpp.
References Decoder< IType >::DecodeTensor(), DataLayoutIndexed::GetChannelsIndex(), DataLayoutIndexed::GetDataLayout(), DataLayoutIndexed::GetDepthIndex(), DataLayoutIndexed::GetHeightIndex(), DataLayoutIndexed::GetWidthIndex(), NDHWC, and Encoder< IType >::Set().
uint32_t armnn::CoordinatesToIndex | ( | TensorShape & | shape, |
std::vector< uint32_t > & | coordinates | ||
) |
Definition at line 32 of file Tile.cpp.
References TensorShape::GetNumDimensions().
Referenced by Tile().
void CopyArmComputeClTensorData | ( | arm_compute::CLTensor & | dstTensor, |
const T * | srcData | ||
) |
Definition at line 64 of file ClWorkloadUtils.hpp.
References ARMNN_SCOPED_PROFILING_EVENT_CL.
Referenced by ClConstantWorkload::Execute(), and GpuFsaConstantWorkload::Execute().
void armnn::CopyArmComputeTensorData | ( | arm_compute::Tensor & | dstTensor, |
const T * | srcData | ||
) |
Definition at line 63 of file NeonWorkloadUtils.hpp.
Referenced by InitializeArmComputeTensorData().
void armnn::CopyTensorContentsGeneric | ( | const ITensorHandle * | srcTensor, |
ITensorHandle * | dstTensor, | ||
CopyFunc | copy | ||
) |
Definition at line 46 of file WorkloadUtils.hpp.
References ARMNN_SCOPED_PROFILING_EVENT, TensorShape::GetNumDimensions(), ITensorHandle::GetShape(), ITensorHandle::GetStrides(), ITensorHandle::Map(), MaxNumOfTensorDimensions, Undefined, and ITensorHandle::Unmap().
Referenced by CopyToOutputTensor(), CopyMemGenericWorkload::Execute(), NeonConvertFp32ToFp16Workload::Execute(), NeonConvertFp16ToFp32Workload::Execute(), and CopyMemGenericWorkload::ExecuteAsync().
void armnn::CopyToOutputTensor | ( | const Tensor & | outputTensor, |
ITensorHandle * | outputTensorHandle | ||
) |
Definition at line 1444 of file LoadedNetwork.cpp.
References ARMNN_SCOPED_PROFILING_EVENT, CopyTensorContentsGeneric(), BaseTensor< MemoryType >::GetInfo(), BaseTensor< MemoryType >::GetMemoryArea(), and Undefined.
Referenced by LoadedNetwork::Execute().
|
inline |
Definition at line 29 of file ArmComputeUtils.hpp.
References TensorInfo::GetShape(), and NCHW.
|
inline |
Definition at line 57 of file ClContextSchema_generated.h.
References ClContextBuilder::add_programs(), and ClContextBuilder::Finish().
Referenced by CreateClContextDirect(), and ClContextSerializer::Serialize().
|
inline |
Definition at line 65 of file ClContextSchema_generated.h.
References CreateClContext().
OriginsDescriptor armnn::CreateDescriptorForConcatenation | ( | TensorShapeIt | first, |
TensorShapeIt | last, | ||
unsigned int | concatenationDimension | ||
) |
Convenience template to create an OriginsDescriptor to use when creating a ConcatLayer for performing concatenation of a number of input tensors.
Definition at line 300 of file Descriptors.hpp.
References OriginsDescriptor::SetConcatAxis(), and OriginsDescriptor::SetViewOriginCoord().
|
inline |
Definition at line 42 of file GpuFsaBackend.cpp.
References Layer::BeginInputSlots(), and Layer::EndInputSlots().
Referenced by GpuFsaBackend::OptimizeSubgraphView().
|
inline |
Definition at line 52 of file GpuFsaBackend.cpp.
References Layer::BeginOutputSlots(), and Layer::EndOutputSlots().
Referenced by GpuFsaBackend::OptimizeSubgraphView().
|
inline |
Definition at line 118 of file ClContextSchema_generated.h.
References ProgramBuilder::add_binary(), ProgramBuilder::add_name(), and ProgramBuilder::Finish().
Referenced by CreateProgramDirect(), and ClContextSerializer::Serialize().
|
inline |
|
inline |
Definition at line 62 of file GpuFsaBackend.cpp.
Referenced by GpuFsaBackend::OptimizeSubgraphView().
BackendsMap CreateSupportedBackends | ( | TensorHandleFactoryRegistry & | handleFactoryRegistry, |
BackendSettings & | backendSettings | ||
) |
Definition at line 1302 of file Network.cpp.
References BackendRegistryInstance(), and BackendSettings::m_SupportedBackends.
Referenced by Optimize().
void Debug | ( | const TensorInfo & | inputInfo, |
const T * | inputData, | ||
LayerGuid | guid, | ||
const std::string & | layerName, | ||
unsigned int | slotIndex, | ||
bool | outputsToFile | ||
) |
template void armnn::Debug< BFloat16 > | ( | const TensorInfo & | inputInfo, |
const BFloat16 * | inputData, | ||
LayerGuid | guid, | ||
const std::string & | layerName, | ||
unsigned int | slotIndex, | ||
bool | outputsToFile | ||
) |
template void armnn::Debug< float > | ( | const TensorInfo & | inputInfo, |
const float * | inputData, | ||
LayerGuid | guid, | ||
const std::string & | layerName, | ||
unsigned int | slotIndex, | ||
bool | outputsToFile | ||
) |
template void armnn::Debug< Half > | ( | const TensorInfo & | inputInfo, |
const Half * | inputData, | ||
LayerGuid | guid, | ||
const std::string & | layerName, | ||
unsigned int | slotIndex, | ||
bool | outputsToFile | ||
) |
template void armnn::Debug< int16_t > | ( | const TensorInfo & | inputInfo, |
const int16_t * | inputData, | ||
LayerGuid | guid, | ||
const std::string & | layerName, | ||
unsigned int | slotIndex, | ||
bool | outputsToFile | ||
) |
template void armnn::Debug< int32_t > | ( | const TensorInfo & | inputInfo, |
const int32_t * | inputData, | ||
LayerGuid | guid, | ||
const std::string & | layerName, | ||
unsigned int | slotIndex, | ||
bool | outputsToFile | ||
) |
template void armnn::Debug< int64_t > | ( | const TensorInfo & | inputInfo, |
const int64_t * | inputData, | ||
LayerGuid | guid, | ||
const std::string & | layerName, | ||
unsigned int | slotIndex, | ||
bool | outputsToFile | ||
) |
template void armnn::Debug< int8_t > | ( | const TensorInfo & | inputInfo, |
const int8_t * | inputData, | ||
LayerGuid | guid, | ||
const std::string & | layerName, | ||
unsigned int | slotIndex, | ||
bool | outputsToFile | ||
) |
template void armnn::Debug< uint8_t > | ( | const TensorInfo & | inputInfo, |
const uint8_t * | inputData, | ||
LayerGuid | guid, | ||
const std::string & | layerName, | ||
unsigned int | slotIndex, | ||
bool | outputsToFile | ||
) |
|
inline |
Definition at line 37 of file GpuFsaBackend.cpp.
void DepthToSpace | ( | const TensorInfo & | inputInfo, |
const DepthToSpaceDescriptor & | descriptor, | ||
const void * | inputData, | ||
void * | outputData, | ||
unsigned int | dataTypeSize | ||
) |
Definition at line 16 of file DepthToSpace.cpp.
References DepthToSpace(), DataLayoutIndexed::GetChannelsIndex(), DataLayoutIndexed::GetHeightIndex(), TensorShape::GetNumElements(), TensorInfo::GetShape(), DataLayoutIndexed::GetWidthIndex(), SpaceToDepthDescriptor::m_BlockSize, SpaceToDepthDescriptor::m_DataLayout, and armnnUtils::Permute().
Referenced by DepthToSpace().
|
inline |
std::vector<float> armnn::Dequantize | ( | const T * | quant, |
const TensorInfo & | info | ||
) |
u8 helpers
Definition at line 102 of file RefWorkloadUtils.hpp.
References Dequantize(), and info.
void Dequantize | ( | Decoder< float > & | inputDecoder, |
Encoder< float > & | outputEncoder, | ||
const TensorInfo & | inputInfo, | ||
const TensorInfo & | outputInfo | ||
) |
Definition at line 11 of file Dequantize.cpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, Decoder< IType >::Get(), TensorInfo::GetNumElements(), and Encoder< IType >::Set().
float Dequantize | ( | QuantizedType | value, |
float | scale, | ||
int32_t | offset | ||
) |
Dequantize an 8-bit data type into a floating point data type.
value | - The value to dequantize. |
scale | - The scale (must be non-zero). |
offset | - The offset. |
Definition at line 48 of file TypesUtils.cpp.
Referenced by QASymm8Decoder::DecodeTensor(), QASymmS8Decoder::DecodeTensor(), QSymmS8Decoder::DecodeTensor(), QSymm16Decoder::DecodeTensor(), QSymm8PerAxisDecoder::DecodeTensor(), ScaledInt32PerAxisDecoder::DecodeTensor(), SelectiveQuantizer< T, DoQuantize >::Dequantize(), Dequantize(), QASymm8Decoder::Get(), QASymmS8Decoder::Get(), QSymmS8Decoder::Get(), QSymm16Decoder::Get(), QASymm8Encoder::Get(), QASymmS8Encoder::Get(), QSymmS8Encoder::Get(), QSymm16Encoder::Get(), QSymm8PerAxisDecoder::Get(), QSymm8PerAxisEncoder::Get(), ScaledInt32PerAxisDecoder::Get(), and QSymm16PerAxisEncoder::Get().
void DetectionPostProcess | ( | const TensorInfo & | boxEncodingsInfo, |
const TensorInfo & | scoresInfo, | ||
const TensorInfo & | , | ||
const TensorInfo & | detectionBoxesInfo, | ||
const TensorInfo & | , | ||
const TensorInfo & | , | ||
const TensorInfo & | , | ||
const DetectionPostProcessDescriptor & | desc, | ||
Decoder< float > & | boxEncodings, | ||
Decoder< float > & | scores, | ||
Decoder< float > & | anchors, | ||
float * | detectionBoxes, | ||
float * | detectionClasses, | ||
float * | detectionScores, | ||
float * | numDetections | ||
) |
Definition at line 139 of file DetectionPostProcess.cpp.
References AllocateOutputData(), GenerateRangeK(), Decoder< IType >::Get(), TensorInfo::GetNumElements(), TensorInfo::GetShape(), DetectionPostProcessDescriptor::m_DetectionsPerClass, DetectionPostProcessDescriptor::m_MaxClassesPerDetection, DetectionPostProcessDescriptor::m_MaxDetections, DetectionPostProcessDescriptor::m_NmsIouThreshold, DetectionPostProcessDescriptor::m_NmsScoreThreshold, DetectionPostProcessDescriptor::m_NumClasses, DetectionPostProcessDescriptor::m_ScaleH, DetectionPostProcessDescriptor::m_ScaleW, DetectionPostProcessDescriptor::m_ScaleX, DetectionPostProcessDescriptor::m_ScaleY, DetectionPostProcessDescriptor::m_UseRegularNms, NonMaxSuppression(), and TopKSort().
void armnn::ExecuteFunction | ( | std::vector< ITensorHandle * > | inputs, |
std::vector< ITensorHandle * > | outputs, | ||
BinaryOperation | operation, | ||
const std::string & | layerName = "" |
||
) |
Definition at line 27 of file RefElementwiseBinaryWorkload.cpp.
References Add, CHECK_LOCATION, Div, GetBinaryOperationAsCString(), TensorInfo::GetShape(), GetTensorInfo(), Map, Maximum, Minimum, Mul, Power, SqDiff, and Sub.
void armnn::ExtractJsonObjects | ( | unsigned int | inferenceIndex, |
const Event * | parentEvent, | ||
JsonChildObject & | parentObject, | ||
std::map< const Event *, std::vector< const Event * >> | descendantsMap | ||
) |
Definition at line 314 of file Profiling.cpp.
References JsonChildObject::AddChild(), JsonChildObject::AddMeasurement(), ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, Event, JsonChildObject::GetChild(), Event::GetInstruments(), Event::GetMeasurements(), Event::GetProfilingGuid(), OptionalBase::has_value(), JsonChildObject::m_Children, Measurement, JsonChildObject::NumChildren(), JsonChildObject::SetGuid(), JsonChildObject::SetType(), JsonChildObject::SetUnit(), and OptionalReferenceSwitch< IsReference, T >::value().
Referenced by ProfilerImpl::Print().
void armnn::FakeQuantization | ( | const float * | inputData, |
float * | outputData, | ||
uint32_t | numElements, | ||
float | min, | ||
float | max | ||
) |
Definition at line 17 of file RefFakeQuantizationFloat32Workload.cpp.
bool armnn::FalseFunc | ( | Optional< std::string & > | reasonIfUnsupported, |
Params &&... | params | ||
) |
Definition at line 62 of file LayerSupportCommon.hpp.
References IgnoreUnused().
bool armnn::FalseFuncF16 | ( | Optional< std::string & > | reasonIfUnsupported, |
Params &&... | params | ||
) |
Definition at line 70 of file LayerSupportCommon.hpp.
References IgnoreUnused(), and SetValueChecked().
bool armnn::FalseFuncF32 | ( | Optional< std::string & > | reasonIfUnsupported, |
Params &&... | params | ||
) |
Definition at line 78 of file LayerSupportCommon.hpp.
References IgnoreUnused(), and SetValueChecked().
bool armnn::FalseFuncI32 | ( | Optional< std::string & > | reasonIfUnsupported, |
Params &&... | params | ||
) |
Definition at line 94 of file LayerSupportCommon.hpp.
References IgnoreUnused(), and SetValueChecked().
bool armnn::FalseFuncU8 | ( | Optional< std::string & > | reasonIfUnsupported, |
Params &&... | params | ||
) |
Definition at line 86 of file LayerSupportCommon.hpp.
References IgnoreUnused(), and SetValueChecked().
bool armnn::FalseInputFuncF16 | ( | Optional< std::string & > | reasonIfUnsupported, |
Params &&... | params | ||
) |
Definition at line 110 of file LayerSupportCommon.hpp.
References IgnoreUnused(), and SetValueChecked().
bool armnn::FalseInputFuncF32 | ( | Optional< std::string & > | reasonIfUnsupported, |
Params &&... | params | ||
) |
Definition at line 102 of file LayerSupportCommon.hpp.
References IgnoreUnused(), and SetValueChecked().
bool armnn::FalseOutputFuncF16 | ( | Optional< std::string & > | reasonIfUnsupported, |
Params &&... | params | ||
) |
Definition at line 126 of file LayerSupportCommon.hpp.
References IgnoreUnused(), and SetValueChecked().
bool armnn::FalseOutputFuncF32 | ( | Optional< std::string & > | reasonIfUnsupported, |
Params &&... | params | ||
) |
Definition at line 118 of file LayerSupportCommon.hpp.
References IgnoreUnused(), and SetValueChecked().
void Fill | ( | Encoder< float > & | output, |
const TensorShape & | desiredOutputShape, | ||
const float | value | ||
) |
Creates a tensor and fills it with a scalar value.
Definition at line 13 of file Fill.cpp.
References TensorShape::GetNumElements(), and Encoder< IType >::Set().
std::vector<Measurement> armnn::FindKernelMeasurements | ( | const Event * | event | ) |
Definition at line 62 of file Profiling.cpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, and Event::GetMeasurements().
Measurement armnn::FindMeasurement | ( | const std::string & | name, |
const Event * | event | ||
) |
Definition at line 43 of file Profiling.cpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, and Event::GetMeasurements().
Referenced by ProfilerImpl::AnalyzeEventSequenceAndWriteResults(), and ProfilerImpl::CalculateProfilingEventStats().
|
inline |
Definition at line 171 of file ClContextSchema_generated.h.
References ClContextIdentifier().
|
inline |
Definition at line 177 of file ClContextSchema_generated.h.
References ClContextIdentifier().
LayerType* armnn::FoldPadIntoAveragePool2d | ( | OptimizationViews & | optimizationViews, |
Pooling2dLayer * | baseLayer, | ||
Pooling2dDescriptor & | poolDescriptor, | ||
PadLayer * | padLayer | ||
) |
Definition at line 341 of file SubgraphUtils.hpp.
References INetwork::AddPooling2dLayer(), FoldPadLayer(), and OptimizationViews::GetINetwork().
LayerType* armnn::FoldPadLayer | ( | OptimizationViews & | optimizationViews, |
LayerType * | baseLayer, | ||
LayerType * | replacementLayer, | ||
PadLayer * | padLayer | ||
) |
Definition at line 234 of file SubgraphUtils.hpp.
References OptimizationViews::AddSubstitution().
Referenced by FoldPadIntoAveragePool2d().
void armnn::ForEachLayerInput | ( | LayerSelectionInfo::LayerInfoContainer & | layerInfos, |
LayerSelectionInfo & | layerInfo, | ||
Delegate | function | ||
) |
Definition at line 277 of file SubgraphViewSelector.cpp.
References Layer::GetInputSlots().
Referenced by AssignSplitId(), and IsReadyForSplitAssignment().
void armnn::ForEachLayerOutput | ( | LayerSelectionInfo::LayerInfoContainer & | layerInfos, |
LayerSelectionInfo & | layerInfo, | ||
Delegate | function | ||
) |
Definition at line 301 of file SubgraphViewSelector.cpp.
References Layer::GetOutputSlots().
Referenced by SubgraphViewSelector::SelectSubgraphs().
void FullyConnected | ( | const TensorShape & | rInputShape, |
Decoder< float > & | rInputDecoder, | ||
const TensorShape & | rOutputShape, | ||
Encoder< float > & | rOutputEncoder, | ||
const TensorShape & | rWeightsShape, | ||
Decoder< float > & | rWeightDecoder, | ||
Decoder< float > * | pBiasDecoder, | ||
const bool | biasEnabled, | ||
const unsigned int | K, | ||
const bool | transposeWeights | ||
) |
Performs a matrix multiplication and optionally adds a bias.
Definition at line 13 of file FullyConnected.cpp.
References Decoder< IType >::DecodeTensor(), and Encoder< IType >::Set().
LayerType* armnn::FuseAdditionLayer | ( | OptimizationViews & | optimizationViews, |
LayerType * | baseLayer, | ||
ActivationLayer * | activationLayer, | ||
ActivationDescriptor & | activationDesc, | ||
std::string | name | ||
) |
Definition at line 73 of file ArmComputeSubgraphUtils.hpp.
References INetwork::AddAdditionLayer(), ARMNN_NO_DEPRECATE_WARN_BEGIN, ARMNN_NO_DEPRECATE_WARN_END, FuseLayer(), and OptimizationViews::GetINetwork().
LayerType* armnn::FuseBatchNormalizationLayer | ( | OptimizationViews & | optimizationViews, |
LayerType * | baseLayer, | ||
ActivationLayer * | activationLayer, | ||
ActivationDescriptor & | activationDesc, | ||
std::string | name | ||
) |
Definition at line 178 of file ArmComputeSubgraphUtils.hpp.
References INetwork::AddBatchNormalizationLayer(), FuseLayer(), and OptimizationViews::GetINetwork().
LayerType* armnn::FuseConvolution2dLayer | ( | OptimizationViews & | optimizationViews, |
LayerType * | baseLayer, | ||
ActivationLayer * | activationLayer, | ||
ActivationDescriptor & | activationDesc, | ||
std::string | name | ||
) |
Definition at line 208 of file ArmComputeSubgraphUtils.hpp.
References INetwork::AddConvolution2dLayer(), FuseLayer(), and OptimizationViews::GetINetwork().
LayerType* armnn::FuseDepthwiseConvolution2dLayer | ( | OptimizationViews & | optimizationViews, |
LayerType * | baseLayer, | ||
ActivationLayer * | activationLayer, | ||
ActivationDescriptor & | activationDesc, | ||
std::string | name | ||
) |
Definition at line 230 of file ArmComputeSubgraphUtils.hpp.
References INetwork::AddDepthwiseConvolution2dLayer(), FuseLayer(), and OptimizationViews::GetINetwork().
LayerType* armnn::FuseDivisionLayer | ( | OptimizationViews & | optimizationViews, |
LayerType * | baseLayer, | ||
ActivationLayer * | activationLayer, | ||
ActivationDescriptor & | activationDesc, | ||
std::string | name | ||
) |
Definition at line 115 of file ArmComputeSubgraphUtils.hpp.
References INetwork::AddDivisionLayer(), ARMNN_NO_DEPRECATE_WARN_BEGIN, ARMNN_NO_DEPRECATE_WARN_END, FuseLayer(), and OptimizationViews::GetINetwork().
LayerType* armnn::FuseElementwiseBinaryLayer | ( | OptimizationViews & | optimizationViews, |
LayerType * | baseLayer, | ||
ActivationLayer * | activationLayer, | ||
ActivationDescriptor & | activationDesc, | ||
BinaryOperation | operation, | ||
std::string | name | ||
) |
Definition at line 157 of file ArmComputeSubgraphUtils.hpp.
References INetwork::AddElementwiseBinaryLayer(), FuseLayer(), and OptimizationViews::GetINetwork().
LayerType* armnn::FuseFullyConnectedLayer | ( | OptimizationViews & | optimizationViews, |
LayerType * | baseLayer, | ||
ActivationLayer * | activationLayer, | ||
ActivationDescriptor & | activationDesc, | ||
std::string | name | ||
) |
Definition at line 252 of file ArmComputeSubgraphUtils.hpp.
References INetwork::AddFullyConnectedLayer(), FuseLayer(), and OptimizationViews::GetINetwork().
LayerType* armnn::FuseLayer | ( | OptimizationViews & | optimizationViews, |
LayerType * | baseLayer, | ||
LayerType * | replacementLayer, | ||
ActivationLayer * | activationLayer, | ||
ActivationDescriptor & | activationDesc | ||
) |
Definition at line 53 of file ArmComputeSubgraphUtils.hpp.
References OptimizationViews::AddSubstitution().
Referenced by FuseAdditionLayer(), FuseBatchNormalizationLayer(), FuseConvolution2dLayer(), FuseDepthwiseConvolution2dLayer(), FuseDivisionLayer(), FuseElementwiseBinaryLayer(), FuseFullyConnectedLayer(), FuseMultiplicationLayer(), and FuseSubtractionLayer().
LayerType* armnn::FuseMultiplicationLayer | ( | OptimizationViews & | optimizationViews, |
LayerType * | baseLayer, | ||
ActivationLayer * | activationLayer, | ||
ActivationDescriptor & | activationDesc, | ||
std::string | name | ||
) |
Definition at line 136 of file ArmComputeSubgraphUtils.hpp.
References INetwork::AddMultiplicationLayer(), ARMNN_NO_DEPRECATE_WARN_BEGIN, ARMNN_NO_DEPRECATE_WARN_END, FuseLayer(), and OptimizationViews::GetINetwork().
LayerType* armnn::FuseSubtractionLayer | ( | OptimizationViews & | optimizationViews, |
LayerType * | baseLayer, | ||
ActivationLayer * | activationLayer, | ||
ActivationDescriptor & | activationDesc, | ||
std::string | name | ||
) |
Definition at line 94 of file ArmComputeSubgraphUtils.hpp.
References INetwork::AddSubtractionLayer(), ARMNN_NO_DEPRECATE_WARN_BEGIN, ARMNN_NO_DEPRECATE_WARN_END, FuseLayer(), and OptimizationViews::GetINetwork().
void Gather | ( | const TensorInfo & | paramsInfo, |
const TensorInfo & | indicesInfo, | ||
const TensorInfo & | outputInfo, | ||
Decoder< float > & | params, | ||
const int32_t * | indices, | ||
Encoder< float > & | output, | ||
const int32_t | axis_int | ||
) |
Definition at line 15 of file Gather.cpp.
References Decoder< IType >::Get(), TensorInfo::GetNumDimensions(), TensorInfo::GetNumElements(), TensorInfo::GetShape(), IgnoreUnused(), and Encoder< IType >::Set().
void armnn::GatherTensorHandlePairs | ( | const DescriptorType & | descriptor, |
std::vector< std::pair< SrcTensorHandleType *, DstTensorHandleType * >> & | tensorHandlePairs | ||
) |
armnn::PermutationVector GeneratePermutationVectorOnLastTwoDimensions | ( | unsigned int | rank | ) |
Generates a permutation vector of size rank that permutes the 2 most right dimensions.
rank | - Tensor rank, i.e. number of dimensions in the tensors |
Definition at line 357 of file WorkloadUtils.cpp.
std::vector<unsigned int> armnn::GenerateRangeK | ( | unsigned int | k | ) |
Definition at line 16 of file DetectionPostProcess.cpp.
Referenced by DetectionPostProcess(), and NonMaxSuppression().
|
constexpr |
|
constexpr |
Definition at line 52 of file TypesUtils.hpp.
Definition at line 28 of file WorkloadData.cpp.
References BFloat16.
Referenced by FullyConnectedQueueDescriptor::Validate(), Convolution2dQueueDescriptor::Validate(), Convolution3dQueueDescriptor::Validate(), DepthwiseConvolution2dQueueDescriptor::Validate(), and TransposeConvolution2dQueueDescriptor::Validate().
|
inline |
Definition at line 13 of file LayerSupportRules.hpp.
References Float16, Float32, QAsymmS8, QAsymmU8, QSymmS16, QSymmS8, Signed32, and OptionalReferenceSwitch< std::is_reference< T >::value, T >::value().
Referenced by BiasAndWeightsTypesMatch::BiasAndWeightsTypesMatch().
|
constexpr |
Definition at line 76 of file TypesUtils.hpp.
References Add, Div, Maximum, Minimum, Mul, Power, SqDiff, and Sub.
Referenced by ExecuteFunction(), and StringifyLayerParameters< ElementwiseBinaryDescriptor >::Serialize().
Optional< const BackendOptions::BackendOption > GetCapability | ( | const std::string & | backendCapabilityName, |
const armnn::BackendId & | backend | ||
) |
Returns a BackendCapability if the backend lists the capability The BackendCapability must then be inspected to check whether or not that BackendCapability is supported Otherwise returns an EmptyOptional if the BackendCapability is unlisted.
Definition at line 51 of file BackendHelper.cpp.
References BackendRegistryInstance(), and GetCapability().
Optional< const BackendOptions::BackendOption > GetCapability | ( | const std::string & | backendCapabilityName, |
const BackendCapabilities & | capabilities | ||
) |
Returns a BackendCapability if the backend lists the capability The BackendCapability must then be inspected to check whether or not that BackendCapability is supported Otherwise returns an EmptyOptional if the BackendCapability is unlisted.
Definition at line 37 of file BackendHelper.cpp.
References BackendOptions::GetOption(), and BackendOptions::GetOptionCount().
Referenced by GetCapability(), HasCapability(), LayerSupportHandle::IsConvolution2dSupported(), LayerSupportHandle::IsDepthwiseConvolutionSupported(), LayerSupportHandle::IsDilatedDepthwiseConvolutionSupported(), and LayerSupportHandle::IsFullyConnectedSupported().
|
inline |
Definition at line 140 of file ClContextSchema_generated.h.
Referenced by ClContextDeserializer::DeserializeFromBinary().
|
constexpr |
Definition at line 62 of file TypesUtils.hpp.
References Equal, Greater, GreaterOrEqual, Less, LessOrEqual, and NotEqual.
Referenced by StringifyLayerParameters< ComparisonDescriptor >::Serialize().
|
constexpr |
Deprecated function that will be removed together with the Compute enum.
Definition at line 34 of file BackendId.hpp.
References CpuAcc, CpuRef, and GpuAcc.
Referenced by BackendId::BackendId(), BackendId::IsCpuAcc(), BackendId::IsCpuRef(), BackendId::IsGpuAcc(), BackendId::IsUndefined(), and operator<<().
|
inline |
Definition at line 46 of file ClWorkloadUtils.hpp.
|
constexpr |
Definition at line 253 of file TypesUtils.hpp.
References NCDHW, NCHW, NDHWC, and NHWC.
Referenced by Convert1HWOTensorInfoToAcl(), Convert1HWOTensorToAcl(), MakeTensorShape(), StringifyLayerParameters< BatchNormalizationDescriptor >::Serialize(), StringifyLayerParameters< BatchToSpaceNdDescriptor >::Serialize(), StringifyLayerParameters< Convolution2dDescriptor >::Serialize(), StringifyLayerParameters< Convolution3dDescriptor >::Serialize(), StringifyLayerParameters< DepthwiseConvolution2dDescriptor >::Serialize(), StringifyLayerParameters< L2NormalizationDescriptor >::Serialize(), StringifyLayerParameters< NormalizationDescriptor >::Serialize(), StringifyLayerParameters< Pooling2dDescriptor >::Serialize(), StringifyLayerParameters< Pooling3dDescriptor >::Serialize(), StringifyLayerParameters< ResizeDescriptor >::Serialize(), StringifyLayerParameters< SpaceToBatchNdDescriptor >::Serialize(), StringifyLayerParameters< SpaceToDepthDescriptor >::Serialize(), StringifyLayerParameters< StridedSliceDescriptor >::Serialize(), and StringifyLayerParameters< TransposeConvolution2dDescriptor >::Serialize().
|
constexpr |
Definition at line 233 of file TypesUtils.hpp.
References BFloat16, Boolean, Float16, Float32, QAsymmS8, QAsymmU8, QSymmS16, QSymmS8, Signed32, and Signed64.
Referenced by AttemptBackendAssignment(), RefDebugWorkload< DataType >::GetName(), armnnUtils::GetPerAxisParams(), ConstantLayer::SerializeLayerParameters(), armnnUtils::ToFloatArray(), and VerifyTensorInfoDataType().
|
constexpr |
Definition at line 182 of file TypesUtils.hpp.
References BFloat16, Boolean, Float16, Float32, QAsymmS8, QAsymmU8, QSymmS16, QSymmS8, Signed32, and Signed64.
Referenced by TosaRefTensorHandle::CanBeImported(), RefTensorHandle::CanBeImported(), TensorInfo::GetNumBytes(), GetUnpaddedTensorStrides(), PermuteTensor(), and armnn_driver::SwizzleAndroidNn4dTensorToArmNn().
Definition at line 109 of file Profiling.cpp.
Referenced by ProfilerImpl::AnalyzeEventSequenceAndWriteResults().
Definition at line 110 of file Profiling.cpp.
|
inline |
Definition at line 71 of file NeonBackendOptimizationUtils.hpp.
References Layer::GetNameStr().
Referenced by NeonBackend::OptimizeSubgraphView().
|
constexpr |
Definition at line 119 of file TypesUtils.hpp.
References AddMulAdd.
Referenced by StringifyLayerParameters< FusedDescriptor >::Serialize().
Graph & GetGraphForTesting | ( | IOptimizedNetwork * | optNet | ) |
Definition at line 49 of file TestUtils.cpp.
References IOptimizedNetwork::pOptimizedNetworkImpl.
LayerSupportHandle GetILayerSupportByBackendId | ( | const armnn::BackendId & | backend | ) |
Convenience function to retrieve the ILayerSupportHandle for a backend.
Definition at line 23 of file BackendHelper.cpp.
References BackendRegistryInstance(), BackendRegistry::GetFactory(), and BackendRegistry::IsBackendRegistered().
const armnn::ConstTensor armnn::GetInputTensor | ( | const LayerBindingId | layerId, |
const InputTensors & | inputTensors | ||
) |
Definition at line 1460 of file LoadedNetwork.cpp.
const DataType* armnn::GetInputTensorData | ( | unsigned int | idx, |
const PayloadType & | data | ||
) |
const BFloat16* armnn::GetInputTensorDataBFloat16 | ( | unsigned int | idx, |
const PayloadType & | data | ||
) |
Definition at line 86 of file RefWorkloadUtils.hpp.
const float* armnn::GetInputTensorDataFloat | ( | unsigned int | idx, |
const PayloadType & | data | ||
) |
Definition at line 62 of file RefWorkloadUtils.hpp.
const Half* armnn::GetInputTensorDataHalf | ( | unsigned int | idx, |
const PayloadType & | data | ||
) |
Definition at line 74 of file RefWorkloadUtils.hpp.
Definition at line 1017 of file Network.cpp.
References Float32, InputSlot::GetConnectedOutputSlot(), TensorInfo::GetDataType(), Layer::GetInputSlot(), Layer::GetNumInputSlots(), Layer::GetNumOutputSlots(), Layer::GetOutputSlot(), and OutputSlot::GetTensorInfo().
Referenced by AssignBackends(), and AssignBackendsIConnectable().
char const * GetLayerTypeAsCString | ( | LayerType | type | ) |
Definition at line 13 of file InternalTypes.cpp.
References LIST_OF_LAYER_TYPE.
Referenced by AttemptBackendAssignment(), Connect(), StrategyBase< DefaultStrategy >::ExecuteStrategy(), Layer::InferOutputShapes(), Graph::InferTensorInfos(), Graph::Print(), ReturnWithError(), Layer::SerializeLayerParameters(), Graph::SerializeToDot(), ElementwiseBaseLayer::ValidateTensorShapesFromInputs(), ElementwiseBinaryLayer::ValidateTensorShapesFromInputs(), ElementwiseUnaryLayer::ValidateTensorShapesFromInputs(), Graph::VerifyConstantLayerSetTensorInfo(), and Layer::VerifyLayerConnections().
|
constexpr |
|
constexpr |
Definition at line 295 of file TypesUtils.hpp.
References MultiAxisPacking, and SingleAxisPacking.
Referenced by RuntimeImpl::RuntimeImpl().
|
inline |
Definition at line 36 of file MemoryOptimizerStrategyLibrary.hpp.
Referenced by RuntimeImpl::RuntimeImpl().
|
inline |
Definition at line 47 of file MemoryOptimizerStrategyLibrary.hpp.
ModelOptions & GetModelOptionsForTesting | ( | IOptimizedNetwork * | optNet | ) |
Definition at line 54 of file TestUtils.cpp.
References IOptimizedNetwork::pOptimizedNetworkImpl.
|
constexpr |
Definition at line 265 of file TypesUtils.hpp.
References Across, and Within.
Referenced by StringifyLayerParameters< NormalizationDescriptor >::Serialize().
|
constexpr |
Definition at line 275 of file TypesUtils.hpp.
References LocalBrightness, and LocalContrast.
Referenced by StringifyLayerParameters< NormalizationDescriptor >::Serialize().
unsigned int armnn::GetNumActivations | ( | const TensorInfo & | inputInfo | ) |
Definition at line 16 of file RefFullyConnectedWorkload.cpp.
References TensorInfo::GetNumDimensions(), and TensorInfo::GetShape().
unsigned int GetNumberOfCacheFiles | ( | const armnn::BackendId & | backend | ) |
Returns the number of cached files if backend supports caching.
Definition at line 130 of file BackendHelper.cpp.
References BackendRegistryInstance().
Referenced by ArmnnDriver::getNumberOfCacheFilesNeeded(), ArmnnDriverImpl::PrepareArmnnModel(), and ArmnnDriverImpl::PrepareArmnnModelFromCache().
uint32_t armnn::GetNumInputs | ( | bool | biasEnabled | ) |
Definition at line 455 of file Descriptors.cpp.
Referenced by FullyConnectedDescriptor::GetNumInputs(), Convolution2dDescriptor::GetNumInputs(), Convolution3dDescriptor::GetNumInputs(), and DepthwiseConvolution2dDescriptor::GetNumInputs().
unsigned int armnn::GetOffset | ( | const TensorShape & | shape, |
unsigned int | b, | ||
unsigned int | h, | ||
unsigned int | w, | ||
unsigned int | c, | ||
const DataLayoutIndexed & | dataLayout | ||
) |
Definition at line 15 of file SpaceToBatchNd.cpp.
References CHECK_LOCATION, DataLayoutIndexed::GetChannelsIndex(), DataLayoutIndexed::GetDataLayout(), DataLayoutIndexed::GetHeightIndex(), TensorShape::GetNumDimensions(), and DataLayoutIndexed::GetWidthIndex().
Referenced by SpaceToBatchNd(), and SpaceToDepth().
|
constexpr |
Definition at line 139 of file TypesUtils.hpp.
References Ceiling, and Floor.
Referenced by StringifyLayerParameters< Pooling2dDescriptor >::Serialize(), and StringifyLayerParameters< Pooling3dDescriptor >::Serialize().
const armnn::Tensor armnn::GetOutputTensor | ( | const LayerBindingId | layerId, |
const OutputTensors & | outputTensors | ||
) |
Definition at line 1473 of file LoadedNetwork.cpp.
DataType* armnn::GetOutputTensorData | ( | ITensorHandle * | tensorHandle | ) |
Definition at line 56 of file RefWorkloadUtils.hpp.
References ITensorHandle::Map().
DataType * GetOutputTensorData | ( | unsigned int | idx, |
const PayloadType & | data | ||
) |
BFloat16* armnn::GetOutputTensorDataBFloat16 | ( | unsigned int | idx, |
const PayloadType & | data | ||
) |
Definition at line 92 of file RefWorkloadUtils.hpp.
float* armnn::GetOutputTensorDataFloat | ( | unsigned int | idx, |
const PayloadType & | data | ||
) |
Definition at line 68 of file RefWorkloadUtils.hpp.
Half* armnn::GetOutputTensorDataHalf | ( | unsigned int | idx, |
const PayloadType & | data | ||
) |
Definition at line 80 of file RefWorkloadUtils.hpp.
|
constexpr |
Definition at line 149 of file TypesUtils.hpp.
References Exclude, and IgnoreValue.
Referenced by StringifyLayerParameters< Pooling2dDescriptor >::Serialize(), and StringifyLayerParameters< Pooling3dDescriptor >::Serialize().
|
constexpr |
Definition at line 159 of file TypesUtils.hpp.
References Constant, Reflect, and Symmetric.
Referenced by StringifyLayerParameters< PadDescriptor >::Serialize().
|
constexpr |
Definition at line 128 of file TypesUtils.hpp.
References Average, L2, and Max.
Referenced by StringifyLayerParameters< Pooling2dDescriptor >::Serialize(), and StringifyLayerParameters< Pooling3dDescriptor >::Serialize().
arm::pipe::IProfilingService & GetProfilingService | ( | armnn::RuntimeImpl * | runtime | ) |
Definition at line 59 of file TestUtils.cpp.
|
constexpr |
Definition at line 170 of file TypesUtils.hpp.
References Max, Mean, Min, Prod, and Sum.
Referenced by ConvertReduceToTosaOperator(), and StringifyLayerParameters< ReduceDescriptor >::Serialize().
|
constexpr |
Definition at line 285 of file TypesUtils.hpp.
References Bilinear, and NearestNeighbor.
Referenced by StringifyLayerParameters< ResizeDescriptor >::Serialize().
|
inline |
Definition at line 144 of file ClContextSchema_generated.h.
|
constexpr |
Definition at line 21 of file TypesUtils.hpp.
References Failure, and Success.
Referenced by operator<<().
|
inline |
float32 helpers
Definition at line 33 of file RefWorkloadUtils.hpp.
Referenced by BatchNormImpl(), Concatenate(), ExecuteFunction(), Split(), Splitter(), FillLayer::ValidateTensorShapesFromInputs(), SwitchLayer::ValidateTensorShapesFromInputs(), ConstantLayer::ValidateTensorShapesFromInputs(), DetectionPostProcessLayer::ValidateTensorShapesFromInputs(), SplitterLayer::ValidateTensorShapesFromInputs(), LstmLayer::ValidateTensorShapesFromInputs(), QuantizedLstmLayer::ValidateTensorShapesFromInputs(), and QLstmLayer::ValidateTensorShapesFromInputs().
|
inline |
Definition at line 19 of file Timer.hpp.
References GetTimeNow().
Referenced by RuntimeImpl::EnqueueWorkload(), RuntimeImpl::Execute(), and RuntimeImpl::~RuntimeImpl().
|
inline |
Definition at line 14 of file Timer.hpp.
Referenced by RuntimeImpl::EnqueueWorkload(), RuntimeImpl::Execute(), GetTimeDuration(), RuntimeImpl::RuntimeImpl(), and RuntimeImpl::~RuntimeImpl().
|
constexpr |
Definition at line 92 of file TypesUtils.hpp.
References Abs, Ceil, Exp, Log, LogicalNot, Neg, Rsqrt, Sin, and Sqrt.
Referenced by StringifyLayerParameters< ElementwiseUnaryDescriptor >::Serialize().
TensorShape GetUnpaddedTensorStrides | ( | const TensorInfo & | tensorInfo | ) |
Definition at line 15 of file TensorHandle.cpp.
References TensorInfo::GetDataType(), GetDataTypeSize(), TensorShape::GetNumDimensions(), and TensorInfo::GetShape().
Referenced by TosaRefTensorHandle::GetStrides(), SampleTensorHandle::GetStrides(), RefTensorHandle::GetStrides(), ConstTensorHandle::GetStrides(), and RefTensorHandleDecorator::GetStrides().
const std::string GetVersion | ( | ) |
Definition at line 81 of file Utils.cpp.
References ARMNN_VERSION.
void GpuFsaActivationCreateOp | ( | GpuFsaPreCompiledBlob * | blob, |
const TensorInfo & | input, | ||
const ActivationDescriptor & | descriptor | ||
) |
Definition at line 58 of file GpuFsaActivation.cpp.
References GpuFsaPreCompiledBlob::sketch, and GpuFsaPreCompiledBlob::workloadContext.
Referenced by GpuFsaBackend::OptimizeSubgraphView().
arm_compute::Status GpuFsaActivationValidate | ( | const TensorInfo & | input, |
const ActivationDescriptor & | descriptor | ||
) |
Definition at line 22 of file GpuFsaActivation.cpp.
Referenced by GpuFsaLayerSupport::IsLayerSupported().
|
constexpr |
Definition at line 10 of file GpuFsaBackendId.hpp.
Referenced by GpuFsaBackend::GetIdStatic().
void GpuFsaBatchMatMulCreateOp | ( | GpuFsaPreCompiledBlob * | blob, |
const TensorInfo & | input0, | ||
const TensorInfo & | input1, | ||
const BatchMatMulDescriptor & | descriptor | ||
) |
Definition at line 51 of file GpuFsaBatchMatMul.cpp.
References GpuFsaPreCompiledBlob::sketch, and GpuFsaPreCompiledBlob::workloadContext.
Referenced by GpuFsaBackend::OptimizeSubgraphView().
arm_compute::Status GpuFsaBatchMatMulValidate | ( | const TensorInfo & | input0, |
const TensorInfo & | input1, | ||
const BatchMatMulDescriptor & | descriptor | ||
) |
Definition at line 22 of file GpuFsaBatchMatMul.cpp.
Referenced by GpuFsaLayerSupport::IsLayerSupported().
void GpuFsaCastCreateOp | ( | GpuFsaPreCompiledBlob * | blob, |
const TensorInfo & | input, | ||
const TensorInfo & | output | ||
) |
Definition at line 61 of file GpuFsaCast.cpp.
References GpuFsaPreCompiledBlob::sketch, and GpuFsaPreCompiledBlob::workloadContext.
Referenced by GpuFsaBackend::OptimizeSubgraphView().
arm_compute::Status GpuFsaCastValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 33 of file GpuFsaCast.cpp.
Referenced by GpuFsaLayerSupport::IsLayerSupported().
arm_compute::Status GpuFsaConstantWorkloadValidate | ( | const TensorInfo & | output | ) |
Definition at line 17 of file GpuFsaConstantWorkload.cpp.
void GpuFsaConvolution2dCreateOp | ( | GpuFsaPreCompiledBlob * | blob, |
const TensorInfo & | input, | ||
const Convolution2dDescriptor & | descriptor, | ||
const TensorInfo & | weights, | ||
const Optional< TensorInfo > & | biases | ||
) |
Definition at line 70 of file GpuFsaConvolution2d.cpp.
References GpuFsaPreCompiledBlob::sketch, and GpuFsaPreCompiledBlob::workloadContext.
Referenced by GpuFsaBackend::OptimizeSubgraphView().
arm_compute::Status GpuFsaConvolution2dValidate | ( | const TensorInfo & | input, |
const Convolution2dDescriptor & | descriptor, | ||
const TensorInfo & | weights, | ||
const Optional< TensorInfo > & | biases | ||
) |
Definition at line 24 of file GpuFsaConvolution2d.cpp.
Referenced by GpuFsaLayerSupport::IsLayerSupported().
void GpuFsaDepthwiseConvolution2dCreateOp | ( | GpuFsaPreCompiledBlob * | blob, |
const TensorInfo & | input, | ||
const DepthwiseConvolution2dDescriptor & | descriptor, | ||
const TensorInfo & | weights, | ||
const Optional< TensorInfo > & | biases | ||
) |
Definition at line 89 of file GpuFsaDepthwiseConvolution2d.cpp.
References GpuFsaPreCompiledBlob::sketch, and GpuFsaPreCompiledBlob::workloadContext.
Referenced by GpuFsaBackend::OptimizeSubgraphView().
arm_compute::Status GpuFsaDepthwiseConvolution2dValidate | ( | const TensorInfo & | input, |
const DepthwiseConvolution2dDescriptor & | descriptor, | ||
const TensorInfo & | weights, | ||
const Optional< TensorInfo > & | biases | ||
) |
Definition at line 26 of file GpuFsaDepthwiseConvolution2d.cpp.
Referenced by GpuFsaLayerSupport::IsLayerSupported().
void GpuFsaElementwiseBinaryCreateOp | ( | GpuFsaPreCompiledBlob * | blob, |
const TensorInfo & | input0, | ||
const TensorInfo & | input1, | ||
const ElementwiseBinaryDescriptor & | descriptor | ||
) |
Definition at line 63 of file GpuFsaElementwiseBinary.cpp.
References GpuFsaPreCompiledBlob::sketch, and GpuFsaPreCompiledBlob::workloadContext.
Referenced by GpuFsaBackend::OptimizeSubgraphView().
arm_compute::Status GpuFsaElementwiseBinaryValidate | ( | const TensorInfo & | input0, |
const TensorInfo & | input1, | ||
const ElementwiseBinaryDescriptor & | descriptor | ||
) |
Definition at line 24 of file GpuFsaElementwiseBinary.cpp.
Referenced by GpuFsaLayerSupport::IsLayerSupported().
void GpuFsaPooling2dCreateOp | ( | GpuFsaPreCompiledBlob * | blob, |
const TensorInfo & | input, | ||
const Pooling2dDescriptor & | descriptor | ||
) |
Definition at line 40 of file GpuFsaPooling2d.cpp.
References GpuFsaPreCompiledBlob::sketch, and GpuFsaPreCompiledBlob::workloadContext.
Referenced by GpuFsaBackend::OptimizeSubgraphView().
arm_compute::Status GpuFsaPooling2dValidate | ( | const TensorInfo & | input, |
const Pooling2dDescriptor & | descriptor | ||
) |
Definition at line 22 of file GpuFsaPooling2d.cpp.
Referenced by GpuFsaLayerSupport::IsLayerSupported().
bool armnn::GpuFsaPreCompiledWorkloadValidate | ( | std::string * | reasonIfUnsupported | ) |
void GpuFsaReshapeCreateOp | ( | GpuFsaPreCompiledBlob * | blob, |
const TensorInfo & | input, | ||
const ReshapeDescriptor & | descriptor | ||
) |
Definition at line 49 of file GpuFsaReshape.cpp.
References GpuFsaPreCompiledBlob::sketch, and GpuFsaPreCompiledBlob::workloadContext.
Referenced by GpuFsaBackend::OptimizeSubgraphView().
arm_compute::Status GpuFsaReshapeValidate | ( | const TensorInfo & | input, |
const ReshapeDescriptor & | descriptor | ||
) |
Definition at line 22 of file GpuFsaReshape.cpp.
Referenced by GpuFsaLayerSupport::IsLayerSupported().
void GpuFsaResizeCreateOp | ( | GpuFsaPreCompiledBlob * | blob, |
const TensorInfo & | input, | ||
const ResizeDescriptor & | descriptor | ||
) |
Definition at line 39 of file GpuFsaResize.cpp.
References GpuFsaPreCompiledBlob::sketch, and GpuFsaPreCompiledBlob::workloadContext.
Referenced by GpuFsaBackend::OptimizeSubgraphView().
arm_compute::Status GpuFsaResizeValidate | ( | const TensorInfo & | input, |
const ResizeDescriptor & | descriptor | ||
) |
Definition at line 22 of file GpuFsaResize.cpp.
Referenced by GpuFsaLayerSupport::IsLayerSupported().
void GpuFsaSoftmaxCreateOp | ( | GpuFsaPreCompiledBlob * | blob, |
const TensorInfo & | input, | ||
const TensorInfo & | output, | ||
const SoftmaxDescriptor & | descriptor | ||
) |
Definition at line 63 of file GpuFsaSoftmax.cpp.
References GpuFsaPreCompiledBlob::sketch, and GpuFsaPreCompiledBlob::workloadContext.
Referenced by GpuFsaBackend::OptimizeSubgraphView().
arm_compute::Status GpuFsaSoftmaxValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const SoftmaxDescriptor & | descriptor | ||
) |
Definition at line 22 of file GpuFsaSoftmax.cpp.
Referenced by GpuFsaLayerSupport::IsLayerSupported().
|
constexpr |
Definition at line 14 of file GpuFsaTensorHandleFactory.hpp.
Referenced by GpuFsaTensorHandleFactory::GetIdStatic().
bool HasCapability | ( | const BackendOptions::BackendOption & | backendOption, |
const armnn::BackendId & | backend | ||
) |
Convenience function to check if a given capability matches a capability in a backend.
Definition at line 80 of file BackendHelper.cpp.
References HasMatchingCapability().
bool HasCapability | ( | const BackendOptions::BackendOption & | capability, |
const BackendCapabilities & | capabilities | ||
) |
Convenience function to check if a given capability matches a capability in a BackendCapabilities struct.
Definition at line 75 of file BackendHelper.cpp.
References HasMatchingCapability().
bool HasCapability | ( | const std::string & | name, |
const armnn::BackendId & | backend | ||
) |
Convenience function to check if a capability exists in a backend.
Definition at line 70 of file BackendHelper.cpp.
References GetCapability().
bool HasCapability | ( | const std::string & | name, |
const BackendCapabilities & | capabilities | ||
) |
Convenience function to check if a capability exists in a BackendCapabilites struct.
Definition at line 65 of file BackendHelper.cpp.
References GetCapability().
bool HasMatchingCapability | ( | const BackendOptions::BackendOption & | backendOption, |
const armnn::BackendId & | backend | ||
) |
Convenience function to check if a given capability matches a capability in a backend.
Definition at line 117 of file BackendHelper.cpp.
References BackendRegistryInstance(), and HasMatchingCapability().
bool HasMatchingCapability | ( | const BackendOptions::BackendOption & | capability, |
const BackendCapabilities & | capabilities | ||
) |
Convenience function to check if a given capability matches a capability in a BackendCapabilities struct.
Definition at line 85 of file BackendHelper.cpp.
References BackendOptions::Var::AsBool(), BackendOptions::Var::AsFloat(), BackendOptions::Var::AsInt(), BackendOptions::Var::AsString(), BackendOptions::Var::AsUnsignedInt(), BackendOptions::BackendOption::GetName(), BackendOptions::GetOption(), BackendOptions::GetOptionCount(), BackendOptions::BackendOption::GetValue(), BackendOptions::Var::IsBool(), BackendOptions::Var::IsFloat(), BackendOptions::Var::IsInt(), BackendOptions::Var::IsString(), and BackendOptions::Var::IsUnsignedInt().
Referenced by ArmnnDevice::ArmnnDevice(), CheckFp16Support(), HasCapability(), HasMatchingCapability(), LoadedNetwork::ImportInputs(), LoadedNetwork::ImportOutputs(), and RuntimeImpl::RuntimeImpl().
|
inline |
Definition at line 14 of file IgnoreUnused.hpp.
Referenced by ClBackendDefaultAllocator::allocate(), DefaultAllocator::allocate(), GpuFsaBackendDefaultAllocator::allocate(), ArgMinMax(), CalculateSlotOptionForOutput(), ITensorHandle::CanBeImported(), ClTensorHandle::CanBeImported(), ClConvolution3dWorkload::ClConvolution3dWorkload(), SpaceToDepthLayer::Clone(), SpaceToBatchNdLayer::Clone(), DynamicBackendUtils::CloseHandle(), ConstTensorPin::ConstTensorPin(), IBackendInternal::CreateExecutionData(), RefTensorHandleFactory::CreateSubTensorHandle(), TosaRefTensorHandleFactory::CreateSubTensorHandle(), SampleDynamicTensorHandleFactory::CreateSubTensorHandle(), SampleDynamicWorkloadFactory::CreateSubTensorHandle(), TosaRefWorkloadFactory::CreateSubTensorHandle(), RefWorkloadFactory::CreateSubTensorHandle(), RefTensorHandleFactory::CreateTensorHandle(), TosaRefTensorHandleFactory::CreateTensorHandle(), SampleDynamicTensorHandleFactory::CreateTensorHandle(), MockTensorHandleFactory::CreateTensorHandle(), ClWorkloadFactory::CreateTensorHandle(), TosaRefWorkloadFactory::CreateTensorHandle(), RefWorkloadFactory::CreateTensorHandle(), ITensorHandleFactory::CreateTensorHandle(), OutputLayer::CreateTensorHandles(), OutputLayer::CreateWorkload(), MergeLayer::CreateWorkload(), UnmapLayer::CreateWorkload(), InputLayer::CreateWorkload(), MapLayer::CreateWorkload(), MemCopyLayer::CreateWorkload(), MemImportLayer::CreateWorkload(), StandInLayer::CreateWorkload(), IBackendInternal::CreateWorkloadFactory(), ITensorHandle::DecorateTensorHandle(), SelectiveQuantizer< T, false >::Dequantize(), SelectiveQuantizer< armnn::Half, false >::Dequantize(), SelectiveQuantizer< armnn::BFloat16, false >::Dequantize(), SerializerStrategy::ExecuteStrategy(), UnmapLayer::ExecuteStrategy(), MapLayer::ExecuteStrategy(), MemImportLayer::ExecuteStrategy(), FakeQuantizationLayer::ExecuteStrategy(), StrategyBase< DefaultStrategy >::ExecuteStrategy(), ExecutionFrame::ExecuteWorkloads(), FalseFunc(), FalseFuncF16(), FalseFuncF32(), FalseFuncI32(), FalseFuncU8(), FalseInputFuncF16(), FalseInputFuncF32(), FalseOutputFuncF16(), FalseOutputFuncF32(), Gather(), ClImportTensorHandleFactory::GetCapabilities(), NeonTensorHandleFactory::GetCapabilities(), ITensorHandleFactory::GetCapabilities(), DynamicBackendUtils::GetEntryPoint(), DefaultAllocator::GetMemoryRegionAtOffset(), ClBackendDefaultAllocator::GetMemoryRegionAtOffset(), GpuFsaBackendDefaultAllocator::GetMemoryRegionAtOffset(), ICustomAllocator::GetMemoryRegionAtOffset(), IDeserializer::DeserializerImpl::GetNetworkInputBindingInfo(), IDeserializer::DeserializerImpl::GetNetworkOutputBindingInfo(), IDeserializer::DeserializerImpl::GetNormalizationDescriptor(), IDeserializer::DeserializerImpl::GetPooling2dDescriptor(), IDeserializer::DeserializerImpl::GetPooling3dDescriptor(), DynamicBackendUtils::GetSharedObjects(), ITensorHandle::Import(), ClTensorHandle::Import(), SliceLayer::InferOutputShapes(), StandInLayer::InferOutputShapes(), ReshapeLayer::InferOutputShapes(), NeonLayerSupport::IsActivationSupported(), RefLayerSupport::IsArgMinMaxSupported(), RefLayerSupport::IsBatchMatMulSupported(), RefLayerSupport::IsBatchNormalizationSupported(), RefLayerSupport::IsBatchToSpaceNdSupported(), RefLayerSupport::IsBroadcastToSupported(), RefLayerSupport::IsChannelShuffleSupported(), RefLayerSupport::IsComparisonSupported(), RefLayerSupport::IsConcatSupported(), RefLayerSupport::IsConvolution2dSupported(), RefLayerSupport::IsConvolution3dSupported(), RefLayerSupport::IsDepthToSpaceSupported(), RefLayerSupport::IsDepthwiseConvolutionSupported(), RefLayerSupport::IsDetectionPostProcessSupported(), RefLayerSupport::IsElementwiseUnarySupported(), RefLayerSupport::IsFakeQuantizationSupported(), NeonLayerSupport::IsFillSupported(), ClLayerSupport::IsFillSupported(), RefLayerSupport::IsFillSupported(), NeonLayerSupport::IsFloorSupported(), RefLayerSupport::IsFloorSupported(), RefLayerSupport::IsGatherSupported(), IsGpuFsaBackendSupported(), RefLayerSupport::IsInstanceNormalizationSupported(), RefLayerSupport::IsL2NormalizationSupported(), TosaRefLayerSupport::IsLayerSupported(), GpuFsaLayerSupport::IsLayerSupported(), ILayerSupport::IsLayerSupported(), ClLayerSupport::IsLogicalBinarySupported(), RefLayerSupport::IsLogicalBinarySupported(), RefLayerSupport::IsLogSoftmaxSupported(), RefLayerSupport::IsLstmSupported(), RefLayerSupport::IsNormalizationSupported(), RefLayerSupport::IsPadSupported(), RefLayerSupport::IsPermuteSupported(), RefLayerSupport::IsPooling2dSupported(), RefLayerSupport::IsPooling3dSupported(), RefLayerSupport::IsQLstmSupported(), RefLayerSupport::IsRankSupported(), RefLayerSupport::IsReduceSupported(), ClLayerSupport::IsReshapeSupported(), NeonLayerSupport::IsReshapeSupported(), RefLayerSupport::IsReshapeSupported(), RefLayerSupport::IsResizeSupported(), RefLayerSupport::IsScatterNdSupported(), RefLayerSupport::IsShapeSupported(), RefLayerSupport::IsSliceSupported(), RefLayerSupport::IsSoftmaxSupported(), RefLayerSupport::IsSpaceToBatchNdSupported(), RefLayerSupport::IsSpaceToDepthSupported(), NeonLayerSupport::IsSplitterSupported(), ClLayerSupport::IsSplitterSupported(), RefLayerSupport::IsSplitterSupported(), RefLayerSupport::IsStackSupported(), RefLayerSupport::IsStridedSliceSupported(), RefLayerSupport::IsTileSupported(), RefLayerSupport::IsTransposeConvolution2dSupported(), RefLayerSupport::IsTransposeSupported(), RefLayerSupport::IsUnidirectionalSequenceLstmSupported(), Layer::Layer(), ClImportTensorHandle::Map(), ClBackend::ClBackendCustomAllocatorMemoryRegion::map(), GpuFsaBackend::ClBackendCustomAllocatorMemoryRegion::map(), ClImportSubTensorHandle::Map(), NeonConvertFp16ToFp32WorkloadValidate(), NeonConvertFp32ToFp16WorkloadValidate(), NeonConvolution3dWorkload::NeonConvolution3dWorkload(), DynamicBackendUtils::OpenHandle(), SelectiveQuantizer< T, false >::Quantize(), SelectiveQuantizer< armnn::Half, false >::Quantize(), SelectiveQuantizer< armnn::BFloat16, false >::Quantize(), BaseWorkload< SplitterQueueDescriptor >::ReplaceInputTensorHandle(), BaseWorkload< SplitterQueueDescriptor >::ReplaceOutputTensorHandle(), OptimizeInverseConversionsImpl::Run(), OptimizeInversePermutesImpl< PermuteType >::Run(), SquashEqualSiblingsImpl< Comparable >::Run(), FuseBatchNorm< ConvLayer, ArmnnType, T >::Run(), ConvertConstants< Converter, Predicate >::Run(), ClImportTensorHandle::SetMemoryGroup(), ClImportSubTensorHandle::SetMemoryGroup(), OpenClTimer::Start(), MemoryManager::StoreMemToAllocate(), TrueFunc(), ClBackend::ClBackendCustomAllocatorMemoryRegion::unmap(), GpuFsaBackend::ClBackendCustomAllocatorMemoryRegion::unmap(), IBackendInternal::UpdateExecutionData(), ClBackend::UseCustomMemoryAllocator(), IBackendInternal::UseCustomMemoryAllocator(), WorkingMemHandle::WorkingMemHandle(), Graph::LayerInGraph< InputLayer >::~LayerInGraph(), and Graph::LayerInGraph< OutputLayer >::~LayerInGraph().
std::vector<uint32_t> armnn::IndexToCoordinates | ( | std::vector< uint32_t > & | shape, |
uint32_t | index | ||
) |
|
inline |
Definition at line 124 of file ClWorkloadUtils.hpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE.
|
inline |
Definition at line 104 of file NeonWorkloadUtils.hpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, BFloat16, CopyArmComputeTensorData(), Float16, Float32, ConstTensorHandle::GetConstTensor(), TensorInfo::GetDataType(), ConstTensorHandle::GetTensorInfo(), QAsymmS8, QAsymmU8, QSymmS16, QSymmS8, and Signed32.
|
inline |
Definition at line 69 of file NeonWorkloadUtils.hpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, BFloat16, CopyArmComputeTensorData(), Float16, Float32, TensorInfo::GetDataType(), ITensorHandle::Map(), QAsymmS8, QAsymmU8, QSymmS16, QSymmS8, and Signed32.
Referenced by NeonFullyConnectedWorkload::Execute(), and NeonConvolution2dWorkload::Execute().
std::vector< ConvertFp16ToFp32Layer * > InsertConvertFp16ToFp32LayersBefore | ( | Graph & | graph, |
Layer & | layer, | ||
bool | expectCorrectInputType | ||
) |
Definition at line 40 of file NetworkUtils.cpp.
References Layer::BeginInputSlots(), Layer::EndInputSlots(), Float16, Float32, InputSlot::GetConnectedOutputSlot(), TensorInfo::GetDataType(), Layer::GetInputSlot(), Layer::GetName(), Layer::GetNumInputSlots(), Layer::GetOutputSlot(), OutputSlot::GetTensorInfo(), Graph::InsertNewLayer(), TensorInfo::SetDataType(), and OutputSlot::SetTensorInfo().
Referenced by AttemptBackendAssignment(), and ConvertFp32NetworkToFp16Impl::Run().
std::vector< ConvertFp32ToFp16Layer * > InsertConvertFp32ToFp16LayersAfter | ( | Graph & | graph, |
Layer & | layer | ||
) |
Definition at line 79 of file NetworkUtils.cpp.
References Float16, Float32, InputSlot::GetConnectedOutputSlot(), TensorInfo::GetDataType(), Layer::GetInputSlot(), Layer::GetName(), Layer::GetNumOutputSlots(), Layer::GetOutputSlot(), OutputSlot::GetTensorInfo(), Graph::InsertNewLayer(), TensorInfo::SetDataType(), and OutputSlot::SetTensorInfo().
Referenced by AttemptBackendAssignment(), and ConvertFp32NetworkToFp16Impl::Run().
std::vector< DebugLayer * > InsertDebugLayerAfter | ( | Graph & | graph, |
Layer & | layer, | ||
bool | toFile | ||
) |
Definition at line 112 of file NetworkUtils.cpp.
References Layer::BeginOutputSlots(), CpuRef, Layer::EndOutputSlots(), InputSlot::GetConnectedOutputSlot(), Layer::GetInputSlot(), Layer::GetNameStr(), Layer::GetNumOutputSlots(), Layer::GetOutputSlot(), OutputSlot::GetTensorInfo(), Graph::InsertNewLayer(), Layer::SetBackendId(), and OutputSlot::SetTensorInfo().
Referenced by AddDebugImpl::Run(), and AddDebugToFileImpl::Run().
void InstanceNorm | ( | const InstanceNormalizationQueueDescriptor & | data, |
const TensorInfo & | inputInfo, | ||
Decoder< float > & | inputDecoder, | ||
Encoder< float > & | outputEncoder | ||
) |
Definition at line 18 of file InstanceNorm.cpp.
References Decoder< IType >::Get(), DataLayoutIndexed::GetChannelsIndex(), DataLayoutIndexed::GetHeightIndex(), DataLayoutIndexed::GetIndex(), TensorInfo::GetShape(), DataLayoutIndexed::GetWidthIndex(), InstanceNormalizationDescriptor::m_Beta, InstanceNormalizationDescriptor::m_DataLayout, InstanceNormalizationDescriptor::m_Eps, InstanceNormalizationDescriptor::m_Gamma, QueueDescriptorWithParameters< LayerDescriptor >::m_Parameters, and Encoder< IType >::Set().
float IntersectionOverUnion | ( | const float * | boxI, |
const float * | boxJ | ||
) |
bool armnn::IsBFloat16 | ( | const WorkloadInfo & | info | ) |
Definition at line 56 of file RefWorkloadFactory.cpp.
References info.
Referenced by RefWorkloadFactory::CreateWorkload().
bool armnn::IsBoolean | ( | const WorkloadInfo & | info | ) |
Definition at line 80 of file RefWorkloadFactory.cpp.
References info.
Referenced by RefWorkloadFactory::CreateWorkload().
bool IsDataType | ( | const WorkloadInfo & | info | ) |
Definition at line 32 of file GpuFsaWorkloadFactory.cpp.
References info.
bool armnn::IsFloat16 | ( | const WorkloadInfo & | info | ) |
Definition at line 60 of file RefWorkloadFactory.cpp.
References info.
Referenced by RefWorkloadFactory::CreateWorkload().
bool armnn::IsGpuFsaBackendSupported | ( | Optional< std::string & > | reasonIfUnsupported, |
Args... | args | ||
) |
Definition at line 31 of file GpuFsaLayerSupport.cpp.
References IgnoreUnused(), and OptionalReferenceSwitch< std::is_reference< T >::value, T >::value().
Referenced by GpuFsaLayerSupport::IsLayerSupported().
bool armnn::IsLayerOptimizable | ( | const armnn::Layer & | layer | ) |
Definition at line 99 of file MockBackend.cpp.
bool armnn::IsLayerOptimizable | ( | const armnn::Layer * | layer | ) |
bool armnn::IsLayerSequence | ( | Layer & | currentLayer, |
TYPE | first, | ||
TYPE | second, | ||
TYPE | third, | ||
Layer * | layerList[4], | ||
bool | handleValidActivates, | ||
const std::vector< ActivationFunction > & | validActivates | ||
) |
Definition at line 375 of file SubgraphUtils.hpp.
bool armnn::IsLayerSupported | ( | const armnn::Layer & | layer | ) |
Definition at line 83 of file MockBackend.cpp.
bool armnn::IsLayerSupported | ( | const armnn::Layer * | layer | ) |
Definition at line 62 of file MockBackend.cpp.
References Addition, ARMNN_ASSERT, Constant, Convolution2d, ElementwiseBinary, Layer::GetType(), Input, and Output.
Referenced by SampleDynamicWorkloadFactory::IsLayerSupported().
bool armnn::IsLayerTypeSupported | ( | const LayerType & | type, |
const std::vector< TensorInfo > & | infos, | ||
const BaseDescriptor & | descriptor, | ||
const Optional< LstmInputParamsInfo > & | lstmParamsInfo, | ||
const Optional< QuantizedLstmInputParamsInfo > & | quantizedLstmParamsInfo, | ||
Optional< std::string & > | reasonIfUnsupported, | ||
const NeonLayerSupport & | support | ||
) |
Definition at line 184 of file NeonLayerSupport.cpp.
References Activation, Add, Addition, ArgMinMax, BatchMatMul, BatchNormalization, BatchToSpaceNd, Cast, ChannelShuffle, Comparison, Concat, Constant, ConvertFp16ToFp32, ConvertFp32ToFp16, Convolution2d, Convolution3d, DepthToSpace, DepthwiseConvolution2d, Dequantize, DetectionPostProcess, Div, Division, ElementwiseBinary, ElementwiseUnary, Fill, Floor, FORWARD_WORKLOAD_VALIDATE_FUNC, FullyConnected, Fused, Gather, GatherNd, Input, InstanceNormalization, NeonLayerSupport::IsActivationSupported(), NeonLayerSupport::IsAdditionSupported(), NeonLayerSupport::IsArgMinMaxSupported(), NeonLayerSupport::IsBatchMatMulSupported(), NeonLayerSupport::IsBatchNormalizationSupported(), NeonLayerSupport::IsBatchToSpaceNdSupported(), NeonLayerSupport::IsCastSupported(), NeonLayerSupport::IsChannelShuffleSupported(), NeonLayerSupport::IsComparisonSupported(), NeonLayerSupport::IsConcatSupported(), NeonLayerSupport::IsConstantSupported(), NeonLayerSupport::IsConvertFp16ToFp32Supported(), NeonLayerSupport::IsConvertFp32ToFp16Supported(), NeonLayerSupport::IsConvolution2dSupported(), NeonLayerSupport::IsConvolution3dSupported(), NeonLayerSupport::IsDepthToSpaceSupported(), NeonLayerSupport::IsDepthwiseConvolutionSupported(), NeonLayerSupport::IsDequantizeSupported(), LayerSupportBase::IsDetectionPostProcessSupported(), NeonLayerSupport::IsDivisionSupported(), NeonLayerSupport::IsElementwiseUnarySupported(), NeonLayerSupport::IsFillSupported(), NeonLayerSupport::IsFloorSupported(), NeonLayerSupport::IsFullyConnectedSupported(), NeonLayerSupport::IsFusedSupported(), NeonLayerSupport::IsGatherNdSupported(), NeonLayerSupport::IsGatherSupported(), NeonLayerSupport::IsInputSupported(), NeonLayerSupport::IsInstanceNormalizationSupported(), NeonLayerSupport::IsL2NormalizationSupported(), NeonLayerSupport::IsLogicalBinarySupported(), NeonLayerSupport::IsLogSoftmaxSupported(), NeonLayerSupport::IsLstmSupported(), NeonLayerSupport::IsMaximumSupported(), NeonLayerSupport::IsMeanSupported(), LayerSupportBase::IsMemCopySupported(), LayerSupportBase::IsMemImportSupported(), LayerSupportBase::IsMergeSupported(), NeonLayerSupport::IsMinimumSupported(), NeonLayerSupport::IsMultiplicationSupported(), NeonLayerSupport::IsNormalizationSupported(), NeonLayerSupport::IsOutputSupported(), NeonLayerSupport::IsPadSupported(), NeonLayerSupport::IsPermuteSupported(), NeonLayerSupport::IsPooling2dSupported(), NeonLayerSupport::IsPooling3dSupported(), NeonLayerSupport::IsPreluSupported(), NeonLayerSupport::IsQLstmSupported(), NeonLayerSupport::IsQuantizedLstmSupported(), NeonLayerSupport::IsQuantizeSupported(), NeonLayerSupport::IsReduceSupported(), NeonLayerSupport::IsReshapeSupported(), NeonLayerSupport::IsResizeSupported(), NeonLayerSupport::IsReverseV2Supported(), LayerSupportBase::IsShapeSupported(), NeonLayerSupport::IsSliceSupported(), NeonLayerSupport::IsSoftmaxSupported(), NeonLayerSupport::IsSpaceToBatchNdSupported(), NeonLayerSupport::IsSpaceToDepthSupported(), NeonLayerSupport::IsSplitterSupported(), NeonLayerSupport::IsStackSupported(), NeonLayerSupport::IsStridedSliceSupported(), NeonLayerSupport::IsSubtractionSupported(), NeonLayerSupport::IsTileSupported(), NeonLayerSupport::IsTransposeConvolution2dSupported(), NeonLayerSupport::IsTransposeSupported(), NeonLayerSupport::IsUnidirectionalSequenceLstmSupported(), L2Normalization, LogicalBinary, LogSoftmax, Lstm, Map, Maximum, Mean, MemCopy, MemImport, Merge, Minimum, Mul, Multiplication, NeonAdditionWorkloadValidate(), NeonDivisionWorkloadValidate(), NeonElementwiseBinaryWorkloadValidate(), NeonMaximumWorkloadValidate(), NeonMinimumWorkloadValidate(), NeonMultiplicationWorkloadValidate(), NeonSubtractionWorkloadValidate(), Normalization, Output, Pad, Permute, PolymorphicDowncast(), Pooling2d, Pooling3d, Power, Prelu, QLstm, Quantize, QuantizedLstm, Rank, Reduce, Reshape, Resize, ReverseV2, Shape, Slice, Softmax, SpaceToBatchNd, SpaceToDepth, Splitter, SqDiff, Stack, StridedSlice, Sub, Subtraction, Tile, Transpose, TransposeConvolution2d, UnidirectionalSequenceLstm, Unmap, and OptionalReferenceSwitch< std::is_reference< T >::value, T >::value().
Referenced by NeonLayerSupport::IsLayerSupported().
|
inline |
Definition at line 213 of file SubgraphUtils.hpp.
References Layer::ExecuteStrategy().
Referenced by ConnectedToLayerWithNCHW().
|
constexpr |
Definition at line 22 of file RefWorkloadFactory.hpp.
|
constexpr |
Definition at line 20 of file RefWorkloadFactory.hpp.
|
constexpr |
Definition at line 24 of file RefWorkloadFactory.hpp.
|
constexpr |
Definition at line 18 of file RefWorkloadFactory.hpp.
bool armnn::IsQAsymmS8 | ( | const WorkloadInfo & | info | ) |
Definition at line 72 of file RefWorkloadFactory.cpp.
References info.
Referenced by RefWorkloadFactory::CreateWorkload().
bool armnn::IsQAsymmU8 | ( | const WorkloadInfo & | info | ) |
Definition at line 76 of file RefWorkloadFactory.cpp.
References info.
Referenced by RefWorkloadFactory::CreateWorkload().
bool armnn::IsQSymmS16 | ( | const WorkloadInfo & | info | ) |
Definition at line 64 of file RefWorkloadFactory.cpp.
References info.
Referenced by RefWorkloadFactory::CreateWorkload().
bool armnn::IsQSymmS8 | ( | const WorkloadInfo & | info | ) |
Definition at line 68 of file RefWorkloadFactory.cpp.
References info.
Referenced by RefWorkloadFactory::CreateWorkload().
|
constexpr |
Definition at line 316 of file TypesUtils.hpp.
References QAsymmS8, QAsymmU8, and QSymmS8.
Referenced by ConvertSoftmaxToTosaOperator(), RefLayerSupport::IsConvolution2dSupported(), RefLayerSupport::IsConvolution3dSupported(), RefLayerSupport::IsDepthwiseConvolutionSupported(), IsQuantizedType(), and RefLayerSupport::IsTransposeConvolution2dSupported().
|
constexpr |
Definition at line 311 of file TypesUtils.hpp.
Referenced by ClMultiplicationWorkload::ClMultiplicationWorkload(), RefWorkloadFactory::CreateWorkload(), TensorInfo::IsQuantized(), NeonMultiplicationWorkload::NeonMultiplicationWorkload(), and QuantizeQueueDescriptor::Validate().
|
constexpr |
Definition at line 323 of file TypesUtils.hpp.
References IsQuantized8BitType(), and QSymmS16.
bool armnn::IsReadyForSplitAssignment | ( | LayerSelectionInfo::LayerInfoContainer & | layerInfos, |
LayerSelectionInfo & | layerInfo | ||
) |
Definition at line 387 of file SubgraphViewSelector.cpp.
References ForEachLayerInput().
Referenced by SubgraphViewSelector::SelectSubgraphs().
|
inline |
Definition at line 367 of file SubgraphUtils.hpp.
References ElementwiseBinary, and Layer::GetType().
Definition at line 362 of file SubgraphUtils.hpp.
References Layer::GetType().
Referenced by BuildAddMulAddTensorInfoLists().
bool armnn::IsSigned32 | ( | const WorkloadInfo & | info | ) |
Definition at line 52 of file RefWorkloadFactory.cpp.
References info.
Referenced by RefWorkloadFactory::CreateWorkload().
bool armnn::IsSigned64 | ( | const WorkloadInfo & | info | ) |
Definition at line 48 of file RefWorkloadFactory.cpp.
References info.
Referenced by RefWorkloadFactory::CreateWorkload().
bool armnn::IsSupportedForDataTypeGeneric | ( | Optional< std::string & > | reasonIfUnsupported, |
DataType | dataType, | ||
Float16Func | float16FuncPtr, | ||
Float32Func | float32FuncPtr, | ||
Uint8Func | uint8FuncPtr, | ||
Int32Func | int32FuncPtr, | ||
BooleanFunc | booleanFuncPtr, | ||
Params &&... | params | ||
) |
Definition at line 27 of file LayerSupportCommon.hpp.
References Boolean, Float16, Float32, QAsymmU8, and Signed32.
Referenced by RefLayerSupport::IsConvertFp16ToFp32Supported(), RefLayerSupport::IsConvertFp32ToFp16Supported(), and NeonLayerSupport::IsFloorSupported().
|
constexpr |
Definition at line 115 of file LayersFwd.hpp.
|
constexpr |
Definition at line 116 of file LayersFwd.hpp.
|
constexpr |
Definition at line 117 of file LayersFwd.hpp.
|
constexpr |
Definition at line 118 of file LayersFwd.hpp.
|
constexpr |
Definition at line 119 of file LayersFwd.hpp.
|
constexpr |
Definition at line 120 of file LayersFwd.hpp.
|
constexpr |
Definition at line 121 of file LayersFwd.hpp.
Definition at line 122 of file LayersFwd.hpp.
|
constexpr |
Definition at line 123 of file LayersFwd.hpp.
|
constexpr |
Definition at line 124 of file LayersFwd.hpp.
|
constexpr |
Definition at line 125 of file LayersFwd.hpp.
|
constexpr |
Definition at line 126 of file LayersFwd.hpp.
|
constexpr |
Definition at line 127 of file LayersFwd.hpp.
|
constexpr |
Definition at line 128 of file LayersFwd.hpp.
|
constexpr |
Definition at line 129 of file LayersFwd.hpp.
|
constexpr |
Definition at line 130 of file LayersFwd.hpp.
|
constexpr |
Definition at line 131 of file LayersFwd.hpp.
|
constexpr |
Definition at line 132 of file LayersFwd.hpp.
|
constexpr |
Definition at line 133 of file LayersFwd.hpp.
|
constexpr |
Definition at line 134 of file LayersFwd.hpp.
|
constexpr |
Definition at line 135 of file LayersFwd.hpp.
|
constexpr |
Definition at line 136 of file LayersFwd.hpp.
|
constexpr |
Definition at line 137 of file LayersFwd.hpp.
|
constexpr |
Definition at line 138 of file LayersFwd.hpp.
|
constexpr |
Definition at line 139 of file LayersFwd.hpp.
Definition at line 140 of file LayersFwd.hpp.
|
constexpr |
Definition at line 141 of file LayersFwd.hpp.
|
constexpr |
Definition at line 142 of file LayersFwd.hpp.
|
constexpr |
Definition at line 143 of file LayersFwd.hpp.
|
constexpr |
Definition at line 144 of file LayersFwd.hpp.
|
constexpr |
Definition at line 145 of file LayersFwd.hpp.
|
constexpr |
Definition at line 146 of file LayersFwd.hpp.
|
constexpr |
Definition at line 147 of file LayersFwd.hpp.
|
constexpr |
Definition at line 148 of file LayersFwd.hpp.
|
constexpr |
Definition at line 149 of file LayersFwd.hpp.
|
constexpr |
Definition at line 150 of file LayersFwd.hpp.
Definition at line 151 of file LayersFwd.hpp.
Definition at line 152 of file LayersFwd.hpp.
|
constexpr |
Definition at line 153 of file LayersFwd.hpp.
Definition at line 154 of file LayersFwd.hpp.
|
constexpr |
Definition at line 155 of file LayersFwd.hpp.
|
constexpr |
Definition at line 156 of file LayersFwd.hpp.
|
constexpr |
Definition at line 157 of file LayersFwd.hpp.
|
constexpr |
Definition at line 158 of file LayersFwd.hpp.
|
constexpr |
Definition at line 159 of file LayersFwd.hpp.
|
constexpr |
Definition at line 160 of file LayersFwd.hpp.
|
constexpr |
Definition at line 161 of file LayersFwd.hpp.
Definition at line 162 of file LayersFwd.hpp.
|
constexpr |
Definition at line 163 of file LayersFwd.hpp.
|
constexpr |
Definition at line 164 of file LayersFwd.hpp.
|
constexpr |
Definition at line 165 of file LayersFwd.hpp.
|
constexpr |
Definition at line 166 of file LayersFwd.hpp.
|
constexpr |
Definition at line 167 of file LayersFwd.hpp.
|
constexpr |
Definition at line 169 of file LayersFwd.hpp.
|
constexpr |
Definition at line 170 of file LayersFwd.hpp.
|
constexpr |
Definition at line 168 of file LayersFwd.hpp.
Definition at line 171 of file LayersFwd.hpp.
|
constexpr |
Definition at line 172 of file LayersFwd.hpp.
|
constexpr |
Definition at line 173 of file LayersFwd.hpp.
|
constexpr |
Definition at line 174 of file LayersFwd.hpp.
|
constexpr |
Definition at line 175 of file LayersFwd.hpp.
|
constexpr |
Definition at line 176 of file LayersFwd.hpp.
|
constexpr |
Definition at line 177 of file LayersFwd.hpp.
|
constexpr |
Definition at line 178 of file LayersFwd.hpp.
|
constexpr |
Definition at line 179 of file LayersFwd.hpp.
|
constexpr |
Definition at line 180 of file LayersFwd.hpp.
|
constexpr |
Definition at line 181 of file LayersFwd.hpp.
|
constexpr |
Definition at line 182 of file LayersFwd.hpp.
|
constexpr |
Definition at line 183 of file LayersFwd.hpp.
|
constexpr |
Definition at line 184 of file LayersFwd.hpp.
|
constexpr |
Definition at line 185 of file LayersFwd.hpp.
|
constexpr |
Definition at line 186 of file LayersFwd.hpp.
|
constexpr |
Definition at line 187 of file LayersFwd.hpp.
|
constexpr |
Definition at line 188 of file LayersFwd.hpp.
|
constexpr |
Definition at line 190 of file LayersFwd.hpp.
|
constexpr |
Definition at line 189 of file LayersFwd.hpp.
|
constexpr |
Definition at line 191 of file LayersFwd.hpp.
|
constexpr |
Definition at line 192 of file LayersFwd.hpp.
|
inline |
Definition at line 22 of file Logging.hpp.
References Debug, Error, Fatal, Info, Trace, and Warning.
Referenced by ScopedRecord::ScopedRecord().
void LogSoftmax | ( | Decoder< float > & | input, |
Encoder< float > & | output, | ||
const TensorInfo & | inputInfo, | ||
const LogSoftmaxDescriptor & | descriptor | ||
) |
Definition at line 27 of file LogSoftmax.cpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, Decoder< IType >::Get(), TensorShape::GetNumDimensions(), TensorInfo::GetNumDimensions(), armnnUtils::GetNumElementsBetween(), TensorInfo::GetShape(), SoftmaxDescriptor::m_Axis, SoftmaxDescriptor::m_Beta, and Encoder< IType >::Set().
void LstmImpl | ( | const LstmDescriptor & | descriptor, |
const TensorInfo & | inputInfo, | ||
const TensorInfo & | outputInfo, | ||
const TensorShape & | inputToOutputWeightsShape, | ||
const TensorShape & | recurrentToOutputWeightsShape, | ||
std::unique_ptr< Decoder< float >> & | inputData, | ||
std::unique_ptr< Decoder< float >> & | outputStateIn, | ||
std::unique_ptr< Decoder< float >> & | cellStateIn, | ||
std::unique_ptr< Encoder< float >> & | outputStateOut, | ||
std::unique_ptr< Encoder< float >> & | cellStateOut, | ||
std::unique_ptr< Encoder< float >> & | output, | ||
std::unique_ptr< Decoder< float >> & | cellStateOutDecoder, | ||
std::unique_ptr< Decoder< float >> & | outputDecoder, | ||
std::unique_ptr< Decoder< float >> & | inputToInputWeightsTensor, | ||
std::unique_ptr< Decoder< float >> & | inputToForgetWeightsTensor, | ||
std::unique_ptr< Decoder< float >> & | inputToCellWeightsTensor, | ||
std::unique_ptr< Decoder< float >> & | inputToOutputWeightsTensor, | ||
std::unique_ptr< Decoder< float >> & | recurrentToInputWeightsTensor, | ||
std::unique_ptr< Decoder< float >> & | recurrentToForgetWeightsTensor, | ||
std::unique_ptr< Decoder< float >> & | recurrentToCellWeightsTensor, | ||
std::unique_ptr< Decoder< float >> & | recurrentToOutputWeightsTensor, | ||
std::unique_ptr< Decoder< float >> & | cellToInputWeightsTensor, | ||
std::unique_ptr< Decoder< float >> & | cellToForgetWeightsTensor, | ||
std::unique_ptr< Decoder< float >> & | cellToOutputWeightsTensor, | ||
std::unique_ptr< Decoder< float >> & | inputGateBiasTensor, | ||
std::unique_ptr< Decoder< float >> & | forgetGateBiasTensor, | ||
std::unique_ptr< Decoder< float >> & | cellBiasTensor, | ||
std::unique_ptr< Decoder< float >> & | outputGateBiasTensor, | ||
std::unique_ptr< Decoder< float >> & | projectionWeightsTensor, | ||
std::unique_ptr< Decoder< float >> & | projectionBiasTensor, | ||
std::unique_ptr< Decoder< float >> & | inputLayerNormWeights, | ||
std::unique_ptr< Decoder< float >> & | forgetLayerNormWeights, | ||
std::unique_ptr< Decoder< float >> & | cellLayerNormWeights, | ||
std::unique_ptr< Decoder< float >> & | outputLayerNormWeights, | ||
std::unique_ptr< Encoder< float >> & | inputGateScratch, | ||
std::unique_ptr< Encoder< float >> & | cellScratch, | ||
std::unique_ptr< Encoder< float >> & | forgetGateScratch, | ||
std::unique_ptr< Encoder< float >> & | outputGateScratch, | ||
std::unique_ptr< Decoder< float >> & | inputGateScratchDecoder, | ||
std::unique_ptr< Decoder< float >> & | cellScratchDecoder, | ||
std::unique_ptr< Decoder< float >> & | forgetGateScratchDecoder, | ||
std::unique_ptr< Decoder< float >> & | outputGateScratchDecoder, | ||
float | layerNormEpsilon | ||
) |
Definition at line 13 of file Lstm.cpp.
References Activation(), ClipVector(), CopyVector(), TensorInfo::GetDataType(), TensorInfo::GetShape(), LstmDescriptor::m_ActivationFunc, LstmDescriptor::m_CifgEnabled, LstmDescriptor::m_ClippingThresCell, LstmDescriptor::m_ClippingThresProj, LstmDescriptor::m_LayerNormEnabled, LstmDescriptor::m_PeepholeEnabled, LstmDescriptor::m_ProjectionEnabled, MatrixBatchVectorMultiplyAccumulate(), MeanStddevNormalization(), SetActivationParameters(), Sigmoid, Sub1Vector(), VectorBatchVectorAdd(), VectorBatchVectorAssign(), VectorBatchVectorCwiseProduct(), VectorBatchVectorCwiseProductAccumulate(), VectorVectorCwiseProduct(), VectorVectorCwiseProductAccumulate(), and ZeroVector().
|
inline |
|
inline |
|
inline |
|
inline |
arm_compute::DetectionPostProcessLayerInfo armnn::MakeInfo | ( | const DetectionPostProcessDescriptor & | descriptor | ) |
Definition at line 17 of file NeonDetectionPostProcessWorkload.cpp.
References DetectionPostProcessDescriptor::m_DetectionsPerClass, DetectionPostProcessDescriptor::m_MaxClassesPerDetection, DetectionPostProcessDescriptor::m_MaxDetections, DetectionPostProcessDescriptor::m_NmsIouThreshold, DetectionPostProcessDescriptor::m_NmsScoreThreshold, DetectionPostProcessDescriptor::m_NumClasses, and DetectionPostProcessDescriptor::m_UseRegularNms.
Referenced by NeonDetectionPostProcessValidate().
Optimizer::Optimizations armnn::MakeOptimizations | ( | Args &&... | args | ) |
Definition at line 43 of file Optimizer.hpp.
References Append().
Referenced by ApplyBackendOptimizations(), and Optimize().
Optional<T> armnn::MakeOptional | ( | Args &&... | args | ) |
Utility template that constructs an object of type T in-place and wraps it inside an Optional<T> object.
Definition at line 305 of file Optional.hpp.
References CONSTRUCT_IN_PLACE.
|
constexpr |
Definition at line 90 of file TransformIterator.hpp.
void MirrorPad | ( | const TensorInfo & | inputInfo, |
const TensorInfo & | outputInfo, | ||
const ITensorHandle * | inputHandle, | ||
ITensorHandle * | outputHandle, | ||
const PadQueueDescriptor & | data | ||
) |
Definition at line 59 of file MirrorPad.cpp.
References Decoder< IType >::Get(), TensorShape::GetNumDimensions(), TensorInfo::GetNumElements(), TensorInfo::GetShape(), PadDescriptor::m_PaddingMode, PadDescriptor::m_PadList, QueueDescriptorWithParameters< LayerDescriptor >::m_Parameters, ITensorHandle::Map(), Reflect, Encoder< IType >::Set(), and Symmetric.
|
constexpr |
Definition at line 14 of file MockTensorHandleFactory.hpp.
Referenced by MockTensorHandleFactory::GetIdStatic().
arm_compute::Status NeonAbsWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 17 of file NeonAbsWorkload.cpp.
Referenced by NeonLayerSupport::IsElementwiseUnarySupported().
arm_compute::Status NeonActivationWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const ActivationDescriptor & | descriptor | ||
) |
Definition at line 17 of file NeonActivationWorkload.cpp.
Referenced by NeonLayerSupport::IsActivationSupported().
arm_compute::Status NeonAdditionWorkloadValidate | ( | const TensorInfo & | input0, |
const TensorInfo & | input1, | ||
const TensorInfo & | output, | ||
const ActivationDescriptor * | activationDescriptor | ||
) |
Definition at line 20 of file NeonAdditionWorkload.cpp.
Referenced by NeonLayerSupport::IsAdditionSupported(), IsLayerTypeSupported(), and NeonBackend::OptimizeSubgraphView().
arm_compute::Status NeonArgMinMaxWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const ArgMinMaxDescriptor & | descriptor | ||
) |
Definition at line 31 of file NeonArgMinMaxWorkload.cpp.
Referenced by NeonLayerSupport::IsArgMinMaxSupported().
|
constexpr |
Definition at line 10 of file NeonBackendId.hpp.
Referenced by NeonBackend::GetIdStatic().
arm_compute::Status NeonBatchMatMulValidate | ( | const TensorInfo & | inputInfoX, |
const TensorInfo & | inputInfoY, | ||
const TensorInfo & | outputInfo, | ||
const BatchMatMulDescriptor & | descriptor, | ||
const bool | isFastMathEnabled, | ||
const ActivationDescriptor * | activationDescriptor | ||
) |
Definition at line 19 of file NeonBatchMatMulWorkload.cpp.
References BatchMatMulDescriptor::m_AdjointX, BatchMatMulDescriptor::m_AdjointY, BatchMatMulDescriptor::m_DataLayoutX, BatchMatMulDescriptor::m_DataLayoutY, and NCHW.
Referenced by NeonLayerSupport::IsBatchMatMulSupported().
arm_compute::Status NeonBatchNormalizationValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const TensorInfo & | mean, | ||
const TensorInfo & | var, | ||
const TensorInfo & | beta, | ||
const TensorInfo & | gamma, | ||
const BatchNormalizationDescriptor & | descriptor, | ||
const ActivationDescriptor * | activationDescriptor | ||
) |
Definition at line 24 of file NeonBatchNormalizationWorkload.cpp.
Referenced by NeonLayerSupport::IsBatchNormalizationSupported(), and NeonBackend::OptimizeSubgraphView().
arm_compute::Status NeonBatchToSpaceNdWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const BatchToSpaceNdDescriptor & | descriptor | ||
) |
Definition at line 15 of file NeonBatchToSpaceNdWorkload.cpp.
Referenced by NeonLayerSupport::IsBatchToSpaceNdSupported().
arm_compute::Status NeonCastValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 19 of file NeonCastWorkload.cpp.
Referenced by NeonLayerSupport::IsCastSupported().
arm_compute::Status NeonChannelShuffleValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const ChannelShuffleDescriptor & | descriptor | ||
) |
Definition at line 17 of file NeonChannelShuffleWorkload.cpp.
Referenced by NeonLayerSupport::IsChannelShuffleSupported().
arm_compute::Status NeonComparisonWorkloadValidate | ( | const TensorInfo & | input0, |
const TensorInfo & | input1, | ||
const TensorInfo & | output, | ||
const ComparisonDescriptor & | descriptor | ||
) |
Definition at line 16 of file NeonComparisonWorkload.cpp.
Referenced by NeonLayerSupport::IsComparisonSupported().
arm_compute::Status NeonConcatWorkloadValidate | ( | const std::vector< const TensorInfo * > & | inputs, |
const TensorInfo & | output, | ||
const OriginsDescriptor & | descriptor | ||
) |
Definition at line 27 of file NeonConcatWorkload.cpp.
Referenced by NeonLayerSupport::IsConcatSupported().
arm_compute::Status NeonConstantWorkloadValidate | ( | const TensorInfo & | output | ) |
Definition at line 20 of file NeonConstantWorkload.cpp.
Referenced by NeonLayerSupport::IsConstantSupported().
arm_compute::Status NeonConvertFp16ToFp32WorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 19 of file NeonConvertFp16ToFp32Workload.cpp.
References IgnoreUnused().
Referenced by NeonLayerSupport::IsConvertFp16ToFp32Supported().
arm_compute::Status NeonConvertFp32ToFp16WorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 21 of file NeonConvertFp32ToFp16Workload.cpp.
References IgnoreUnused().
Referenced by NeonLayerSupport::IsConvertFp32ToFp16Supported().
arm_compute::Status NeonConvolution2dWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const Convolution2dDescriptor & | descriptor, | ||
const TensorInfo & | weights, | ||
const Optional< TensorInfo > & | biases, | ||
bool | isFastMathEnabled, | ||
const ActivationDescriptor * | activationDescriptor | ||
) |
Definition at line 24 of file NeonConvolution2dWorkload.cpp.
Referenced by NeonLayerSupport::IsConvolution2dSupported(), and NeonBackend::OptimizeSubgraphView().
arm_compute::Status NeonConvolution3dWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const Convolution3dDescriptor & | descriptor, | ||
const TensorInfo & | weights, | ||
const Optional< TensorInfo > & | biases, | ||
bool | isFastMathEnabled, | ||
const ActivationDescriptor * | activationDescriptor | ||
) |
Definition at line 24 of file NeonConvolution3dWorkload.cpp.
Referenced by NeonLayerSupport::IsConvolution3dSupported().
arm_compute::Status NeonDepthToSpaceWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const DepthToSpaceDescriptor & | descriptor | ||
) |
Definition at line 19 of file NeonDepthToSpaceWorkload.cpp.
References SpaceToDepthDescriptor::m_DataLayout.
Referenced by NeonLayerSupport::IsDepthToSpaceSupported().
arm_compute::Status NeonDepthwiseConvolutionWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const DepthwiseConvolution2dDescriptor & | descriptor, | ||
const TensorInfo & | weights, | ||
const Optional< TensorInfo > & | biases, | ||
const ActivationDescriptor * | activationDescriptor | ||
) |
Definition at line 29 of file NeonDepthwiseConvolutionWorkload.cpp.
Referenced by NeonLayerSupport::IsDepthwiseConvolutionSupported(), NeonLayerSupport::IsDilatedDepthwiseConvolutionSupported(), and NeonBackend::OptimizeSubgraphView().
arm_compute::Status NeonDequantizeWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 22 of file NeonDequantizeWorkload.cpp.
Referenced by NeonLayerSupport::IsDequantizeSupported().
bool NeonDetected | ( | ) |
arm_compute::Status NeonDetectionPostProcessValidate | ( | const TensorInfo & | boxEncodings, |
const TensorInfo & | scores, | ||
const TensorInfo & | anchors, | ||
const TensorInfo & | detectionBoxes, | ||
const TensorInfo & | detectionClasses, | ||
const TensorInfo & | detectionScores, | ||
const TensorInfo & | numDetections, | ||
const DetectionPostProcessDescriptor & | descriptor | ||
) |
Definition at line 32 of file NeonDetectionPostProcessWorkload.cpp.
References info, and MakeInfo().
arm_compute::Status NeonDivisionWorkloadValidate | ( | const TensorInfo & | input0, |
const TensorInfo & | input1, | ||
const TensorInfo & | output, | ||
const ActivationDescriptor * | activationDescriptor | ||
) |
Definition at line 18 of file NeonDivisionWorkload.cpp.
Referenced by NeonLayerSupport::IsDivisionSupported(), IsLayerTypeSupported(), and NeonBackend::OptimizeSubgraphView().
arm_compute::Status NeonElementwiseBinaryWorkloadValidate | ( | const TensorInfo & | input0, |
const TensorInfo & | input1, | ||
const TensorInfo & | output, | ||
const ElementwiseBinaryDescriptor & | descriptor, | ||
const ActivationDescriptor * | activationDescriptor | ||
) |
Definition at line 20 of file NeonElementwiseBinaryWorkload.cpp.
Referenced by IsLayerTypeSupported().
arm_compute::Status NeonExpWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 17 of file NeonExpWorkload.cpp.
Referenced by NeonLayerSupport::IsElementwiseUnarySupported().
arm_compute::Status NeonFullyConnectedWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const TensorInfo & | weights, | ||
const Optional< TensorInfo > & | biases, | ||
const FullyConnectedDescriptor & | descriptor, | ||
const ActivationDescriptor * | activationDescriptor | ||
) |
Definition at line 24 of file NeonFullyConnectedWorkload.cpp.
Referenced by NeonLayerSupport::IsFullyConnectedSupported(), and NeonBackend::OptimizeSubgraphView().
arm_compute::Status NeonFusedWorkloadValidate | ( | const std::vector< std::reference_wrapper< TensorInfo >> & | inputInfos, |
const std::vector< std::reference_wrapper< TensorInfo >> & | outputInfos, | ||
const FusedDescriptor & | fusedDescriptor, | ||
const ActivationDescriptor * | activationDescriptor | ||
) |
Definition at line 22 of file NeonFusedWorkload.cpp.
Referenced by NeonLayerSupport::IsFusedSupported(), and NeonBackend::OptimizeSubgraphView().
arm_compute::Status NeonGatherNdWorkloadValidate | ( | const TensorInfo & | paramsInfo, |
const TensorInfo & | indicesInfo, | ||
const TensorInfo & | outputInfo | ||
) |
Validate Mul
Validate ReduceSum
Validate Gather
Validate Reshape
Return OK if all the layers are valid
Definition at line 14 of file NeonGatherNdWorkload.cpp.
References CalculateGatherNdKeyIndices(), and TensorInfo::SetShape().
Referenced by NeonLayerSupport::IsGatherNdSupported().
arm_compute::Status NeonGatherWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | indices, | ||
const TensorInfo & | output, | ||
const GatherDescriptor & | descriptor | ||
) |
Definition at line 13 of file NeonGatherWorkload.cpp.
Referenced by NeonLayerSupport::IsGatherSupported().
arm_compute::Status NeonInstanceNormalizationWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const InstanceNormalizationDescriptor & | descriptor | ||
) |
Definition at line 19 of file NeonInstanceNormalizationWorkload.cpp.
Referenced by NeonLayerSupport::IsInstanceNormalizationSupported().
arm_compute::Status NeonL2NormalizationWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const L2NormalizationDescriptor & | descriptor | ||
) |
Definition at line 19 of file NeonL2NormalizationFloatWorkload.cpp.
Referenced by NeonLayerSupport::IsL2NormalizationSupported().
arm_compute::Status NeonLogicalAndWorkloadValidate | ( | const TensorInfo & | input0, |
const TensorInfo & | input1, | ||
const TensorInfo & | output | ||
) |
Definition at line 18 of file NeonLogicalAndWorkload.cpp.
Referenced by NeonLayerSupport::IsLogicalBinarySupported().
arm_compute::Status NeonLogicalNotWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 19 of file NeonLogicalNotWorkload.cpp.
Referenced by NeonLayerSupport::IsElementwiseUnarySupported().
arm_compute::Status NeonLogicalOrWorkloadValidate | ( | const TensorInfo & | input0, |
const TensorInfo & | input1, | ||
const TensorInfo & | output | ||
) |
Definition at line 18 of file NeonLogicalOrWorkload.cpp.
Referenced by NeonLayerSupport::IsLogicalBinarySupported().
arm_compute::Status NeonLogSoftmaxWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const LogSoftmaxDescriptor & | descriptor | ||
) |
Definition at line 19 of file NeonLogSoftmaxWorkload.cpp.
Referenced by NeonLayerSupport::IsLogSoftmaxSupported().
arm_compute::Status NeonLogWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 17 of file NeonLogWorkload.cpp.
Referenced by NeonLayerSupport::IsElementwiseUnarySupported().
arm_compute::Status NeonLstmFloatWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | outputStateIn, | ||
const TensorInfo & | cellStateIn, | ||
const TensorInfo & | scratchBuffer, | ||
const TensorInfo & | outputStateOut, | ||
const TensorInfo & | cellStateOut, | ||
const TensorInfo & | output, | ||
const LstmDescriptor & | descriptor, | ||
const LstmInputParamsInfo & | paramsInfo | ||
) |
Definition at line 253 of file NeonLstmFloatWorkload.cpp.
Referenced by NeonLayerSupport::IsLstmSupported().
arm_compute::Status NeonMaximumWorkloadValidate | ( | const TensorInfo & | input0, |
const TensorInfo & | input1, | ||
const TensorInfo & | output | ||
) |
Definition at line 14 of file NeonMaximumWorkload.cpp.
Referenced by IsLayerTypeSupported(), and NeonLayerSupport::IsMaximumSupported().
arm_compute::Status NeonMeanWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const MeanDescriptor & | descriptor | ||
) |
Definition at line 18 of file NeonMeanWorkload.cpp.
Referenced by NeonLayerSupport::IsMeanSupported().
arm_compute::Status NeonMinimumWorkloadValidate | ( | const TensorInfo & | input0, |
const TensorInfo & | input1, | ||
const TensorInfo & | output | ||
) |
Validate function for validating the inputs and output.
[in] | input0 | The input0 value to be validated. |
[in] | input1 | The input1 value to be validated. |
[in] | output | The output value to be validated. |
Definition at line 15 of file NeonMinimumWorkload.cpp.
Referenced by IsLayerTypeSupported(), and NeonLayerSupport::IsMinimumSupported().
arm_compute::Status NeonMultiplicationWorkloadValidate | ( | const TensorInfo & | input0, |
const TensorInfo & | input1, | ||
const TensorInfo & | output, | ||
const ActivationDescriptor * | activationDescriptor | ||
) |
Definition at line 19 of file NeonMultiplicationWorkload.cpp.
Referenced by IsLayerTypeSupported(), NeonLayerSupport::IsMultiplicationSupported(), and NeonBackend::OptimizeSubgraphView().
arm_compute::Status NeonNegWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 17 of file NeonNegWorkload.cpp.
Referenced by NeonLayerSupport::IsElementwiseUnarySupported().
arm_compute::Status NeonNormalizationWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const NormalizationDescriptor & | descriptor | ||
) |
Definition at line 49 of file NeonNormalizationFloatWorkload.cpp.
Referenced by NeonLayerSupport::IsNormalizationSupported().
arm_compute::Status NeonPadWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const PadDescriptor & | descriptor | ||
) |
Definition at line 59 of file NeonPadWorkload.cpp.
Referenced by NeonLayerSupport::IsPadSupported().
arm_compute::Status NeonPermuteWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const PermuteDescriptor & | descriptor | ||
) |
Definition at line 15 of file NeonPermuteWorkload.cpp.
Referenced by NeonLayerSupport::IsPermuteSupported().
arm_compute::Status NeonPooling2dWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const Pooling2dDescriptor & | descriptor | ||
) |
Definition at line 22 of file NeonPooling2dWorkload.cpp.
Referenced by NeonLayerSupport::IsPooling2dSupported().
arm_compute::Status NeonPooling3dWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const Pooling3dDescriptor & | descriptor | ||
) |
Definition at line 15 of file NeonPooling3dWorkload.cpp.
Referenced by NeonLayerSupport::IsPooling3dSupported().
arm_compute::Status NeonPreluWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | alpha, | ||
const TensorInfo & | output | ||
) |
Definition at line 17 of file NeonPreluWorkload.cpp.
Referenced by NeonLayerSupport::IsPreluSupported().
arm_compute::Status NeonQLstmWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | cellStateIn, | ||
const TensorInfo & | outputStateIn, | ||
const TensorInfo & | cellStateOut, | ||
const TensorInfo & | outputStateOut, | ||
const TensorInfo & | output, | ||
const QLstmDescriptor & | descriptor, | ||
const LstmInputParamsInfo & | paramsInfo | ||
) |
Definition at line 243 of file NeonQLstmWorkload.cpp.
Referenced by NeonLayerSupport::IsQLstmSupported().
arm_compute::Status NeonQuantizedLstmWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | cellStateIn, | ||
const TensorInfo & | outputStateIn, | ||
const TensorInfo & | cellStateOut, | ||
const TensorInfo & | outputStateOut, | ||
const QuantizedLstmInputParamsInfo & | paramsInfo | ||
) |
Definition at line 131 of file NeonQuantizedLstmWorkload.cpp.
Referenced by NeonLayerSupport::IsQuantizedLstmSupported().
arm_compute::Status NeonQuantizeWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 18 of file NeonQuantizeWorkload.cpp.
Referenced by NeonLayerSupport::IsQuantizeSupported().
arm_compute::Status NeonReduceWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const ReduceDescriptor & | descriptor | ||
) |
Definition at line 19 of file NeonReduceWorkload.cpp.
References ReduceDescriptor::m_vAxis.
Referenced by NeonLayerSupport::IsReduceSupported().
arm_compute::Status NeonReshapeWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 17 of file NeonReshapeWorkload.cpp.
Referenced by NeonLayerSupport::IsReshapeSupported().
arm_compute::Status NeonResizeWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const ResizeDescriptor & | descriptor | ||
) |
Definition at line 22 of file NeonResizeWorkload.cpp.
Referenced by NeonLayerSupport::IsResizeSupported().
arm_compute::Status NeonReverseV2WorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | axis, | ||
const TensorInfo & | output | ||
) |
Definition at line 14 of file NeonReverseV2Workload.cpp.
Referenced by NeonLayerSupport::IsReverseV2Supported().
arm_compute::Status NeonRsqrtWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 18 of file NeonRsqrtWorkload.cpp.
Referenced by NeonLayerSupport::IsElementwiseUnarySupported().
arm_compute::Status NeonSinWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 17 of file NeonSinWorkload.cpp.
Referenced by NeonLayerSupport::IsElementwiseUnarySupported().
arm_compute::Status NeonSliceWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const SliceDescriptor & | descriptor | ||
) |
Definition at line 21 of file NeonSliceWorkload.cpp.
Referenced by NeonLayerSupport::IsSliceSupported().
arm_compute::Status NeonSoftmaxWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const SoftmaxDescriptor & | descriptor | ||
) |
Definition at line 19 of file NeonSoftmaxWorkload.cpp.
Referenced by NeonLayerSupport::IsSoftmaxSupported().
arm_compute::Status NeonSpaceToBatchNdWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const SpaceToBatchNdDescriptor & | descriptor | ||
) |
Definition at line 15 of file NeonSpaceToBatchNdWorkload.cpp.
Referenced by NeonLayerSupport::IsSpaceToBatchNdSupported().
arm_compute::Status NeonSpaceToDepthWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const SpaceToDepthDescriptor & | descriptor | ||
) |
Definition at line 19 of file NeonSpaceToDepthWorkload.cpp.
References SpaceToDepthDescriptor::m_DataLayout.
Referenced by NeonLayerSupport::IsSpaceToDepthSupported().
arm_compute::Status NeonSplitterWorkloadValidate | ( | const TensorInfo & | input, |
const std::vector< std::reference_wrapper< TensorInfo >> & | outputs, | ||
unsigned int | splitAxis | ||
) |
Definition at line 33 of file NeonSplitterWorkload.cpp.
Referenced by NeonLayerSupport::IsSplitterSupported().
arm_compute::Status NeonSqrtWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output | ||
) |
Definition at line 18 of file NeonSqrtWorkload.cpp.
Referenced by NeonLayerSupport::IsElementwiseUnarySupported().
arm_compute::Status NeonStackWorkloadValidate | ( | const std::vector< const TensorInfo * > & | inputs, |
const TensorInfo & | output, | ||
const StackDescriptor & | descriptor | ||
) |
Definition at line 27 of file NeonStackWorkload.cpp.
Referenced by NeonLayerSupport::IsStackSupported().
arm_compute::Status NeonStridedSliceWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const StridedSliceDescriptor & | descriptor | ||
) |
Definition at line 19 of file NeonStridedSliceWorkload.cpp.
Referenced by NeonLayerSupport::IsStridedSliceSupported().
arm_compute::Status NeonSubtractionWorkloadValidate | ( | const TensorInfo & | input0, |
const TensorInfo & | input1, | ||
const TensorInfo & | output, | ||
const ActivationDescriptor * | activationDescriptor | ||
) |
Definition at line 22 of file NeonSubtractionWorkload.cpp.
Referenced by IsLayerTypeSupported(), NeonLayerSupport::IsSubtractionSupported(), and NeonBackend::OptimizeSubgraphView().
|
constexpr |
Definition at line 14 of file NeonTensorHandleFactory.hpp.
Referenced by NeonTensorHandleFactory::GetIdStatic().
arm_compute::Status NeonTileWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const TileDescriptor & | descriptor | ||
) |
Definition at line 14 of file NeonTileWorkload.cpp.
Referenced by NeonLayerSupport::IsTileSupported().
arm_compute::Status NeonTransposeConvolution2dWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const TransposeConvolution2dDescriptor & | descriptor, | ||
const TensorInfo & | weights, | ||
const Optional< TensorInfo > & | biases | ||
) |
Definition at line 25 of file NeonTransposeConvolution2dWorkload.cpp.
Referenced by NeonLayerSupport::IsTransposeConvolution2dSupported().
arm_compute::Status NeonTransposeWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | output, | ||
const TransposeDescriptor & | descriptor | ||
) |
Definition at line 15 of file NeonTransposeWorkload.cpp.
Referenced by NeonLayerSupport::IsTransposeSupported().
arm_compute::Status NeonUnidirectionalSequenceLstmFloatWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | outputStateIn, | ||
const TensorInfo & | cellStateIn, | ||
const TensorInfo & | outputStateOut, | ||
const TensorInfo & | cellStateOut, | ||
const TensorInfo & | output, | ||
const UnidirectionalSequenceLstmDescriptor & | descriptor, | ||
const LstmInputParamsInfo & | paramsInfo | ||
) |
Definition at line 510 of file NeonUnidirectionalSequenceLstmFloatWorkload.cpp.
References TensorShape::GetNumDimensions(), TensorInfo::GetShape(), and LstmDescriptor::m_TimeMajor.
Referenced by NeonLayerSupport::IsUnidirectionalSequenceLstmSupported().
arm_compute::Status NeonUnidirectionalSequenceLstmWorkloadValidate | ( | const TensorInfo & | input, |
const TensorInfo & | outputStateIn, | ||
const TensorInfo & | cellStateIn, | ||
const TensorInfo & | outputStateOut, | ||
const TensorInfo & | cellStateOut, | ||
const TensorInfo & | output, | ||
const UnidirectionalSequenceLstmDescriptor & | descriptor, | ||
const LstmInputParamsInfo & | paramsInfo | ||
) |
Definition at line 491 of file NeonUnidirectionalSequenceLstmWorkload.cpp.
References TensorShape::GetNumDimensions(), TensorInfo::GetShape(), and LstmDescriptor::m_TimeMajor.
Referenced by NeonLayerSupport::IsUnidirectionalSequenceLstmSupported().
bool armnn::NextIndex | ( | const unsigned int | numDims, |
const armnn::TensorShape & | dims, | ||
std::vector< unsigned int > & | current | ||
) |
std::vector< unsigned int > NonMaxSuppression | ( | unsigned int | numBoxes, |
const std::vector< float > & | boxCorners, | ||
const std::vector< float > & | scores, | ||
float | nmsScoreThreshold, | ||
unsigned int | maxDetection, | ||
float | nmsIouThreshold | ||
) |
Definition at line 48 of file DetectionPostProcess.cpp.
References GenerateRangeK(), IntersectionOverUnion(), and TopKSort().
Referenced by DetectionPostProcess().
std::enable_if_t< std::is_unsigned<Source>::value && std::is_unsigned<Dest>::value, Dest> armnn::numeric_cast | ( | Source | source | ) |
std::enable_if_t< std::is_signed<Source>::value && std::is_integral<Source>::value && std::is_signed<Dest>::value && std::is_integral<Dest>::value, Dest> armnn::numeric_cast | ( | Source | source | ) |
Definition at line 58 of file NumericCast.hpp.
References ARMNN_NUMERIC_CAST_CHECK.
std::enable_if_t< std::is_floating_point<Source>::value && std::is_floating_point<Dest>::value, Dest> armnn::numeric_cast | ( | Source | source | ) |
Definition at line 83 of file NumericCast.hpp.
References ARMNN_NUMERIC_CAST_CHECK.
std::enable_if_t< std::is_floating_point<Source>::value && std::is_signed<Dest>::value && std::is_integral<Dest>::value, Dest> armnn::numeric_cast | ( | Source | source | ) |
Definition at line 109 of file NumericCast.hpp.
References ARMNN_NUMERIC_CAST_CHECK.
std::enable_if_t< std::is_signed<Source>::value && std::is_integral<Source>::value && std::is_floating_point<Dest>::value, Dest> armnn::numeric_cast | ( | Source | source | ) |
Definition at line 135 of file NumericCast.hpp.
References ARMNN_NUMERIC_CAST_CHECK.
std::enable_if_t< std::is_signed<Dest>::value && std::is_integral<Dest>::value && std::is_unsigned<Source>::value, Dest> armnn::numeric_cast | ( | Source | sValue | ) |
Definition at line 165 of file NumericCast.hpp.
References ARMNN_NUMERIC_CAST_CHECK.
std::enable_if_t< std::is_floating_point<Dest>::value && std::is_unsigned<Source>::value, Dest> armnn::numeric_cast | ( | Source | sValue | ) |
Definition at line 184 of file NumericCast.hpp.
References ARMNN_NUMERIC_CAST_CHECK.
std::enable_if_t< std::is_unsigned<Dest>::value && std::is_signed<Source>::value && std::is_integral<Source>::value, Dest> armnn::numeric_cast | ( | Source | sValue | ) |
Definition at line 206 of file NumericCast.hpp.
References ARMNN_NUMERIC_CAST_CHECK.
std::enable_if_t< std::is_unsigned<Dest>::value && std::is_floating_point<Source>::value, Dest> armnn::numeric_cast | ( | Source | sValue | ) |
Definition at line 230 of file NumericCast.hpp.
References ARMNN_NUMERIC_CAST_CHECK.
unsigned int armnn::Offset | ( | const TensorShape & | shape, |
unsigned int | batch, | ||
unsigned int | height, | ||
unsigned int | width, | ||
unsigned int | channels, | ||
const DataLayoutIndexed & | dataLayout | ||
) |
Definition at line 15 of file BatchToSpaceNd.cpp.
References CHECK_LOCATION, DataLayoutIndexed::GetChannelsIndex(), DataLayoutIndexed::GetDataLayout(), DataLayoutIndexed::GetHeightIndex(), TensorShape::GetNumDimensions(), and DataLayoutIndexed::GetWidthIndex().
Referenced by BatchToSpaceNd().
|
inline |
Definition at line 335 of file TypesUtils.hpp.
References TensorShape::GetDimensionality(), TensorShape::GetDimensionSpecificity(), TensorShape::GetNumDimensions(), and NotSpecified.
|
inline |
Definition at line 176 of file BackendId.hpp.
|
inline |
Definition at line 70 of file IBackendInternal.hpp.
References BackendVersion::m_Major, and BackendVersion::m_Minor.
|
inline |
Definition at line 122 of file BFloat16.hpp.
References BFloat16::ToFloat32(), and BFloat16::Val().
|
inline |
Deprecated function that will be removed together with the Compute enum.
Definition at line 69 of file BackendId.hpp.
References GetComputeDeviceAsCString().
|
inline |
Deprecated function that will be removed together with the Compute enum.
Definition at line 58 of file BackendId.hpp.
References GetComputeDeviceAsCString().
|
inline |
Deprecated function that will be removed together with the Compute enum.
Definition at line 47 of file BackendId.hpp.
References GetComputeDeviceAsCString().
std::ostream& armnn::operator<< | ( | std::ostream & | os, |
const TContainer< BackendId, TContainerTemplateArgs... > & | ids | ||
) |
Definition at line 183 of file BackendId.hpp.
|
inline |
Definition at line 328 of file TypesUtils.hpp.
References GetStatusAsCString().
IOptimizedNetworkPtr Optimize | ( | const Graph & | inGraph, |
const std::vector< BackendId > & | backendPreferences, | ||
const IDeviceSpec & | deviceSpec, | ||
const OptimizerOptions & | options, | ||
Optional< std::vector< std::string > & > | messages = EmptyOptional() |
||
) |
IOptimizedNetworkPtr Optimize | ( | const Graph & | inGraph, |
const std::vector< BackendId > & | backendPreferences, | ||
const IDeviceSpec & | deviceSpec, | ||
const OptimizerOptionsOpaque & | options, | ||
Optional< std::vector< std::string > & > | messages = EmptyOptional() |
||
) |
Create an optimized version of the network.
inGraph | Graph to be optimized. |
backendPreferences | The choice of the backend ordered by user preferences. |
deviceSpec | DeviceSpec object as queried from the runtime. See IRuntime::GetDeviceSpec() |
messages | If there are failures or warnings a string describing same will be added to the vector |
options | OptimizerOptions object with optimizer configuration options |
Definition at line 1906 of file Network.cpp.
References Graph::AddCompatibilityLayers(), ApplyBackendOptimizations(), ARMNN_LOG, ARMNN_SCOPED_PROFILING_EVENT, AssignBackends(), Graph::begin(), CheckFp16Support(), armnnUtils::Filesystem::CreateDirectory(), CreateSupportedBackends(), debug, Graph::end(), BackendSettings::GetAvailablePreferredBackends(), OptimizerOptionsOpaque::GetDebugEnabled(), OptimizerOptionsOpaque::GetDebugToFileEnabled(), OptimizerOptionsOpaque::GetExportEnabled(), OptimizerOptionsOpaque::GetImportEnabled(), OptimizerOptionsOpaque::GetModelOptions(), Graph::GetProfiler(), OptimizerOptionsOpaque::GetProfilingEnabled(), OptimizerOptionsOpaque::GetReduceFp32ToBf16(), OptimizerOptionsOpaque::GetReduceFp32ToFp16(), OptimizerOptionsOpaque::GetShapeInferenceMethod(), Graph::InferTensorInfos(), info, OptimizationResult::m_Error, BackendSettings::m_SupportedBackends, MakeOptimizations(), IOptimizedNetwork::pOptimizedNetworkImpl, ReportError(), SelectTensorHandleStrategy(), OptimizerOptionsOpaque::ToString(), Graph::VerifyConstantLayerSetTensorInfo(), warning, and Exception::what().
IOptimizedNetworkPtr Optimize | ( | const INetwork & | network, |
const std::vector< BackendId > & | backendPreferences, | ||
const IDeviceSpec & | deviceSpec, | ||
const OptimizerOptions & | options, | ||
Optional< std::vector< std::string > & > | messages = EmptyOptional() |
||
) |
IOptimizedNetworkPtr Optimize | ( | const INetwork & | network, |
const std::vector< BackendId > & | backendPreferences, | ||
const IDeviceSpec & | deviceSpec, | ||
const OptimizerOptionsOpaque & | options = OptimizerOptionsOpaque() , |
||
Optional< std::vector< std::string > & > | messages = EmptyOptional() |
||
) |
Create an optimized version of the network.
network | INetwork description of the network to be optimized. |
backendPreferences | The choice of the backend ordered by user preferences. |
deviceSpec | DeviceSpec object as queried from the runtime. See IRuntime::GetDeviceSpec() |
messages | If there are failures or warnings a string describing same will be added to the vector |
options | OptimizerOptions object with optimizer configuration options |
Definition at line 2145 of file Network.cpp.
References INetwork::pNetworkImpl.
Referenced by Optimize(), ArmnnDriverImpl::PrepareArmnnModel(), ArmnnDriverImpl::PrepareArmnnModelFromCache(), ParserPrototxtFixture< TParser >::Setup(), and ParserPrototxtFixture< TParser >::SetupOptimizedNetwork().
void Pad | ( | const TensorInfo & | inputInfo, |
const TensorInfo & | outputInfo, | ||
const ITensorHandle * | inputHandle, | ||
ITensorHandle * | outputHandle, | ||
const PadQueueDescriptor & | data | ||
) |
Definition at line 39 of file Pad.cpp.
References Decoder< IType >::Get(), TensorInfo::GetDataType(), TensorShape::GetNumDimensions(), TensorInfo::GetNumElements(), TensorInfo::GetShape(), TensorInfo::IsQuantized(), PadDescriptor::m_PadList, PadDescriptor::m_PadValue, QueueDescriptorWithParameters< LayerDescriptor >::m_Parameters, ITensorHandle::Map(), and Encoder< IType >::Set().
|
inline |
Definition at line 312 of file BackendOptions.hpp.
References BackendOptions::Var::AsBool(), and BackendOptions::Var::IsBool().
|
constexpr |
Deprecated function that will be removed together with the Compute enum.
Definition at line 213 of file TypesUtils.hpp.
References CpuAcc, CpuRef, GpuAcc, StrEqual(), and Undefined.
|
inline |
Definition at line 330 of file BackendOptions.hpp.
References BackendOptions::Var::AsInt(), and BackendOptions::Var::IsInt().
Referenced by ClBackendModelContext::ClBackendModelContext().
void armnn::ParseOptions | ( | const std::vector< BackendOptions > & | options, |
BackendId | backend, | ||
F | f | ||
) |
Definition at line 297 of file BackendOptions.hpp.
References BackendOptions::BackendOption::GetName(), and BackendOptions::BackendOption::GetValue().
Referenced by ClBackendContext::ClBackendContext(), ClBackendModelContext::ClBackendModelContext(), GpuFsaBackendContext::GpuFsaBackendContext(), NeonBackendModelContext::NeonBackendModelContext(), and RuntimeImpl::RuntimeImpl().
|
inline |
Definition at line 321 of file BackendOptions.hpp.
References BackendOptions::Var::AsString(), and BackendOptions::Var::IsString().
Referenced by ClBackendContext::ClBackendContext(), and GpuFsaBackendContext::GpuFsaBackendContext().
|
inline |
Definition at line 26 of file ArmComputeTuningUtils.hpp.
References ARMNN_LOG, BackendOptions::Var::AsInt(), Exhaustive, BackendOptions::Var::IsInt(), None, and warning.
Referenced by ClBackendContext::ClBackendContext(), and GpuFsaBackendContext::GpuFsaBackendContext().
armnn::ConstTensor PermuteTensor | ( | const ConstTensorHandle * | tensor, |
const PermutationVector & | permutationVector, | ||
void * | permuteBuffer | ||
) |
Definition at line 19 of file WorkloadUtils.cpp.
References ConstTensorHandle::GetConstTensor(), TensorInfo::GetDataType(), GetDataTypeSize(), TensorInfo::GetNumBytes(), TensorInfo::GetShape(), PermutationVector::GetSize(), ConstTensorHandle::GetTensorInfo(), armnnUtils::Permute(), armnnUtils::Permuted(), and TensorInfo::SetConstant().
Referenced by Convert1HWOTensorToAcl(), Convert1HWOtoMIHW(), and ConvertWeightTensorFromArmnnToAcl().
DestType armnn::PolymorphicDowncast | ( | SourceType * | value | ) |
Polymorphic downcast for build in pointers only.
Usage: Child* pChild = PolymorphicDowncast<Child*>(pBase);
DestType | Pointer type to the target object (Child pointer type) |
SourceType | Pointer type to the source object (Base pointer type) |
value | Pointer to the source object |
Definition at line 74 of file PolymorphicDowncast.hpp.
References ARMNN_POLYMORPHIC_CAST_CHECK.
Referenced by ClLayerSupport::IsLayerSupported(), and IsLayerTypeSupported().
auto armnn::PolymorphicPointerDowncast | ( | const SourceType & | value | ) |
Polymorphic downcast for shared pointers and build in pointers.
Usage: auto pChild = PolymorphicPointerDowncast<Child>(pBase)
DestType | Type of the target object (Child type) |
SourceType | Pointer type to the source object (Base (shared) pointer type) |
value | Pointer to the source object |
Definition at line 93 of file PolymorphicDowncast.hpp.
References ARMNN_POLYMORPHIC_CAST_CHECK.
void Pooling2d | ( | Decoder< float > & | rInputDecoder, |
Encoder< float > & | rOutputEncoder, | ||
const TensorInfo & | inputInfo, | ||
const TensorInfo & | outputInfo, | ||
const Pooling2dDescriptor & | params | ||
) |
Computes the Pooling2d operation.
Definition at line 142 of file Pooling2d.cpp.
References Decoder< IType >::DecodeTensor(), DataLayoutIndexed::GetChannelsIndex(), DataLayoutIndexed::GetDataLayout(), DataLayoutIndexed::GetHeightIndex(), TensorInfo::GetShape(), DataLayoutIndexed::GetWidthIndex(), Pooling2dDescriptor::m_DataLayout, Pooling2dDescriptor::m_PadBottom, Pooling2dDescriptor::m_PaddingMethod, Pooling2dDescriptor::m_PadLeft, Pooling2dDescriptor::m_PadRight, Pooling2dDescriptor::m_PadTop, Pooling2dDescriptor::m_PoolHeight, Pooling2dDescriptor::m_PoolType, Pooling2dDescriptor::m_PoolWidth, Pooling2dDescriptor::m_StrideX, Pooling2dDescriptor::m_StrideY, Pooling2d(), and Encoder< IType >::Set().
Referenced by Pooling2d(), and Pooling2dLayer::Pooling2dLayer().
void Pooling3d | ( | Decoder< float > & | rInputDecoder, |
Encoder< float > & | rOutputEncoder, | ||
const TensorInfo & | inputInfo, | ||
const TensorInfo & | outputInfo, | ||
const Pooling3dDescriptor & | params | ||
) |
Computes the Pooling3d operation.
Definition at line 172 of file Pooling3d.cpp.
References Decoder< IType >::DecodeTensor(), DataLayoutIndexed::GetChannelsIndex(), DataLayoutIndexed::GetDepthIndex(), DataLayoutIndexed::GetHeightIndex(), TensorInfo::GetShape(), DataLayoutIndexed::GetWidthIndex(), Pooling3dDescriptor::m_DataLayout, Pooling3dDescriptor::m_PadBack, Pooling3dDescriptor::m_PadBottom, Pooling3dDescriptor::m_PaddingMethod, Pooling3dDescriptor::m_PadFront, Pooling3dDescriptor::m_PadLeft, Pooling3dDescriptor::m_PadRight, Pooling3dDescriptor::m_PadTop, Pooling3dDescriptor::m_PoolDepth, Pooling3dDescriptor::m_PoolHeight, Pooling3dDescriptor::m_PoolType, Pooling3dDescriptor::m_PoolWidth, Pooling3dDescriptor::m_StrideX, Pooling3dDescriptor::m_StrideY, Pooling3dDescriptor::m_StrideZ, Pooling3d(), and Encoder< IType >::Set().
Referenced by Pooling3d(), and Pooling3dLayer::Pooling3dLayer().
void PreluImpl | ( | const TensorInfo & | inputInfo, |
const TensorInfo & | alphaInfo, | ||
const TensorInfo & | outputInfo, | ||
Decoder< float > & | inputData, | ||
Decoder< float > & | alphaData, | ||
Encoder< float > & | outputData | ||
) |
Definition at line 13 of file PreluImpl.cpp.
References TensorInfo::GetShape(), and BroadcastLoop::Unroll().
void armnn::PrintOutput | ( | const TensorInfo & | inputInfo, |
const T * | inputData, | ||
LayerGuid | guid, | ||
const std::string & | layerName, | ||
unsigned int | slotIndex, | ||
std::ostream & | os | ||
) |
Definition at line 23 of file Debug.cpp.
References TensorInfo::GetNumDimensions(), TensorInfo::GetNumElements(), and TensorInfo::GetShape().
|
inline |
< Profiler used
Definition at line 180 of file Profiling.hpp.
References ProfilerManager::GetInstance(), and IProfiler::IsProfilingEnabled().
template int32_t Quantize< int32_t > | ( | float | value, |
float | scale, | ||
int32_t | offset | ||
) |
Quantize a floating point data type into an 8-bit data type.
Explicit specialization of Quantize for int32_t.
Explicit specialization of Quantize for int16_t.
Explicit specialization of Quantize for uint8_t.
Explicit specialization of Quantize for int8_t.
value | - The value to quantize. |
scale | - The scale (must be non-zero). |
offset | - The offset. |
Definition at line 30 of file TypesUtils.cpp.
|
inline |
void Reduce | ( | const TensorInfo & | inputInfo, |
const TensorInfo & | outputInfo, | ||
Decoder< float > & | input, | ||
Encoder< float > & | output, | ||
const std::vector< uint32_t > | axis, | ||
const ReduceOperation | reduceOperation | ||
) |
Definition at line 70 of file Reduce.cpp.
References Decoder< IType >::Get(), TensorInfo::GetNumDimensions(), TensorInfo::GetNumElements(), TensorInfo::GetShape(), Max, Mean, Min, NextIndex(), Prod, ReducedOutputOffset(), Encoder< IType >::Set(), and Sum.
unsigned int armnn::ReducedOutputOffset | ( | const unsigned int | numDims, |
const armnn::TensorShape & | dims, | ||
std::vector< unsigned int > & | index, | ||
const unsigned int | numAxis, | ||
const std::vector< unsigned int > & | axis | ||
) |
|
constexpr |
Definition at line 10 of file RefBackendId.hpp.
Referenced by RefBackend::GetIdStatic().
|
constexpr |
Definition at line 15 of file RefTensorHandleFactory.hpp.
Referenced by RefTensorHandleFactory::GetIdStatic().
|
inline |
Definition at line 293 of file SubgraphUtils.hpp.
References OptimizationViews::AddDeletedSubgraph(), Constant, InputSlot::GetConnectedOutputSlot(), OutputSlot::GetConnection(), Layer::GetGuid(), Layer::GetInputSlot(), Layer::GetOutputSlot(), InputSlot::GetOwningLayer(), OutputSlot::GetOwningLayer(), LayerWithParameters< Parameters >::GetParameters(), InputSlot::GetSlotIndex(), OutputSlot::GetTensorInfo(), Layer::GetType(), Input, ReshapeDescriptor::m_TargetShape, Output, TensorInfo::SetShape(), and InputSlot::SetTensorInfo().
Referenced by RefBackend::OptimizeSubgraphView(), NeonBackend::OptimizeSubgraphView(), and ClBackend::OptimizeSubgraphView().
ConstTensor armnn::ReorderWeightChannelsForAcl | ( | const ConstTensor & | weightHandle, |
DataLayout | dataLayout, | ||
void * | permuteBuffer | ||
) |
Definition at line 74 of file WorkloadUtils.cpp.
References BaseTensor< MemoryType >::GetInfo(), TensorInfo::GetNumBytes(), BaseTensor< MemoryType >::GetShape(), NCHW, and NHWC.
void armnn::ReplaceLayers | ( | OptimizationViews & | optimizationViews, |
LayerType * | baseLayer, | ||
std::vector< IConnectableLayer * > & | layers | ||
) |
Definition at line 339 of file ArmComputeSubgraphUtils.hpp.
References OptimizationViews::AddSubstitution().
void armnn::ReplaceMultipleLayers | ( | OptimizationViews & | optimizationViews, |
std::vector< IConnectableLayer * > & | originalLayers, | ||
LayerType * | baseLayer, | ||
const std::vector< SlotList > | inputLayersSlotLists, | ||
const std::vector< SlotList > | outputLayersSlotLists | ||
) |
Definition at line 357 of file ArmComputeSubgraphUtils.hpp.
References OptimizationViews::AddSubstitution().
void armnn::ReportError | ( | const std::string & | errorMessage, |
Optional< std::vector< std::string > & > | errorMessages | ||
) |
Definition at line 762 of file Network.cpp.
References ARMNN_LOG, and warning.
Referenced by AssignBackends(), CheckScaleSetOnQuantizedType(), Optimize(), and ReturnWithError().
|
inline |
Definition at line 220 of file SubgraphUtils.hpp.
References OptimizationViews::AddUntouchedSubgraph().
Referenced by RefBackend::OptimizeSubgraphView(), NeonBackend::OptimizeSubgraphView(), ClBackend::OptimizeSubgraphView(), and GpuFsaBackend::OptimizeSubgraphView().
void armnn::ReportWarning | ( | const std::string & | warningMessage, |
Optional< std::vector< std::string > & > | warningMessages | ||
) |
Definition at line 774 of file Network.cpp.
References ARMNN_LOG, and warning.
Referenced by ApplyBackendOptimizations(), and AttemptBackendAssignment().
bool armnn::RequiresCopy | ( | ITensorHandleFactory::FactoryId | src, |
ITensorHandleFactory::FactoryId | dst, | ||
TensorHandleFactoryRegistry & | registry | ||
) |
Definition at line 1454 of file Network.cpp.
References ITensorHandleFactory::GetExportFlags(), TensorHandleFactoryRegistry::GetFactory(), and ITensorHandleFactory::GetImportFlags().
Referenced by CalculateSlotOption().
void ReshapeWeightsForAcl | ( | TensorInfo & | weightInfo, |
DataLayout | dataLayout | ||
) |
Definition at line 48 of file WorkloadUtils.cpp.
References TensorInfo::GetShape(), NCHW, NHWC, and TensorInfo::SetShape().
Referenced by ConvertWeightTensorFromArmnnToAcl(), and ConvertWeightTensorInfoFromArmnnToAcl().
void Resize | ( | Decoder< float > & | in, |
const TensorInfo & | inputInfo, | ||
Encoder< float > & | out, | ||
const TensorInfo & | outputInfo, | ||
DataLayoutIndexed | dataLayout, | ||
ResizeMethod | resizeMethod, | ||
bool | alignCorners, | ||
bool | halfPixelCenters | ||
) |
Definition at line 65 of file Resize.cpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, Decoder< IType >::Get(), DataLayoutIndexed::GetChannelsIndex(), DataLayoutIndexed::GetHeightIndex(), DataLayoutIndexed::GetIndex(), TensorInfo::GetShape(), DataLayoutIndexed::GetWidthIndex(), Resize(), roundf(), and Encoder< IType >::Set().
Referenced by Resize(), and ResizeLayer::ResizeLayer().
OptimizationResult armnn::ReturnWithError | ( | OptimizationResult | res, |
const Layer * | layer, | ||
const BackendSettings & | backendSettings, | ||
Optional< std::vector< std::string > & > | errMessages | ||
) |
Definition at line 786 of file Network.cpp.
References GetLayerTypeAsCString(), Layer::GetType(), OptimizationResult::m_Error, BackendSettings::m_PreferredBackends, and ReportError().
Referenced by AssignBackendsIConnectable(), and AttemptBackendAssignment().
unsigned int armnn::ReverseGetFlatIdx | ( | const std::vector< unsigned int > & | idxList, |
unsigned int | inputRank, | ||
std::vector< unsigned int > & | elementNumInner | ||
) |
std::vector<unsigned int> armnn::ReverseGetMultIdx | ( | const unsigned int | idx, |
unsigned int | inputRank, | ||
std::vector< unsigned int > & | elementNumInner | ||
) |
unsigned int armnn::ReverseRelocateIdx | ( | unsigned int | idx, |
unsigned int | inputRank, | ||
std::vector< bool > & | axisFlag, | ||
std::vector< unsigned int > & | dimSize, | ||
std::vector< unsigned int > & | elementNumInner | ||
) |
Definition at line 49 of file ReverseV2Impl.cpp.
References ReverseGetFlatIdx(), and ReverseGetMultIdx().
Referenced by ReverseV2().
void ReverseV2 | ( | const TensorInfo & | inputInfo, |
const TensorInfo & | axisInfo, | ||
Decoder< float > & | inputDecoder, | ||
Decoder< int > & | axisDecoder, | ||
Encoder< float > & | outputEncoder | ||
) |
Definition at line 78 of file ReverseV2Impl.cpp.
References Decoder< IType >::Get(), TensorInfo::GetNumDimensions(), TensorInfo::GetNumElements(), TensorInfo::GetShape(), ReverseRelocateIdx(), and Encoder< IType >::Set().
bool armnn::RevertConstantWeightsToFP32 | ( | Layer * | layer | ) |
|
inline |
|
inline |
Definition at line 167 of file ClWorkloadUtils.hpp.
References error, and WrapClError().
Referenced by ClFillWorkload::Execute(), ClPadWorkload::Execute(), ClAdditionWorkload::Execute(), ClSubtractionWorkload::Execute(), ClActivationWorkload::Execute(), ClNegWorkload::Execute(), ClCastWorkload::Execute(), ClExpWorkload::Execute(), ClPreluWorkload::Execute(), ClConvertFp16ToFp32Workload::Execute(), ClQuantizeWorkload::Execute(), ClRsqrtWorkload::Execute(), ClSinWorkload::Execute(), ClAbsWorkload::Execute(), ClSqrtWorkload::Execute(), ClConvertFp32ToFp16Workload::Execute(), ClLogWorkload::Execute(), ClLstmFloatWorkload::Execute(), ClNormalizationFloatWorkload::Execute(), ClFloorFloatWorkload::Execute(), ClReshapeWorkload::Execute(), ClResizeWorkload::Execute(), ClGatherWorkload::Execute(), ClSpaceToDepthWorkload::Execute(), ClInstanceNormalizationWorkload::Execute(), ClMaximumWorkload::Execute(), ClMinimumWorkload::Execute(), ClArgMinMaxWorkload::Execute(), ClChannelShuffleWorkload::Execute(), ClL2NormalizationFloatWorkload::Execute(), ClComparisonWorkload::Execute(), ClBatchMatMulWorkload::Execute(), ClSliceWorkload::Execute(), ClSpaceToBatchNdWorkload::Execute(), ClDepthToSpaceWorkload::Execute(), ClDivisionWorkload::Execute(), ClPooling2dWorkload::Execute(), ClPooling3dWorkload::Execute(), ClGatherNdWorkload::Execute(), ClStridedSliceWorkload::Execute(), ClBatchToSpaceNdWorkload::Execute(), ClMultiplicationWorkload::Execute(), ClPermuteWorkload::Execute(), ClScatterNdWorkload::Execute(), ClTransposeWorkload::Execute(), ClQuantizedLstmWorkload::Execute(), ClLogSoftmaxWorkload::Execute(), ClSoftmaxWorkload::Execute(), ClDepthwiseConvolutionWorkload::Execute(), ClBatchNormalizationFloatWorkload::Execute(), ClConvolution3dWorkload::Execute(), ClFullyConnectedWorkload::Execute(), ClTransposeConvolution2dWorkload::Execute(), and ClConvolution2dWorkload::Execute().
void ScatterNd | ( | const TensorInfo & | indicesInfo, |
const TensorInfo & | updatesInfo, | ||
const TensorInfo & | shapeInfo, | ||
Decoder< int > & | indices, | ||
Decoder< float > & | updates, | ||
Decoder< int > & | shape, | ||
Encoder< float > & | output, | ||
ScatterNdDescriptor | descriptor | ||
) |
Definition at line 181 of file ScatterNd.cpp.
References Decoder< IType >::DecodeTensor(), Decoder< IType >::Get(), TensorInfo::GetNumDimensions(), TensorInfo::GetNumElements(), TensorInfo::GetShape(), ScatterNdDescriptor::m_AxisEnabled, ScatterNdDescriptor::m_Function, ScatterOperation(), Encoder< IType >::Set(), and Update.
void ScatterNd | ( | const TensorInfo & | inputInfo, |
const TensorInfo & | indicesInfo, | ||
const TensorInfo & | updatesInfo, | ||
Decoder< float > & | input, | ||
Decoder< int > & | indices, | ||
Decoder< float > & | updates, | ||
Encoder< float > & | output, | ||
ScatterNdDescriptor | descriptor | ||
) |
Definition at line 41 of file ScatterNd.cpp.
References Decoder< IType >::Get(), TensorInfo::GetNumDimensions(), TensorInfo::GetNumElements(), TensorInfo::GetShape(), ScatterNdDescriptor::m_AxisEnabled, ScatterNdDescriptor::m_Function, ScatterOperation(), Encoder< IType >::Set(), and Update.
float armnn::ScatterOperation | ( | ScatterNdFunction | operation, |
float | input, | ||
float | update | ||
) |
Definition at line 18 of file ScatterNd.cpp.
References Add, Max, Min, Mul, Sub, and Update.
Referenced by ScatterNd().
OptimizationResult SelectTensorHandleStrategy | ( | Graph & | optGraph, |
BackendsMap & | backends, | ||
TensorHandleFactoryRegistry & | registry, | ||
bool | importEnabled, | ||
bool | exportEnabled, | ||
Optional< std::vector< std::string > & > | errMessages | ||
) |
Definition at line 1821 of file Network.cpp.
References ARMNN_SCOPED_PROFILING_EVENT, Graph::ForEachLayer(), Layer::GetBackendId(), OutputSlot::GetConnections(), Layer::GetNumOutputSlots(), Layer::GetOutputSlot(), Layer::GetType(), ITensorHandleFactory::LegacyFactoryId, OutputSlot::SetEdgeStrategy(), OutputSlot::SetTensorHandleFactory(), and Undefined.
Referenced by Optimize().
void SetAllLoggingSinks | ( | bool | standardOut, |
bool | debugOut, | ||
bool | coloured | ||
) |
|
inline |
Definition at line 100 of file ClWorkloadUtils.hpp.
Referenced by ClSliceWorkload::ClSliceWorkload().
|
inline |
Definition at line 79 of file ClWorkloadUtils.hpp.
Referenced by ClStridedSliceWorkload::ClStridedSliceWorkload().
void SetLogFilter | ( | LogSeverity | level | ) |
Definition at line 73 of file Logging.cpp.
References ARMNN_FALLTHROUGH, Debug, SimpleLogger< Level >::Enable(), Error, Fatal, SimpleLogger< Level >::Get(), Info, Trace, and Warning.
Referenced by ConfigureLogging(), and TEST_SUITE().
|
inline |
Definition at line 167 of file Logging.cpp.
References SimpleLogger< Level >::AddSink(), SimpleLogger< Level >::Get(), and SimpleLogger< Level >::RemoveAllSinks().
|
inline |
Definition at line 160 of file NeonWorkloadUtils.hpp.
Referenced by NeonSliceWorkload::NeonSliceWorkload().
|
inline |
Definition at line 138 of file NeonWorkloadUtils.hpp.
Referenced by NeonStridedSliceWorkload::NeonStridedSliceWorkload().
void armnn::SetValueChecked | ( | Optional< T & > | optionalRef, |
V && | val | ||
) |
Definition at line 17 of file LayerSupportCommon.hpp.
References OptionalReferenceSwitch< std::is_reference< T >::value, T >::value().
Referenced by FalseFuncF16(), FalseFuncF32(), FalseFuncI32(), FalseFuncU8(), FalseInputFuncF16(), FalseInputFuncF32(), FalseOutputFuncF16(), FalseOutputFuncF32(), NeonLayerSupport::IsConcatSupported(), ClLayerSupport::IsConcatSupported(), NeonLayerSupport::IsSplitterSupported(), and ClLayerSupport::IsSplitterSupported().
void Slice | ( | const TensorInfo & | inputInfo, |
const SliceDescriptor & | descriptor, | ||
const void * | inputData, | ||
void * | outputData, | ||
unsigned int | dataTypeSize | ||
) |
Definition at line 14 of file Slice.cpp.
References TensorShape::GetNumDimensions(), TensorInfo::GetShape(), SliceDescriptor::m_Begin, and SliceDescriptor::m_Size.
void Softmax | ( | Decoder< float > & | in, |
Encoder< float > & | out, | ||
const TensorInfo & | inputTensorInfo, | ||
float | beta, | ||
int | axis | ||
) |
Computes the softmax function on some inputs, into outputs, with a shape given by tensorInfo.
Definition at line 17 of file Softmax.cpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, Decoder< IType >::Get(), TensorShape::GetNumDimensions(), TensorInfo::GetNumDimensions(), armnnUtils::GetNumElementsBetween(), TensorInfo::GetShape(), and Encoder< IType >::Set().
void SpaceToBatchNd | ( | const TensorInfo & | inputInfo, |
const TensorInfo & | outputInfo, | ||
const SpaceToBatchNdDescriptor & | params, | ||
Decoder< float > & | inputData, | ||
Encoder< float > & | outputData | ||
) |
Definition at line 48 of file SpaceToBatchNd.cpp.
References CHECK_LOCATION, Decoder< IType >::Get(), DataLayoutIndexed::GetChannelsIndex(), DataLayoutIndexed::GetHeightIndex(), TensorInfo::GetNumDimensions(), GetOffset(), TensorInfo::GetShape(), DataLayoutIndexed::GetWidthIndex(), SpaceToBatchNdDescriptor::m_BlockShape, SpaceToBatchNdDescriptor::m_DataLayout, SpaceToBatchNdDescriptor::m_PadList, Encoder< IType >::Set(), and SpaceToBatchNd().
Referenced by SpaceToBatchNd(), and SpaceToBatchNdLayer::SpaceToBatchNdLayer().
void SpaceToDepth | ( | const TensorInfo & | inputInfo, |
const TensorInfo & | outputInfo, | ||
const SpaceToDepthDescriptor & | params, | ||
Decoder< float > & | inputData, | ||
Encoder< float > & | outputData | ||
) |
Definition at line 36 of file SpaceToDepth.cpp.
References Decoder< IType >::Get(), DataLayoutIndexed::GetChannelsIndex(), DataLayoutIndexed::GetHeightIndex(), GetOffset(), TensorInfo::GetShape(), DataLayoutIndexed::GetWidthIndex(), SpaceToDepthDescriptor::m_BlockSize, SpaceToDepthDescriptor::m_DataLayout, Encoder< IType >::Set(), and SpaceToDepth().
Referenced by SpaceToDepth(), and SpaceToDepthLayer::SpaceToDepthLayer().
void Split | ( | const SplitterQueueDescriptor & | data, |
std::vector< ITensorHandle * > | inputs, | ||
std::vector< ITensorHandle * > | outputs | ||
) |
Definition at line 20 of file Splitter.cpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, Decoder< IType >::Get(), TensorInfo::GetNumDimensions(), TensorInfo::GetNumElements(), TensorInfo::GetShape(), GetTensorInfo(), SplitterQueueDescriptor::ViewOrigin::m_Origin, SplitterQueueDescriptor::m_ViewOrigins, Map, MaxNumOfTensorDimensions, and Encoder< IType >::Set().
void armnn::Splitter | ( | const SplitterQueueDescriptor & | data, |
std::vector< ITensorHandle * > | inputs, | ||
std::vector< ITensorHandle * > | outputs | ||
) |
Definition at line 17 of file Splitter.hpp.
References ARMNN_THROW_INVALIDARG_MSG_IF_FALSE, TensorInfo::GetNumDimensions(), TensorInfo::GetNumElements(), TensorInfo::GetShape(), GetTensorInfo(), SplitterQueueDescriptor::ViewOrigin::m_Origin, SplitterQueueDescriptor::m_ViewOrigins, and MaxNumOfTensorDimensions.
void Stack | ( | const StackQueueDescriptor & | data, |
std::vector< std::unique_ptr< Decoder< float >>> & | inputs, | ||
Encoder< float > & | output, | ||
const TensorInfo & | inputInfo, | ||
const TensorInfo & | outputInfo | ||
) |
Definition at line 12 of file Stack.cpp.
References TensorInfo::GetNumDimensions(), TensorInfo::GetNumElements(), TensorInfo::GetShape(), StackDescriptor::m_Axis, QueueDescriptor::m_Inputs, StackDescriptor::m_NumInputs, QueueDescriptorWithParameters< LayerDescriptor >::m_Parameters, and Encoder< IType >::Set().
|
constexpr |
void StridedSlice | ( | const TensorInfo & | inputInfo, |
const StridedSliceDescriptor & | params, | ||
const void * | inputData, | ||
void * | outputData, | ||
unsigned int | dataTypeSize | ||
) |
|
inline |
Definition at line 43 of file Logging.hpp.
void armnn::swap | ( | OriginsDescriptor & | first, |
OriginsDescriptor & | second | ||
) |
Definition at line 357 of file Descriptors.cpp.
References swap().
Referenced by BackendId::operator=(), SquashEqualSiblingsImpl< Comparable >::Run(), BackendRegistry::Swap(), and swap().
void armnn::swap | ( | ViewsDescriptor & | first, |
ViewsDescriptor & | second | ||
) |
void Tile | ( | const TileDescriptor & | params, |
const TensorInfo & | inputInfo, | ||
Decoder< float > & | inputDecoder, | ||
Encoder< float > & | outputEncoder | ||
) |
Definition at line 45 of file Tile.cpp.
References CoordinatesToIndex(), Decoder< IType >::DecodeTensor(), Decoder< IType >::Get(), TensorInfo::GetNumDimensions(), TensorInfo::GetNumElements(), TensorInfo::GetShape(), IndexToCoordinates(), TileDescriptor::m_Multiples, and Encoder< IType >::Set().
void TopKSort | ( | unsigned int | k, |
unsigned int * | indices, | ||
const float * | values, | ||
unsigned int | numElement | ||
) |
Definition at line 23 of file DetectionPostProcess.cpp.
Referenced by DetectionPostProcess(), and NonMaxSuppression().
|
constexpr |
Definition at line 10 of file TosaRefBackendId.hpp.
Referenced by TosaRefBackend::GetIdStatic().
bool TosaRefPreCompiledWorkloadValidate | ( | std::string * | ) |
Definition at line 166 of file TosaRefPreCompiledWorkload.cpp.
|
constexpr |
Definition at line 15 of file TosaRefTensorHandleFactory.hpp.
Referenced by TosaRefTensorHandleFactory::GetIdStatic().
void TransposeConvolution2dImpl | ( | const TransposeConvolution2dDescriptor & | descriptor, |
const TensorShape & | inputShape, | ||
Decoder< float > & | inputDecoder, | ||
const TensorShape & | outputShape, | ||
Encoder< float > & | outputEncoder, | ||
const TensorShape & | weightsShape, | ||
Decoder< float > & | weightsDecoder, | ||
Decoder< float > * | biasesDecoder | ||
) |
Definition at line 15 of file TransposeConvolution2d.cpp.
References Decoder< IType >::DecodeTensor(), Decoder< IType >::Get(), DataLayoutIndexed::GetChannelsIndex(), DataLayoutIndexed::GetHeightIndex(), DataLayoutIndexed::GetIndex(), TensorShape::GetNumElements(), DataLayoutIndexed::GetWidthIndex(), TransposeConvolution2dDescriptor::m_BiasEnabled, TransposeConvolution2dDescriptor::m_DataLayout, TransposeConvolution2dDescriptor::m_PadLeft, TransposeConvolution2dDescriptor::m_PadTop, TransposeConvolution2dDescriptor::m_StrideX, TransposeConvolution2dDescriptor::m_StrideY, NHWC, and Encoder< IType >::Set().
bool armnn::TrueFunc | ( | Optional< std::string & > | reasonIfUnsupported, |
Params &&... | params | ||
) |
Definition at line 54 of file LayerSupportCommon.hpp.
References IgnoreUnused().
void armnn::ValidateSourcesMatchOptimizedNetwork | ( | std::vector< BackendOptions > | optimizedOptions, |
const INetworkProperties & | networkProperties | ||
) |
This function performs a sanity check to ensure that the combination of input and output memory source matches the values for importEnabled and exportEnabled that were specified during optimization.
During optimization the tensor handle factories are chosen based on whether import and export are enabled. If the user then specifies something incompatible here it can lead to problems.
optimizedOptions | |
networkProperties |
Definition at line 101 of file LoadedNetwork.cpp.
|
inline |
Definition at line 157 of file ClContextSchema_generated.h.
References ClContextIdentifier().
|
inline |
Definition at line 162 of file ClContextSchema_generated.h.
References ClContextIdentifier().
|
inline |
|
inline |
Definition at line 159 of file ClWorkloadUtils.hpp.
References Exception::what().
Referenced by RunClFunction().
const BackendCapabilities cpuAccCapabilities("CpuAcc", { {"NonConstWeights", true}, {"AsyncExecution", false}, {"ProtectedContentAllocation", false}, {"ConstantTensorsAsInputs", true}, {"PreImportIOTensors", false}, {"ExternallyManagedMemory", true}, {"MultiAxisPacking", false}, {"SingleAxisPacking", true}, {"HasFp16", arm_compute::CPUInfo::get().has_fp16()} }) |
Referenced by NeonBackend::GetCapabilities().
const BackendCapabilities cpuRefCapabilities("CpuRef", { {"NonConstWeights", true}, {"AsyncExecution", true}, {"ProtectedContentAllocation", false}, {"ConstantTensorsAsInputs", true}, {"PreImportIOTensors", true}, {"ExternallyManagedMemory", true}, {"MultiAxisPacking", false}, {"SingleAxisPacking", true}, {"HasFp16", true} }) |
Referenced by RefBackend::GetCapabilities().
|
constexpr |
|
constexpr |
Definition at line 37 of file Profiling.cpp.
Referenced by ProfilerImpl::AnalyzeEventsAndWriteResults().
|
constexpr |
Definition at line 29 of file Profiling.cpp.
Referenced by ProfilerImpl::ProfilerImpl().
|
constexpr |
Definition at line 32 of file Profiling.cpp.
Referenced by ProfilerImpl::AnalyzeEventSequenceAndWriteResults().
|
constexpr |
Definition at line 41 of file Profiling.cpp.
Referenced by ProfilerImpl::~ProfilerImpl().
const BackendCapabilities gpuFsaCapabilities("GpuFsa", { {"NonConstWeights", false}, {"AsyncExecution", false}, {"ProtectedContentAllocation", false}, {"ConstantTensorsAsInputs", true}, {"PreImportIOTensors", false}, {"ExternallyManagedMemory", false}, {"MultiAxisPacking", false}, {"SingleAxisPacking", false} }) |
Referenced by GpuFsaBackend::GetCapabilities().
|
constexpr |
|
constexpr |
Definition at line 31 of file Types.hpp.
Referenced by Concatenate(), CopyTensorContentsGeneric(), PermutationVector::PermutationVector(), armnnUtils::Permuted(), Split(), Splitter(), TosaTensorNumDimensionsWithinBounds::TosaTensorNumDimensionsWithinBounds(), armnnDeserializer::ToTensorInfo(), and armnnUtils::TransposeTensorShape().
const std::set<armnn::BackendCapability> oldCpuRefCapabilities |
Definition at line 25 of file RefBackend.hpp.
const std::set<armnn::LayerType> paddingRequiredLayers |
Definition at line 16 of file NeonTensorHandleFactory.hpp.
Referenced by NeonTensorHandleFactory::GetCapabilities().
thread_local IProfiler* tl_Profiler = nullptr |
Definition at line 600 of file Profiling.cpp.
Referenced by ProfilerManager::GetProfiler(), and ProfilerManager::RegisterProfiler().
|
constexpr |
Definition at line 22 of file SingleAxisPriorityList.cpp.