Backends should implement their own CreateWorkload function with a switch statement.
250{
251 switch(type)
252 {
253 case LayerType::Activation :
254 {
255 auto activationQueueDescriptor = PolymorphicDowncast<const ActivationQueueDescriptor*>(&descriptor);
256 return MakeWorkload<ClActivationWorkload>(*activationQueueDescriptor, info, m_CLCompileContext);
257 }
258 case LayerType::Addition :
259 {
260 auto additionQueueDescriptor = PolymorphicDowncast<const AdditionQueueDescriptor*>(&descriptor);
261 return MakeWorkload<ClAdditionWorkload>(*additionQueueDescriptor, info, m_CLCompileContext);
262 }
263 case LayerType::ArgMinMax :
264 {
265 auto argMinMaxQueueDescriptor = PolymorphicDowncast<const ArgMinMaxQueueDescriptor*>(&descriptor);
266 return MakeWorkload<ClArgMinMaxWorkload>(*argMinMaxQueueDescriptor, info, m_CLCompileContext);
267 }
268 case LayerType::BatchMatMul :
269 {
270 auto batchMatMulQueueDescriptor = PolymorphicDowncast<const BatchMatMulQueueDescriptor*>(&descriptor);
271 return std::make_unique<ClBatchMatMulWorkload>(*batchMatMulQueueDescriptor, info, m_CLCompileContext);
272 }
273 case LayerType::BatchNormalization :
274 {
275 auto batchNormalizationQueueDescriptor
276 = PolymorphicDowncast<const BatchNormalizationQueueDescriptor*>(&descriptor);
277 return MakeWorkload<ClBatchNormalizationFloatWorkload, NullWorkload>
278 (*batchNormalizationQueueDescriptor, info, m_CLCompileContext);
279 }
280 case LayerType::BatchToSpaceNd :
281 {
282 auto batchToSpaceNdQueueDescriptor
283 = PolymorphicDowncast<const BatchToSpaceNdQueueDescriptor*>(&descriptor);
284 return MakeWorkload<ClBatchToSpaceNdWorkload>(*batchToSpaceNdQueueDescriptor, info, m_CLCompileContext);
285 }
286 case LayerType::Cast :
287 {
288 auto castQueueDescriptor = PolymorphicDowncast<const CastQueueDescriptor*>(&descriptor);
289 return MakeWorkload<ClCastWorkload>(*castQueueDescriptor, info, m_CLCompileContext);
290 }
291 case LayerType::ChannelShuffle :
292 {
293 auto channelShuffleQueueDescriptor
294 = PolymorphicDowncast<const ChannelShuffleQueueDescriptor*>(&descriptor);
295 return MakeWorkload<ClChannelShuffleWorkload>(*channelShuffleQueueDescriptor, info, m_CLCompileContext);
296 }
297 case LayerType::Comparison :
298 {
299 auto comparisonQueueDescriptor = PolymorphicDowncast<const ComparisonQueueDescriptor*>(&descriptor);
300 return MakeWorkload<ClComparisonWorkload>(*comparisonQueueDescriptor, info, m_CLCompileContext);
301 }
302 case LayerType::Concat :
303 {
304 auto concatQueueDescriptor = PolymorphicDowncast<const ConcatQueueDescriptor*>(&descriptor);
305 return MakeWorkload<ClConcatWorkload>(*concatQueueDescriptor, info, m_CLCompileContext);
306 }
307 case LayerType::Constant :
308 {
309 auto constantQueueDescriptor = PolymorphicDowncast<const ConstantQueueDescriptor*>(&descriptor);
310 return MakeWorkload<ClConstantWorkload>(*constantQueueDescriptor, info, m_CLCompileContext);
311 }
312 case LayerType::ConvertFp16ToFp32 :
313 {
314 auto convertFp16ToFp32QueueDescriptor
315 = PolymorphicDowncast<const ConvertFp16ToFp32QueueDescriptor*>(&descriptor);
316 return MakeWorkload<ClConvertFp16ToFp32Workload>(*convertFp16ToFp32QueueDescriptor,
317 info,
318 m_CLCompileContext);
319 }
320 case LayerType::ConvertFp32ToFp16 :
321 {
322 auto convertFp32ToFp16QueueDescriptor
323 = PolymorphicDowncast<const ConvertFp32ToFp16QueueDescriptor*>(&descriptor);
324 return MakeWorkload<ClConvertFp32ToFp16Workload>(*convertFp32ToFp16QueueDescriptor,
325 info,
326 m_CLCompileContext);
327 }
328 case LayerType::Convolution2d :
329 {
330 auto convolution2dQueueDescriptor = PolymorphicDowncast<const Convolution2dQueueDescriptor*>(&descriptor);
331 bool isFastMathEnabled = false;
332 if (m_ModelContextPtr)
333 {
334 if (m_ModelContextPtr.get() != nullptr)
335 {
336 auto modelOptions = dynamic_cast<ClBackendModelContext*>(m_ModelContextPtr.get());
337 if (modelOptions)
338 {
339 isFastMathEnabled = modelOptions->IsFastMathEnabled();
340 }
341 }
342 }
343 return MakeWorkload<ClConvolution2dWorkload>(*convolution2dQueueDescriptor,
344 info,
345 m_MemoryManager->GetIntraLayerManager(),
346 m_CLCompileContext,
347 isFastMathEnabled);
348 }
349 case LayerType::Convolution3d :
350 {
351 auto convolution3dQueueDescriptor = PolymorphicDowncast<const Convolution3dQueueDescriptor*>(&descriptor);
352 bool isFastMathEnabled = false;
353 if (m_ModelContextPtr)
354 {
355 if (m_ModelContextPtr.get() != nullptr)
356 {
357 auto modelOptions = dynamic_cast<ClBackendModelContext*>(m_ModelContextPtr.get());
358 if (modelOptions)
359 {
360 isFastMathEnabled = modelOptions->IsFastMathEnabled();
361 }
362 }
363 }
364 return MakeWorkload<ClConvolution3dWorkload>(*convolution3dQueueDescriptor,
365 info,
366 m_MemoryManager->GetIntraLayerManager(),
367 m_CLCompileContext,
368 isFastMathEnabled);
369 }
370 case LayerType::Debug :
371 {
372 auto debugQueueDescriptor = PolymorphicDowncast<const DebugQueueDescriptor*>(&descriptor);
373 return MakeWorkload<NullWorkload, NullWorkload>(*debugQueueDescriptor, info, m_CLCompileContext);
374 }
375 case LayerType::DepthToSpace :
376 {
377 auto depthToSpaceQueueDescriptor = PolymorphicDowncast<const DepthToSpaceQueueDescriptor*>(&descriptor);
378 return MakeWorkload<ClDepthToSpaceWorkload>(*depthToSpaceQueueDescriptor, info, m_CLCompileContext);
379 }
380 case LayerType::DepthwiseConvolution2d :
381 {
382 auto depthwiseConvolution2dQueueDescriptor
383 = PolymorphicDowncast<const DepthwiseConvolution2dQueueDescriptor*>(&descriptor);
384 return MakeWorkload<ClDepthwiseConvolutionWorkload>(*depthwiseConvolution2dQueueDescriptor,
385 info,
386 m_CLCompileContext);
387 }
388 case LayerType::Dequantize :
389 {
390 auto dequantizeQueueDescriptor = PolymorphicDowncast<const DequantizeQueueDescriptor*>(&descriptor);
391 return MakeWorkload<ClDequantizeWorkload>(*dequantizeQueueDescriptor, info, m_CLCompileContext);
392 }
393 case LayerType::DetectionPostProcess :
394 {
395 auto detectionPostProcessQueueDescriptor
396 = PolymorphicDowncast<const DetectionPostProcessQueueDescriptor*>(&descriptor);
397 return MakeWorkload<NullWorkload, NullWorkload>(*detectionPostProcessQueueDescriptor,
398 info,
399 m_CLCompileContext);
400 }
401 case LayerType::Division :
402 {
403 auto divisionQueueDescriptor = PolymorphicDowncast<const DivisionQueueDescriptor*>(&descriptor);
404 return std::make_unique<ClDivisionWorkload>(*divisionQueueDescriptor, info, m_CLCompileContext);
405 }
406 case LayerType::ElementwiseBinary :
407 {
408 auto elementwiseBinaryQueueDescriptor
409 = PolymorphicDowncast<const ElementwiseBinaryQueueDescriptor*>(&descriptor);
410 switch (elementwiseBinaryQueueDescriptor->m_Parameters.m_Operation)
411 {
412 case BinaryOperation::Add:
413 {
414 AdditionQueueDescriptor additionQueueDescriptor;
415 additionQueueDescriptor.m_Inputs = descriptor.m_Inputs;
416 additionQueueDescriptor.m_Outputs = descriptor.m_Outputs;
417 additionQueueDescriptor.m_AdditionalInfoObject =
418 elementwiseBinaryQueueDescriptor->m_AdditionalInfoObject;
419 return std::make_unique<ClAdditionWorkload>(additionQueueDescriptor, info, m_CLCompileContext);
420 }
421 case BinaryOperation::Div:
422 {
423 DivisionQueueDescriptor divisionQueueDescriptor;
424 divisionQueueDescriptor.m_Inputs = descriptor.m_Inputs;
425 divisionQueueDescriptor.m_Outputs = descriptor.m_Outputs;
426 divisionQueueDescriptor.m_AdditionalInfoObject =
427 elementwiseBinaryQueueDescriptor->m_AdditionalInfoObject;
428 return std::make_unique<ClDivisionWorkload>(divisionQueueDescriptor, info, m_CLCompileContext);
429 }
430 case BinaryOperation::FloorDiv:
431 {
432 DivisionQueueDescriptor divisionQueueDescriptor;
433 divisionQueueDescriptor.m_Inputs = descriptor.m_Inputs;
434 divisionQueueDescriptor.m_Outputs = descriptor.m_Outputs;
435 divisionQueueDescriptor.m_AdditionalInfoObject =
436 elementwiseBinaryQueueDescriptor->m_AdditionalInfoObject;
437 return std::make_unique<ClFloorDivWorkload>(divisionQueueDescriptor, info, m_CLCompileContext);
438 }
439 case BinaryOperation::Maximum:
440 {
441 MaximumQueueDescriptor maximumQueueDescriptor;
442 maximumQueueDescriptor.m_Inputs = descriptor.m_Inputs;
443 maximumQueueDescriptor.m_Outputs = descriptor.m_Outputs;
444 maximumQueueDescriptor.m_AdditionalInfoObject =
445 elementwiseBinaryQueueDescriptor->m_AdditionalInfoObject;
446 return std::make_unique<ClMaximumWorkload>(maximumQueueDescriptor, info, m_CLCompileContext);
447 }
448 case BinaryOperation::Minimum:
449 {
450 MinimumQueueDescriptor minimumQueueDescriptor;
451 minimumQueueDescriptor.m_Inputs = descriptor.m_Inputs;
452 minimumQueueDescriptor.m_Outputs = descriptor.m_Outputs;
453 minimumQueueDescriptor.m_AdditionalInfoObject =
454 elementwiseBinaryQueueDescriptor->m_AdditionalInfoObject;
455 return std::make_unique<ClMinimumWorkload>(minimumQueueDescriptor, info, m_CLCompileContext);
456 }
457 case BinaryOperation::Mul:
458 {
459 MultiplicationQueueDescriptor multiplicationQueueDescriptor;
460 multiplicationQueueDescriptor.m_Inputs = descriptor.m_Inputs;
461 multiplicationQueueDescriptor.m_Outputs = descriptor.m_Outputs;
462 multiplicationQueueDescriptor.m_AdditionalInfoObject =
463 elementwiseBinaryQueueDescriptor->m_AdditionalInfoObject;
464 return std::make_unique<ClMultiplicationWorkload>(multiplicationQueueDescriptor,
465 info,
466 m_CLCompileContext);
467 }
468 case BinaryOperation::Power:
469 case BinaryOperation::SqDiff:
470 {
471 return std::make_unique<ClElementwiseBinaryWorkload>(*elementwiseBinaryQueueDescriptor,
472 info,
473 m_CLCompileContext);
474 }
475 case BinaryOperation::Sub:
476 {
477 SubtractionQueueDescriptor subtractionQueueDescriptor;
478 subtractionQueueDescriptor.m_Inputs = descriptor.m_Inputs;
479 subtractionQueueDescriptor.m_Outputs = descriptor.m_Outputs;
480 subtractionQueueDescriptor.m_AdditionalInfoObject =
481 elementwiseBinaryQueueDescriptor->m_AdditionalInfoObject;
482 return std::make_unique<ClSubtractionWorkload>(subtractionQueueDescriptor,
483 info,
484 m_CLCompileContext);
485 }
486 default:
487 return nullptr;
488 }
489 }
490 case LayerType::ElementwiseUnary :
491 {
492 auto elementwiseUnaryQueueDescriptor
493 = PolymorphicDowncast<const ElementwiseUnaryQueueDescriptor*>(&descriptor);
494 switch(elementwiseUnaryQueueDescriptor->m_Parameters.m_Operation)
495 {
496 case UnaryOperation::Abs:
497 {
498 AbsQueueDescriptor absQueueDescriptor;
499 absQueueDescriptor.m_Inputs = elementwiseUnaryQueueDescriptor->m_Inputs;
500 absQueueDescriptor.m_Outputs = elementwiseUnaryQueueDescriptor->m_Outputs;
501 return std::make_unique<ClAbsWorkload>(absQueueDescriptor, info, m_CLCompileContext);
502 }
503 case UnaryOperation::Exp:
504 return std::make_unique<ClExpWorkload>(*elementwiseUnaryQueueDescriptor, info, m_CLCompileContext);
505 case UnaryOperation::Log:
506 return std::make_unique<ClLogWorkload>(*elementwiseUnaryQueueDescriptor, info, m_CLCompileContext);
507 case UnaryOperation::LogicalNot:
508 return std::make_unique<ClLogicalNotWorkload>(*elementwiseUnaryQueueDescriptor,
509 info,
510 m_CLCompileContext);
511 case UnaryOperation::Neg:
512 return std::make_unique<ClNegWorkload>(*elementwiseUnaryQueueDescriptor, info, m_CLCompileContext);
513 case UnaryOperation::Rsqrt:
514 {
515 RsqrtQueueDescriptor rsqrtQueueDescriptor;
516 rsqrtQueueDescriptor.m_Inputs = elementwiseUnaryQueueDescriptor->m_Inputs;
517 rsqrtQueueDescriptor.m_Outputs = elementwiseUnaryQueueDescriptor->m_Outputs;
518 return std::make_unique<ClRsqrtWorkload>(rsqrtQueueDescriptor, info, m_CLCompileContext);
519 }
520 case UnaryOperation::Sin:
521 return std::make_unique<ClSinWorkload>(*elementwiseUnaryQueueDescriptor, info, m_CLCompileContext);
522 case UnaryOperation::Sqrt:
523 return std::make_unique<ClSqrtWorkload>(*elementwiseUnaryQueueDescriptor, info, m_CLCompileContext);
524 default:
525 return nullptr;
526 }
527 }
528 case LayerType::Fill :
529 {
530 auto fillQueueDescriptor = PolymorphicDowncast<const FillQueueDescriptor*>(&descriptor);
531 return std::make_unique<ClFillWorkload>(*fillQueueDescriptor, info, m_CLCompileContext);
532 }
533 case LayerType::Floor :
534 {
535 auto floorQueueDescriptor = PolymorphicDowncast<const FloorQueueDescriptor*>(&descriptor);
536 return MakeWorkload<ClFloorFloatWorkload, NullWorkload>(*floorQueueDescriptor, info, m_CLCompileContext);
537 }
538 case LayerType::FullyConnected :
539 {
540 auto fullyConnectedQueueDescriptor
541 = PolymorphicDowncast<const FullyConnectedQueueDescriptor*>(&descriptor);
542 return MakeWorkload<ClFullyConnectedWorkload>(*fullyConnectedQueueDescriptor,
543 info,
544 m_MemoryManager->GetIntraLayerManager(),
545 m_CLCompileContext);
546 }
547 case LayerType::Gather :
548 {
549 auto gatherQueueDescriptor = PolymorphicDowncast<const GatherQueueDescriptor*>(&descriptor);
550 return MakeWorkload<ClGatherWorkload>(*gatherQueueDescriptor, info, m_CLCompileContext);
551 }
552 case LayerType::GatherNd :
553 {
554 auto gatherNdQueueDescriptor = PolymorphicDowncast<const GatherNdQueueDescriptor*>(&descriptor);
555 return MakeWorkload<ClGatherNdWorkload>(*gatherNdQueueDescriptor, info, m_CLCompileContext);
556 }
557 case LayerType::Input :
558 {
559 auto inputQueueDescriptor = PolymorphicDowncast<const InputQueueDescriptor*>(&descriptor);
560 return std::make_unique<CopyMemGenericWorkload>(*inputQueueDescriptor, info);
561 }
562 case LayerType::InstanceNormalization :
563 {
564 auto instanceNormalizationQueueDescriptor
565 = PolymorphicDowncast<const InstanceNormalizationQueueDescriptor*>(&descriptor);
566 return MakeWorkload<ClInstanceNormalizationWorkload>(*instanceNormalizationQueueDescriptor,
567 info,
568 m_CLCompileContext);
569 }
570 case LayerType::L2Normalization :
571 {
572 auto l2NormalizationQueueDescriptor
573 = PolymorphicDowncast<const L2NormalizationQueueDescriptor*>(&descriptor);
574 return MakeWorkload<ClL2NormalizationFloatWorkload, NullWorkload>(*l2NormalizationQueueDescriptor,
575 info,
576 m_CLCompileContext);
577 }
578 case LayerType::LogicalBinary :
579 {
580 auto logicalBinaryQueueDescriptor = PolymorphicDowncast<const LogicalBinaryQueueDescriptor*>(&descriptor);
581 switch(logicalBinaryQueueDescriptor->m_Parameters.m_Operation)
582 {
583 case LogicalBinaryOperation::LogicalAnd:
584 return std::make_unique<ClLogicalAndWorkload>(*logicalBinaryQueueDescriptor,
585 info,
586 m_CLCompileContext);
587 case LogicalBinaryOperation::LogicalOr:
588 return std::make_unique<ClLogicalOrWorkload>(*logicalBinaryQueueDescriptor,
589 info,
590 m_CLCompileContext);
591 default:
592 return nullptr;
593 }
594 }
595 case LayerType::LogSoftmax :
596 {
597 auto logSoftmaxQueueDescriptor = PolymorphicDowncast<const LogSoftmaxQueueDescriptor*>(&descriptor);
598 return MakeWorkload<ClLogSoftmaxWorkload>(*logSoftmaxQueueDescriptor,
599 info,
600 m_MemoryManager->GetIntraLayerManager(),
601 m_CLCompileContext);
602 }
603 case LayerType::Lstm :
604 {
605 auto lstmQueueDescriptor = PolymorphicDowncast<const LstmQueueDescriptor*>(&descriptor);
606 return MakeWorkload<ClLstmFloatWorkload, NullWorkload>(*lstmQueueDescriptor, info, m_CLCompileContext);
607 }
608 case LayerType::Maximum :
609 {
610 auto maximumQueueDescriptor = PolymorphicDowncast<const MaximumQueueDescriptor*>(&descriptor);
611 return MakeWorkload<ClMaximumWorkload>(*maximumQueueDescriptor, info, m_CLCompileContext);
612 }
613 case LayerType::Mean :
614 {
615 auto meanQueueDescriptor = PolymorphicDowncast<const MeanQueueDescriptor*>(&descriptor);
616 return MakeWorkload<ClMeanWorkload>(*meanQueueDescriptor, info, m_CLCompileContext);
617 }
618 case LayerType::MemCopy :
619 {
620 auto memCopyQueueDescriptor = PolymorphicDowncast<const MemCopyQueueDescriptor*>(&descriptor);
621 if (memCopyQueueDescriptor->m_Inputs.empty() || !memCopyQueueDescriptor->m_Inputs[0])
622 {
623 throw InvalidArgumentException("ClWorkloadFactory: Invalid null input for MemCopy workload");
624 }
625 return MakeWorkload<CopyMemGenericWorkload>(*memCopyQueueDescriptor, info);
626 }
627 case LayerType::MemImport :
628 {
629 auto memImportQueueDescriptor = PolymorphicDowncast<const MemImportQueueDescriptor*>(&descriptor);
630 if (memImportQueueDescriptor->m_Inputs.empty() || !memImportQueueDescriptor->m_Inputs[0])
631 {
632 throw InvalidArgumentException("ClWorkloadFactory: Invalid null input for MemImport workload");
633 }
634 return std::make_unique<ImportMemGenericWorkload>(*memImportQueueDescriptor, info);
635 }
636 case LayerType::Minimum :
637 {
638 auto minimumQueueDescriptor = PolymorphicDowncast<const MinimumQueueDescriptor*>(&descriptor);
639 return MakeWorkload<ClMinimumWorkload>(*minimumQueueDescriptor, info, m_CLCompileContext);
640 }
641 case LayerType::Multiplication :
642 {
643 auto multiplicationQueueDescriptor = PolymorphicDowncast<const MultiplicationQueueDescriptor*>(&descriptor);
644 return MakeWorkload<ClMultiplicationWorkload>(*multiplicationQueueDescriptor, info, m_CLCompileContext);
645 }
646 case LayerType::Normalization :
647 {
648 auto normalizationQueueDescriptor = PolymorphicDowncast<const NormalizationQueueDescriptor*>(&descriptor);
649 return MakeWorkload<ClNormalizationFloatWorkload, NullWorkload>(*normalizationQueueDescriptor,
650 info,
651 m_CLCompileContext);
652 }
653 case LayerType::Output :
654 {
655 auto outputQueueDescriptor = PolymorphicDowncast<const OutputQueueDescriptor*>(&descriptor);
656 return std::make_unique<CopyMemGenericWorkload>(*outputQueueDescriptor, info);
657 }
658 case LayerType::Pad :
659 {
660 auto padQueueDescriptor = PolymorphicDowncast<const PadQueueDescriptor*>(&descriptor);
661 return MakeWorkload<ClPadWorkload>(*padQueueDescriptor, info, m_CLCompileContext);
662 }
663 case LayerType::Permute :
664 {
665 auto permuteQueueDescriptor = PolymorphicDowncast<const PermuteQueueDescriptor*>(&descriptor);
666 return MakeWorkload<ClPermuteWorkload>(*permuteQueueDescriptor, info, m_CLCompileContext);
667 }
668 case LayerType::Pooling2d :
669 {
670 auto pooling2dQueueDescriptor = PolymorphicDowncast<const Pooling2dQueueDescriptor*>(&descriptor);
671 return MakeWorkload<ClPooling2dWorkload>(*pooling2dQueueDescriptor, info, m_CLCompileContext);
672 }
673 case LayerType::Pooling3d :
674 {
675 auto pooling3dQueueDescriptor = PolymorphicDowncast<const Pooling3dQueueDescriptor*>(&descriptor);
676 return MakeWorkload<ClPooling3dWorkload>(*pooling3dQueueDescriptor, info, m_CLCompileContext);
677 }
678 case LayerType::PreCompiled :
679 {
680 auto preCompiledQueueDescriptor = PolymorphicDowncast<const PreCompiledQueueDescriptor*>(&descriptor);
681 return MakeWorkload<NullWorkload, NullWorkload>(*preCompiledQueueDescriptor, info, m_CLCompileContext);
682 }
683 case LayerType::Prelu :
684 {
685 auto preluQueueDescriptor = PolymorphicDowncast<const PreluQueueDescriptor*>(&descriptor);
686 return MakeWorkload<ClPreluWorkload>(*preluQueueDescriptor, info, m_CLCompileContext);
687 }
688 case LayerType::QLstm :
689 {
690 auto qLstmQueueDescriptor = PolymorphicDowncast<const QLstmQueueDescriptor*>(&descriptor);
691 return std::make_unique<ClQLstmWorkload>(*qLstmQueueDescriptor, info, m_CLCompileContext);
692 }
693 case LayerType::Quantize :
694 {
695 auto quantizeQueueDescriptor = PolymorphicDowncast<const QuantizeQueueDescriptor*>(&descriptor);
696 return MakeWorkload<ClQuantizeWorkload>(*quantizeQueueDescriptor, info, m_CLCompileContext);
697 }
698 case LayerType::QuantizedLstm :
699 {
700 auto quantizedLstmQueueDescriptor = PolymorphicDowncast<const QuantizedLstmQueueDescriptor*>(&descriptor);
701 return MakeWorkload<ClQuantizedLstmWorkload>(*quantizedLstmQueueDescriptor, info, m_CLCompileContext);
702 }
703 case LayerType::Rank :
704 {
705 auto rankQueueDescriptor = PolymorphicDowncast<const RankQueueDescriptor*>(&descriptor);
706 return std::make_unique<ClRankWorkload>(*rankQueueDescriptor, info);
707 }
708 case LayerType::Reduce :
709 {
710 auto reduceQueueDescriptor = PolymorphicDowncast<const ReduceQueueDescriptor*>(&descriptor);
711 return std::make_unique<ClReduceWorkload>(*reduceQueueDescriptor, info);
712 }
713 case LayerType::Reshape :
714 {
715 auto reshapeQueueDescriptor = PolymorphicDowncast<const ReshapeQueueDescriptor*>(&descriptor);
716 return MakeWorkload<ClReshapeWorkload>(*reshapeQueueDescriptor, info, m_CLCompileContext);
717 }
718 case LayerType::Resize :
719 {
720 auto resizeQueueDescriptor = PolymorphicDowncast<const ResizeQueueDescriptor*>(&descriptor);
721 return MakeWorkload<ClResizeWorkload>(*resizeQueueDescriptor, info, m_CLCompileContext);
722 }
723 case LayerType::ReverseV2:
724 {
725 auto reverseV2QueueDescriptor = PolymorphicDowncast<const ReverseV2QueueDescriptor*>(&descriptor);
726 return MakeWorkload<ClReverseV2Workload>(*reverseV2QueueDescriptor, info, m_CLCompileContext);
727 }
728 case LayerType::ScatterNd :
729 {
730 auto scatterNdQueueDescriptor = PolymorphicDowncast<const ScatterNdQueueDescriptor*>(&descriptor);
731 return MakeWorkload<ClScatterNdWorkload>(*scatterNdQueueDescriptor, info, m_CLCompileContext);
732 }
733 case LayerType::Slice :
734 {
735 auto sliceQueueDescriptor = PolymorphicDowncast<const SliceQueueDescriptor*>(&descriptor);
736 return MakeWorkload<ClSliceWorkload>(*sliceQueueDescriptor, info, m_CLCompileContext);
737 }
738 case LayerType::Softmax :
739 {
740 auto softmaxQueueDescriptor = PolymorphicDowncast<const SoftmaxQueueDescriptor*>(&descriptor);
741 return std::make_unique<ClSoftmaxWorkload>(*softmaxQueueDescriptor,
742 info,
743 m_MemoryManager->GetIntraLayerManager(),
744 m_CLCompileContext);
745 }
746 case LayerType::SpaceToBatchNd :
747 {
748 auto spaceToBatchNdQueueDescriptor
749 = PolymorphicDowncast<const SpaceToBatchNdQueueDescriptor*>(&descriptor);
750 return MakeWorkload<ClSpaceToBatchNdWorkload>(*spaceToBatchNdQueueDescriptor, info, m_CLCompileContext);
751 }
752 case LayerType::SpaceToDepth :
753 {
754 auto spaceToDepthQueueDescriptor = PolymorphicDowncast<const SpaceToDepthQueueDescriptor*>(&descriptor);
755 return MakeWorkload<ClSpaceToDepthWorkload>(*spaceToDepthQueueDescriptor, info, m_CLCompileContext);
756 }
757 case LayerType::Splitter :
758 {
759 auto splitterQueueDescriptor = PolymorphicDowncast<const SplitterQueueDescriptor*>(&descriptor);
760 return MakeWorkload<ClSplitterWorkload>(*splitterQueueDescriptor, info, m_CLCompileContext);
761 }
762 case LayerType::Stack :
763 {
764 auto stackQueueDescriptor = PolymorphicDowncast<const StackQueueDescriptor*>(&descriptor);
765 return MakeWorkload<ClStackWorkload>(*stackQueueDescriptor, info, m_CLCompileContext);
766 }
767 case LayerType::StridedSlice :
768 {
769 auto stridedSliceQueueDescriptor = PolymorphicDowncast<const StridedSliceQueueDescriptor*>(&descriptor);
770 return MakeWorkload<ClStridedSliceWorkload>(*stridedSliceQueueDescriptor, info, m_CLCompileContext);
771 }
772 case LayerType::Subtraction :
773 {
774 auto subtractionQueueDescriptor = PolymorphicDowncast<const SubtractionQueueDescriptor*>(&descriptor);
775 return MakeWorkload<ClSubtractionWorkload>(*subtractionQueueDescriptor, info, m_CLCompileContext);
776 }
777 case LayerType::Tile:
778 {
779 auto tileQueueDescriptor = PolymorphicDowncast<const TileQueueDescriptor*>(&descriptor);
780 return MakeWorkload<ClTileWorkload>(*tileQueueDescriptor, info, m_CLCompileContext);
781 }
782 case LayerType::Transpose :
783 {
784 auto transposeQueueDescriptor = PolymorphicDowncast<const TransposeQueueDescriptor*>(&descriptor);
785 return MakeWorkload<ClTransposeWorkload>(*transposeQueueDescriptor, info, m_CLCompileContext);
786 }
787 case LayerType::TransposeConvolution2d :
788 {
789 auto transposeConvolution2dQueueDescriptor
790 = PolymorphicDowncast<const TransposeConvolution2dQueueDescriptor*>(&descriptor);
791 return MakeWorkload<ClTransposeConvolution2dWorkload>(*transposeConvolution2dQueueDescriptor,
792 info,
793 m_MemoryManager->GetIntraLayerManager(),
794 m_CLCompileContext);
795 }
796 case LayerType::UnidirectionalSequenceLstm :
797 {
798 auto desc = PolymorphicDowncast<const UnidirectionalSequenceLstmQueueDescriptor*>(&descriptor);
799 return MakeWorkloadHelper<ClUnidirectionalSequenceLstmFloatWorkload, NullWorkload>(*desc,
800 info,
801 m_CLCompileContext);
802 }
803 default:
804 return nullptr;
805 }
806}