Lines Matching refs:operands

75                         RunTimeOperandInfo* operands) {  in updateForArguments()  argument
80 RunTimeOperandInfo& to = operands[operandIndex]; in updateForArguments()
111 const size_t count = subgraph.operands.size(); in initializeRunTimeInfo()
112 std::vector<RunTimeOperandInfo> operands(count); in initializeRunTimeInfo() local
114 const V1_3::Operand& from = subgraph.operands[i]; in initializeRunTimeInfo()
115 RunTimeOperandInfo& to = operands[i]; in initializeRunTimeInfo()
149 return operands; in initializeRunTimeInfo()
157 std::vector<RunTimeOperandInfo>& operands, in Create() argument
168 /*external_value_ids=*/operands.size(), /*flags=*/0, &subgraphPtr); in Create()
180 std::vector<int> tensors(operands.size(), -1); in Create()
201 if (isScalarType(operands[ins[k]].type)) continue; in Create()
207 if (isScalarType(operands[outs[k]].type)) continue; in Create()
214 std::vector<uint32_t> xnnpackTensors(operands.size()); in Create()
217 if (operands[tensors[t]].type != OperandType::TENSOR_FLOAT32) { in Create()
224 if (operands[tensors[t]].lifetime == Operand::LifeTime::CONSTANT_COPY || in Create()
225 operands[tensors[t]].lifetime == Operand::LifeTime::CONSTANT_REFERENCE || in Create()
226 operands[tensors[t]].lifetime == Operand::LifeTime::POINTER) { in Create()
227 data = operands[tensors[t]].buffer; in Create()
239 std::vector<size_t> dims(operands[tensors[t]].dimensions.size()); in Create()
241 dims[i] = operands[tensors[t]].dimensions[i]; in Create()
255 if (VisitNode(subgraph.get(), operation, operands.data(), xnnpackTensors) != in Create()
273 V1_3::ErrorStatus Invoke(RunTimeOperandInfo* operands) { in Invoke() argument
282 value.data = operands[t].buffer; in Invoke()
497 RunTimeOperandInfo* operands, in VisitNode() argument
501 return VisitAbsNode(subgraph, operation, operands, xnnpackTensors); in VisitNode()
503 return VisitAddNode(subgraph, operation, operands, xnnpackTensors); in VisitNode()
505 return VisitAveragePool2DNode(subgraph, operation, operands, xnnpackTensors); in VisitNode()
507 return VisitConv2DNode(subgraph, operation, operands, xnnpackTensors); in VisitNode()
509 return VisitDepthwiseConv2DNode(subgraph, operation, operands, xnnpackTensors); in VisitNode()
511 return VisitDivNode(subgraph, operation, operands, xnnpackTensors); in VisitNode()
513 return VisitFloorNode(subgraph, operation, operands, xnnpackTensors); in VisitNode()
515 return VisitFullyConnectedNode(subgraph, operation, operands, xnnpackTensors); in VisitNode()
517 return VisitHardSwishNode(subgraph, operation, operands, xnnpackTensors); in VisitNode()
519 return VisitLogisticNode(subgraph, operation, operands, xnnpackTensors); in VisitNode()
521 return VisitMaxPool2DNode(subgraph, operation, operands, xnnpackTensors); in VisitNode()
523 return VisitMaximumNode(subgraph, operation, operands, xnnpackTensors); in VisitNode()
525 return VisitMeanNode(subgraph, operation, operands, xnnpackTensors); in VisitNode()
527 return VisitMinimumNode(subgraph, operation, operands, xnnpackTensors); in VisitNode()
529 return VisitMulNode(subgraph, operation, operands, xnnpackTensors); in VisitNode()
531 return VisitNegNode(subgraph, operation, operands, xnnpackTensors); in VisitNode()
533 return VisitPadNode(subgraph, operation, operands, 0.0f, xnnpackTensors); in VisitNode()
535 return VisitPadV2Node(subgraph, operation, operands, xnnpackTensors); in VisitNode()
537 return VisitReshapeNode(subgraph, operation, operands, xnnpackTensors); in VisitNode()
539 return VisitResizeBilinearNode(subgraph, operation, operands, xnnpackTensors); in VisitNode()
541 return VisitPreluNode(subgraph, operation, operands, xnnpackTensors); in VisitNode()
543 return VisitReluNode(subgraph, operation, operands, 0.0f, in VisitNode()
546 return VisitReluNode(subgraph, operation, operands, -1.0f, 1.0f, xnnpackTensors); in VisitNode()
548 return VisitReluNode(subgraph, operation, operands, 0.0f, 6.0f, xnnpackTensors); in VisitNode()
550 return VisitSqrtNode(subgraph, operation, operands, xnnpackTensors); in VisitNode()
552 return VisitSubNode(subgraph, operation, operands, xnnpackTensors); in VisitNode()
554 return VisitSoftmaxNode(subgraph, operation, operands, xnnpackTensors); in VisitNode()
561 RunTimeOperandInfo* operands, in VisitAbsNode() argument
565 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[0]].type)); in VisitAbsNode()
566 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[outs[0]].type)); in VisitAbsNode()
581 RunTimeOperandInfo* operands, in VisitAddNode() argument
585 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[0]].type)); in VisitAddNode()
586 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[1]].type)); in VisitAddNode()
587 NN_DRIVER_RETURN_IF_ERROR(CheckTensorStaticAllocation(operands[ins[2]].lifetime)); in VisitAddNode()
588 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[outs[0]].type)); in VisitAddNode()
592 int activation = getScalarData<int32_t>(operands[ins[2]]); in VisitAddNode()
612 RunTimeOperandInfo* operands, in VisitAveragePool2DNode() argument
616 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[0]].type)); in VisitAveragePool2DNode()
617 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[outs[0]].type)); in VisitAveragePool2DNode()
620 NN_DRIVER_RETURN_IF_ERROR(CheckTensorStaticAllocation(operands[ins[i]].lifetime)); in VisitAveragePool2DNode()
625 use_nchw = getScalarData<bool>(operands[ins[7]]); in VisitAveragePool2DNode()
628 use_nchw = getScalarData<bool>(operands[ins[10]]); in VisitAveragePool2DNode()
643 input_padding_left = static_cast<uint32_t>(getScalarData<int32_t>(operands[ins[1]])); in VisitAveragePool2DNode()
644 input_padding_right = static_cast<uint32_t>(getScalarData<int32_t>(operands[ins[2]])); in VisitAveragePool2DNode()
645 input_padding_top = static_cast<uint32_t>(getScalarData<int32_t>(operands[ins[3]])); in VisitAveragePool2DNode()
646 input_padding_bottom = static_cast<uint32_t>(getScalarData<int32_t>(operands[ins[4]])); in VisitAveragePool2DNode()
647 stride_width = getScalarData<int32_t>(operands[ins[5]]); in VisitAveragePool2DNode()
648 stride_height = getScalarData<int32_t>(operands[ins[6]]); in VisitAveragePool2DNode()
649 filter_width = getScalarData<int32_t>(operands[ins[7]]); in VisitAveragePool2DNode()
650 filter_height = getScalarData<int32_t>(operands[ins[8]]); in VisitAveragePool2DNode()
651 activation = getScalarData<int32_t>(operands[ins[9]]); in VisitAveragePool2DNode()
654 int padding_implicit = getScalarData<int32_t>(operands[ins[1]]); in VisitAveragePool2DNode()
656 stride_width = getScalarData<int32_t>(operands[ins[2]]); in VisitAveragePool2DNode()
657 stride_height = getScalarData<int32_t>(operands[ins[3]]); in VisitAveragePool2DNode()
658 filter_width = getScalarData<int32_t>(operands[ins[4]]); in VisitAveragePool2DNode()
659 filter_height = getScalarData<int32_t>(operands[ins[5]]); in VisitAveragePool2DNode()
660 activation = getScalarData<int32_t>(operands[ins[6]]); in VisitAveragePool2DNode()
695 RunTimeOperandInfo* operands, in VisitConv2DNode() argument
699 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[0]].type)); in VisitConv2DNode()
700 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[1]].type)); in VisitConv2DNode()
701 NN_DRIVER_RETURN_IF_ERROR(CheckTensorStaticAllocation(operands[ins[1]].lifetime)); in VisitConv2DNode()
702 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[2]].type)); in VisitConv2DNode()
703 NN_DRIVER_RETURN_IF_ERROR(CheckTensorStaticAllocation(operands[ins[2]].lifetime)); in VisitConv2DNode()
704 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[outs[0]].type)); in VisitConv2DNode()
707 NN_DRIVER_RETURN_IF_ERROR(CheckTensorStaticAllocation(operands[ins[i]].lifetime)); in VisitConv2DNode()
711 if (ins.size() >= 8 && operands[ins[7]].type == OperandType::BOOL) { in VisitConv2DNode()
712 use_nchw = getScalarData<bool>(operands[ins[7]]); in VisitConv2DNode()
715 use_nchw = getScalarData<bool>(operands[ins[10]]); in VisitConv2DNode()
730 if (ins.size() >= 10 && operands[ins[7]].type != OperandType::BOOL) { in VisitConv2DNode()
732 input_padding_left = static_cast<uint32_t>(getScalarData<int32_t>(operands[ins[3]])); in VisitConv2DNode()
733 input_padding_right = static_cast<uint32_t>(getScalarData<int32_t>(operands[ins[4]])); in VisitConv2DNode()
734 input_padding_top = static_cast<uint32_t>(getScalarData<int32_t>(operands[ins[5]])); in VisitConv2DNode()
735 input_padding_bottom = static_cast<uint32_t>(getScalarData<int32_t>(operands[ins[6]])); in VisitConv2DNode()
736 stride_width = getScalarData<int32_t>(operands[ins[7]]); in VisitConv2DNode()
737 stride_height = getScalarData<int32_t>(operands[ins[8]]); in VisitConv2DNode()
738 activation = getScalarData<int32_t>(operands[ins[9]]); in VisitConv2DNode()
740 dilation_width_factor = getScalarData<int32_t>(operands[ins[11]]); in VisitConv2DNode()
741 dilation_height_factor = getScalarData<int32_t>(operands[ins[12]]); in VisitConv2DNode()
745 int padding_implicit = getScalarData<int32_t>(operands[ins[3]]); in VisitConv2DNode()
747 stride_width = getScalarData<int32_t>(operands[ins[4]]); in VisitConv2DNode()
748 stride_height = getScalarData<int32_t>(operands[ins[5]]); in VisitConv2DNode()
749 activation = getScalarData<int32_t>(operands[ins[6]]); in VisitConv2DNode()
751 dilation_width_factor = getScalarData<int32_t>(operands[ins[8]]); in VisitConv2DNode()
752 dilation_height_factor = getScalarData<int32_t>(operands[ins[9]]); in VisitConv2DNode()
763 const RunTimeOperandInfo& filter = operands[ins[1]]; in VisitConv2DNode()
794 RunTimeOperandInfo* operands, in VisitDepthwiseConv2DNode() argument
798 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[0]].type)); in VisitDepthwiseConv2DNode()
799 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[1]].type)); in VisitDepthwiseConv2DNode()
800 NN_DRIVER_RETURN_IF_ERROR(CheckTensorStaticAllocation(operands[ins[1]].lifetime)); in VisitDepthwiseConv2DNode()
801 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[2]].type)); in VisitDepthwiseConv2DNode()
802 NN_DRIVER_RETURN_IF_ERROR(CheckTensorStaticAllocation(operands[ins[2]].lifetime)); in VisitDepthwiseConv2DNode()
803 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[outs[0]].type)); in VisitDepthwiseConv2DNode()
806 NN_DRIVER_RETURN_IF_ERROR(CheckTensorStaticAllocation(operands[ins[i]].lifetime)); in VisitDepthwiseConv2DNode()
810 if (ins.size() >= 9 && operands[ins[8]].type == OperandType::BOOL) { in VisitDepthwiseConv2DNode()
811 use_nchw = getScalarData<bool>(operands[ins[8]]); in VisitDepthwiseConv2DNode()
814 use_nchw = getScalarData<bool>(operands[ins[11]]); in VisitDepthwiseConv2DNode()
830 if (ins.size() >= 11 && operands[ins[8]].type != OperandType::BOOL) { in VisitDepthwiseConv2DNode()
832 input_padding_left = static_cast<uint32_t>(getScalarData<int32_t>(operands[ins[3]])); in VisitDepthwiseConv2DNode()
833 input_padding_right = static_cast<uint32_t>(getScalarData<int32_t>(operands[ins[4]])); in VisitDepthwiseConv2DNode()
834 input_padding_top = static_cast<uint32_t>(getScalarData<int32_t>(operands[ins[5]])); in VisitDepthwiseConv2DNode()
835 input_padding_bottom = static_cast<uint32_t>(getScalarData<int32_t>(operands[ins[6]])); in VisitDepthwiseConv2DNode()
836 stride_width = getScalarData<int32_t>(operands[ins[7]]); in VisitDepthwiseConv2DNode()
837 stride_height = getScalarData<int32_t>(operands[ins[8]]); in VisitDepthwiseConv2DNode()
838 depth_multiplier = getScalarData<int32_t>(operands[ins[9]]); in VisitDepthwiseConv2DNode()
839 activation = getScalarData<int32_t>(operands[ins[10]]); in VisitDepthwiseConv2DNode()
841 dilation_width_factor = getScalarData<int32_t>(operands[ins[12]]); in VisitDepthwiseConv2DNode()
842 dilation_height_factor = getScalarData<int32_t>(operands[ins[13]]); in VisitDepthwiseConv2DNode()
846 int padding_implicit = getScalarData<int32_t>(operands[ins[3]]); in VisitDepthwiseConv2DNode()
848 stride_width = getScalarData<int32_t>(operands[ins[4]]); in VisitDepthwiseConv2DNode()
849 stride_height = getScalarData<int32_t>(operands[ins[5]]); in VisitDepthwiseConv2DNode()
850 depth_multiplier = getScalarData<int32_t>(operands[ins[6]]); in VisitDepthwiseConv2DNode()
851 activation = getScalarData<int32_t>(operands[ins[7]]); in VisitDepthwiseConv2DNode()
853 dilation_width_factor = getScalarData<int32_t>(operands[ins[9]]); in VisitDepthwiseConv2DNode()
854 dilation_height_factor = getScalarData<int32_t>(operands[ins[10]]); in VisitDepthwiseConv2DNode()
862 const RunTimeOperandInfo& filter = operands[ins[1]]; in VisitDepthwiseConv2DNode()
894 RunTimeOperandInfo* operands, in VisitDivNode() argument
898 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[0]].type)); in VisitDivNode()
899 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[1]].type)); in VisitDivNode()
900 NN_DRIVER_RETURN_IF_ERROR(CheckTensorStaticAllocation(operands[ins[2]].lifetime)); in VisitDivNode()
901 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[outs[0]].type)); in VisitDivNode()
905 int activation = getScalarData<int32_t>(operands[ins[2]]); in VisitDivNode()
925 RunTimeOperandInfo* operands, in VisitFullyConnectedNode() argument
929 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[0]].type)); in VisitFullyConnectedNode()
930 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[1]].type)); in VisitFullyConnectedNode()
931 NN_DRIVER_RETURN_IF_ERROR(CheckTensorStaticAllocation(operands[ins[1]].lifetime)); in VisitFullyConnectedNode()
932 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[2]].type)); in VisitFullyConnectedNode()
933 NN_DRIVER_RETURN_IF_ERROR(CheckTensorStaticAllocation(operands[ins[2]].lifetime)); in VisitFullyConnectedNode()
934 NN_DRIVER_RETURN_IF_ERROR(CheckTensorStaticAllocation(operands[ins[3]].lifetime)); in VisitFullyConnectedNode()
935 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[outs[0]].type)); in VisitFullyConnectedNode()
939 int activation = getScalarData<int32_t>(operands[ins[3]]); in VisitFullyConnectedNode()
961 RunTimeOperandInfo* operands, in VisitFloorNode() argument
965 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[0]].type)); in VisitFloorNode()
966 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[outs[0]].type)); in VisitFloorNode()
983 RunTimeOperandInfo* operands, in VisitHardSwishNode() argument
987 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[0]].type)); in VisitHardSwishNode()
988 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[outs[0]].type)); in VisitHardSwishNode()
1004 RunTimeOperandInfo* operands, in VisitLogisticNode() argument
1008 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[0]].type)); in VisitLogisticNode()
1009 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[outs[0]].type)); in VisitLogisticNode()
1025 RunTimeOperandInfo* operands, in VisitMaxPool2DNode() argument
1029 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[0]].type)); in VisitMaxPool2DNode()
1030 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[outs[0]].type)); in VisitMaxPool2DNode()
1033 NN_DRIVER_RETURN_IF_ERROR(CheckTensorStaticAllocation(operands[ins[i]].lifetime)); in VisitMaxPool2DNode()
1038 use_nchw = getScalarData<bool>(operands[ins[7]]); in VisitMaxPool2DNode()
1041 use_nchw = getScalarData<bool>(operands[ins[10]]); in VisitMaxPool2DNode()
1056 input_padding_left = static_cast<uint32_t>(getScalarData<int32_t>(operands[ins[1]])); in VisitMaxPool2DNode()
1057 input_padding_right = static_cast<uint32_t>(getScalarData<int32_t>(operands[ins[2]])); in VisitMaxPool2DNode()
1058 input_padding_top = static_cast<uint32_t>(getScalarData<int32_t>(operands[ins[3]])); in VisitMaxPool2DNode()
1059 input_padding_bottom = static_cast<uint32_t>(getScalarData<int32_t>(operands[ins[4]])); in VisitMaxPool2DNode()
1060 stride_width = getScalarData<int32_t>(operands[ins[5]]); in VisitMaxPool2DNode()
1061 stride_height = getScalarData<int32_t>(operands[ins[6]]); in VisitMaxPool2DNode()
1062 filter_width = getScalarData<int32_t>(operands[ins[7]]); in VisitMaxPool2DNode()
1063 filter_height = getScalarData<int32_t>(operands[ins[8]]); in VisitMaxPool2DNode()
1064 activation = getScalarData<int32_t>(operands[ins[9]]); in VisitMaxPool2DNode()
1067 int padding_implicit = getScalarData<int32_t>(operands[ins[1]]); in VisitMaxPool2DNode()
1069 stride_width = getScalarData<int32_t>(operands[ins[2]]); in VisitMaxPool2DNode()
1070 stride_height = getScalarData<int32_t>(operands[ins[3]]); in VisitMaxPool2DNode()
1071 filter_width = getScalarData<int32_t>(operands[ins[4]]); in VisitMaxPool2DNode()
1072 filter_height = getScalarData<int32_t>(operands[ins[5]]); in VisitMaxPool2DNode()
1073 activation = getScalarData<int32_t>(operands[ins[6]]); in VisitMaxPool2DNode()
1109 RunTimeOperandInfo* operands, in VisitMaximumNode() argument
1113 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[0]].type)); in VisitMaximumNode()
1114 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[1]].type)); in VisitMaximumNode()
1115 NN_DRIVER_RETURN_IF_ERROR(CheckTensorStaticAllocation(operands[ins[2]].lifetime)); in VisitMaximumNode()
1116 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[outs[0]].type)); in VisitMaximumNode()
1120 int activation = getScalarData<int32_t>(operands[ins[2]]); in VisitMaximumNode()
1140 RunTimeOperandInfo* operands, in VisitMeanNode() argument
1144 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[0]].type)); in VisitMeanNode()
1145 NN_DRIVER_RETURN_IF_ERROR(CheckTensorShape(operands[ins[0]].dimensions, 4)); in VisitMeanNode()
1146 NN_DRIVER_RETURN_IF_ERROR(CheckAxesTensorShape(operands[ins[1]].dimensions)); in VisitMeanNode()
1147 NN_DRIVER_RETURN_IF_ERROR(CheckTensorStaticAllocation(operands[ins[1]].lifetime)); in VisitMeanNode()
1148 NN_DRIVER_RETURN_IF_ERROR(CheckTensorStaticAllocation(operands[ins[2]].lifetime)); in VisitMeanNode()
1149 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[outs[0]].type)); in VisitMeanNode()
1150 NN_DRIVER_RETURN_IF_ERROR(CheckTensorShape(operands[outs[0]].dimensions, 4)); in VisitMeanNode()
1152 int keep_dims = getScalarData<int32_t>(operands[ins[2]]); in VisitMeanNode()
1157 const int32_t* axes_buffer = reinterpret_cast<const int32_t*>(operands[ins[1]].buffer); in VisitMeanNode()
1158 if (operands[ins[1]].dimensions[0] != 2) { in VisitMeanNode()
1184 RunTimeOperandInfo* operands, in VisitMinimumNode() argument
1188 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[0]].type)); in VisitMinimumNode()
1189 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[1]].type)); in VisitMinimumNode()
1190 NN_DRIVER_RETURN_IF_ERROR(CheckTensorStaticAllocation(operands[ins[2]].lifetime)); in VisitMinimumNode()
1191 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[outs[0]].type)); in VisitMinimumNode()
1195 int activation = getScalarData<int32_t>(operands[ins[2]]); in VisitMinimumNode()
1214 RunTimeOperandInfo* operands, in VisitMulNode() argument
1218 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[0]].type)); in VisitMulNode()
1219 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[1]].type)); in VisitMulNode()
1220 NN_DRIVER_RETURN_IF_ERROR(CheckTensorStaticAllocation(operands[ins[2]].lifetime)); in VisitMulNode()
1221 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[outs[0]].type)); in VisitMulNode()
1223 int activation = getScalarData<int32_t>(operands[ins[2]]); in VisitMulNode()
1244 RunTimeOperandInfo* operands, in VisitNegNode() argument
1248 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[0]].type)); in VisitNegNode()
1249 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[outs[0]].type)); in VisitNegNode()
1266 RunTimeOperandInfo* operands, in VisitPreluNode() argument
1270 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[0]].type)); in VisitPreluNode()
1272 CheckTensorShape(operands[ins[0]].dimensions, 1, XNN_MAX_TENSOR_DIMS)); in VisitPreluNode()
1273 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[1]].type)); in VisitPreluNode()
1274 NN_DRIVER_RETURN_IF_ERROR(CheckSlopeTensorShape(operands[ins[1]].dimensions)); in VisitPreluNode()
1275 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[outs[0]].type)); in VisitPreluNode()
1277 CheckTensorShape(operands[outs[0]].dimensions, 1, XNN_MAX_TENSOR_DIMS)); in VisitPreluNode()
1293 RunTimeOperandInfo* operands, float padding_value, in VisitPadNode() argument
1297 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[0]].type)); in VisitPadNode()
1299 CheckTensorShape(operands[ins[0]].dimensions, 1, XNN_MAX_TENSOR_DIMS)); in VisitPadNode()
1300 NN_DRIVER_RETURN_IF_ERROR(CheckTensorStaticAllocation(operands[ins[1]].lifetime)); in VisitPadNode()
1301 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[outs[0]].type)); in VisitPadNode()
1303 CheckTensorShape(operands[outs[0]].dimensions, 1, XNN_MAX_TENSOR_DIMS)); in VisitPadNode()
1305 const int32_t* paddings_data = reinterpret_cast<const int32_t*>(operands[ins[1]].buffer); in VisitPadNode()
1306 for (size_t i = 0; i < operands[ins[1]].dimensions.size() * 2; i++) { in VisitPadNode()
1312 for (size_t i = 0; i < operands[ins[1]].dimensions.size(); i++) { in VisitPadNode()
1330 RunTimeOperandInfo* operands, in VisitPadV2Node() argument
1333 if (operands[ins[2]].type != OperandType::FLOAT32) { in VisitPadV2Node()
1336 float padding_value = getScalarData<float>(operands[ins[2]]); in VisitPadV2Node()
1337 return VisitPadNode(subgraph, operation, operands, padding_value, xnnpackTensors); in VisitPadV2Node()
1342 RunTimeOperandInfo* operands, in VisitReshapeNode() argument
1346 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[0]].type)); in VisitReshapeNode()
1348 CheckTensorShape(operands[ins[0]].dimensions, 0, XNN_MAX_TENSOR_DIMS)); in VisitReshapeNode()
1349 NN_DRIVER_RETURN_IF_ERROR(CheckTensorStaticAllocation(operands[ins[1]].lifetime)); in VisitReshapeNode()
1350 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[outs[0]].type)); in VisitReshapeNode()
1352 CheckTensorShape(operands[outs[0]].dimensions, 0, XNN_MAX_TENSOR_DIMS)); in VisitReshapeNode()
1356 for (uint32_t i = 0; i < operands[outs[0]].dimensions.size(); i++) { in VisitReshapeNode()
1357 new_shape[i] = static_cast<size_t>(operands[outs[0]].dimensions[i]); in VisitReshapeNode()
1360 subgraph, static_cast<size_t>(operands[outs[0]].dimensions.size()), in VisitReshapeNode()
1374 RunTimeOperandInfo* operands, in VisitResizeBilinearNode() argument
1378 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[0]].type)); in VisitResizeBilinearNode()
1379 NN_DRIVER_RETURN_IF_ERROR(CheckTensorShape(operands[ins[0]].dimensions, 4)); in VisitResizeBilinearNode()
1380 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[outs[0]].type)); in VisitResizeBilinearNode()
1381 NN_DRIVER_RETURN_IF_ERROR(CheckTensorShape(operands[outs[0]].dimensions, 4)); in VisitResizeBilinearNode()
1384 NN_DRIVER_RETURN_IF_ERROR(CheckTensorStaticAllocation(operands[ins[i]].lifetime)); in VisitResizeBilinearNode()
1388 bool use_nchw = getScalarData<bool>(operands[ins[3]]); in VisitResizeBilinearNode()
1397 if (operands[ins[1]].type == OperandType::INT32) { in VisitResizeBilinearNode()
1399 new_width = static_cast<size_t>(getScalarData<int32_t>(operands[ins[1]])); in VisitResizeBilinearNode()
1400 new_height = static_cast<size_t>(getScalarData<int32_t>(operands[ins[2]])); in VisitResizeBilinearNode()
1401 } else if (operands[ins[1]].type == OperandType::FLOAT32) { in VisitResizeBilinearNode()
1403 float width_scale = getScalarData<float>(operands[ins[1]]); in VisitResizeBilinearNode()
1404 float height_scale = getScalarData<float>(operands[ins[2]]); in VisitResizeBilinearNode()
1408 new_height = static_cast<size_t>(operands[ins[0]].dimensions[1] * height_scale); in VisitResizeBilinearNode()
1409 new_width = static_cast<size_t>(operands[ins[0]].dimensions[2] * width_scale); in VisitResizeBilinearNode()
1417 align_corners = getScalarData<bool>(operands[ins[4]]); in VisitResizeBilinearNode()
1418 half_pixel_centers = getScalarData<bool>(operands[ins[5]]); in VisitResizeBilinearNode()
1444 RunTimeOperandInfo* operands, float outputMin, in VisitReluNode() argument
1449 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[0]].type)); in VisitReluNode()
1450 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[outs[0]].type)); in VisitReluNode()
1467 RunTimeOperandInfo* operands, in VisitSqrtNode() argument
1471 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[0]].type)); in VisitSqrtNode()
1472 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[outs[0]].type)); in VisitSqrtNode()
1488 RunTimeOperandInfo* operands, in VisitSubNode() argument
1492 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[0]].type)); in VisitSubNode()
1493 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[1]].type)); in VisitSubNode()
1494 NN_DRIVER_RETURN_IF_ERROR(CheckTensorStaticAllocation(operands[ins[2]].lifetime)); in VisitSubNode()
1495 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[outs[0]].type)); in VisitSubNode()
1499 int activation = getScalarData<int32_t>(operands[ins[2]]); in VisitSubNode()
1519 RunTimeOperandInfo* operands, in VisitSoftmaxNode() argument
1523 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[ins[0]].type)); in VisitSoftmaxNode()
1524 NN_DRIVER_RETURN_IF_ERROR(CheckTensorStaticAllocation(operands[ins[1]].lifetime)); in VisitSoftmaxNode()
1525 NN_DRIVER_RETURN_IF_ERROR(CheckTensorFloatType(operands[outs[0]].type)); in VisitSoftmaxNode()
1527 float beta = getScalarData<float>(operands[ins[1]]); in VisitSoftmaxNode()
1533 NN_DRIVER_RETURN_IF_ERROR(CheckTensorStaticAllocation(operands[ins[2]].lifetime)); in VisitSoftmaxNode()
1534 int axis = getScalarData<int32_t>(operands[ins[2]]); in VisitSoftmaxNode()
1644 void asyncExecuteXNNPACK(Subgraph* subgraph, RunTimeOperandInfo* operands, in asyncExecuteXNNPACK() argument
1653 updateForArguments(model.main.inputIndexes, request.inputs, requestPoolInfos, operands); in asyncExecuteXNNPACK()
1654 updateForArguments(model.main.outputIndexes, request.outputs, requestPoolInfos, operands); in asyncExecuteXNNPACK()
1655 auto status = subgraph->Invoke(operands); in asyncExecuteXNNPACK()
1667 V1_3::ErrorStatus executeXNNPACKBase(Subgraph* subgraph, RunTimeOperandInfo* operands, in executeXNNPACKBase() argument
1691 std::thread([&subgraph, &operands, &model, request, measure, deadline, loopTimeoutDuration, in executeXNNPACKBase()
1693 asyncExecuteXNNPACK(subgraph, operands, request, measure, model, deadline, in executeXNNPACKBase()
1729 executeSynchronouslyXNNPACKBase(Subgraph* subgraph, RunTimeOperandInfo* operands, in executeSynchronouslyXNNPACKBase() argument
1748 updateForArguments(model.main.inputIndexes, request.inputs, requestPoolInfos, operands); in executeSynchronouslyXNNPACKBase()
1749 updateForArguments(model.main.outputIndexes, request.outputs, requestPoolInfos, operands); in executeSynchronouslyXNNPACKBase()
1751 auto status = subgraph->Invoke(operands); in executeSynchronouslyXNNPACKBase()
1958 auto operands = initializeRunTimeInfo(model.main, poolInfos, &model.operandValues); in getSupportedOperationsImpl() local
1964 if (Subgraph::VisitNode(/*subgraph=*/nullptr, operation, operands.data(), {}) == in getSupportedOperationsImpl()