/aosp12/packages/modules/NeuralNetworks/driver/sample/ |
H A D | CanonicalPreparedModel.cpp | 41 std::vector<RunTimePoolInfo> requestPoolInfos; in createRunTimePoolInfos() local 43 requestPoolInfos.reserve(request.pools.size()); in createRunTimePoolInfos() 53 requestPoolInfos.push_back(std::move(*buffer)); in createRunTimePoolInfos() 65 requestPoolInfos.push_back(bufferWrapper->createRunTimePoolInfo()); in createRunTimePoolInfos() 69 return std::make_pair(std::move(requestPoolInfos), std::move(bufferWrappers)); in createRunTimePoolInfos() 154 const auto [requestPoolInfos, bufferWrappers] = in execute() 168 int n = executor.run(kModel, request, kPoolInfos, requestPoolInfos); in execute() 236 const auto [requestPoolInfos, bufferWrappers] = in executeFenced() 249 int n = executor.run(kModel, request, kPoolInfos, requestPoolInfos); in executeFenced()
|
H A D | SampleDriver.cpp | 384 std::vector<RunTimePoolInfo> requestPoolInfos; in createRunTimePoolInfos() local 386 requestPoolInfos.reserve(request.pools.size()); in createRunTimePoolInfos() 398 requestPoolInfos.push_back(std::move(*buffer)); in createRunTimePoolInfos() 469 const auto [poolStatus, requestPoolInfos, bufferWrappers] = in asyncExecute() 489 requestPoolInfos); in asyncExecute() 603 const auto [poolStatus, requestPoolInfos, bufferWrappers] = in executeSynchronouslyBase() 621 requestPoolInfos); in executeSynchronouslyBase() 715 const auto [poolStatus, requestPoolInfos, bufferWrappers] = in executeFenced() 734 requestPoolInfos); in executeFenced() 826 std::vector<RunTimePoolInfo> requestPoolInfos; in execute() local [all …]
|
H A D | SampleDriverFloatXNNPACK.cpp | 74 const std::vector<RunTimePoolInfo>& requestPoolInfos, in updateForArguments() argument 95 CHECK_LT(poolIndex, requestPoolInfos.size()); in updateForArguments() 96 auto& r = requestPoolInfos[poolIndex]; in updateForArguments() 1649 std::vector<RunTimePoolInfo> requestPoolInfos; in asyncExecuteXNNPACK() local 1659 for (auto& runtimeInfo : requestPoolInfos) { in asyncExecuteXNNPACK() 1744 std::vector<RunTimePoolInfo> requestPoolInfos; in executeSynchronouslyXNNPACKBase() local 1755 for (auto& runtimeInfo : requestPoolInfos) { in executeSynchronouslyXNNPACKBase() 1813 std::vector<RunTimePoolInfo> requestPoolInfos; in executeFenced() local 1817 updateForArguments(model->main.inputIndexes, request.inputs, requestPoolInfos, in executeFenced() 1819 updateForArguments(model->main.outputIndexes, request.outputs, requestPoolInfos, in executeFenced() [all …]
|
/aosp12/packages/modules/NeuralNetworks/runtime/ |
H A D | Manager.cpp | 944 std::vector<RunTimePoolInfo> requestPoolInfos, in CpuExecution() argument 948 kRequestPoolInfos(std::move(requestPoolInfos)), in CpuExecution() 1040 int err = executor.run(model, request, modelPoolInfos, requestPoolInfos); in computeOnCpu() 1079 std::vector<RunTimePoolInfo> requestPoolInfos; in createCpuRequest() local 1080 requestPoolInfos.reserve(memories.size()); in createCpuRequest() 1083 requestPoolInfos.emplace_back(*poolInfo); in createCpuRequest() 1090 [&requestPoolInfos](const std::vector<ModelArgumentInfo>& argumentInfos) { in createCpuRequest() 1111 return {ANEURALNETWORKS_NO_ERROR, std::move(request), std::move(requestPoolInfos)}; in createCpuRequest() 1132 std::vector<RunTimePoolInfo> requestPoolInfos; in execute() local 1133 std::tie(nCreateRequest, request, requestPoolInfos) = in execute() [all …]
|
/aosp12/packages/modules/NeuralNetworks/driver/sample_aidl/ |
H A D | SampleDriver.cpp | 310 std::vector<RunTimePoolInfo> requestPoolInfos; in createRunTimePoolInfos() local 312 requestPoolInfos.reserve(request.pools.size()); in createRunTimePoolInfos() 322 requestPoolInfos.push_back(std::move(*buffer)); in createRunTimePoolInfos() 336 requestPoolInfos.push_back(bufferWrapper->createRunTimePoolInfo()); in createRunTimePoolInfos() 343 return {aidl_hal::ErrorStatus::NONE, std::move(requestPoolInfos), std::move(bufferWrappers)}; in createRunTimePoolInfos() 426 const auto [poolStatus, requestPoolInfos, bufferWrappers] = in executeSynchronously() 442 int n = executor.run(model, request, mPoolInfos, requestPoolInfos); in executeSynchronously() 541 const auto [poolStatus, requestPoolInfos, bufferWrappers] = in executeFenced() 557 int n = executor.run(model, request, mPoolInfos, requestPoolInfos); in executeFenced()
|
/aosp12/packages/modules/NeuralNetworks/common/include/ |
H A D | CpuExecutor.h | 151 const std::vector<RunTimePoolInfo>& requestPoolInfos); 168 const std::vector<RunTimePoolInfo>& requestPoolInfos,
|
/aosp12/packages/modules/NeuralNetworks/common/ |
H A D | CpuExecutor.cpp | 560 const std::vector<RunTimePoolInfo>& requestPoolInfos) { in run() argument 573 updateForArguments(model.main.inputIndexes, request.inputs, requestPoolInfos, operands.data()); in run() 574 updateForArguments(model.main.outputIndexes, request.outputs, requestPoolInfos, in run() 581 for (auto& runtimeInfo : requestPoolInfos) { in run() 668 const std::vector<RunTimePoolInfo>& requestPoolInfos, in updateForArguments() argument 692 CHECK_LT(poolIndex, requestPoolInfos.size()); in updateForArguments() 693 auto& r = requestPoolInfos[poolIndex]; in updateForArguments()
|