Lines Matching refs:requestPoolInfos

944                  std::vector<RunTimePoolInfo> requestPoolInfos,  in CpuExecution()  argument
948 kRequestPoolInfos(std::move(requestPoolInfos)), in CpuExecution()
1030 const std::vector<RunTimePoolInfo>& requestPoolInfos, const OptionalTimePoint& deadline, in computeOnCpu() argument
1040 int err = executor.run(model, request, modelPoolInfos, requestPoolInfos); in computeOnCpu()
1079 std::vector<RunTimePoolInfo> requestPoolInfos; in createCpuRequest() local
1080 requestPoolInfos.reserve(memories.size()); in createCpuRequest()
1083 requestPoolInfos.emplace_back(*poolInfo); in createCpuRequest()
1090 [&requestPoolInfos](const std::vector<ModelArgumentInfo>& argumentInfos) { in createCpuRequest()
1095 {.poolIndex = static_cast<uint32_t>(requestPoolInfos.size()), in createCpuRequest()
1099 requestPoolInfos.emplace_back(RunTimePoolInfo::createFromExistingBuffer( in createCpuRequest()
1111 return {ANEURALNETWORKS_NO_ERROR, std::move(request), std::move(requestPoolInfos)}; in createCpuRequest()
1132 std::vector<RunTimePoolInfo> requestPoolInfos; in execute() local
1133 std::tie(nCreateRequest, request, requestPoolInfos) = in execute()
1144 std::thread([this, &request, &requestPoolInfos, &deadline, &loopTimeoutDuration, &result] { in execute()
1145 result = computeOnCpu(mModel, request, mModelPoolInfos, requestPoolInfos, deadline, in execute()
1151 return computeOnCpu(mModel, request, mModelPoolInfos, requestPoolInfos, deadline, in execute()
1159 auto [nCreateRequest, request, requestPoolInfos] = createCpuRequest(inputs, outputs, memories); in createReusableExecution()
1164 *this, std::move(request), std::move(requestPoolInfos), loopTimeoutDuration); in createReusableExecution()