1 /*
2  * Copyright (c) 2022 Huawei Device Co., Ltd.
3  * Licensed under the Apache License, Version 2.0 (the "License");
4  * you may not use this file except in compliance with the License.
5  * You may obtain a copy of the License at
6  *
7  *     http://www.apache.org/licenses/LICENSE-2.0
8  *
9  * Unless required by applicable law or agreed to in writing, software
10  * distributed under the License is distributed on an "AS IS" BASIS,
11  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12  * See the License for the specific language governing permissions and
13  * limitations under the License.
14  */
15 
16 #ifndef NEURAL_NETWORK_RUNTIME_HDI_DEVICE_V1_0_H
17 #define NEURAL_NETWORK_RUNTIME_HDI_DEVICE_V1_0_H
18 
19 #include <v1_0/nnrt_types.h>
20 #include <v1_0/innrt_device.h>
21 #include <v1_0/iprepared_model.h>
22 #include "refbase.h"
23 
24 #include "device.h"
25 
26 namespace OHOS {
27 namespace NeuralNetworkRuntime {
28 namespace V1_0 = OHOS::HDI::Nnrt::V1_0;
29 class HDIDeviceV1_0 : public Device {
30 public:
31     explicit HDIDeviceV1_0(OHOS::sptr<V1_0::INnrtDevice> device);
32 
33     OH_NN_ReturnCode GetDeviceName(std::string& name) override;
34     OH_NN_ReturnCode GetVendorName(std::string& name) override;
35     OH_NN_ReturnCode GetVersion(std::string& version) override;
36     OH_NN_ReturnCode GetDeviceType(OH_NN_DeviceType& deviceType) override;
37     OH_NN_ReturnCode GetDeviceStatus(DeviceStatus& status) override;
38     OH_NN_ReturnCode GetSupportedOperation(std::shared_ptr<const mindspore::lite::LiteGraph> model,
39                                            std::vector<bool>& ops) override;
40 
41     OH_NN_ReturnCode IsFloat16PrecisionSupported(bool& isSupported) override;
42     OH_NN_ReturnCode IsPerformanceModeSupported(bool& isSupported) override;
43     OH_NN_ReturnCode IsPrioritySupported(bool& isSupported) override;
44     OH_NN_ReturnCode IsDynamicInputSupported(bool& isSupported) override;
45     OH_NN_ReturnCode IsModelCacheSupported(bool& isSupported) override;
46 
47     OH_NN_ReturnCode PrepareModel(std::shared_ptr<const mindspore::lite::LiteGraph> model,
48                                   const ModelConfig& config,
49                                   std::shared_ptr<PreparedModel>& preparedModel) override;
50     OH_NN_ReturnCode PrepareModel(const void* metaGraph,
51                                   const ModelConfig& config,
52                                   std::shared_ptr<PreparedModel>& preparedModel) override;
53     OH_NN_ReturnCode PrepareModelFromModelCache(const std::vector<Buffer>& modelCache,
54                                                 const ModelConfig& config,
55                                                 std::shared_ptr<PreparedModel>& preparedModel,
56                                                 bool& isUpdatable) override;
57     OH_NN_ReturnCode PrepareOfflineModel(std::shared_ptr<const mindspore::lite::LiteGraph> model,
58                                          const ModelConfig& config,
59                                          std::shared_ptr<PreparedModel>& preparedModel) override;
60 
61     void* AllocateBuffer(size_t length) override;
62     void* AllocateTensorBuffer(size_t length, std::shared_ptr<TensorDesc> tensor) override;
63     void* AllocateTensorBuffer(size_t length, std::shared_ptr<NNTensor> tensor) override;
64     OH_NN_ReturnCode ReleaseBuffer(const void* buffer) override;
65 
66     OH_NN_ReturnCode AllocateBuffer(size_t length, int& fd) override;
67     OH_NN_ReturnCode ReleaseBuffer(int fd, size_t length) override;
68 
69 private:
70     OH_NN_ReturnCode ReleaseSharedBuffer(const V1_0::SharedBuffer& buffer);
71 
72 private:
73     // first: major version, second: minor version
74     std::pair<uint32_t, uint32_t> m_hdiVersion;
75     OHOS::sptr<V1_0::INnrtDevice> m_iDevice {nullptr};
76 };
77 } // namespace NeuralNetworkRuntime
78 } // namespace OHOS
79 #endif // NEURAL_NETWORK_RUNTIME_HDI_DEVICE_V1_0_H
80