FastDeploy  latest
Fast & Easy to Deploy!
paddle_backend.h
1 // Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved.
2 //
3 // Licensed under the Apache License, Version 2.0 (the "License");
4 // you may not use this file except in compliance with the License.
5 // You may obtain a copy of the License at
6 //
7 // http://www.apache.org/licenses/LICENSE-2.0
8 //
9 // Unless required by applicable law or agreed to in writing, software
10 // distributed under the License is distributed on an "AS IS" BASIS,
11 // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 // See the License for the specific language governing permissions and
13 // limitations under the License.
14 
15 #pragma once
16 
17 #include <iostream>
18 #include <memory>
19 #include <string>
20 #include <vector>
21 
22 #include "fastdeploy/runtime/backends/backend.h"
23 #include "fastdeploy/runtime/backends/paddle/option.h"
24 #ifdef ENABLE_PADDLE2ONNX
25 #include "paddle2onnx/converter.h"
26 #endif
27 #include "fastdeploy/utils/unique_ptr.h"
28 #include "paddle_inference_api.h" // NOLINT
29 
30 namespace fastdeploy {
31 
32 // convert FD device to paddle place type
33 paddle_infer::PlaceType ConvertFDDeviceToPlace(Device device);
34 
35 // Share memory buffer with paddle_infer::Tensor from fastdeploy::FDTensor
36 void ShareTensorFromFDTensor(paddle_infer::Tensor* tensor, FDTensor& fd_tensor);
37 
38 void ShareOutTensorFromFDTensor(paddle_infer::Tensor* tensor,
39  FDTensor& fd_tensor);
40 
41 // convert paddle_infer::Tensor to fastdeploy::FDTensor
42 // if copy_to_fd is true, copy memory data to FDTensor
44 void PaddleTensorToFDTensor(std::unique_ptr<paddle_infer::Tensor>& tensor,
45  FDTensor* fd_tensor, bool copy_to_fd);
46 
47 // Convert data type from paddle inference to fastdeploy
48 FDDataType PaddleDataTypeToFD(const paddle_infer::DataType& dtype);
49 
50 // Convert data type from paddle2onnx::PaddleReader to fastdeploy
51 FDDataType ReaderDataTypeToFD(int32_t dtype);
52 
53 class PaddleBackend : public BaseBackend {
54  public:
55  PaddleBackend() {}
56  virtual ~PaddleBackend() = default;
57  bool Init(const RuntimeOption& option);
58  bool Infer(std::vector<FDTensor>& inputs, std::vector<FDTensor>* outputs,
59  bool copy_to_fd = true) override;
60 
61  int NumInputs() const override { return inputs_desc_.size(); }
62 
63  int NumOutputs() const override { return outputs_desc_.size(); }
64 
65  std::unique_ptr<BaseBackend> Clone(RuntimeOption &runtime_option,
66  void* stream = nullptr,
67  int device_id = -1) override;
68 
69  TensorInfo GetInputInfo(int index) override;
70  TensorInfo GetOutputInfo(int index) override;
71  std::vector<TensorInfo> GetInputInfos() override;
72  std::vector<TensorInfo> GetOutputInfos() override;
73 
74  private:
75  void BuildOption(const PaddleBackendOption& option);
76 
77  bool InitFromPaddle(const std::string& model,
78  const std::string& params,
79  bool model_from_memory,
80  const PaddleBackendOption& option = PaddleBackendOption());
81 
82  void
83  CollectShapeRun(paddle_infer::Predictor* predictor,
84  const std::map<std::string, std::vector<int>>& shape) const;
85  void GetDynamicShapeFromOption(
86  const PaddleBackendOption& option,
87  std::map<std::string, std::vector<int>>* max_shape,
88  std::map<std::string, std::vector<int>>* min_shape,
89  std::map<std::string, std::vector<int>>* opt_shape) const;
90  void SetTRTDynamicShapeToConfig(const PaddleBackendOption& option);
91  PaddleBackendOption option_;
92  paddle_infer::Config config_;
93  std::shared_ptr<paddle_infer::Predictor> predictor_;
94  std::vector<TensorInfo> inputs_desc_;
95  std::vector<TensorInfo> outputs_desc_;
96 };
97 } // namespace fastdeploy
void PaddleTensorToFDTensor(std::unique_ptr< paddle_infer::Tensor > &tensor, FDTensor *fd_tensor, bool copy_to_fd)
else share memory to FDTensor
Definition: util.cc:101
All C++ FastDeploy APIs are defined inside this namespace.
Definition: option.h:16