FastDeploy
latest
Fast & Easy to Deploy!
|
Option object to configure Paddle Inference backend. More...
#include <option.h>
Public Member Functions | |
void | DisableTrtOps (const std::vector< std::string > &ops) |
Disable type of operators run on TensorRT. | |
void | DeletePass (const std::string &pass_name) |
Delete pass by name. | |
Public Attributes | |
bool | enable_log_info = false |
Print log information while initialize Paddle Inference backend. | |
bool | enable_mkldnn = true |
Enable MKLDNN while inference on CPU. | |
bool | enable_trt = false |
Use Paddle Inference + TensorRT to inference model on GPU. | |
bool | enable_memory_optimize = true |
Whether enable memory optimize, default true. | |
bool | switch_ir_debug = false |
Whether enable ir debug, default false. | |
bool | collect_trt_shape = false |
Collect shape for model while enable_trt is true. | |
int | mkldnn_cache_size = -1 |
Cache input shape for mkldnn while the input data will change dynamiclly. | |
int | gpu_mem_init_size = 100 |
initialize memory size(MB) for GPU | |
bool | enable_fixed_size_opt = false |
The option to enable fixed size optimization for transformer model. | |
Option object to configure Paddle Inference backend.