|
@@ -48,21 +48,23 @@ public class DemoModel implements Model {
|
|
|
|
|
|
@Override
|
|
|
public boolean loadFromStream(InputStream in) throws Exception {
|
|
|
- String modelDir = PropertiesUtil.getString("model.dir");
|
|
|
- CompressUtil.decompressGzFile(in, modelDir + "/demo");
|
|
|
+ String modelDir = PropertiesUtil.getString("model.dir") + "/demo";
|
|
|
+ CompressUtil.decompressGzFile(in, modelDir);
|
|
|
|
|
|
- String modelFile = "";
|
|
|
- String paramFile = "";
|
|
|
+ String modelFile = "inference.pdmodel";
|
|
|
+ String paramFile = "inference.pdiparams";
|
|
|
|
|
|
+ log.info("start init Config");
|
|
|
Config config = new Config();
|
|
|
- config.setCppModel(modelFile, paramFile);
|
|
|
+ config.setCppModel(modelDir + "/" + modelFile, modelDir + "/" + paramFile);
|
|
|
config.enableMemoryOptim(true);
|
|
|
config.enableProfile();
|
|
|
config.enableMKLDNN();
|
|
|
config.getCpuMathLibraryNumThreads();
|
|
|
config.getFractionOfGpuMemoryForPool();
|
|
|
config.switchIrDebug(false);
|
|
|
-
|
|
|
+
|
|
|
+ log.info("start init Predictor");
|
|
|
Predictor predictor = Predictor.createPaddlePredictor(config);
|
|
|
Predictor temp = predictor;
|
|
|
this.sourcePredictor = predictor;
|