diff --git a/test_benchmark/TensorFlow/ResNetV1/deploy_infer.py b/test_benchmark/TensorFlow/ResNetV1/deploy_infer.py index 5bedd8de4..ab79ef59f 100644 --- a/test_benchmark/TensorFlow/ResNetV1/deploy_infer.py +++ b/test_benchmark/TensorFlow/ResNetV1/deploy_infer.py @@ -4,7 +4,6 @@ import numpy as np import paddle -import paddle.fluid as fluid from paddle.inference import Config from paddle.inference import create_predictor diff --git a/test_benchmark/TensorFlow/ResNetV1/pd_infer.py b/test_benchmark/TensorFlow/ResNetV1/pd_infer.py index 4b639d86f..bb5d95afe 100644 --- a/test_benchmark/TensorFlow/ResNetV1/pd_infer.py +++ b/test_benchmark/TensorFlow/ResNetV1/pd_infer.py @@ -1,5 +1,4 @@ import paddle -import paddle.fluid as fluid import numpy as np import sys import os @@ -11,11 +10,8 @@ exe = paddle.static.Executor(paddle.CPUPlace()) # test dygraph - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_dygraph/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_dygraph/inference_model/model", executor=exe) data = np.load('../dataset/ResNetV1/input.npy') result = exe.run(prog, feed={inputs[0]: data}, fetch_list=outputs) diff --git a/test_benchmark/TensorFlow/ResNetV2/deploy_infer.py b/test_benchmark/TensorFlow/ResNetV2/deploy_infer.py index e3415d29e..b2f750b55 100644 --- a/test_benchmark/TensorFlow/ResNetV2/deploy_infer.py +++ b/test_benchmark/TensorFlow/ResNetV2/deploy_infer.py @@ -4,7 +4,6 @@ import numpy as np import paddle -import paddle.fluid as fluid from paddle.inference import Config from paddle.inference import create_predictor diff --git a/test_benchmark/TensorFlow/ResNetV2/pd_infer.py b/test_benchmark/TensorFlow/ResNetV2/pd_infer.py index d8bc9d8a2..f5445275b 100644 --- a/test_benchmark/TensorFlow/ResNetV2/pd_infer.py +++ b/test_benchmark/TensorFlow/ResNetV2/pd_infer.py @@ -1,5 +1,4 @@ import paddle -import paddle.fluid as fluid import numpy as np import sys import os @@ -11,11 +10,8 @@ exe = paddle.static.Executor(paddle.CPUPlace()) # test dygraph - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_dygraph/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_dygraph/inference_model/model", executor=exe) data = np.load('../dataset/ResNetV2/input.npy') result = exe.run(prog, feed={inputs[0]: data}, fetch_list=outputs) @@ -33,3 +29,5 @@ f.write("!!!!!Dygraph Failed\n") except: f.write("!!!!!Failed\n") + + raise diff --git a/test_benchmark/TensorFlow/RetinaFace/deploy_infer.py b/test_benchmark/TensorFlow/RetinaFace/deploy_infer.py index c4d68c1fb..79c05982e 100644 --- a/test_benchmark/TensorFlow/RetinaFace/deploy_infer.py +++ b/test_benchmark/TensorFlow/RetinaFace/deploy_infer.py @@ -4,7 +4,6 @@ import numpy as np import paddle -import paddle.fluid as fluid from paddle.inference import Config from paddle.inference import create_predictor diff --git a/test_benchmark/TensorFlow/RetinaFace/pd_infer.py b/test_benchmark/TensorFlow/RetinaFace/pd_infer.py index 8d75187c8..8e06fb8cd 100644 --- a/test_benchmark/TensorFlow/RetinaFace/pd_infer.py +++ b/test_benchmark/TensorFlow/RetinaFace/pd_infer.py @@ -1,5 +1,4 @@ import paddle -import paddle.fluid as fluid import numpy as np import pickle import sys @@ -12,11 +11,8 @@ exe = paddle.static.Executor(paddle.CPUPlace()) # test dygraph - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_dygraph/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_dygraph/inference_model/model", executor=exe) data = np.load('../dataset/RetinaFace/input.npy') result = exe.run(prog, feed={inputs[0]: data}, fetch_list=outputs) diff --git a/test_benchmark/TensorFlow/ShuffleNet/deploy_infer.py b/test_benchmark/TensorFlow/ShuffleNet/deploy_infer.py index bbaaa91e1..2c24eca6f 100644 --- a/test_benchmark/TensorFlow/ShuffleNet/deploy_infer.py +++ b/test_benchmark/TensorFlow/ShuffleNet/deploy_infer.py @@ -4,7 +4,6 @@ import numpy as np import paddle -import paddle.fluid as fluid from paddle.inference import Config from paddle.inference import create_predictor diff --git a/test_benchmark/TensorFlow/ShuffleNet/pd_infer.py b/test_benchmark/TensorFlow/ShuffleNet/pd_infer.py index 9a36b2579..12e4d870c 100644 --- a/test_benchmark/TensorFlow/ShuffleNet/pd_infer.py +++ b/test_benchmark/TensorFlow/ShuffleNet/pd_infer.py @@ -1,5 +1,4 @@ import paddle -import paddle.fluid as fluid import numpy as np import sys import os @@ -11,12 +10,8 @@ try: paddle.enable_static() exe = paddle.static.Executor(paddle.CPUPlace()) - - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_dygraph/inference_model", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_dygraph/inference_model/model", executor=exe) result = exe.run(prog, feed={inputs[0]: data}, fetch_list=outputs) diff = result[0] - tf_result diff --git a/test_benchmark/TensorFlow/SqueezeNet/deploy_infer.py b/test_benchmark/TensorFlow/SqueezeNet/deploy_infer.py index 5fa32a8ec..026ee48a6 100644 --- a/test_benchmark/TensorFlow/SqueezeNet/deploy_infer.py +++ b/test_benchmark/TensorFlow/SqueezeNet/deploy_infer.py @@ -4,7 +4,6 @@ import numpy as np import paddle -import paddle.fluid as fluid from paddle.inference import Config from paddle.inference import create_predictor diff --git a/test_benchmark/TensorFlow/SqueezeNet/pd_infer.py b/test_benchmark/TensorFlow/SqueezeNet/pd_infer.py index f1d20f4fe..9da5db643 100644 --- a/test_benchmark/TensorFlow/SqueezeNet/pd_infer.py +++ b/test_benchmark/TensorFlow/SqueezeNet/pd_infer.py @@ -1,5 +1,4 @@ import paddle -import paddle.fluid as fluid import numpy as np import sys import os @@ -11,11 +10,8 @@ exe = paddle.static.Executor(paddle.CPUPlace()) # test dygraph - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_dygraph/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_dygraph/inference_model/model", executor=exe) data = np.load('../dataset/SqueezeNet/input.npy') result = exe.run(prog, feed={inputs[0]: data}, fetch_list=outputs) diff --git a/test_benchmark/TensorFlow/ToyUNet-UNet/deploy_infer.py b/test_benchmark/TensorFlow/ToyUNet-UNet/deploy_infer.py index a2d68a87e..c7c3d8996 100644 --- a/test_benchmark/TensorFlow/ToyUNet-UNet/deploy_infer.py +++ b/test_benchmark/TensorFlow/ToyUNet-UNet/deploy_infer.py @@ -4,7 +4,6 @@ import numpy as np import paddle -import paddle.fluid as fluid from paddle.inference import Config from paddle.inference import create_predictor diff --git a/test_benchmark/TensorFlow/ToyUNet-UNet/pd_infer.py b/test_benchmark/TensorFlow/ToyUNet-UNet/pd_infer.py index 321d49ce3..4694a09d1 100644 --- a/test_benchmark/TensorFlow/ToyUNet-UNet/pd_infer.py +++ b/test_benchmark/TensorFlow/ToyUNet-UNet/pd_infer.py @@ -1,5 +1,4 @@ import paddle -import paddle.fluid as fluid import numpy as np import pickle import sys @@ -12,11 +11,8 @@ exe = paddle.static.Executor(paddle.CPUPlace()) # test dygraph - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_dygraph/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_dygraph/inference_model/model", executor=exe) with open('../dataset/ToyUNet-UNet/input.pkl', 'rb') as f1: data = pickle.load(f1) diff --git a/test_benchmark/TensorFlow/UNet-GanUNet/deploy_infer.py b/test_benchmark/TensorFlow/UNet-GanUNet/deploy_infer.py index 288df8d63..bdf013fe4 100644 --- a/test_benchmark/TensorFlow/UNet-GanUNet/deploy_infer.py +++ b/test_benchmark/TensorFlow/UNet-GanUNet/deploy_infer.py @@ -4,7 +4,6 @@ import numpy as np import paddle -import paddle.fluid as fluid from paddle.inference import Config from paddle.inference import create_predictor diff --git a/test_benchmark/TensorFlow/UNet-GanUNet/pd_infer.py b/test_benchmark/TensorFlow/UNet-GanUNet/pd_infer.py index e108bd152..544cae0c2 100644 --- a/test_benchmark/TensorFlow/UNet-GanUNet/pd_infer.py +++ b/test_benchmark/TensorFlow/UNet-GanUNet/pd_infer.py @@ -1,5 +1,4 @@ import paddle -import paddle.fluid as fluid import numpy as np import pickle import sys @@ -12,11 +11,8 @@ exe = paddle.static.Executor(paddle.CPUPlace()) # test dygraph - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_dygraph/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_dygraph/inference_model/model", executor=exe) data = [np.load('../dataset/UNet-GanUNet/input.npy')] result = exe.run(prog, feed={inputs[0]: data[0]}, fetch_list=outputs) print(result[0].shape, result[0].mean(), result[0].max(), result[0].min()) diff --git a/test_benchmark/TensorFlow/UNet-UNet/deploy_infer.py b/test_benchmark/TensorFlow/UNet-UNet/deploy_infer.py index 660863240..ffd77bead 100644 --- a/test_benchmark/TensorFlow/UNet-UNet/deploy_infer.py +++ b/test_benchmark/TensorFlow/UNet-UNet/deploy_infer.py @@ -4,7 +4,6 @@ import numpy as np import paddle -import paddle.fluid as fluid from paddle.inference import Config from paddle.inference import create_predictor diff --git a/test_benchmark/TensorFlow/UNet-UNet/pd_infer.py b/test_benchmark/TensorFlow/UNet-UNet/pd_infer.py index fdcd66204..d2497c9a7 100644 --- a/test_benchmark/TensorFlow/UNet-UNet/pd_infer.py +++ b/test_benchmark/TensorFlow/UNet-UNet/pd_infer.py @@ -1,5 +1,4 @@ import paddle -import paddle.fluid as fluid import numpy as np import pickle import sys @@ -12,11 +11,8 @@ exe = paddle.static.Executor(paddle.CPUPlace()) # test dygraph - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_dygraph/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_dygraph/inference_model/model", executor=exe) data = [np.load('../dataset/UNet-UNet/input.npy')] result = exe.run(prog, feed={inputs[0]: data[0]}, fetch_list=outputs) diff --git a/test_benchmark/TensorFlow/VGG16/deploy_infer.py b/test_benchmark/TensorFlow/VGG16/deploy_infer.py index db5d4bcdb..dbdedd5ec 100644 --- a/test_benchmark/TensorFlow/VGG16/deploy_infer.py +++ b/test_benchmark/TensorFlow/VGG16/deploy_infer.py @@ -4,7 +4,6 @@ import numpy as np import paddle -import paddle.fluid as fluid from paddle.inference import Config from paddle.inference import create_predictor diff --git a/test_benchmark/TensorFlow/VGG16/pd_infer.py b/test_benchmark/TensorFlow/VGG16/pd_infer.py index f8dd4f10e..a4e75f74b 100644 --- a/test_benchmark/TensorFlow/VGG16/pd_infer.py +++ b/test_benchmark/TensorFlow/VGG16/pd_infer.py @@ -1,5 +1,4 @@ import paddle -import paddle.fluid as fluid import numpy as np import sys import os @@ -11,11 +10,8 @@ exe = paddle.static.Executor(paddle.CPUPlace()) # test dygraph - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_dygraph/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_dygraph/inference_model/model", executor=exe) data = np.load('../dataset/VGG16/input.npy') result = exe.run(prog, feed={inputs[0]: data}, fetch_list=outputs) diff --git a/test_benchmark/TensorFlow/YOLOv3/deploy_infer.py b/test_benchmark/TensorFlow/YOLOv3/deploy_infer.py index 50ca9d64d..a97f7029b 100644 --- a/test_benchmark/TensorFlow/YOLOv3/deploy_infer.py +++ b/test_benchmark/TensorFlow/YOLOv3/deploy_infer.py @@ -4,7 +4,6 @@ import numpy as np import paddle -import paddle.fluid as fluid from paddle.inference import Config from paddle.inference import create_predictor diff --git a/test_benchmark/TensorFlow/YOLOv3/pd_infer.py b/test_benchmark/TensorFlow/YOLOv3/pd_infer.py index a5af74b20..d343bf0b6 100644 --- a/test_benchmark/TensorFlow/YOLOv3/pd_infer.py +++ b/test_benchmark/TensorFlow/YOLOv3/pd_infer.py @@ -1,5 +1,4 @@ import paddle -import paddle.fluid as fluid import numpy as np import pickle import sys @@ -12,11 +11,8 @@ exe = paddle.static.Executor(paddle.CPUPlace()) # test dygraph - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_dygraph/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_dygraph/inference_model/model", executor=exe) data = [np.load('../dataset/YOLOv3/input.npy')] result = exe.run(prog, feed={inputs[0]: data[0]}, fetch_list=outputs) diff --git a/test_benchmark/TensorFlow/YOLOv3_darknet/deploy_infer.py b/test_benchmark/TensorFlow/YOLOv3_darknet/deploy_infer.py index d88643da9..c2eb0079e 100644 --- a/test_benchmark/TensorFlow/YOLOv3_darknet/deploy_infer.py +++ b/test_benchmark/TensorFlow/YOLOv3_darknet/deploy_infer.py @@ -4,7 +4,6 @@ import numpy as np import paddle -import paddle.fluid as fluid from paddle.inference import Config from paddle.inference import create_predictor diff --git a/test_benchmark/TensorFlow/YOLOv3_darknet/pd_infer.py b/test_benchmark/TensorFlow/YOLOv3_darknet/pd_infer.py index 596b95517..10e5b5c64 100644 --- a/test_benchmark/TensorFlow/YOLOv3_darknet/pd_infer.py +++ b/test_benchmark/TensorFlow/YOLOv3_darknet/pd_infer.py @@ -1,5 +1,4 @@ import paddle -import paddle.fluid as fluid import numpy as np import pickle import sys @@ -14,11 +13,8 @@ exe = paddle.static.Executor(paddle.CPUPlace()) # test dygraph - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_dygraph/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_dygraph/inference_model/model", executor=exe) result = exe.run(prog, feed={inputs[0]: data}, fetch_list=outputs) diff = result[0] - tf_result diff --git a/test_benchmark/TensorFlow/black.list b/test_benchmark/TensorFlow/black.list index 49a3ebfab..a852ece08 100644 --- a/test_benchmark/TensorFlow/black.list +++ b/test_benchmark/TensorFlow/black.list @@ -1,19 +1,6 @@ EfficientNet KerasBert -ResNetV1 -ResNetV2 -RetinaFace -ShuffleNet -SqueezeNet ToyUNet-ToyUnet -ToyUNet-UNet -UNet-GanUNet -UNet-UNet -VGG16 -YOLOv3 -YOLOv3_darknet -frozen -frozen_MobileNetV1 pix2pix dataset tools diff --git a/test_benchmark/TensorFlow/frozen/deploy_infer.py b/test_benchmark/TensorFlow/frozen/deploy_infer.py index 9e9d30273..3a30d2404 100644 --- a/test_benchmark/TensorFlow/frozen/deploy_infer.py +++ b/test_benchmark/TensorFlow/frozen/deploy_infer.py @@ -4,7 +4,6 @@ import numpy as np import paddle -import paddle.fluid as fluid from paddle.inference import Config from paddle.inference import create_predictor diff --git a/test_benchmark/TensorFlow/frozen/pd_infer.py b/test_benchmark/TensorFlow/frozen/pd_infer.py index 051963754..82d0c89b5 100644 --- a/test_benchmark/TensorFlow/frozen/pd_infer.py +++ b/test_benchmark/TensorFlow/frozen/pd_infer.py @@ -1,5 +1,4 @@ import paddle -import paddle.fluid as fluid import numpy as np import numpy import sys @@ -17,11 +16,8 @@ exe = paddle.static.Executor(paddle.CPUPlace()) # test dygraph -[prog, inputs, outputs - ] = fluid.io.load_inference_model(dirname="pd_model_dygraph/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") +[prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_dygraph/inference_model/model", executor=exe) result = exe.run(prog, feed={ inputs[0]: input_data["ipt0"], diff --git a/test_benchmark/TensorFlow/frozen_MobileNetV1/deploy_infer.py b/test_benchmark/TensorFlow/frozen_MobileNetV1/deploy_infer.py index 4f72e28fb..ad03a4809 100644 --- a/test_benchmark/TensorFlow/frozen_MobileNetV1/deploy_infer.py +++ b/test_benchmark/TensorFlow/frozen_MobileNetV1/deploy_infer.py @@ -4,7 +4,6 @@ import numpy as np import paddle -import paddle.fluid as fluid from paddle.inference import Config from paddle.inference import create_predictor diff --git a/test_benchmark/TensorFlow/frozen_MobileNetV1/pd_infer.py b/test_benchmark/TensorFlow/frozen_MobileNetV1/pd_infer.py index c3b242894..b85416831 100644 --- a/test_benchmark/TensorFlow/frozen_MobileNetV1/pd_infer.py +++ b/test_benchmark/TensorFlow/frozen_MobileNetV1/pd_infer.py @@ -1,6 +1,5 @@ from __future__ import print_function import paddle -import paddle.fluid as fluid import sys import os import numpy as np @@ -14,11 +13,9 @@ input_data = np.random.rand(1, 224, 224, 3).astype("float32") # test dygrah -[inference_program, feed_target_names, fetch_targets - ] = fluid.io.load_inference_model(dirname="pd_model_dygraph/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") +[inference_program, feed_target_names, + fetch_targets] = paddle.static.load_inference_model( + path_prefix="pd_model_dygraph/inference_model/model", executor=exe) result = exe.run(inference_program, feed={feed_target_names[0]: input_data},