From 5a6729fddf346234e89c2c3e944a3fa65af89a6e Mon Sep 17 00:00:00 2001 From: megemini Date: Sun, 13 Oct 2024 10:21:16 +0000 Subject: [PATCH 01/25] [Fix] version check --- requirements.txt | 1 + x2paddle/convert.py | 19 +++++++------------ x2paddle/utils.py | 29 ++++++++++++++++++++++------- 3 files changed, 30 insertions(+), 19 deletions(-) diff --git a/requirements.txt b/requirements.txt index ded0ee752..7127bf47d 100644 --- a/requirements.txt +++ b/requirements.txt @@ -1 +1,2 @@ sympy +packaging diff --git a/x2paddle/convert.py b/x2paddle/convert.py index 3372c5bc1..0a7e83a29 100644 --- a/x2paddle/convert.py +++ b/x2paddle/convert.py @@ -14,7 +14,7 @@ from six import text_type as _text_type from x2paddle import program -from x2paddle.utils import ConverterCheck +from x2paddle.utils import ConverterCheck, check_version import argparse import sys import logging @@ -449,20 +449,15 @@ def main(): assert args.save_dir is not None, "--save_dir is not defined" try: - import platform - v0, v1, v2 = platform.python_version().split('.') - if not (int(v0) >= 3 and int(v1) >= 5): - logging.info("[ERROR] python>=3.5 is required") + if not sys.version_info >= (3, 8): + logging.error("[ERROR] python>=3.8 is required") return + import paddle - v0, v1, v2 = paddle.__version__.split('.') - logging.info("paddle.__version__ = {}".format(paddle.__version__)) - if v0 == '0' and v1 == '0' and v2 == '0': - logging.info( - "[WARNING] You are use develop version of paddlepaddle") - elif int(v0) != 2 or int(v1) < 0: - logging.info("[ERROR] paddlepaddle>=2.0.0 is required") + if not check_version('2.0.0'): + logging.error("[ERROR] paddlepaddle>=2.0.0 is required") return + except: logging.info( "[ERROR] paddlepaddle not installed, use \"pip install paddlepaddle\"" diff --git a/x2paddle/utils.py b/x2paddle/utils.py index 2117a40c7..b51dda7f4 100644 --- a/x2paddle/utils.py +++ b/x2paddle/utils.py @@ -13,6 +13,10 @@ # See the License for the specific language governing permissions and # limitations under the License. +import logging + +from packaging.version import Version + import paddle import x2paddle import hashlib @@ -30,15 +34,26 @@ def string(param): return "\'{}\'".format(param) -def check_version(): - version = paddle.__version__ - v0, v1, v2 = version.split('.') - if not ((v0 == '0' and v1 == '0' and v2 == '0') or - (int(v0) >= 2 and int(v1) >= 1)): - return False - else: +def check_version(base_version: str = '2.1.0') -> bool: + """ + Return `True` if the current version is equal or bigger than `base_version`. + The default version `2.1.0` is used for checking `is_new_version`. + """ + is_new = False + + dev_version = Version('0.0.0') + cur_version = Version(paddle.__version__) + + if cur_version == dev_version: + logging.info("[WARNING] You are use develop version of paddlepaddle") + return True + if cur_version >= Version(base_version): + return True + + return False + def _md5(text: str): '''Calculate the md5 value of the input text.''' From 2b90922cd76493f029568d993633d0d0fb935ae1 Mon Sep 17 00:00:00 2001 From: megemini Date: Sun, 13 Oct 2024 18:29:53 +0800 Subject: [PATCH 02/25] [Fix] pytorch alexnet --- test_benchmark/PyTorch/AlexNet/pd_infer.py | 15 ++++----------- 1 file changed, 4 insertions(+), 11 deletions(-) diff --git a/test_benchmark/PyTorch/AlexNet/pd_infer.py b/test_benchmark/PyTorch/AlexNet/pd_infer.py index b041af0cf..d9aec6812 100644 --- a/test_benchmark/PyTorch/AlexNet/pd_infer.py +++ b/test_benchmark/PyTorch/AlexNet/pd_infer.py @@ -1,5 +1,4 @@ from __future__ import print_function -import paddle.fluid as fluid import paddle import sys import os @@ -16,11 +15,8 @@ # trace paddle.enable_static() exe = paddle.static.Executor(paddle.CPUPlace()) - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_trace/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_trace/inference_model/model", executor=exe) result = exe.run(prog, feed={inputs[0]: input_data}, fetch_list=outputs) df = pytorch_output - result if numpy.max(numpy.fabs(df)) > 1e-04: @@ -31,11 +27,8 @@ # script paddle.enable_static() exe = paddle.static.Executor(paddle.CPUPlace()) - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_script/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_script/inference_model/model", executor=exe) result = exe.run(prog, feed={inputs[0]: input_data}, fetch_list=outputs) df = pytorch_output - result if numpy.max(numpy.fabs(df)) > 1e-05: From afd0953cead7d441646e0816be5cd885a4fc0836 Mon Sep 17 00:00:00 2001 From: megemini Date: Sun, 13 Oct 2024 18:33:14 +0800 Subject: [PATCH 03/25] [Fix] pytorch basnet --- test_benchmark/PyTorch/BASNet/deploy_infer.py | 1 - test_benchmark/PyTorch/BASNet/pd_infer.py | 9 +++------ 2 files changed, 3 insertions(+), 7 deletions(-) diff --git a/test_benchmark/PyTorch/BASNet/deploy_infer.py b/test_benchmark/PyTorch/BASNet/deploy_infer.py index d86690068..76d951a12 100644 --- a/test_benchmark/PyTorch/BASNet/deploy_infer.py +++ b/test_benchmark/PyTorch/BASNet/deploy_infer.py @@ -4,7 +4,6 @@ import numpy as np import paddle -import paddle.fluid as fluid from paddle.inference import Config from paddle.inference import create_predictor diff --git a/test_benchmark/PyTorch/BASNet/pd_infer.py b/test_benchmark/PyTorch/BASNet/pd_infer.py index bf018c407..d46564325 100644 --- a/test_benchmark/PyTorch/BASNet/pd_infer.py +++ b/test_benchmark/PyTorch/BASNet/pd_infer.py @@ -1,5 +1,5 @@ from __future__ import print_function -import paddle.fluid as fluid + import paddle import sys import os @@ -16,11 +16,8 @@ # trace paddle.enable_static() exe = paddle.static.Executor(paddle.CPUPlace()) - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_trace/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_trace/inference_model/model", executor=exe) result = exe.run(prog, feed={inputs[0]: input_data}, fetch_list=outputs) result_string = "Trace Successed" for i in range(len(result)): From 82a29998296cf5f1ac5745c1feefbd5248a07eee Mon Sep 17 00:00:00 2001 From: megemini Date: Sun, 13 Oct 2024 19:32:35 +0800 Subject: [PATCH 04/25] [Fix] pytorch dbface --- test_benchmark/PyTorch/DBFace/deploy_infer.py | 1 - test_benchmark/PyTorch/DBFace/pd_infer.py | 9 +++------ 2 files changed, 3 insertions(+), 7 deletions(-) diff --git a/test_benchmark/PyTorch/DBFace/deploy_infer.py b/test_benchmark/PyTorch/DBFace/deploy_infer.py index fb2f0bde2..d53b6fc9f 100644 --- a/test_benchmark/PyTorch/DBFace/deploy_infer.py +++ b/test_benchmark/PyTorch/DBFace/deploy_infer.py @@ -5,7 +5,6 @@ import numpy as np import paddle -import paddle.fluid as fluid from paddle.inference import Config from paddle.inference import create_predictor diff --git a/test_benchmark/PyTorch/DBFace/pd_infer.py b/test_benchmark/PyTorch/DBFace/pd_infer.py index f6aeb5499..60f1671fb 100644 --- a/test_benchmark/PyTorch/DBFace/pd_infer.py +++ b/test_benchmark/PyTorch/DBFace/pd_infer.py @@ -1,5 +1,5 @@ from __future__ import print_function -import paddle.fluid as fluid + import paddle import common import numpy as np @@ -23,11 +23,8 @@ # trace paddle.enable_static() exe = paddle.static.Executor(paddle.CPUPlace()) - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_trace/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_trace/inference_model/model", executor=exe) result = exe.run(prog, feed={inputs[0]: image}, fetch_list=outputs) df0 = pytorch_output["hm"] - result[0] From 37dcbf13d297d604c79e8e6fb48c979b4ab0d25e Mon Sep 17 00:00:00 2001 From: megemini Date: Sun, 13 Oct 2024 19:41:05 +0800 Subject: [PATCH 05/25] [Fix] pytorch edsr --- test_benchmark/PyTorch/EDSR/deploy_infer.py | 1 - test_benchmark/PyTorch/EDSR/pd_infer.py | 16 +++++----------- 2 files changed, 5 insertions(+), 12 deletions(-) diff --git a/test_benchmark/PyTorch/EDSR/deploy_infer.py b/test_benchmark/PyTorch/EDSR/deploy_infer.py index 1b583962f..1975bd8c2 100644 --- a/test_benchmark/PyTorch/EDSR/deploy_infer.py +++ b/test_benchmark/PyTorch/EDSR/deploy_infer.py @@ -4,7 +4,6 @@ import numpy as np import paddle -import paddle.fluid as fluid from paddle.inference import Config from paddle.inference import create_predictor diff --git a/test_benchmark/PyTorch/EDSR/pd_infer.py b/test_benchmark/PyTorch/EDSR/pd_infer.py index d15f86a11..249b5fd96 100644 --- a/test_benchmark/PyTorch/EDSR/pd_infer.py +++ b/test_benchmark/PyTorch/EDSR/pd_infer.py @@ -1,5 +1,5 @@ from __future__ import print_function -import paddle.fluid as fluid + import paddle import sys import os @@ -13,11 +13,8 @@ # trace paddle.enable_static() exe = paddle.static.Executor(paddle.CPUPlace()) - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_trace/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_trace/inference_model/model", executor=exe) result = exe.run(prog, feed={inputs[0]: img}, fetch_list=outputs) df = pytorch_output - result if np.max(np.fabs(df)) > 1e-03: @@ -28,11 +25,8 @@ # script paddle.enable_static() exe = paddle.static.Executor(paddle.CPUPlace()) - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_script/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_script/inference_model/model", executor=exe) result = exe.run(prog, feed={inputs[0]: img}, fetch_list=outputs) df = pytorch_output - result if np.max(np.fabs(df)) > 1e-03: From 88e36454d36e35244286e405027430489dfe7902 Mon Sep 17 00:00:00 2001 From: megemini Date: Sun, 13 Oct 2024 20:15:38 +0800 Subject: [PATCH 06/25] [Fix] pytorch electramodel --- test_benchmark/PyTorch/ElectraModel/deploy_infer.py | 1 - test_benchmark/PyTorch/ElectraModel/pd_infer.py | 10 ++++------ 2 files changed, 4 insertions(+), 7 deletions(-) diff --git a/test_benchmark/PyTorch/ElectraModel/deploy_infer.py b/test_benchmark/PyTorch/ElectraModel/deploy_infer.py index 4f04524d7..dc8d247a5 100644 --- a/test_benchmark/PyTorch/ElectraModel/deploy_infer.py +++ b/test_benchmark/PyTorch/ElectraModel/deploy_infer.py @@ -4,7 +4,6 @@ import numpy as np import paddle -import paddle.fluid as fluid from paddle.inference import Config from paddle.inference import create_predictor diff --git a/test_benchmark/PyTorch/ElectraModel/pd_infer.py b/test_benchmark/PyTorch/ElectraModel/pd_infer.py index f00efff16..a624664a6 100644 --- a/test_benchmark/PyTorch/ElectraModel/pd_infer.py +++ b/test_benchmark/PyTorch/ElectraModel/pd_infer.py @@ -1,5 +1,5 @@ from __future__ import print_function -import paddle.fluid as fluid + import paddle import sys import os @@ -22,11 +22,9 @@ def rel_err(x, y): try: paddle.enable_static() exe = paddle.static.Executor(paddle.CPUPlace()) - [prog, inputs, outputs - ] = fluid.io.load_inference_model(dirname="pd_model/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model/inference_model/model", executor=exe) + result = exe.run(prog, feed={ inputs[0]: input_data["input_ids"], From 68937f62a39d93bd02e665166df142ae45c321d4 Mon Sep 17 00:00:00 2001 From: megemini Date: Sun, 13 Oct 2024 20:18:18 +0800 Subject: [PATCH 07/25] [Fix] pytorch flaubertmodel --- test_benchmark/PyTorch/FlaubertModel/deploy_infer.py | 1 - test_benchmark/PyTorch/FlaubertModel/pd_infer.py | 10 ++++------ 2 files changed, 4 insertions(+), 7 deletions(-) diff --git a/test_benchmark/PyTorch/FlaubertModel/deploy_infer.py b/test_benchmark/PyTorch/FlaubertModel/deploy_infer.py index 0897bff48..0bfdf69b6 100644 --- a/test_benchmark/PyTorch/FlaubertModel/deploy_infer.py +++ b/test_benchmark/PyTorch/FlaubertModel/deploy_infer.py @@ -4,7 +4,6 @@ import numpy as np import paddle -import paddle.fluid as fluid from paddle.inference import Config from paddle.inference import create_predictor diff --git a/test_benchmark/PyTorch/FlaubertModel/pd_infer.py b/test_benchmark/PyTorch/FlaubertModel/pd_infer.py index 34364fc1e..3c53bca42 100644 --- a/test_benchmark/PyTorch/FlaubertModel/pd_infer.py +++ b/test_benchmark/PyTorch/FlaubertModel/pd_infer.py @@ -1,5 +1,5 @@ from __future__ import print_function -import paddle.fluid as fluid + import paddle import sys import os @@ -22,11 +22,9 @@ def rel_err(x, y): try: paddle.enable_static() exe = paddle.static.Executor(paddle.CPUPlace()) - [prog, inputs, outputs - ] = fluid.io.load_inference_model(dirname="pd_model/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model/inference_model/model", executor=exe) + result = exe.run(prog, feed={ inputs[0]: input_data["input_ids"], From 13bab783cbb4425faaf757f6413361f5ad38fe97 Mon Sep 17 00:00:00 2001 From: megemini Date: Sun, 13 Oct 2024 20:20:41 +0800 Subject: [PATCH 08/25] [Fix] pytorch inceptionv3 --- .../PyTorch/InceptionV3/deploy_infer.py | 1 - test_benchmark/PyTorch/InceptionV3/pd_infer.py | 16 +++++----------- 2 files changed, 5 insertions(+), 12 deletions(-) diff --git a/test_benchmark/PyTorch/InceptionV3/deploy_infer.py b/test_benchmark/PyTorch/InceptionV3/deploy_infer.py index ef3b88a7c..1eb859880 100644 --- a/test_benchmark/PyTorch/InceptionV3/deploy_infer.py +++ b/test_benchmark/PyTorch/InceptionV3/deploy_infer.py @@ -4,7 +4,6 @@ import numpy as np import paddle -import paddle.fluid as fluid from paddle.inference import Config from paddle.inference import create_predictor diff --git a/test_benchmark/PyTorch/InceptionV3/pd_infer.py b/test_benchmark/PyTorch/InceptionV3/pd_infer.py index a1cd035aa..f48f30f10 100644 --- a/test_benchmark/PyTorch/InceptionV3/pd_infer.py +++ b/test_benchmark/PyTorch/InceptionV3/pd_infer.py @@ -1,5 +1,5 @@ from __future__ import print_function -import paddle.fluid as fluid + import paddle import sys import os @@ -22,11 +22,8 @@ def rel_err(x, y): # trace paddle.enable_static() exe = paddle.static.Executor(paddle.CPUPlace()) - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_trace/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_trace/inference_model/model", executor=exe) result = exe.run(prog, feed={inputs[0]: input_data}, fetch_list=outputs) df = pytorch_output - result[0] if numpy.max(numpy.fabs(df)) > 1e-04: @@ -37,11 +34,8 @@ def rel_err(x, y): # script paddle.enable_static() exe = paddle.static.Executor(paddle.CPUPlace()) - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_script/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_script/inference_model/model", executor=exe) result = exe.run(prog, feed={inputs[0]: input_data}, fetch_list=outputs) df = pytorch_output - result[0] if numpy.max(numpy.fabs(df)) > 1e-04: From dde4745843a10763dd72b11e4fa51b163615e39a Mon Sep 17 00:00:00 2001 From: megemini Date: Sun, 13 Oct 2024 20:24:33 +0800 Subject: [PATCH 09/25] [Fix] pytorch mnasnet --- test_benchmark/PyTorch/MNasNet/deploy_infer.py | 1 - test_benchmark/PyTorch/MNasNet/pd_infer.py | 16 +++++----------- 2 files changed, 5 insertions(+), 12 deletions(-) diff --git a/test_benchmark/PyTorch/MNasNet/deploy_infer.py b/test_benchmark/PyTorch/MNasNet/deploy_infer.py index 2113a59b7..47220a1f4 100644 --- a/test_benchmark/PyTorch/MNasNet/deploy_infer.py +++ b/test_benchmark/PyTorch/MNasNet/deploy_infer.py @@ -4,7 +4,6 @@ import numpy as np import paddle -import paddle.fluid as fluid from paddle.inference import Config from paddle.inference import create_predictor diff --git a/test_benchmark/PyTorch/MNasNet/pd_infer.py b/test_benchmark/PyTorch/MNasNet/pd_infer.py index 294c440dc..f8fac93ef 100644 --- a/test_benchmark/PyTorch/MNasNet/pd_infer.py +++ b/test_benchmark/PyTorch/MNasNet/pd_infer.py @@ -1,5 +1,5 @@ from __future__ import print_function -import paddle.fluid as fluid + import paddle import sys import os @@ -22,11 +22,8 @@ def rel_err(x, y): # trace paddle.enable_static() exe = paddle.static.Executor(paddle.CPUPlace()) - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_trace/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_trace/inference_model/model", executor=exe) result = exe.run(prog, feed={inputs[0]: input_data}, fetch_list=outputs) df = pytorch_output - result[0] if numpy.max(numpy.fabs(df)) > 1e-04: @@ -37,11 +34,8 @@ def rel_err(x, y): # script paddle.enable_static() exe = paddle.static.Executor(paddle.CPUPlace()) - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_script/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_script/inference_model/model", executor=exe) result = exe.run(prog, feed={inputs[0]: input_data}, fetch_list=outputs) df = pytorch_output - result[0] if numpy.max(numpy.fabs(df)) > 1e-04: From c0a7813d34fff1f34ca3ab5d422e5a371893a2c7 Mon Sep 17 00:00:00 2001 From: megemini Date: Sun, 13 Oct 2024 20:29:35 +0800 Subject: [PATCH 10/25] [Fix] pytorch mobilenetv2 --- .../PyTorch/MobileNetV2/deploy_infer.py | 1 - test_benchmark/PyTorch/MobileNetV2/pd_infer.py | 16 +++++----------- 2 files changed, 5 insertions(+), 12 deletions(-) diff --git a/test_benchmark/PyTorch/MobileNetV2/deploy_infer.py b/test_benchmark/PyTorch/MobileNetV2/deploy_infer.py index c098217ae..569fa7e04 100644 --- a/test_benchmark/PyTorch/MobileNetV2/deploy_infer.py +++ b/test_benchmark/PyTorch/MobileNetV2/deploy_infer.py @@ -4,7 +4,6 @@ import numpy as np import paddle -import paddle.fluid as fluid from paddle.inference import Config from paddle.inference import create_predictor diff --git a/test_benchmark/PyTorch/MobileNetV2/pd_infer.py b/test_benchmark/PyTorch/MobileNetV2/pd_infer.py index 18758f965..7606b6047 100644 --- a/test_benchmark/PyTorch/MobileNetV2/pd_infer.py +++ b/test_benchmark/PyTorch/MobileNetV2/pd_infer.py @@ -1,5 +1,5 @@ from __future__ import print_function -import paddle.fluid as fluid + import paddle import sys import os @@ -22,11 +22,8 @@ def rel_err(x, y): # trace paddle.enable_static() exe = paddle.static.Executor(paddle.CPUPlace()) - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_trace/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_trace/inference_model/model", executor=exe) result = exe.run(prog, feed={inputs[0]: input_data}, fetch_list=outputs) df = pytorch_output - result[0] if numpy.max(numpy.fabs(df)) > 1e-04: @@ -37,11 +34,8 @@ def rel_err(x, y): # script paddle.enable_static() exe = paddle.static.Executor(paddle.CPUPlace()) - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_script/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_script/inference_model/model", executor=exe) result = exe.run(prog, feed={inputs[0]: input_data}, fetch_list=outputs) df = pytorch_output - result[0] if numpy.max(numpy.fabs(df)) > 1e-04: From ae94b6ca78eb787a0f2b7cbdaf181585abc86aa3 Mon Sep 17 00:00:00 2001 From: megemini Date: Sun, 13 Oct 2024 20:41:19 +0800 Subject: [PATCH 11/25] [Fix] pytorch resnet18 --- test_benchmark/PyTorch/ResNet18/deploy_infer.py | 1 - test_benchmark/PyTorch/ResNet18/pd_infer.py | 16 +++++----------- 2 files changed, 5 insertions(+), 12 deletions(-) diff --git a/test_benchmark/PyTorch/ResNet18/deploy_infer.py b/test_benchmark/PyTorch/ResNet18/deploy_infer.py index fe49099ac..e86853f4e 100644 --- a/test_benchmark/PyTorch/ResNet18/deploy_infer.py +++ b/test_benchmark/PyTorch/ResNet18/deploy_infer.py @@ -4,7 +4,6 @@ import numpy as np import paddle -import paddle.fluid as fluid from paddle.inference import Config from paddle.inference import create_predictor diff --git a/test_benchmark/PyTorch/ResNet18/pd_infer.py b/test_benchmark/PyTorch/ResNet18/pd_infer.py index fa20652f1..718ffe5b7 100644 --- a/test_benchmark/PyTorch/ResNet18/pd_infer.py +++ b/test_benchmark/PyTorch/ResNet18/pd_infer.py @@ -1,5 +1,5 @@ from __future__ import print_function -import paddle.fluid as fluid + import paddle import sys import os @@ -22,11 +22,8 @@ def rel_err(x, y): # trace paddle.enable_static() exe = paddle.static.Executor(paddle.CPUPlace()) - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_trace/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_trace/inference_model/model", executor=exe) result = exe.run(prog, feed={inputs[0]: input_data}, fetch_list=outputs) df = pytorch_output - result[0] if numpy.max(numpy.fabs(df)) > 1e-04: @@ -37,11 +34,8 @@ def rel_err(x, y): # script paddle.enable_static() exe = paddle.static.Executor(paddle.CPUPlace()) - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_script/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_script/inference_model/model", executor=exe) result = exe.run(prog, feed={inputs[0]: input_data}, fetch_list=outputs) df = pytorch_output - result[0] if numpy.max(numpy.fabs(df)) > 1e-04: From 738264be1eb8ec49bc02ad3144e50b67d3da85be Mon Sep 17 00:00:00 2001 From: megemini Date: Sun, 13 Oct 2024 20:43:18 +0800 Subject: [PATCH 12/25] [Fix] pytorch resnet18_2 --- test_benchmark/PyTorch/ResNet18_2/deploy_infer.py | 1 - test_benchmark/PyTorch/ResNet18_2/pd_infer.py | 9 +++------ 2 files changed, 3 insertions(+), 7 deletions(-) diff --git a/test_benchmark/PyTorch/ResNet18_2/deploy_infer.py b/test_benchmark/PyTorch/ResNet18_2/deploy_infer.py index 9ccfb9ade..cd66b803d 100644 --- a/test_benchmark/PyTorch/ResNet18_2/deploy_infer.py +++ b/test_benchmark/PyTorch/ResNet18_2/deploy_infer.py @@ -4,7 +4,6 @@ import numpy as np import paddle -import paddle.fluid as fluid from paddle.inference import Config from paddle.inference import create_predictor diff --git a/test_benchmark/PyTorch/ResNet18_2/pd_infer.py b/test_benchmark/PyTorch/ResNet18_2/pd_infer.py index 0e0be1741..dfc2ba640 100644 --- a/test_benchmark/PyTorch/ResNet18_2/pd_infer.py +++ b/test_benchmark/PyTorch/ResNet18_2/pd_infer.py @@ -1,5 +1,5 @@ import paddle -import paddle.fluid as fluid + import numpy as np import pickle import sys @@ -11,11 +11,8 @@ paddle.enable_static() exe = paddle.static.Executor(paddle.CPUPlace()) # test trace - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_trace/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_trace/inference_model/model", executor=exe) data = np.load('../dataset/ResNet18_2/input.npy') result = exe.run(prog, feed={inputs[0]: data}, fetch_list=outputs) From 74129bafc093772792207c7215e5928e73d77e35 Mon Sep 17 00:00:00 2001 From: megemini Date: Sun, 13 Oct 2024 20:47:22 +0800 Subject: [PATCH 13/25] [Fix] pytorch shufflenetv2 --- .../PyTorch/ShuffleNetV2/deploy_infer.py | 1 - test_benchmark/PyTorch/ShuffleNetV2/pd_infer.py | 16 +++++----------- 2 files changed, 5 insertions(+), 12 deletions(-) diff --git a/test_benchmark/PyTorch/ShuffleNetV2/deploy_infer.py b/test_benchmark/PyTorch/ShuffleNetV2/deploy_infer.py index 35752a997..b011be2b3 100644 --- a/test_benchmark/PyTorch/ShuffleNetV2/deploy_infer.py +++ b/test_benchmark/PyTorch/ShuffleNetV2/deploy_infer.py @@ -4,7 +4,6 @@ import numpy as np import paddle -import paddle.fluid as fluid from paddle.inference import Config from paddle.inference import create_predictor diff --git a/test_benchmark/PyTorch/ShuffleNetV2/pd_infer.py b/test_benchmark/PyTorch/ShuffleNetV2/pd_infer.py index 63ec1e58f..7572384e9 100644 --- a/test_benchmark/PyTorch/ShuffleNetV2/pd_infer.py +++ b/test_benchmark/PyTorch/ShuffleNetV2/pd_infer.py @@ -1,5 +1,5 @@ from __future__ import print_function -import paddle.fluid as fluid + import paddle import sys import os @@ -22,11 +22,8 @@ def rel_err(x, y): # trace paddle.enable_static() exe = paddle.static.Executor(paddle.CPUPlace()) - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_trace/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_trace/inference_model/model", executor=exe) result = exe.run(prog, feed={inputs[0]: input_data}, fetch_list=outputs) df = pytorch_output - result[0] if numpy.max(numpy.fabs(df)) > 1e-04: @@ -37,11 +34,8 @@ def rel_err(x, y): # script paddle.enable_static() exe = paddle.static.Executor(paddle.CPUPlace()) - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_script/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_script/inference_model/model", executor=exe) result = exe.run(prog, feed={inputs[0]: input_data}, fetch_list=outputs) df = pytorch_output - result[0] if numpy.max(numpy.fabs(df)) > 1e-04: From 0c36cf74efb6f3fb8e0ddb095ba436cdae9f2ffd Mon Sep 17 00:00:00 2001 From: megemini Date: Sun, 13 Oct 2024 20:49:32 +0800 Subject: [PATCH 14/25] [Fix] pytorch squeezenet --- .../PyTorch/SqueezeNet/deploy_infer.py | 1 - test_benchmark/PyTorch/SqueezeNet/pd_infer.py | 16 +++++----------- 2 files changed, 5 insertions(+), 12 deletions(-) diff --git a/test_benchmark/PyTorch/SqueezeNet/deploy_infer.py b/test_benchmark/PyTorch/SqueezeNet/deploy_infer.py index 85fa1adad..b7d01981b 100644 --- a/test_benchmark/PyTorch/SqueezeNet/deploy_infer.py +++ b/test_benchmark/PyTorch/SqueezeNet/deploy_infer.py @@ -4,7 +4,6 @@ import numpy as np import paddle -import paddle.fluid as fluid from paddle.inference import Config from paddle.inference import create_predictor diff --git a/test_benchmark/PyTorch/SqueezeNet/pd_infer.py b/test_benchmark/PyTorch/SqueezeNet/pd_infer.py index 06f50a893..942f43186 100644 --- a/test_benchmark/PyTorch/SqueezeNet/pd_infer.py +++ b/test_benchmark/PyTorch/SqueezeNet/pd_infer.py @@ -1,5 +1,5 @@ from __future__ import print_function -import paddle.fluid as fluid + import paddle import sys import os @@ -22,11 +22,8 @@ def rel_err(x, y): # trace paddle.enable_static() exe = paddle.static.Executor(paddle.CPUPlace()) - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_trace/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_trace/inference_model/model", executor=exe) result = exe.run(prog, feed={inputs[0]: input_data}, fetch_list=outputs) df = pytorch_output - result[0] if numpy.max(numpy.fabs(df)) > 1e-04: @@ -37,11 +34,8 @@ def rel_err(x, y): # script paddle.enable_static() exe = paddle.static.Executor(paddle.CPUPlace()) - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_script/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_script/inference_model/model", executor=exe) result = exe.run(prog, feed={inputs[0]: input_data}, fetch_list=outputs) df = pytorch_output - result[0] if numpy.max(numpy.fabs(df)) > 1e-04: From bebdfaac1068f49a8eacd327407e66143d466ff1 Mon Sep 17 00:00:00 2001 From: megemini Date: Sun, 13 Oct 2024 20:52:26 +0800 Subject: [PATCH 15/25] [Fix] pytorch vgg16 --- test_benchmark/PyTorch/VGG16/deploy_infer.py | 1 - test_benchmark/PyTorch/VGG16/pd_infer.py | 16 +++++----------- 2 files changed, 5 insertions(+), 12 deletions(-) diff --git a/test_benchmark/PyTorch/VGG16/deploy_infer.py b/test_benchmark/PyTorch/VGG16/deploy_infer.py index ff61a9611..3c8b8d611 100644 --- a/test_benchmark/PyTorch/VGG16/deploy_infer.py +++ b/test_benchmark/PyTorch/VGG16/deploy_infer.py @@ -4,7 +4,6 @@ import numpy as np import paddle -import paddle.fluid as fluid from paddle.inference import Config from paddle.inference import create_predictor diff --git a/test_benchmark/PyTorch/VGG16/pd_infer.py b/test_benchmark/PyTorch/VGG16/pd_infer.py index d452fe283..30da19671 100644 --- a/test_benchmark/PyTorch/VGG16/pd_infer.py +++ b/test_benchmark/PyTorch/VGG16/pd_infer.py @@ -1,5 +1,5 @@ from __future__ import print_function -import paddle.fluid as fluid + import paddle import sys import os @@ -22,11 +22,8 @@ def rel_err(x, y): # trace paddle.enable_static() exe = paddle.static.Executor(paddle.CPUPlace()) - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_trace/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_trace/inference_model/model", executor=exe) result = exe.run(prog, feed={inputs[0]: input_data}, fetch_list=outputs) df = pytorch_output - result[0] if numpy.max(numpy.fabs(df)) > 1e-04: @@ -37,11 +34,8 @@ def rel_err(x, y): # script paddle.enable_static() exe = paddle.static.Executor(paddle.CPUPlace()) - [prog, inputs, outputs] = fluid.io.load_inference_model( - dirname="pd_model_script/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model_script/inference_model/model", executor=exe) result = exe.run(prog, feed={inputs[0]: input_data}, fetch_list=outputs) df = pytorch_output - result[0] if numpy.max(numpy.fabs(df)) > 1e-04: From 23fd0e7a133c67ced2fae3761adcb8f4e8ab09fe Mon Sep 17 00:00:00 2001 From: megemini Date: Sun, 13 Oct 2024 20:55:03 +0800 Subject: [PATCH 16/25] [Fix] pytorch mobilenetv3 --- test_benchmark/PyTorch/mobilenetv3/pd_infer.py | 8 ++------ 1 file changed, 2 insertions(+), 6 deletions(-) diff --git a/test_benchmark/PyTorch/mobilenetv3/pd_infer.py b/test_benchmark/PyTorch/mobilenetv3/pd_infer.py index 036f60171..afe5fb7a0 100644 --- a/test_benchmark/PyTorch/mobilenetv3/pd_infer.py +++ b/test_benchmark/PyTorch/mobilenetv3/pd_infer.py @@ -1,4 +1,3 @@ -import paddle.fluid as fluid import paddle import numpy as np import sys @@ -11,11 +10,8 @@ exe = paddle.static.Executor(paddle.CPUPlace()) # test dygraph - [prog, inputs, outputs - ] = fluid.io.load_inference_model(dirname="pd_model/inference_model/", - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix="pd_model/inference_model/model", executor=exe) data = np.load('../dataset/mobilenetv3/input.npy') result = exe.run(prog, feed={inputs[0]: data}, fetch_list=outputs) From efc235d56e4867b2c8e811c1ce30f4b86ff259c8 Mon Sep 17 00:00:00 2001 From: megemini Date: Mon, 14 Oct 2024 22:09:06 +0800 Subject: [PATCH 17/25] [Fix] pytorch tools --- test_benchmark/PyTorch/tools/predict.py | 9 +++------ 1 file changed, 3 insertions(+), 6 deletions(-) diff --git a/test_benchmark/PyTorch/tools/predict.py b/test_benchmark/PyTorch/tools/predict.py index 8b2e99ce4..5104a2882 100644 --- a/test_benchmark/PyTorch/tools/predict.py +++ b/test_benchmark/PyTorch/tools/predict.py @@ -10,7 +10,6 @@ import yaml import numpy as np import paddle -import paddle.fluid as fluid from paddle.inference import create_predictor, PrecisionType from paddle.inference import Config from benchmark_utils import PaddleInferBenchmark @@ -194,11 +193,9 @@ def analysis_operators(self, model_dir): paddle.enable_static() exe = paddle.static.Executor(paddle.CPUPlace()) # test Dygraph - [prog, inputs, outputs - ] = fluid.io.load_inference_model(dirname=model_dir, - executor=exe, - model_filename="model.pdmodel", - params_filename="model.pdiparams") + [prog, inputs, outputs] = paddle.static.load_inference_model( + path_prefix=model_dir + "/model", executor=exe) + #test op nums op_dict = dict() op_nums = 0 From 273ab4f1275d7b09d146659623c99a4290683072 Mon Sep 17 00:00:00 2001 From: megemini Date: Tue, 22 Oct 2024 17:05:34 +0800 Subject: [PATCH 18/25] [Update] black.list --- test_benchmark/PyTorch/black.list | 15 --------------- 1 file changed, 15 deletions(-) diff --git a/test_benchmark/PyTorch/black.list b/test_benchmark/PyTorch/black.list index fae6e32d4..28f2493d5 100644 --- a/test_benchmark/PyTorch/black.list +++ b/test_benchmark/PyTorch/black.list @@ -1,35 +1,20 @@ ACG_UnitTest -AlexNet -BASNet BertForMaskedLM_dccuchile BertModel_SpanBert CamembertForQuestionAnswering -DBFace DPRContextEncoder DeepLabv3_ResNet50 -EDSR EasyOCR_detector EasyOCR_recognizer -ElectraModel FCN_ResNet50 -FlaubertModel GRU -InceptionV3 -MNasNet MiniFasNet -MobileNetV2 Mobilestereonet MockingBird -ResNet18 -ResNet18_2 Roberta Saicinpainting_LaMa -ShuffleNetV2 -SqueezeNet SwinTransformer -VGG16 XLMRobertaForTokenClassification -mobilenetv3 opadd dataset tools From f045fb41b8bf59e4e743bf679244cd36b2575894 Mon Sep 17 00:00:00 2001 From: megemini Date: Tue, 22 Oct 2024 17:29:36 +0800 Subject: [PATCH 19/25] [Debug] mobilenetv3 --- test_benchmark/PyTorch/black.list | 14 ++++++++++++++ test_benchmark/PyTorch/mobilenetv3/pd_infer.py | 2 ++ 2 files changed, 16 insertions(+) diff --git a/test_benchmark/PyTorch/black.list b/test_benchmark/PyTorch/black.list index 28f2493d5..c1a9b1a53 100644 --- a/test_benchmark/PyTorch/black.list +++ b/test_benchmark/PyTorch/black.list @@ -1,19 +1,33 @@ ACG_UnitTest +AlexNet +BASNet BertForMaskedLM_dccuchile BertModel_SpanBert CamembertForQuestionAnswering +DBFace DPRContextEncoder DeepLabv3_ResNet50 +EDSR EasyOCR_detector EasyOCR_recognizer +ElectraModel FCN_ResNet50 +FlaubertModel GRU +InceptionV3 +MNasNet MiniFasNet +MobileNetV2 Mobilestereonet MockingBird +ResNet18 +ResNet18_2 Roberta Saicinpainting_LaMa +ShuffleNetV2 +SqueezeNet SwinTransformer +VGG16 XLMRobertaForTokenClassification opadd dataset diff --git a/test_benchmark/PyTorch/mobilenetv3/pd_infer.py b/test_benchmark/PyTorch/mobilenetv3/pd_infer.py index afe5fb7a0..6d22cd6f9 100644 --- a/test_benchmark/PyTorch/mobilenetv3/pd_infer.py +++ b/test_benchmark/PyTorch/mobilenetv3/pd_infer.py @@ -32,3 +32,5 @@ f.write("!!!!!Dygraph Failed\n") except: f.write("!!!!!Failed\n") + + raise From f13775ad5558277f4803f70edcb126f9c50c8a1c Mon Sep 17 00:00:00 2001 From: megemini Date: Tue, 22 Oct 2024 17:36:01 +0800 Subject: [PATCH 20/25] [Debug] mobilenetv3 --- test_benchmark/PyTorch/mobilenetv3/pd_infer.py | 6 ++++-- 1 file changed, 4 insertions(+), 2 deletions(-) diff --git a/test_benchmark/PyTorch/mobilenetv3/pd_infer.py b/test_benchmark/PyTorch/mobilenetv3/pd_infer.py index 6d22cd6f9..1ed0663e8 100644 --- a/test_benchmark/PyTorch/mobilenetv3/pd_infer.py +++ b/test_benchmark/PyTorch/mobilenetv3/pd_infer.py @@ -3,6 +3,8 @@ import sys import pickle +import traceback + f = open('result.txt', 'w') f.write("======Mobilenetv3: \n") try: @@ -30,7 +32,7 @@ f.write("Dygraph Successed\n") else: f.write("!!!!!Dygraph Failed\n") -except: +except Exception as e: f.write("!!!!!Failed\n") - raise + f.write(traceback.format_exc()) From 0119e916e241f67a9a61de97932286a39c574f95 Mon Sep 17 00:00:00 2001 From: megemini Date: Tue, 22 Oct 2024 17:40:45 +0800 Subject: [PATCH 21/25] [Debug] mobilenetv3 --- x2paddle/convert.py | 26 +++++++++++++------------- 1 file changed, 13 insertions(+), 13 deletions(-) diff --git a/x2paddle/convert.py b/x2paddle/convert.py index 56ff967c5..b9cd52350 100644 --- a/x2paddle/convert.py +++ b/x2paddle/convert.py @@ -379,14 +379,14 @@ def pytorch2paddle(module, v2 = v2.split('+')[0] version_sum = int(v0) * 100 + int(v1) * 10 + int(v2) if version_sum < 150: - logging.info("[ERROR] PyTorch>=1.5.0 is required") + logger.info("[ERROR] PyTorch>=1.5.0 is required") return except: - logging.info( + logger.info( "[ERROR] PyTorch is not installed, use \"pip install torch torchvision\"." ) return - logging.info("Now translating model from PyTorch to Paddle.") + logger.info("Now translating model from PyTorch to Paddle.") from x2paddle.decoder.pytorch_decoder import ScriptDecoder, TraceDecoder from x2paddle.op_mapper.pytorch2paddle.pytorch_op_mapper import PyTorchOpMapper @@ -397,39 +397,39 @@ def pytorch2paddle(module, model = ScriptDecoder(module, input_examples) mapper = PyTorchOpMapper(model) mapper.paddle_graph.build() - logging.info("Model optimizing ...") + logger.info("Model optimizing ...") from x2paddle.optimizer.optimizer import GraphOptimizer graph_opt = GraphOptimizer(source_frame="pytorch", jit_type=jit_type) graph_opt.optimize(mapper.paddle_graph) - logging.info("Model optimized!") + logger.info("Model optimized!") mapper.paddle_graph.gen_model(save_dir, jit_type=jit_type, enable_code_optim=enable_code_optim) - logging.info("Successfully exported Paddle static graph model!") + logger.info("Successfully exported Paddle static graph model!") if not disable_feedback: ConverterCheck(task="PyTorch", time_info=time_info, convert_state="Success").start() if convert_to_lite: - logging.info("Now translating model from Paddle to Paddle Lite ...") + logger.info("Now translating model from Paddle to Paddle Lite ...") if not disable_feedback: ConverterCheck(task="PyTorch", time_info=time_info, lite_state="Start").start() convert2lite(save_dir, lite_valid_places, lite_model_type) - logging.info("Successfully exported Paddle Lite support model!") + logger.info("Successfully exported Paddle Lite support model!") if not disable_feedback: ConverterCheck(task="PyTorch", time_info=time_info, lite_state="Success").start() # for convert survey - logging.info("================================================") - logging.info("") - logging.info( + logger.info("================================================") + logger.info("") + logger.info( "Model Converted! Fill this survey to help X2Paddle better, https://iwenjuan.baidu.com/?code=npyd51 " ) - logging.info("") - logging.info("================================================") + logger.info("") + logger.info("================================================") def main(): From 0f3a4fbbb0742a998f28a1ebe01451918c9d1f56 Mon Sep 17 00:00:00 2001 From: megemini Date: Tue, 22 Oct 2024 17:43:29 +0800 Subject: [PATCH 22/25] [Debug] mobilenetv3 --- test_benchmark/PyTorch/convert.sh | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/test_benchmark/PyTorch/convert.sh b/test_benchmark/PyTorch/convert.sh index 4c80dfed0..0c62c7138 100644 --- a/test_benchmark/PyTorch/convert.sh +++ b/test_benchmark/PyTorch/convert.sh @@ -32,7 +32,9 @@ do touch result.txt echo $model ">>>Failed"> result.txt - sh run_convert.sh 1>run.log 2>run.err & + # TODO(megemini): debug + # sh run_convert.sh 1>run.log 2>run.err & + sh run_convert.sh cd .. counter=$(($counter+1)) step=$(( $counter % 1 )) From 115ce620a2b418b767c8644e40d77e9a44bb88b5 Mon Sep 17 00:00:00 2001 From: megemini Date: Tue, 22 Oct 2024 17:46:46 +0800 Subject: [PATCH 23/25] [Debug] mobilenetv3 --- test_benchmark/PyTorch/mobilenetv3/convert.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/test_benchmark/PyTorch/mobilenetv3/convert.py b/test_benchmark/PyTorch/mobilenetv3/convert.py index 23990d22c..e4ef08e2d 100644 --- a/test_benchmark/PyTorch/mobilenetv3/convert.py +++ b/test_benchmark/PyTorch/mobilenetv3/convert.py @@ -8,7 +8,7 @@ modelFile = '../dataset/mobilenetv3/MobileNetV3_large.pth' torch_model = mobilenet_v3_large(num_classes=n_classes) #模型结构 #加载模型 -checkpoint = torch.load(modelFile, map_location='cuda:0') +checkpoint = torch.load(modelFile) torch_model.load_state_dict(checkpoint) #加载预训练参数 # 设置为eval模式 torch_model.eval() From 5dcc700146db2c228939f10a9991a08b2b7172aa Mon Sep 17 00:00:00 2001 From: megemini Date: Tue, 22 Oct 2024 17:48:53 +0800 Subject: [PATCH 24/25] [Debug] mobilenetv3 --- test_benchmark/PyTorch/mobilenetv3/convert.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/test_benchmark/PyTorch/mobilenetv3/convert.py b/test_benchmark/PyTorch/mobilenetv3/convert.py index e4ef08e2d..ba5277f7d 100644 --- a/test_benchmark/PyTorch/mobilenetv3/convert.py +++ b/test_benchmark/PyTorch/mobilenetv3/convert.py @@ -8,7 +8,7 @@ modelFile = '../dataset/mobilenetv3/MobileNetV3_large.pth' torch_model = mobilenet_v3_large(num_classes=n_classes) #模型结构 #加载模型 -checkpoint = torch.load(modelFile) +checkpoint = torch.load(modelFile, map_location='cpu') torch_model.load_state_dict(checkpoint) #加载预训练参数 # 设置为eval模式 torch_model.eval() From 7d10903cf036d74fa8030925f04e854a652513a5 Mon Sep 17 00:00:00 2001 From: megemini Date: Tue, 22 Oct 2024 17:53:10 +0800 Subject: [PATCH 25/25] [Update] black.list --- test_benchmark/PyTorch/black.list | 14 -------------- 1 file changed, 14 deletions(-) diff --git a/test_benchmark/PyTorch/black.list b/test_benchmark/PyTorch/black.list index c1a9b1a53..28f2493d5 100644 --- a/test_benchmark/PyTorch/black.list +++ b/test_benchmark/PyTorch/black.list @@ -1,33 +1,19 @@ ACG_UnitTest -AlexNet -BASNet BertForMaskedLM_dccuchile BertModel_SpanBert CamembertForQuestionAnswering -DBFace DPRContextEncoder DeepLabv3_ResNet50 -EDSR EasyOCR_detector EasyOCR_recognizer -ElectraModel FCN_ResNet50 -FlaubertModel GRU -InceptionV3 -MNasNet MiniFasNet -MobileNetV2 Mobilestereonet MockingBird -ResNet18 -ResNet18_2 Roberta Saicinpainting_LaMa -ShuffleNetV2 -SqueezeNet SwinTransformer -VGG16 XLMRobertaForTokenClassification opadd dataset