【发布时间】:2021-04-06 07:56:28
【问题描述】:
我正在尝试将预训练的 Torch 模型转换为 ONNX,但收到以下错误:
RuntimeError: step!=1 is currently not supported
我正在预训练的着色模型上尝试这个:https://github.com/richzhang/colorization
这是我在 Google Colab 中运行的代码:
!git clone https://github.com/richzhang/colorization.git
cd colorization/
import colorizers
model = colorizer_siggraph17 = colorizers.siggraph17(pretrained=True).eval()
input_names = [ "input" ]
output_names = [ "output" ]
dummy_input = torch.randn(1, 1, 256, 256, device='cpu')
torch.onnx.export(model, dummy_input, "test_converted_model.onnx", verbose=True,
input_names=input_names, output_names=output_names)
感谢您的帮助:)
更新 1: @Proko 建议解决了 ONNX 导出问题。现在,当我尝试将 ONNX 转换为 TensorRT 时,我遇到了一个可能相关的新问题。我收到以下错误:
[TensorRT] ERROR: Network must have at least one output
这是我使用的代码:
import torch
import pycuda.driver as cuda
import pycuda.autoinit
import tensorrt as trt
import onnx
TRT_LOGGER = trt.Logger()
def build_engine(onnx_file_path):
# initialize TensorRT engine and parse ONNX model
builder = trt.Builder(TRT_LOGGER)
builder.max_workspace_size = 1 << 25
builder.max_batch_size = 1
if builder.platform_has_fast_fp16:
builder.fp16_mode = True
network = builder.create_network()
parser = trt.OnnxParser(network, TRT_LOGGER)
# parse ONNX
with open(onnx_file_path, 'rb') as model:
print('Beginning ONNX file parsing')
parser.parse(model.read())
print('Completed parsing of ONNX file')
# generate TensorRT engine optimized for the target platform
print('Building an engine...')
engine = builder.build_cuda_engine(network)
context = engine.create_execution_context()
print("Completed creating Engine")
return engine, context
ONNX_FILE_PATH = 'siggraph17.onnx' # Exported using the code above
engine,_ = build_engine(ONNX_FILE_PATH)
我试图通过以下方式强制 build_engine 函数使用网络的输出:
network.mark_output(network.get_layer(network.num_layers-1).get_output(0))
但它不起作用。 我需要任何帮助!
【问题讨论】:
-
此时您将无法导出此模型。
torch.onnx现在根本不支持步长不同于 1 的切片。也许重写模型以使用与n-step 切片不同的东西(但当然给出相同的结果)可能会对您有所帮助
标签: pytorch google-colaboratory onnx