Mxnet速查_CPU和GPU的mnist预测训练_模型导出_模型导入再预测_导出onnx并预测

需要做点什么 方便广大烟酒生研究生、人工智障炼丹师算法工程师快速使用mxnet,所以特写此文章,默认使用者已有基本的深度学习概念、数据集概念。
系统环境 python 3.7.4
mxnet 1.9.0
mxnet-cu112 1.9.0
onnx 1.9.0
onnxruntime-gpu 1.9.0
数据准备 MNIST数据集csv文件是一个42000x785的矩阵
42000表示有42000张图片
785中第一列是图片的类别(0,1,2,..,9),第二列到最后一列是图片数据向量 (28x28的图片张成784的向量), 数据集长这个样子:
1 0 0 0 0 0 0 0 0 0 ..
0 0 0 0 0 0 0 0 0 0 ..
1 0 0 0 0 0 0 0 0 0 ..
1. 导入需要的包

import time import copy import onnx import logging import platform import mxnet as mx import numpy as np import pandas as pd import onnxruntime as ort from sklearn.metrics import accuracy_scorelogger = logging.getLogger() logger.setLevel(logging.DEBUG)# Mxnet Chcek if platform.system().lower() != 'windows': print(mx.runtime.feature_list()) print(mx.context.num_gpus()) a = mx.nd.ones((2, 3), mx.cpu()) b = a * 2 + 1 print(b)

运行输出
[? CUDA, ? CUDNN, ? NCCL, ? CUDA_RTC, ? TENSORRT, ? CPU_SSE, ? CPU_SSE2, ? CPU_SSE3, ? CPU_SSE4_1, ? CPU_SSE4_2, ? CPU_SSE4A, ? CPU_AVX, ? CPU_AVX2, ? OPENMP, ? SSE, ? F16C, ? JEMALLOC, ? BLAS_OPEN, ? BLAS_ATLAS, ? BLAS_MKL, ? BLAS_APPLE, ? LAPACK, ? MKLDNN, ? OPENCV, ? CAFFE, ? PROFILER, ? DIST_KVSTORE, ? CXX14, ? INT64_TENSOR_SIZE, ? SIGNAL_HANDLER, ? DEBUG, ? TVM_OP] 1[[3. 3. 3.] [3. 3. 3.]]

2. 参数准备
N_EPOCH = 1 N_BATCH = 32 N_BATCH_NUM = 900 S_DATA_PATH = r"mnist_train.csv" S_MODEL_PATH = r"mxnet_cnn" S_SYM_PATH = './mxnet_cnn-symbol.json' S_PARAMS_PATH = './mxnet_cnn-0001.params' S_ONNX_MODEL_PATH = './mxnet_cnn.onnx' S_DEVICE, N_DEVICE_ID, S_DEVICE_FULL = "cuda", 0, "cuda:0" # S_DEVICE, N_DEVICE_ID, S_DEVICE_FULL = "cpu", 0, "cpu" CTX = mx.cpu() if S_DEVICE == "cpu" else mx.gpu(N_DEVICE_ID) B_IS_UNIX = True

3. 读取数据
df = pd.read_csv(S_DATA_PATH, header=None) print(df.shape) np_mat = np.array(df) print(np_mat.shape) X = np_mat[:, 1:] Y = np_mat[:, 0] X = X.astype(np.float32) / 255 X_train = X[:N_BATCH * N_BATCH_NUM] X_test = X[N_BATCH * N_BATCH_NUM:] Y_train = Y[:N_BATCH * N_BATCH_NUM] Y_test = Y[N_BATCH * N_BATCH_NUM:] X_train = X_train.reshape(X_train.shape[0], 1, 28, 28) X_test = X_test.reshape(X_test.shape[0], 1, 28, 28) print(X_train.shape) print(Y_train.shape) print(X_test.shape) print(Y_test.shape) train_iter = mx.io.NDArrayIter(X_train, Y_train, batch_size=N_BATCH) test_iter = mx.io.NDArrayIter(X_test, Y_test, batch_size=N_BATCH) test_iter_2 = copy.copy(test_iter)

运行输出
(37800, 785) (37800, 785) (28800, 1, 28, 28) (28800,) (9000, 1, 28, 28) (9000,)

4. 模型构建
net = mx.gluon.nn.HybridSequential() with net.name_scope(): net.add(mx.gluon.nn.Conv2D(channels=32, kernel_size=3, activation='relu'))# bx28x28 ==> net.add(mx.gluon.nn.MaxPool2D(pool_size=2, strides=2)) net.add(mx.gluon.nn.Flatten()) net.add(mx.gluon.nn.Dense(128, activation="relu")) net.add(mx.gluon.nn.Dense(10)) net.hybridize() print(net) net.collect_params().initialize(mx.init.Xavier(), ctx=CTX) softmax_cross_entropy = mx.gluon.loss.SoftmaxCrossEntropyLoss() trainer = mx.gluon.Trainer(net.collect_params(), 'adam', {'learning_rate': .001})

运行输出
HybridSequential( (0): Conv2D(None -> 32, kernel_size=(3, 3), stride=(1, 1), Activation(relu)) (1): MaxPool2D(size=(2, 2), stride=(2, 2), padding=(0, 0), ceil_mode=False, global_pool=False, pool_type=max, layout=NCHW) (2): Flatten (3): Dense(None -> 128, Activation(relu)) (4): Dense(None -> 10, linear) )

5. 模型训练
for epoch in range(N_EPOCH): for batch_num, itr in enumerate(train_iter): data = https://www.it610.com/article/itr.data[0].as_in_context(CTX) label = itr.label[0].as_in_context(CTX) with mx.autograd.record(): output = net(data)# Run the forward pass loss = softmax_cross_entropy(output, label)# Compute the loss loss.backward() trainer.step(data.shape[0]) if batch_num % 50 == 0:# Print loss once in a while curr_loss = mx.nd.mean(loss)# .asscalar() pred = mx.nd.argmax(output, axis=1) np_pred, np_lable = pred.asnumpy(), label.asnumpy() f_acc = accuracy_score(np_lable, np_pred) print(f"Epoch: {epoch}; Batch {batch_num}; ACC {f_acc}") print(f"loss: {curr_loss}") print() # print("Epoch: %d; Batch %d; Loss %s; ACC %f" % #(epoch, batch_num, str(curr_loss), f_acc)) print()

运行输出
Epoch: 0; Batch 0; ACC 0.09375 loss: [2.2868602] Epoch: 0; Batch 50; ACC 0.875 loss: [0.512461] Epoch: 0; Batch 100; ACC 0.90625 loss: [0.43415746] Epoch: 0; Batch 150; ACC 0.84375 loss: [0.3854709] Epoch: 0; Batch 200; ACC 1.0 loss: [0.04192135] Epoch: 0; Batch 250; ACC 0.90625 loss: [0.21156572] Epoch: 0; Batch 300; ACC 0.9375 loss: [0.15938525] Epoch: 0; Batch 350; ACC 1.0 loss: [0.0379494] Epoch: 0; Batch 400; ACC 0.96875 loss: [0.17104594] Epoch: 0; Batch 450; ACC 0.96875 loss: [0.12192786] Epoch: 0; Batch 500; ACC 0.96875 loss: [0.09210478] Epoch: 0; Batch 550; ACC 0.9375 loss: [0.13728428] Epoch: 0; Batch 600; ACC 0.96875 loss: [0.0762211] Epoch: 0; Batch 650; ACC 0.96875 loss: [0.12162409] Epoch: 0; Batch 700; ACC 1.0 loss: [0.04334489] Epoch: 0; Batch 750; ACC 1.0 loss: [0.06458903] Epoch: 0; Batch 800; ACC 0.96875 loss: [0.07410634] Epoch: 0; Batch 850; ACC 0.96875 loss: [0.14233188]

6.模型预测
for batch_num, itr in enumerate(test_iter_2): data = https://www.it610.com/article/itr.data[0].as_in_context(CTX) label = itr.label[0].as_in_context(CTX)output = net(data)# Run the forward pass loss = softmax_cross_entropy(output, label)# Compute the lossif batch_num % 50 == 0:# Print loss once in a while curr_loss = mx.nd.mean(loss)# .asscalar() pred = mx.nd.argmax(output, axis=1) np_pred, np_lable = pred.asnumpy(), label.asnumpy() f_acc = accuracy_score(np_lable, np_pred) print(f"Epoch: {epoch}; Batch {batch_num}; ACC {f_acc}") print(f"loss: {curr_loss}") print()

运行输出
Epoch: 0; Batch 0; ACC 0.96875 loss: [0.22968824] Epoch: 0; Batch 50; ACC 0.96875 loss: [0.05668993] Epoch: 0; Batch 100; ACC 0.96875 loss: [0.08171713] Epoch: 0; Batch 150; ACC 1.0 loss: [0.02264522] Epoch: 0; Batch 200; ACC 0.96875 loss: [0.080383] Epoch: 0; Batch 250; ACC 1.0 loss: [0.03774196]

7.模型保存
net.export(S_MODEL_PATH, epoch=N_EPOCH)# 保存模型结构和全部参数

8.模型加载和加载模型使用
print("load net and do test") load_net = mx.gluon.nn.SymbolBlock.imports(S_SYM_PATH, ['data'], S_PARAMS_PATH, ctx=CTX)# 加载模型 print("load ok") for batch_num, itr in enumerate(test_iter):# Test data = https://www.it610.com/article/itr.data[0].as_in_context(CTX) label = itr.label[0].as_in_context(CTX)output = load_net(data)# Run the forward pass loss = softmax_cross_entropy(output, label)# Compute the lossif batch_num % 50 == 0:# Print loss once in a while curr_loss = mx.nd.mean(loss)# .asscalar() pred = mx.nd.argmax(output, axis=1) np_pred, np_lable = pred.asnumpy(), label.asnumpy() f_acc = accuracy_score(np_lable, np_pred) print(f"Epoch: {epoch}; Batch {batch_num}; ACC {f_acc}") print(f"loss: {curr_loss}") print() print("finish")

运行输出
load net and do test load ok Epoch: 0; Batch 0; ACC 0.96875 loss: [0.22968824] Epoch: 0; Batch 50; ACC 0.96875 loss: [0.05668993] Epoch: 0; Batch 100; ACC 0.96875 loss: [0.08171713] Epoch: 0; Batch 150; ACC 1.0 loss: [0.02264522] Epoch: 0; Batch 200; ACC 0.96875 loss: [0.080383] Epoch: 0; Batch 250; ACC 1.0 loss: [0.03774196] finish

9.导出ONNX
if platform.system().lower() != 'windows': mx.onnx.export_model(S_SYM_PATH, S_PARAMS_PATH, [(32, 1, 28, 28)], [np.float32], S_ONNX_MODEL_PATH, verbose=True, dynamic=True)

运行输出
INFO:root:Converting json and weight file to sym and params INFO:root:Converting idx: 0, op: null, name: data INFO:root:Converting idx: 1, op: null, name: hybridsequential0_conv0_weight INFO:root:Converting idx: 2, op: null, name: hybridsequential0_conv0_bias INFO:root:Converting idx: 3, op: Convolution, name: hybridsequential0_conv0_fwd INFO:root:Converting idx: 4, op: Activation, name: hybridsequential0_conv0_relu_fwd INFO:root:Converting idx: 5, op: Pooling, name: hybridsequential0_pool0_fwd INFO:root:Converting idx: 6, op: Flatten, name: hybridsequential0_flatten0_flatten0 INFO:root:Converting idx: 7, op: null, name: hybridsequential0_dense0_weight INFO:root:Converting idx: 8, op: null, name: hybridsequential0_dense0_bias INFO:root:Converting idx: 9, op: FullyConnected, name: hybridsequential0_dense0_fwd INFO:root:Converting idx: 10, op: Activation, name: hybridsequential0_dense0_relu_fwd INFO:root:Converting idx: 11, op: null, name: hybridsequential0_dense1_weight INFO:root:Converting idx: 12, op: null, name: hybridsequential0_dense1_bias INFO:root:Converting idx: 13, op: FullyConnected, name: hybridsequential0_dense1_fwd INFO:root:Output node is: hybridsequential0_dense1_fwd INFO:root:Input shape of the model [(32, 1, 28, 28)] INFO:root:Exported ONNX file ./mxnet_cnn.onnx saved to disk

10. 加载ONNX并运行
if platform.system().lower() != 'windows': model = onnx.load(S_ONNX_MODEL_PATH) print(onnx.checker.check_model(model))# Check that the model is well formed # print(onnx.helper.printable_graph(model.graph))# Print a human readable representation of the graph ls_input_name, ls_output_name = [input.name for input in model.graph.input], [output.name for output in model.graph.output] print("input name ", ls_input_name) print("output name ", ls_output_name) s_input_name = ls_input_name[0]x_input = X_train[:N_BATCH*2, :, :, :].astype(np.float32) ort_val = ort.OrtValue.ortvalue_from_numpy(x_input, S_DEVICE, N_DEVICE_ID) print("val device ", ort_val.device_name()) print("val shape ", ort_val.shape()) print("val data type ", ort_val.data_type()) print("is_tensor ", ort_val.is_tensor()) print("array_equal ", np.array_equal(ort_val.numpy(), x_input)) providers = 'CUDAExecutionProvider' if S_DEVICE == "cuda" else 'CPUExecutionProvider' print("providers ", providers) ort_session = ort.InferenceSession(S_ONNX_MODEL_PATH, providers=[providers])# gpu运行 ort_session.set_providers([providers]) outputs = ort_session.run(None, {s_input_name: ort_val}) print("sess env ", ort_session.get_providers()) print(type(outputs)) print(outputs[0])''' For example ['CUDAExecutionProvider', 'CPUExecutionProvider'] means execute a node using CUDAExecutionProvider if capable, otherwise execute using CPUExecutionProvider. '''

运行输出
None input name['data', 'hybridsequential0_conv0_weight', 'hybridsequential0_conv0_bias', 'hybridsequential0_dense0_weight', 'hybridsequential0_dense0_bias', 'hybridsequential0_dense1_weight', 'hybridsequential0_dense1_bias'] output name['hybridsequential0_dense1_fwd'] val devicecuda val shape[64, 1, 28, 28] val data typetensor(float) is_tensorTrue array_equalTrue providersCUDAExecutionProvider sess env['CUDAExecutionProvider', 'CPUExecutionProvider'] [[-2.69336128e+008.42524242e+00 -3.34120363e-01 -1.17912292e+00 3.82278800e-01 -3.60794234e+003.58125120e-01 -2.58064723e+00 1.55215383e+00 -2.03553891e+00] [ 1.02665892e+01 -6.65782404e+00 -2.04501271e-01 -2.25653172e+00 -6.31941366e+001.13084137e+00 -3.83885235e-018.22283030e-01 -1.21192622e+003.33601260e+00] [-3.27186418e+001.00050325e+015.39114550e-02 -1.44938648e+00 -9.89762247e-01 -2.09957671e+00 -1.49389958e+006.52510405e-01 1.73153889e+00 -1.25597775e+00] [ 5.72116375e-01 -3.36192799e+00 -6.68362260e-01 -2.81247520e+00 8.36382389e+00 -3.67477946e-022.23792076e+00 -2.91093756e-02 -4.56922323e-01 -6.77382052e-01] [ 1.18602552e+01 -5.09683752e+004.54203248e-01 -2.55723000e+00 -8.68753910e+006.96948707e-01 -1.50591761e-01 -3.62227589e-01 9.83437955e-017.46711075e-01] [ 7.33289337e+00 -6.65414715e+001.57180679e+00 -2.62657452e+00 4.11511570e-01 -1.35336161e+00 -1.40558392e-013.81030589e-01 1.73799121e+008.02671254e-01] [-3.02898431e+001.26861107e+00 -2.04946566e+00 -2.52499342e-01 -2.73597687e-01 -3.01714039e+00 -7.10914516e+001.10452967e+01 -5.82177579e-011.86712158e+00] [-7.78098392e+00 -6.01984358e+001.23355007e+001.18682652e+01 -9.83472538e+008.27242088e+00 -1.02135544e+013.95661980e-01 6.63226461e+003.33681512e+00] [-2.72245955e+00 -6.74849796e+00 -6.24665642e+003.11165476e+00 -4.71174330e-011.22390661e+01 -1.23519528e+00 -1.24356663e+00 1.26693976e+005.81862879e+00] [-5.65229607e+00 -1.25138938e+003.68380380e+001.24947300e+01 -8.21508980e+001.61641145e+00 -8.01925087e+008.37018967e-01 -2.64613247e+007.92313635e-01] [-3.73405719e+00 -3.41621947e+00 -7.94842839e-014.55352879e+00 -2.28238964e+001.88887548e+00 -5.84129477e+006.03430390e-01 1.05920439e+012.25430655e+00] [-5.44103146e+00 -5.48421431e+00 -3.62234282e+001.20194650e+00 3.48899674e+001.50794566e+00 -6.30612850e+004.01568127e+00 1.61318648e+009.87832165e+00] [-3.34073186e+008.10987663e+00 -6.43497527e-01 -1.64372277e+00 -4.42907363e-01 -1.46176386e+00 -8.56327295e-015.20323329e-02 1.73289025e+00 -8.17061365e-01] [-6.88457203e+001.38391244e+001.33096969e+001.28132534e+01 -6.20939922e+001.48244214e+00 -6.59804583e+00 -1.38118923e+00 4.26289368e+00 -1.22962976e+00] [-6.09051991e+00 -3.15275192e+001.79273260e+009.92699528e+00 -5.97349882e+003.68225765e+00 -6.47421646e+00 -1.99264419e+00 2.15714622e+002.32836318e+00] [-3.25946307e+008.14360428e+00 -1.00535810e+00 -2.37552500e+00 2.38139248e+00 -2.92597318e+00 -1.54173911e+002.25682306e+00 -2.83430189e-01 -1.33554244e+00] [-2.99147058e+003.86941671e+008.82810593e+002.20121431e+00 -8.40485859e+00 -8.66728902e-01 -5.97998762e+00 -5.21699572e+00 5.80638123e+00 -2.57314467e+00] [ 8.64277363e+00 -4.99241495e+002.84688592e+00 -4.15350378e-01 -1.87728360e-01 -2.40291572e+004.42544132e-01 -4.54446167e-01 -1.88113344e+00 -1.23334014e+00] [-2.00169897e+00 -2.65497804e+001.18750989e+009.70900059e-01 -4.53840446e+00 -2.65584946e+00 -8.23472023e+009.93836498e+00 -5.57100773e-013.42955470e+00] [-3.57249069e+00 -5.03176594e+00 -1.79369414e+00 -5.03321826e-01 -1.97100627e+009.01608944e+006.62497377e+00 -5.48222637e+00 6.09256268e+00 -4.71334040e-01] [-5.27715540e+00 -7.84428477e-01 -6.26944721e-013.87298250e+00 -1.88836837e+001.15252662e+00 -2.98473048e+00 -3.10233998e+00 9.71112537e+003.10839200e+00] [-9.50223565e-01 -6.47654009e+002.26750326e+001.95419586e+00 1.68217969e+001.66003108e+009.82697105e+00 -9.94868219e-01 -2.03924966e+00 -1.88321277e-01] [-3.11575246e+003.43664408e+001.19877796e+014.36916590e+00 -1.17812777e+01 -1.69431508e+00 -5.82668829e+00 -5.09948444e+00 4.15738583e+00 -4.30461359e+00] [ 9.72177792e+00 -5.31352401e-01 -1.21784186e+00 -1.07392669e+00 -7.11223555e+001.67838800e+001.01826215e+00 -8.88240516e-01 6.95959151e-012.38748863e-01] [-2.06619406e+001.86608231e+001.12100420e+014.22539425e+00 -1.21493711e+01 -4.57662535e+00 -6.88935089e+00 -9.81215835e-01 3.87611055e+00 -3.28470826e+00] [-6.73031902e+00 -2.54390073e+00 -1.10151446e+001.51524162e+01 -1.10052080e+016.60323954e+00 -7.94388771e+003.31939721e+00 -1.40840662e+002.65730071e+00] [-1.96954179e+00 -1.13817227e+009.40351069e-01 -1.75684047e+00 3.60373807e+002.01377797e+001.00558109e+01 -1.10547984e+00 5.17374456e-01 -3.94047165e+00] [-5.81787634e+00 -1.20211565e+00 -3.53216052e+001.17569458e+00 4.21314144e+00 -2.53644252e+00 -7.64466667e+004.19782829e+00 4.28840429e-011.04579344e+01] [-4.20310974e+00 -3.19272375e+00 -4.62792778e+002.71683741e+00 4.43899345e+003.31357956e-01 -6.24839544e+003.80388188e+00 -1.22620119e-029.65024757e+00] [-8.26945066e-01 -5.25947523e+003.72772887e-022.30585241e+00 -4.95726252e+00 -1.19987357e+00 -1.20395079e+011.53253164e+01 -2.10372299e-011.89387524e+00] [-5.09596729e+00 -7.76027665e-02 -9.53466833e-012.89041376e+00 -1.50858855e+002.27854323e+00 -1.95591903e+00 -3.15785193e+00 1.00103540e+011.08987451e+00] [-4.01680946e-01 -4.62062168e+003.90530303e-02 -1.66790807e+00 5.43311167e+00 -1.78802896e+00 -2.88405848e+002.93439984e+00 -2.16558409e+008.71198368e+00] [-1.29969406e+00 -3.92871022e+00 -3.82151055e+00 -2.93831253e+00 1.03674269e+018.88044477e-017.88922787e-013.86107159e+00 1.60807288e+00 -3.76913965e-01] [-2.25020099e+00 -8.17249107e+00 -1.82360613e+00 -5.90175152e-01 4.72407389e+009.39436078e-01 -3.85674310e+003.95303702e+00 1.83473241e+009.13874054e+00] [-3.26617742e+00 -2.91517663e+008.37770653e+001.61820054e-01 -2.98638320e+00 -2.47211266e+005.08574843e-014.65608168e+00 2.66001201e+00 -4.67262363e+00] [-2.29874635e+007.77097034e+001.11359918e+00 -2.06103897e+00 -7.61267126e-011.00877440e+001.47708499e+00 -1.20483887e+00 1.99922264e+00 -3.81118345e+00] [-6.87821198e+00 -9.18823600e-012.16773844e+001.07671242e+01 -7.48823595e+002.90310860e+00 -1.02075748e+01 -3.83400464e+00 4.76818371e+004.06564474e+00] [-2.06487226e+008.76828384e+001.10449910e+00 -2.29669046e+00 -1.15668392e+00 -2.50351834e+00 -1.69508122e-02 -1.05916834e+00 1.91057479e+00 -2.64592767e+00] [-2.24318981e+009.02024174e+001.38990092e+00 -2.72154903e+00 1.46101296e-01 -4.43454313e+00 -8.21092844e-01 -2.40808502e-01 3.36577922e-01 -2.63193059e+00] [-4.35961342e+00 -7.74704576e-01 -2.74345660e+00 -3.27951574e+00 1.50971518e+01 -2.80669570e+00 -1.28740633e+003.94157290e+00 4.20372874e-018.37333024e-01] [-2.80749345e+00 -3.33036280e+00 -1.00865018e+004.57633829e+00 -5.03952360e+002.93345642e+00 -8.54609489e+002.26549125e+00 4.73208952e+005.93849993e+00] [-3.31042576e+009.97719002e+004.38573778e-01 -1.35296178e+00 -1.21057940e+00 -2.46178842e+00 -8.34564090e-012.19030753e-01 2.03147411e+00 -1.80211437e+00] [-2.14534068e+00 -2.93023801e+004.41405416e-01 -2.29865336e+00 1.47422533e+01 -1.86358702e+004.61042017e-01 -6.20108247e-01 -1.36792552e+002.14018896e-01] [-2.64241481e+00 -2.28332114e+002.01109338e+009.67352509e-01 6.09287119e+00 -2.35626236e-01 -3.02941656e+004.32772923e+00 -4.63955021e+003.73136783e+00] [-4.55847168e+001.04014337e+009.12987328e+002.06433630e+00 -1.67355919e+00 -1.49593079e+004.09124941e-012.41894865e+00 -6.86871633e-02 -4.42179346e+00] [-6.10608578e-01 -2.73860097e+00 -1.09864855e+00 -5.68899512e-01 2.45831108e+005.50326490e+001.22601585e+01 -5.23877192e+00 2.11066246e+00 -2.98584485e+00] [-5.96745872e+00 -1.91458237e+00 -3.10774088e-011.00216856e+01 -3.81997776e+007.14399862e+00 -4.61386251e+00 -5.18248987e+00 4.25162363e+001.18878789e-01] [-4.24126434e+00 -9.63249326e-011.06391013e+004.45316315e+00 -4.47125340e+00 -1.21906054e+00 -9.75789547e+001.10335569e+01 -1.17632782e+008.78942788e-01] [ 3.76867175e-01 -4.75102758e+00 -2.59345794e+00 -3.96257102e-01 -7.50329159e-031.81642962e+00 -6.01041269e+009.97849655e+00 -2.57468176e+005.00644207e+00] [-2.21995306e+00 -4.23465443e+00 -2.19536662e+00 -3.71420813e+00 1.49460211e+01 -2.73240638e+003.03538777e-024.29108334e+00 1.21963143e+006.85284913e-01] [-1.93794692e+00 -3.39166284e+00 -3.41372967e-01 -2.16144085e-01 1.32588074e-01 -3.83050554e-02 -7.32452822e+009.68561840e+00 -1.16044319e+003.63913298e+00] [-3.31972694e+00 -4.84879112e+00 -3.41381001e+001.93332338e+00 -5.16150045e+001.05730495e+011.52961123e+00 -4.64916992e+00 4.11477995e+004.80105543e+00] [-3.22445488e+001.01509037e+016.03635088e-02 -1.48001885e+00 -3.28380197e-01 -2.36782789e+00 -8.66441727e-016.68077886e-01 1.45576179e+00 -1.95271623e+00] [-5.64618587e+00 -3.71156931e+00 -2.31174397e+001.76912701e+00 2.95111752e+002.09562635e+00 -5.34609461e+00 -2.59399921e-01 1.10373080e+001.10444403e+01] [ 1.32743053e+01 -5.11389780e+001.10446036e+00 -5.01595545e+00 -4.87907743e+004.46035750e-024.57046556e+00 -1.76434004e+00 -5.40824793e-02 -1.01547205e+00] [-4.27589798e+001.26832044e+006.49948978e+00 -8.06193352e-02 1.34645328e-016.92090929e-01 -8.92272711e-012.15252757e+00 -9.95365143e-01 -2.46636438e+00] [-5.08349514e+00 -1.79646879e-019.57399654e+003.35643005e+00 -3.42183971e+00 -2.33653522e+00 -1.98645079e+001.50538552e+00 -1.43313253e+00 -9.75638926e-01] [-8.48450565e+00 -4.32870531e+00 -1.54253757e+001.61029205e+01 -8.41084957e+005.45092726e+00 -1.00705996e+012.31331086e+00 1.05400957e-015.19563723e+00] [-4.89484310e+00 -4.33120441e+00 -3.71932483e+00 -1.18670315e-01 4.70187807e+002.67010808e-01 -5.52650118e+005.62736416e+00 5.05499423e-018.69004250e+00] [-1.94048929e+007.98250961e+00 -3.75457823e-01 -2.09433126e+00 -4.63896915e-02 -2.64650702e+00 -5.73348761e-01 -1.34597674e-01 1.26143038e+00 -1.32178509e+00] [-6.36190224e+005.44552183e+001.89667892e+003.91665459e+00 -1.99860716e+002.76620483e+003.57649279e+00 -3.53898597e+00 -1.21411644e-01 -2.60307193e+00] [-3.60333633e+004.92228556e+002.87770915e+001.31902504e+00 -4.28756446e-01 -3.29862523e+00 -2.29294825e+00 -1.10349190e+00 3.81862259e+00 -1.23572731e+00] [ 1.06328213e+00 -6.86575174e+00 -3.61938500e+001.15000987e+00 -2.32747698e+001.32029047e+01 -3.19671059e+008.91836137e-02 4.70500660e+002.51928687e+00] [ 1.00363417e+01 -8.02793884e+009.38569680e-02 -2.78522468e+00 -2.84671545e+002.95867848e+004.23545599e+002.52083421e+00 -3.35666537e+001.45630157e+00]]

你甚至不愿意Start的Github 【Mxnet速查_CPU和GPU的mnist预测训练_模型导出_模型导入再预测_导出onnx并预测】ai_fast_handbook

    推荐阅读