ycdfwzy opened a new issue #7101:
URL: https://github.com/apache/tvm/issues/7101


   Hello!
   I installed TVM on win10. After running 
`tutorials\get_started\relay_quick_start.py`, I got the following outputs, but 
did't know how to solve it
   ```
   #[version = "0.0.5"]
   def @main(%data: Tensor[(1, 3, 224, 224), float32], %bn_data_gamma: 
Tensor[(3), float32], %bn_data_beta: Tensor[(3), float32], 
%bn_data_moving_mean: Tensor[(3), float32], %bn_data_moving_var: Tensor[(3), 
float32], %conv0_weight: Tensor[(64, 3, 7, 7), float32], %bn0_gamma: 
Tensor[(64), float32], %bn0_beta: Tensor[(64), float32], %bn0_moving_mean: 
Tensor[(64), float32], %bn0_moving_var: Tensor[(64), float32], 
%stage1_unit1_bn1_gamma: Tensor[(64), float32], %stage1_unit1_bn1_beta: 
Tensor[(64), float32], %stage1_unit1_bn1_moving_mean: Tensor[(64), float32], 
%stage1_unit1_bn1_moving_var: Tensor[(64), float32], 
%stage1_unit1_conv1_weight: Tensor[(64, 64, 3, 3), float32], 
%stage1_unit1_bn2_gamma: Tensor[(64), float32], %stage1_unit1_bn2_beta: 
Tensor[(64), float32], %stage1_unit1_bn2_moving_mean: Tensor[(64), float32], 
%stage1_unit1_bn2_moving_var: Tensor[(64), float32], 
%stage1_unit1_conv2_weight: Tensor[(64, 64, 3, 3), float32], 
%stage1_unit1_sc_weight: Tensor[(64, 64, 1, 1), float32], 
 %stage1_unit2_bn1_gamma: Tensor[(64), float32], %stage1_unit2_bn1_beta: 
Tensor[(64), float32], %stage1_unit2_bn1_moving_mean: Tensor[(64), float32], 
%stage1_unit2_bn1_moving_var: Tensor[(64), float32], 
%stage1_unit2_conv1_weight: Tensor[(64, 64, 3, 3), float32], 
%stage1_unit2_bn2_gamma: Tensor[(64), float32], %stage1_unit2_bn2_beta: 
Tensor[(64), float32], %stage1_unit2_bn2_moving_mean: Tensor[(64), float32], 
%stage1_unit2_bn2_moving_var: Tensor[(64), float32], 
%stage1_unit2_conv2_weight: Tensor[(64, 64, 3, 3), float32], 
%stage2_unit1_bn1_gamma: Tensor[(64), float32], %stage2_unit1_bn1_beta: 
Tensor[(64), float32], %stage2_unit1_bn1_moving_mean: Tensor[(64), float32], 
%stage2_unit1_bn1_moving_var: Tensor[(64), float32], 
%stage2_unit1_conv1_weight: Tensor[(128, 64, 3, 3), float32], 
%stage2_unit1_bn2_gamma: Tensor[(128), float32], %stage2_unit1_bn2_beta: 
Tensor[(128), float32], %stage2_unit1_bn2_moving_mean: Tensor[(128), float32], 
%stage2_unit1_bn2_moving_var: Tensor[(128), float32], %
 stage2_unit1_conv2_weight: Tensor[(128, 128, 3, 3), float32], 
%stage2_unit1_sc_weight: Tensor[(128, 64, 1, 1), float32], 
%stage2_unit2_bn1_gamma: Tensor[(128), float32], %stage2_unit2_bn1_beta: 
Tensor[(128), float32], %stage2_unit2_bn1_moving_mean: Tensor[(128), float32], 
%stage2_unit2_bn1_moving_var: Tensor[(128), float32], 
%stage2_unit2_conv1_weight: Tensor[(128, 128, 3, 3), float32], 
%stage2_unit2_bn2_gamma: Tensor[(128), float32], %stage2_unit2_bn2_beta: 
Tensor[(128), float32], %stage2_unit2_bn2_moving_mean: Tensor[(128), float32], 
%stage2_unit2_bn2_moving_var: Tensor[(128), float32], 
%stage2_unit2_conv2_weight: Tensor[(128, 128, 3, 3), float32], 
%stage3_unit1_bn1_gamma: Tensor[(128), float32], %stage3_unit1_bn1_beta: 
Tensor[(128), float32], %stage3_unit1_bn1_moving_mean: Tensor[(128), float32], 
%stage3_unit1_bn1_moving_var: Tensor[(128), float32], 
%stage3_unit1_conv1_weight: Tensor[(256, 128, 3, 3), float32], 
%stage3_unit1_bn2_gamma: Tensor[(256), float32], %stage3_unit1_bn2_be
 ta: Tensor[(256), float32], %stage3_unit1_bn2_moving_mean: Tensor[(256), 
float32], %stage3_unit1_bn2_moving_var: Tensor[(256), float32], 
%stage3_unit1_conv2_weight: Tensor[(256, 256, 3, 3), float32], 
%stage3_unit1_sc_weight: Tensor[(256, 128, 1, 1), float32], 
%stage3_unit2_bn1_gamma: Tensor[(256), float32], %stage3_unit2_bn1_beta: 
Tensor[(256), float32], %stage3_unit2_bn1_moving_mean: Tensor[(256), float32], 
%stage3_unit2_bn1_moving_var: Tensor[(256), float32], 
%stage3_unit2_conv1_weight: Tensor[(256, 256, 3, 3), float32], 
%stage3_unit2_bn2_gamma: Tensor[(256), float32], %stage3_unit2_bn2_beta: 
Tensor[(256), float32], %stage3_unit2_bn2_moving_mean: Tensor[(256), float32], 
%stage3_unit2_bn2_moving_var: Tensor[(256), float32], 
%stage3_unit2_conv2_weight: Tensor[(256, 256, 3, 3), float32], 
%stage4_unit1_bn1_gamma: Tensor[(256), float32], %stage4_unit1_bn1_beta: 
Tensor[(256), float32], %stage4_unit1_bn1_moving_mean: Tensor[(256), float32], 
%stage4_unit1_bn1_moving_var: Tensor[(256), flo
 at32], %stage4_unit1_conv1_weight: Tensor[(512, 256, 3, 3), float32], 
%stage4_unit1_bn2_gamma: Tensor[(512), float32], %stage4_unit1_bn2_beta: 
Tensor[(512), float32], %stage4_unit1_bn2_moving_mean: Tensor[(512), float32], 
%stage4_unit1_bn2_moving_var: Tensor[(512), float32], 
%stage4_unit1_conv2_weight: Tensor[(512, 512, 3, 3), float32], 
%stage4_unit1_sc_weight: Tensor[(512, 256, 1, 1), float32], 
%stage4_unit2_bn1_gamma: Tensor[(512), float32], %stage4_unit2_bn1_beta: 
Tensor[(512), float32], %stage4_unit2_bn1_moving_mean: Tensor[(512), float32], 
%stage4_unit2_bn1_moving_var: Tensor[(512), float32], 
%stage4_unit2_conv1_weight: Tensor[(512, 512, 3, 3), float32], 
%stage4_unit2_bn2_gamma: Tensor[(512), float32], %stage4_unit2_bn2_beta: 
Tensor[(512), float32], %stage4_unit2_bn2_moving_mean: Tensor[(512), float32], 
%stage4_unit2_bn2_moving_var: Tensor[(512), float32], 
%stage4_unit2_conv2_weight: Tensor[(512, 512, 3, 3), float32], %bn1_gamma: 
Tensor[(512), float32], %bn1_beta: Tensor[(512),
  float32], %bn1_moving_mean: Tensor[(512), float32], %bn1_moving_var: 
Tensor[(512), float32], %fc1_weight: Tensor[(1000, 512), float32], %fc1_bias: 
Tensor[(1000), float32]) -> Tensor[(1, 1000), float32] {
     %0 = nn.batch_norm(%data, %bn_data_gamma, %bn_data_beta, 
%bn_data_moving_mean, %bn_data_moving_var, epsilon=2e-05f, scale=False) /* 
ty=(Tensor[(1, 3, 224, 224), float32], Tensor[(3), float32], Tensor[(3), 
float32]) */;
     %1 = %0.0;
     %2 = nn.conv2d(%1, %conv0_weight, strides=[2, 2], padding=[3, 3, 3, 3], 
channels=64, kernel_size=[7, 7]) /* ty=Tensor[(1, 64, 112, 112), float32] */;
     %3 = nn.batch_norm(%2, %bn0_gamma, %bn0_beta, %bn0_moving_mean, 
%bn0_moving_var, epsilon=2e-05f) /* ty=(Tensor[(1, 64, 112, 112), float32], 
Tensor[(64), float32], Tensor[(64), float32]) */;
     %4 = %3.0;
     %5 = nn.relu(%4) /* ty=Tensor[(1, 64, 112, 112), float32] */;
     %6 = nn.max_pool2d(%5, pool_size=[3, 3], strides=[2, 2], padding=[1, 1, 1, 
1]) /* ty=Tensor[(1, 64, 56, 56), float32] */;
     %7 = nn.batch_norm(%6, %stage1_unit1_bn1_gamma, %stage1_unit1_bn1_beta, 
%stage1_unit1_bn1_moving_mean, %stage1_unit1_bn1_moving_var, epsilon=2e-05f) /* 
ty=(Tensor[(1, 64, 56, 56), float32], Tensor[(64), float32], Tensor[(64), 
float32]) */;
     %8 = %7.0;
     %9 = nn.relu(%8) /* ty=Tensor[(1, 64, 56, 56), float32] */;
     %10 = nn.conv2d(%9, %stage1_unit1_conv1_weight, padding=[1, 1, 1, 1], 
channels=64, kernel_size=[3, 3]) /* ty=Tensor[(1, 64, 56, 56), float32] */;
     %11 = nn.batch_norm(%10, %stage1_unit1_bn2_gamma, %stage1_unit1_bn2_beta, 
%stage1_unit1_bn2_moving_mean, %stage1_unit1_bn2_moving_var, epsilon=2e-05f) /* 
ty=(Tensor[(1, 64, 56, 56), float32], Tensor[(64), float32], Tensor[(64), 
float32]) */;
     %12 = %11.0;
     %13 = nn.relu(%12) /* ty=Tensor[(1, 64, 56, 56), float32] */;
     %14 = nn.conv2d(%13, %stage1_unit1_conv2_weight, padding=[1, 1, 1, 1], 
channels=64, kernel_size=[3, 3]) /* ty=Tensor[(1, 64, 56, 56), float32] */;
     %15 = nn.conv2d(%9, %stage1_unit1_sc_weight, padding=[0, 0, 0, 0], 
channels=64, kernel_size=[1, 1]) /* ty=Tensor[(1, 64, 56, 56), float32] */;
     %16 = add(%14, %15) /* ty=Tensor[(1, 64, 56, 56), float32] */;
     %17 = nn.batch_norm(%16, %stage1_unit2_bn1_gamma, %stage1_unit2_bn1_beta, 
%stage1_unit2_bn1_moving_mean, %stage1_unit2_bn1_moving_var, epsilon=2e-05f) /* 
ty=(Tensor[(1, 64, 56, 56), float32], Tensor[(64), float32], Tensor[(64), 
float32]) */;
     %18 = %17.0;
     %19 = nn.relu(%18) /* ty=Tensor[(1, 64, 56, 56), float32] */;
     %20 = nn.conv2d(%19, %stage1_unit2_conv1_weight, padding=[1, 1, 1, 1], 
channels=64, kernel_size=[3, 3]) /* ty=Tensor[(1, 64, 56, 56), float32] */;
     %21 = nn.batch_norm(%20, %stage1_unit2_bn2_gamma, %stage1_unit2_bn2_beta, 
%stage1_unit2_bn2_moving_mean, %stage1_unit2_bn2_moving_var, epsilon=2e-05f) /* 
ty=(Tensor[(1, 64, 56, 56), float32], Tensor[(64), float32], Tensor[(64), 
float32]) */;
     %22 = %21.0;
     %23 = nn.relu(%22) /* ty=Tensor[(1, 64, 56, 56), float32] */;
     %24 = nn.conv2d(%23, %stage1_unit2_conv2_weight, padding=[1, 1, 1, 1], 
channels=64, kernel_size=[3, 3]) /* ty=Tensor[(1, 64, 56, 56), float32] */;
     %25 = add(%24, %16) /* ty=Tensor[(1, 64, 56, 56), float32] */;
     %26 = nn.batch_norm(%25, %stage2_unit1_bn1_gamma, %stage2_unit1_bn1_beta, 
%stage2_unit1_bn1_moving_mean, %stage2_unit1_bn1_moving_var, epsilon=2e-05f) /* 
ty=(Tensor[(1, 64, 56, 56), float32], Tensor[(64), float32], Tensor[(64), 
float32]) */;
     %27 = %26.0;
     %28 = nn.relu(%27) /* ty=Tensor[(1, 64, 56, 56), float32] */;
     %29 = nn.conv2d(%28, %stage2_unit1_conv1_weight, strides=[2, 2], 
padding=[1, 1, 1, 1], channels=128, kernel_size=[3, 3]) /* ty=Tensor[(1, 128, 
28, 28), float32] */;
     %30 = nn.batch_norm(%29, %stage2_unit1_bn2_gamma, %stage2_unit1_bn2_beta, 
%stage2_unit1_bn2_moving_mean, %stage2_unit1_bn2_moving_var, epsilon=2e-05f) /* 
ty=(Tensor[(1, 128, 28, 28), float32], Tensor[(128), float32], Tensor[(128), 
float32]) */;
     %31 = %30.0;
     %32 = nn.relu(%31) /* ty=Tensor[(1, 128, 28, 28), float32] */;
     %33 = nn.conv2d(%32, %stage2_unit1_conv2_weight, padding=[1, 1, 1, 1], 
channels=128, kernel_size=[3, 3]) /* ty=Tensor[(1, 128, 28, 28), float32] */;
     %34 = nn.conv2d(%28, %stage2_unit1_sc_weight, strides=[2, 2], padding=[0, 
0, 0, 0], channels=128, kernel_size=[1, 1]) /* ty=Tensor[(1, 128, 28, 28), 
float32] */;
     %35 = add(%33, %34) /* ty=Tensor[(1, 128, 28, 28), float32] */;
     %36 = nn.batch_norm(%35, %stage2_unit2_bn1_gamma, %stage2_unit2_bn1_beta, 
%stage2_unit2_bn1_moving_mean, %stage2_unit2_bn1_moving_var, epsilon=2e-05f) /* 
ty=(Tensor[(1, 128, 28, 28), float32], Tensor[(128), float32], Tensor[(128), 
float32]) */;
     %37 = %36.0;
     %38 = nn.relu(%37) /* ty=Tensor[(1, 128, 28, 28), float32] */;
     %39 = nn.conv2d(%38, %stage2_unit2_conv1_weight, padding=[1, 1, 1, 1], 
channels=128, kernel_size=[3, 3]) /* ty=Tensor[(1, 128, 28, 28), float32] */;
     %40 = nn.batch_norm(%39, %stage2_unit2_bn2_gamma, %stage2_unit2_bn2_beta, 
%stage2_unit2_bn2_moving_mean, %stage2_unit2_bn2_moving_var, epsilon=2e-05f) /* 
ty=(Tensor[(1, 128, 28, 28), float32], Tensor[(128), float32], Tensor[(128), 
float32]) */;
     %41 = %40.0;
     %42 = nn.relu(%41) /* ty=Tensor[(1, 128, 28, 28), float32] */;
     %43 = nn.conv2d(%42, %stage2_unit2_conv2_weight, padding=[1, 1, 1, 1], 
channels=128, kernel_size=[3, 3]) /* ty=Tensor[(1, 128, 28, 28), float32] */;
     %44 = add(%43, %35) /* ty=Tensor[(1, 128, 28, 28), float32] */;
     %45 = nn.batch_norm(%44, %stage3_unit1_bn1_gamma, %stage3_unit1_bn1_beta, 
%stage3_unit1_bn1_moving_mean, %stage3_unit1_bn1_moving_var, epsilon=2e-05f) /* 
ty=(Tensor[(1, 128, 28, 28), float32], Tensor[(128), float32], Tensor[(128), 
float32]) */;
     %46 = %45.0;
     %47 = nn.relu(%46) /* ty=Tensor[(1, 128, 28, 28), float32] */;
     %48 = nn.conv2d(%47, %stage3_unit1_conv1_weight, strides=[2, 2], 
padding=[1, 1, 1, 1], channels=256, kernel_size=[3, 3]) /* ty=Tensor[(1, 256, 
14, 14), float32] */;
     %49 = nn.batch_norm(%48, %stage3_unit1_bn2_gamma, %stage3_unit1_bn2_beta, 
%stage3_unit1_bn2_moving_mean, %stage3_unit1_bn2_moving_var, epsilon=2e-05f) /* 
ty=(Tensor[(1, 256, 14, 14), float32], Tensor[(256), float32], Tensor[(256), 
float32]) */;
     %50 = %49.0;
     %51 = nn.relu(%50) /* ty=Tensor[(1, 256, 14, 14), float32] */;
     %52 = nn.conv2d(%51, %stage3_unit1_conv2_weight, padding=[1, 1, 1, 1], 
channels=256, kernel_size=[3, 3]) /* ty=Tensor[(1, 256, 14, 14), float32] */;
     %53 = nn.conv2d(%47, %stage3_unit1_sc_weight, strides=[2, 2], padding=[0, 
0, 0, 0], channels=256, kernel_size=[1, 1]) /* ty=Tensor[(1, 256, 14, 14), 
float32] */;
     %54 = add(%52, %53) /* ty=Tensor[(1, 256, 14, 14), float32] */;
     %55 = nn.batch_norm(%54, %stage3_unit2_bn1_gamma, %stage3_unit2_bn1_beta, 
%stage3_unit2_bn1_moving_mean, %stage3_unit2_bn1_moving_var, epsilon=2e-05f) /* 
ty=(Tensor[(1, 256, 14, 14), float32], Tensor[(256), float32], Tensor[(256), 
float32]) */;
     %56 = %55.0;
     %57 = nn.relu(%56) /* ty=Tensor[(1, 256, 14, 14), float32] */;
     %58 = nn.conv2d(%57, %stage3_unit2_conv1_weight, padding=[1, 1, 1, 1], 
channels=256, kernel_size=[3, 3]) /* ty=Tensor[(1, 256, 14, 14), float32] */;
     %59 = nn.batch_norm(%58, %stage3_unit2_bn2_gamma, %stage3_unit2_bn2_beta, 
%stage3_unit2_bn2_moving_mean, %stage3_unit2_bn2_moving_var, epsilon=2e-05f) /* 
ty=(Tensor[(1, 256, 14, 14), float32], Tensor[(256), float32], Tensor[(256), 
float32]) */;
     %60 = %59.0;
     %61 = nn.relu(%60) /* ty=Tensor[(1, 256, 14, 14), float32] */;
     %62 = nn.conv2d(%61, %stage3_unit2_conv2_weight, padding=[1, 1, 1, 1], 
channels=256, kernel_size=[3, 3]) /* ty=Tensor[(1, 256, 14, 14), float32] */;
     %63 = add(%62, %54) /* ty=Tensor[(1, 256, 14, 14), float32] */;
     %64 = nn.batch_norm(%63, %stage4_unit1_bn1_gamma, %stage4_unit1_bn1_beta, 
%stage4_unit1_bn1_moving_mean, %stage4_unit1_bn1_moving_var, epsilon=2e-05f) /* 
ty=(Tensor[(1, 256, 14, 14), float32], Tensor[(256), float32], Tensor[(256), 
float32]) */;
     %65 = %64.0;
     %66 = nn.relu(%65) /* ty=Tensor[(1, 256, 14, 14), float32] */;
     %67 = nn.conv2d(%66, %stage4_unit1_conv1_weight, strides=[2, 2], 
padding=[1, 1, 1, 1], channels=512, kernel_size=[3, 3]) /* ty=Tensor[(1, 512, 
7, 7), float32] */;
     %68 = nn.batch_norm(%67, %stage4_unit1_bn2_gamma, %stage4_unit1_bn2_beta, 
%stage4_unit1_bn2_moving_mean, %stage4_unit1_bn2_moving_var, epsilon=2e-05f) /* 
ty=(Tensor[(1, 512, 7, 7), float32], Tensor[(512), float32], Tensor[(512), 
float32]) */;
     %69 = %68.0;
     %70 = nn.relu(%69) /* ty=Tensor[(1, 512, 7, 7), float32] */;
     %71 = nn.conv2d(%70, %stage4_unit1_conv2_weight, padding=[1, 1, 1, 1], 
channels=512, kernel_size=[3, 3]) /* ty=Tensor[(1, 512, 7, 7), float32] */;
     %72 = nn.conv2d(%66, %stage4_unit1_sc_weight, strides=[2, 2], padding=[0, 
0, 0, 0], channels=512, kernel_size=[1, 1]) /* ty=Tensor[(1, 512, 7, 7), 
float32] */;
     %73 = add(%71, %72) /* ty=Tensor[(1, 512, 7, 7), float32] */;
     %74 = nn.batch_norm(%73, %stage4_unit2_bn1_gamma, %stage4_unit2_bn1_beta, 
%stage4_unit2_bn1_moving_mean, %stage4_unit2_bn1_moving_var, epsilon=2e-05f) /* 
ty=(Tensor[(1, 512, 7, 7), float32], Tensor[(512), float32], Tensor[(512), 
float32]) */;
     %75 = %74.0;
     %76 = nn.relu(%75) /* ty=Tensor[(1, 512, 7, 7), float32] */;
     %77 = nn.conv2d(%76, %stage4_unit2_conv1_weight, padding=[1, 1, 1, 1], 
channels=512, kernel_size=[3, 3]) /* ty=Tensor[(1, 512, 7, 7), float32] */;
     %78 = nn.batch_norm(%77, %stage4_unit2_bn2_gamma, %stage4_unit2_bn2_beta, 
%stage4_unit2_bn2_moving_mean, %stage4_unit2_bn2_moving_var, epsilon=2e-05f) /* 
ty=(Tensor[(1, 512, 7, 7), float32], Tensor[(512), float32], Tensor[(512), 
float32]) */;
     %79 = %78.0;
     %80 = nn.relu(%79) /* ty=Tensor[(1, 512, 7, 7), float32] */;
     %81 = nn.conv2d(%80, %stage4_unit2_conv2_weight, padding=[1, 1, 1, 1], 
channels=512, kernel_size=[3, 3]) /* ty=Tensor[(1, 512, 7, 7), float32] */;
     %82 = add(%81, %73) /* ty=Tensor[(1, 512, 7, 7), float32] */;
     %83 = nn.batch_norm(%82, %bn1_gamma, %bn1_beta, %bn1_moving_mean, 
%bn1_moving_var, epsilon=2e-05f) /* ty=(Tensor[(1, 512, 7, 7), float32], 
Tensor[(512), float32], Tensor[(512), float32]) */;
     %84 = %83.0;
     %85 = nn.relu(%84) /* ty=Tensor[(1, 512, 7, 7), float32] */;
     %86 = nn.global_avg_pool2d(%85) /* ty=Tensor[(1, 512, 1, 1), float32] */;
     %87 = nn.batch_flatten(%86) /* ty=Tensor[(1, 512), float32] */;
     %88 = nn.dense(%87, %fc1_weight, units=1000) /* ty=Tensor[(1, 1000), 
float32] */;
     %89 = nn.bias_add(%88, %fc1_bias, axis=-1) /* ty=Tensor[(1, 1000), 
float32] */;
     nn.softmax(%89) /* ty=Tensor[(1, 1000), float32] */
   }
   
   download failed due to URLError(ConnectionResetError(10054, 'An existing 
connection was forcibly closed by the remote host', None, 10054, None)), 
retrying, 2 attempts left
   download failed due to URLError(ConnectionResetError(10054, 'An existing 
connection was forcibly closed by the remote host', None, 10054, None)), 
retrying, 1 attempt left
   WARNING:root:Failed to download tophub package for cuda: <urlopen error 
[WinError 10054] An existing connection was forcibly closed by the remote host>
   WARNING:autotvm:Cannot find config for target=cuda -keys=cuda,gpu 
-max_num_threads=1024 -model=unknown -thread_warp_size=32, 
workload=('conv2d_nchw.cuda', ('TENSOR', (1, 3, 224, 224), 'float32'), 
('TENSOR', (64, 3, 7, 7), 'float32'), (2, 2), (3, 3, 3, 3), (1, 1), 'float32'). 
A fallback configuration is used, which may bring great performance regression.
   WARNING:autotvm:Cannot find config for target=cuda -keys=cuda,gpu 
-max_num_threads=1024 -model=unknown -thread_warp_size=32, 
workload=('conv2d_nchw.cuda', ('TENSOR', (1, 64, 56, 56), 'float32'), 
('TENSOR', (64, 64, 3, 3), 'float32'), (1, 1), (1, 1, 1, 1), (1, 1), 
'float32'). A fallback configuration is used, which may bring great performance 
regression.
   WARNING:autotvm:Cannot find config for target=cuda -keys=cuda,gpu 
-max_num_threads=1024 -model=unknown -thread_warp_size=32, 
workload=('conv2d_nchw.cuda', ('TENSOR', (1, 64, 56, 56), 'float32'), 
('TENSOR', (64, 64, 1, 1), 'float32'), (1, 1), (0, 0, 0, 0), (1, 1), 
'float32'). A fallback configuration is used, which may bring great performance 
regression.
   WARNING:autotvm:Cannot find config for target=cuda -keys=cuda,gpu 
-max_num_threads=1024 -model=unknown -thread_warp_size=32, 
workload=('conv2d_nchw.cuda', ('TENSOR', (1, 64, 56, 56), 'float32'), 
('TENSOR', (128, 64, 3, 3), 'float32'), (2, 2), (1, 1, 1, 1), (1, 1), 
'float32'). A fallback configuration is used, which may bring great performance 
regression.
   WARNING:autotvm:Cannot find config for target=cuda -keys=cuda,gpu 
-max_num_threads=1024 -model=unknown -thread_warp_size=32, 
workload=('conv2d_nchw.cuda', ('TENSOR', (1, 128, 28, 28), 'float32'), 
('TENSOR', (128, 128, 3, 3), 'float32'), (1, 1), (1, 1, 1, 1), (1, 1), 
'float32'). A fallback configuration is used, which may bring great performance 
regression.
   WARNING:autotvm:Cannot find config for target=cuda -keys=cuda,gpu 
-max_num_threads=1024 -model=unknown -thread_warp_size=32, 
workload=('conv2d_nchw.cuda', ('TENSOR', (1, 64, 56, 56), 'float32'), 
('TENSOR', (128, 64, 1, 1), 'float32'), (2, 2), (0, 0, 0, 0), (1, 1), 
'float32'). A fallback configuration is used, which may bring great performance 
regression.
   WARNING:autotvm:Cannot find config for target=cuda -keys=cuda,gpu 
-max_num_threads=1024 -model=unknown -thread_warp_size=32, 
workload=('conv2d_nchw.cuda', ('TENSOR', (1, 128, 28, 28), 'float32'), 
('TENSOR', (256, 128, 3, 3), 'float32'), (2, 2), (1, 1, 1, 1), (1, 1), 
'float32'). A fallback configuration is used, which may bring great performance 
regression.
   WARNING:autotvm:Cannot find config for target=cuda -keys=cuda,gpu 
-max_num_threads=1024 -model=unknown -thread_warp_size=32, 
workload=('conv2d_nchw.cuda', ('TENSOR', (1, 256, 14, 14), 'float32'), 
('TENSOR', (256, 256, 3, 3), 'float32'), (1, 1), (1, 1, 1, 1), (1, 1), 
'float32'). A fallback configuration is used, which may bring great performance 
regression.
   WARNING:autotvm:Cannot find config for target=cuda -keys=cuda,gpu 
-max_num_threads=1024 -model=unknown -thread_warp_size=32, 
workload=('conv2d_nchw.cuda', ('TENSOR', (1, 128, 28, 28), 'float32'), 
('TENSOR', (256, 128, 1, 1), 'float32'), (2, 2), (0, 0, 0, 0), (1, 1), 
'float32'). A fallback configuration is used, which may bring great performance 
regression.
   WARNING:autotvm:Cannot find config for target=cuda -keys=cuda,gpu 
-max_num_threads=1024 -model=unknown -thread_warp_size=32, 
workload=('conv2d_nchw.cuda', ('TENSOR', (1, 256, 14, 14), 'float32'), 
('TENSOR', (512, 256, 3, 3), 'float32'), (2, 2), (1, 1, 1, 1), (1, 1), 
'float32'). A fallback configuration is used, which may bring great performance 
regression.
   WARNING:autotvm:Cannot find config for target=cuda -keys=cuda,gpu 
-max_num_threads=1024 -model=unknown -thread_warp_size=32, 
workload=('conv2d_nchw.cuda', ('TENSOR', (1, 512, 7, 7), 'float32'), ('TENSOR', 
(512, 512, 3, 3), 'float32'), (1, 1), (1, 1, 1, 1), (1, 1), 'float32'). A 
fallback configuration is used, which may bring great performance regression.
   WARNING:autotvm:Cannot find config for target=cuda -keys=cuda,gpu 
-max_num_threads=1024 -model=unknown -thread_warp_size=32, 
workload=('conv2d_nchw.cuda', ('TENSOR', (1, 256, 14, 14), 'float32'), 
('TENSOR', (512, 256, 1, 1), 'float32'), (2, 2), (0, 0, 0, 0), (1, 1), 
'float32'). A fallback configuration is used, which may bring great performance 
regression.
   WARNING:autotvm:Cannot find config for target=cuda -keys=cuda,gpu 
-max_num_threads=1024 -model=unknown -thread_warp_size=32, 
workload=('dense_small_batch.cuda', ('TENSOR', (1, 512), 'float32'), ('TENSOR', 
(1000, 512), 'float32'), None, 'float32'). A fallback configuration is used, 
which may bring great performance regression.download failed due to 
URLError(ConnectionResetError(10054, 'An existing connection was forcibly 
closed by the remote host', None, 10054, None)), retrying, 2 attempts left
   download failed due to URLError(ConnectionResetError(10054, 'An existing 
connection was forcibly closed by the remote host', None, 10054, None)), 
retrying, 1 attempt left
   Traceback (most recent call last):
     File "relay_quick_start.py", line 101, in <module>
       lib = relay.build(mod, target, params=params)
     File 
"C:\Users\ycdfwzy\AppData\Roaming\Python\Python37\site-packages\tvm-0.8.dev338+g28647f2e7-py3.7-win-amd64.egg\tvm\relay\build_module.py",
 line 275, in build
       graph_json, mod, params = bld_mod.build(mod, target, target_host, params)
     File 
"C:\Users\ycdfwzy\AppData\Roaming\Python\Python37\site-packages\tvm-0.8.dev338+g28647f2e7-py3.7-win-amd64.egg\tvm\relay\build_module.py",
 line 138, in build
       self._build(mod, target, target_host)
     File 
"C:\Users\ycdfwzy\AppData\Roaming\Python\Python37\site-packages\tvm-0.8.dev338+g28647f2e7-py3.7-win-amd64.egg\tvm\_ffi\_ctypes\packed_func.py",
 line 237, in __call__
       raise get_last_ffi_error()
   tvm._ffi.base.TVMError: Traceback (most recent call last):
     File "D:\softwares\anaconda3\envs\tvm-dev\lib\urllib\request.py", line 
1350, in do_open
       encode_chunked=req.has_header('Transfer-encoding'))
     File "D:\softwares\anaconda3\envs\tvm-dev\lib\http\client.py", line 1277, 
in request
       self._send_request(method, url, body, headers, encode_chunked)
     File "D:\softwares\anaconda3\envs\tvm-dev\lib\http\client.py", line 1323, 
in _send_request
       self.endheaders(body, encode_chunked=encode_chunked)
     File "D:\softwares\anaconda3\envs\tvm-dev\lib\http\client.py", line 1272, 
in endheaders
       self._send_output(message_body, encode_chunked=encode_chunked)
     File "D:\softwares\anaconda3\envs\tvm-dev\lib\http\client.py", line 1032, 
in _send_output
       self.send(msg)
     File "D:\softwares\anaconda3\envs\tvm-dev\lib\http\client.py", line 972, 
in send
       self.connect()
     File "D:\softwares\anaconda3\envs\tvm-dev\lib\http\client.py", line 1447, 
in connect
       server_hostname=server_hostname)
     File "D:\softwares\anaconda3\envs\tvm-dev\lib\ssl.py", line 423, in 
wrap_socket
       session=session
     File "D:\softwares\anaconda3\envs\tvm-dev\lib\ssl.py", line 870, in _create
       self.do_handshake()
     File "D:\softwares\anaconda3\envs\tvm-dev\lib\ssl.py", line 1139, in 
do_handshake
       self._sslobj.do_handshake()
     File 
"C:\Users\ycdfwzy\AppData\Roaming\Python\Python37\site-packages\tvm-0.8.dev338+g28647f2e7-py3.7-win-amd64.egg\tvm\_ffi\_ctypes\packed_func.py",
 line 81, in cfun
       rv = local_pyfunc(*pyargs)
     File 
"C:\Users\ycdfwzy\AppData\Roaming\Python\Python37\site-packages\tvm-0.8.dev338+g28647f2e7-py3.7-win-amd64.egg\tvm\relay\op\strategy\generic.py",
 line 47, in wrapper
       return topi_schedule(outs)
     File 
"C:\Users\ycdfwzy\AppData\Roaming\Python\Python37\site-packages\tvm-0.8.dev338+g28647f2e7-py3.7-win-amd64.egg\tvm\autotvm\task\topi_integration.py",
 line 235, in wrapper
       return topi_schedule(cfg, outs, *args, **kwargs)
     File 
"C:\Users\ycdfwzy\AppData\Roaming\Python\Python37\site-packages\tvm-0.8.dev338+g28647f2e7-py3.7-win-amd64.egg\tvm\topi\cuda\conv2d.py",
 line 47, in schedule_conv2d_nchw
       traverse_inline(s, outs[0].op, _callback)
     File 
"C:\Users\ycdfwzy\AppData\Roaming\Python\Python37\site-packages\tvm-0.8.dev338+g28647f2e7-py3.7-win-amd64.egg\tvm\topi\utils.py",
 line 70, in traverse_inline
       _traverse(final_op)
     File 
"C:\Users\ycdfwzy\AppData\Roaming\Python\Python37\site-packages\tvm-0.8.dev338+g28647f2e7-py3.7-win-amd64.egg\tvm\topi\utils.py",
 line 67, in _traverse
       _traverse(tensor.op)
     File 
"C:\Users\ycdfwzy\AppData\Roaming\Python\Python37\site-packages\tvm-0.8.dev338+g28647f2e7-py3.7-win-amd64.egg\tvm\topi\utils.py",
 line 67, in _traverse
       _traverse(tensor.op)
     File 
"C:\Users\ycdfwzy\AppData\Roaming\Python\Python37\site-packages\tvm-0.8.dev338+g28647f2e7-py3.7-win-amd64.egg\tvm\topi\utils.py",
 line 67, in _traverse
       _traverse(tensor.op)
     [Previous line repeated 1 more time]
     File 
"C:\Users\ycdfwzy\AppData\Roaming\Python\Python37\site-packages\tvm-0.8.dev338+g28647f2e7-py3.7-win-amd64.egg\tvm\topi\utils.py",
 line 68, in _traverse
       callback(op)
     File 
"C:\Users\ycdfwzy\AppData\Roaming\Python\Python37\site-packages\tvm-0.8.dev338+g28647f2e7-py3.7-win-amd64.egg\tvm\topi\cuda\conv2d.py",
 line 45, in _callback
       schedule_direct_cuda(cfg, s, op.output(0))
     File 
"C:\Users\ycdfwzy\AppData\Roaming\Python\Python37\site-packages\tvm-0.8.dev338+g28647f2e7-py3.7-win-amd64.egg\tvm\topi\cuda\conv2d_direct.py",
 line 48, in schedule_direct_cuda
       target.kind.name, target.model, "conv2d_nchw.cuda"
     File 
"C:\Users\ycdfwzy\AppData\Roaming\Python\Python37\site-packages\tvm-0.8.dev338+g28647f2e7-py3.7-win-amd64.egg\tvm\autotvm\tophub.py",
 line 224, in load_reference_log
       download_package(tophub_location, package_name)
     File 
"C:\Users\ycdfwzy\AppData\Roaming\Python\Python37\site-packages\tvm-0.8.dev338+g28647f2e7-py3.7-win-amd64.egg\tvm\autotvm\tophub.py",
 line 188, in download_package
       download(download_url, os.path.join(rootpath, package_name), True, 
verbose=0)
     File 
"C:\Users\ycdfwzy\AppData\Roaming\Python\Python37\site-packages\tvm-0.8.dev338+g28647f2e7-py3.7-win-amd64.egg\tvm\contrib\download.py",
 line 114, in download
       raise err
     File 
"C:\Users\ycdfwzy\AppData\Roaming\Python\Python37\site-packages\tvm-0.8.dev338+g28647f2e7-py3.7-win-amd64.egg\tvm\contrib\download.py",
 line 100, in download
       urllib2.urlretrieve(url, tempfile, reporthook=_download_progress)
     File "D:\softwares\anaconda3\envs\tvm-dev\lib\urllib\request.py", line 
247, in urlretrieve
       with contextlib.closing(urlopen(url, data)) as fp:
     File "D:\softwares\anaconda3\envs\tvm-dev\lib\urllib\request.py", line 
222, in urlopen
       return opener.open(url, data, timeout)
     File "D:\softwares\anaconda3\envs\tvm-dev\lib\urllib\request.py", line 
525, in open
       response = self._open(req, data)
     File "D:\softwares\anaconda3\envs\tvm-dev\lib\urllib\request.py", line 
543, in _open
       '_open', req)
     File "D:\softwares\anaconda3\envs\tvm-dev\lib\urllib\request.py", line 
503, in _call_chain
       result = func(*args)
     File "D:\softwares\anaconda3\envs\tvm-dev\lib\urllib\request.py", line 
1393, in https_open
       context=self._context, check_hostname=self._check_hostname)
     File "D:\softwares\anaconda3\envs\tvm-dev\lib\urllib\request.py", line 
1352, in do_open
       raise URLError(err)
   ConnectionResetError: [WinError 10054] An existing connection was forcibly 
closed by the remote host
   During handling of the above exception, another exception occurred:
   
   urllib.error.URLError: <urlopen error [WinError 10054] An existing 
connection was forcibly closed by the remote host>
   ```


----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
[email protected]


Reply via email to