Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- graph torch-jit-export (
- %input.1[FLOAT, 1x3x256x128]
- ) initializers (
- %classifier.bias[FLOAT, 1041]
- %classifier.weight[FLOAT, 1041x512]
- %conv1.bn.bias[FLOAT, 64]
- %conv1.bn.weight[FLOAT, 64]
- %conv1.conv.weight[FLOAT, 64x3x7x7]
- %conv2.0.IN.bias[FLOAT, 256]
- %conv2.0.IN.weight[FLOAT, 256]
- %conv2.0.conv1.bn.bias[FLOAT, 64]
- %conv2.0.conv1.bn.num_batches_tracked[INT64, scalar]
- %conv2.0.conv1.bn.running_mean[FLOAT, 64]
- %conv2.0.conv1.bn.running_var[FLOAT, 64]
- %conv2.0.conv1.bn.weight[FLOAT, 64]
- %conv2.0.conv1.conv.weight[FLOAT, 64x64x1x1]
- %conv2.0.conv2.0.layers.0.bn.bias[FLOAT, 64]
- %conv2.0.conv2.0.layers.0.bn.num_batches_tracked[INT64, scalar]
- %conv2.0.conv2.0.layers.0.bn.running_mean[FLOAT, 64]
- %conv2.0.conv2.0.layers.0.bn.running_var[FLOAT, 64]
- %conv2.0.conv2.0.layers.0.bn.weight[FLOAT, 64]
- %conv2.0.conv2.0.layers.0.conv1.weight[FLOAT, 64x64x1x1]
- %conv2.0.conv2.0.layers.0.conv2.weight[FLOAT, 64x1x3x3]
- %conv2.0.conv2.1.layers.0.bn.bias[FLOAT, 64]
- %conv2.0.conv2.1.layers.0.bn.num_batches_tracked[INT64, scalar]
- %conv2.0.conv2.1.layers.0.bn.running_mean[FLOAT, 64]
- %conv2.0.conv2.1.layers.0.bn.running_var[FLOAT, 64]
- %conv2.0.conv2.1.layers.0.bn.weight[FLOAT, 64]
- %conv2.0.conv2.1.layers.0.conv1.weight[FLOAT, 64x64x1x1]
- %conv2.0.conv2.1.layers.0.conv2.weight[FLOAT, 64x1x3x3]
- %conv2.0.conv2.1.layers.1.bn.bias[FLOAT, 64]
- %conv2.0.conv2.1.layers.1.bn.num_batches_tracked[INT64, scalar]
- %conv2.0.conv2.1.layers.1.bn.running_mean[FLOAT, 64]
- %conv2.0.conv2.1.layers.1.bn.running_var[FLOAT, 64]
- %conv2.0.conv2.1.layers.1.bn.weight[FLOAT, 64]
- %conv2.0.conv2.1.layers.1.conv1.weight[FLOAT, 64x64x1x1]
- %conv2.0.conv2.1.layers.1.conv2.weight[FLOAT, 64x1x3x3]
- %conv2.0.conv2.2.layers.0.bn.bias[FLOAT, 64]
- %conv2.0.conv2.2.layers.0.bn.num_batches_tracked[INT64, scalar]
- %conv2.0.conv2.2.layers.0.bn.running_mean[FLOAT, 64]
- %conv2.0.conv2.2.layers.0.bn.running_var[FLOAT, 64]
- %conv2.0.conv2.2.layers.0.bn.weight[FLOAT, 64]
- %conv2.0.conv2.2.layers.0.conv1.weight[FLOAT, 64x64x1x1]
- %conv2.0.conv2.2.layers.0.conv2.weight[FLOAT, 64x1x3x3]
- %conv2.0.conv2.2.layers.1.bn.bias[FLOAT, 64]
- %conv2.0.conv2.2.layers.1.bn.num_batches_tracked[INT64, scalar]
- %conv2.0.conv2.2.layers.1.bn.running_mean[FLOAT, 64]
- %conv2.0.conv2.2.layers.1.bn.running_var[FLOAT, 64]
- %conv2.0.conv2.2.layers.1.bn.weight[FLOAT, 64]
- %conv2.0.conv2.2.layers.1.conv1.weight[FLOAT, 64x64x1x1]
- %conv2.0.conv2.2.layers.1.conv2.weight[FLOAT, 64x1x3x3]
- %conv2.0.conv2.2.layers.2.bn.bias[FLOAT, 64]
- %conv2.0.conv2.2.layers.2.bn.num_batches_tracked[INT64, scalar]
- %conv2.0.conv2.2.layers.2.bn.running_mean[FLOAT, 64]
- %conv2.0.conv2.2.layers.2.bn.running_var[FLOAT, 64]
- %conv2.0.conv2.2.layers.2.bn.weight[FLOAT, 64]
- %conv2.0.conv2.2.layers.2.conv1.weight[FLOAT, 64x64x1x1]
- %conv2.0.conv2.2.layers.2.conv2.weight[FLOAT, 64x1x3x3]
- %conv2.0.conv2.3.layers.0.bn.bias[FLOAT, 64]
- %conv2.0.conv2.3.layers.0.bn.num_batches_tracked[INT64, scalar]
- %conv2.0.conv2.3.layers.0.bn.running_mean[FLOAT, 64]
- %conv2.0.conv2.3.layers.0.bn.running_var[FLOAT, 64]
- %conv2.0.conv2.3.layers.0.bn.weight[FLOAT, 64]
- %conv2.0.conv2.3.layers.0.conv1.weight[FLOAT, 64x64x1x1]
- %conv2.0.conv2.3.layers.0.conv2.weight[FLOAT, 64x1x3x3]
- %conv2.0.conv2.3.layers.1.bn.bias[FLOAT, 64]
- %conv2.0.conv2.3.layers.1.bn.num_batches_tracked[INT64, scalar]
- %conv2.0.conv2.3.layers.1.bn.running_mean[FLOAT, 64]
- %conv2.0.conv2.3.layers.1.bn.running_var[FLOAT, 64]
- %conv2.0.conv2.3.layers.1.bn.weight[FLOAT, 64]
- %conv2.0.conv2.3.layers.1.conv1.weight[FLOAT, 64x64x1x1]
- %conv2.0.conv2.3.layers.1.conv2.weight[FLOAT, 64x1x3x3]
- %conv2.0.conv2.3.layers.2.bn.bias[FLOAT, 64]
- %conv2.0.conv2.3.layers.2.bn.num_batches_tracked[INT64, scalar]
- %conv2.0.conv2.3.layers.2.bn.running_mean[FLOAT, 64]
- %conv2.0.conv2.3.layers.2.bn.running_var[FLOAT, 64]
- %conv2.0.conv2.3.layers.2.bn.weight[FLOAT, 64]
- %conv2.0.conv2.3.layers.2.conv1.weight[FLOAT, 64x64x1x1]
- %conv2.0.conv2.3.layers.2.conv2.weight[FLOAT, 64x1x3x3]
- %conv2.0.conv2.3.layers.3.bn.bias[FLOAT, 64]
- %conv2.0.conv2.3.layers.3.bn.num_batches_tracked[INT64, scalar]
- %conv2.0.conv2.3.layers.3.bn.running_mean[FLOAT, 64]
- %conv2.0.conv2.3.layers.3.bn.running_var[FLOAT, 64]
- %conv2.0.conv2.3.layers.3.bn.weight[FLOAT, 64]
- %conv2.0.conv2.3.layers.3.conv1.weight[FLOAT, 64x64x1x1]
- %conv2.0.conv2.3.layers.3.conv2.weight[FLOAT, 64x1x3x3]
- %conv2.0.conv3.conv.weight[FLOAT, 256x64x1x1]
- %conv2.0.downsample.bn.bias[FLOAT, 256]
- %conv2.0.downsample.bn.num_batches_tracked[INT64, scalar]
- %conv2.0.downsample.bn.running_mean[FLOAT, 256]
- %conv2.0.downsample.bn.running_var[FLOAT, 256]
- %conv2.0.downsample.bn.weight[FLOAT, 256]
- %conv2.0.downsample.conv.weight[FLOAT, 256x64x1x1]
- %conv2.0.gate.fc1.bias[FLOAT, 4]
- %conv2.0.gate.fc1.weight[FLOAT, 4x64x1x1]
- %conv2.0.gate.fc2.bias[FLOAT, 64]
- %conv2.0.gate.fc2.weight[FLOAT, 64x4x1x1]
- %conv2.1.IN.bias[FLOAT, 256]
- %conv2.1.IN.weight[FLOAT, 256]
- %conv2.1.conv1.bn.bias[FLOAT, 64]
- %conv2.1.conv1.bn.num_batches_tracked[INT64, scalar]
- %conv2.1.conv1.bn.running_mean[FLOAT, 64]
- %conv2.1.conv1.bn.running_var[FLOAT, 64]
- %conv2.1.conv1.bn.weight[FLOAT, 64]
- %conv2.1.conv1.conv.weight[FLOAT, 64x256x1x1]
- %conv2.1.conv2.0.layers.0.bn.bias[FLOAT, 64]
- %conv2.1.conv2.0.layers.0.bn.num_batches_tracked[INT64, scalar]
- %conv2.1.conv2.0.layers.0.bn.running_mean[FLOAT, 64]
- %conv2.1.conv2.0.layers.0.bn.running_var[FLOAT, 64]
- %conv2.1.conv2.0.layers.0.bn.weight[FLOAT, 64]
- %conv2.1.conv2.0.layers.0.conv1.weight[FLOAT, 64x64x1x1]
- %conv2.1.conv2.0.layers.0.conv2.weight[FLOAT, 64x1x3x3]
- %conv2.1.conv2.1.layers.0.bn.bias[FLOAT, 64]
- %conv2.1.conv2.1.layers.0.bn.num_batches_tracked[INT64, scalar]
- %conv2.1.conv2.1.layers.0.bn.running_mean[FLOAT, 64]
- %conv2.1.conv2.1.layers.0.bn.running_var[FLOAT, 64]
- %conv2.1.conv2.1.layers.0.bn.weight[FLOAT, 64]
- %conv2.1.conv2.1.layers.0.conv1.weight[FLOAT, 64x64x1x1]
- %conv2.1.conv2.1.layers.0.conv2.weight[FLOAT, 64x1x3x3]
- %conv2.1.conv2.1.layers.1.bn.bias[FLOAT, 64]
- %conv2.1.conv2.1.layers.1.bn.num_batches_tracked[INT64, scalar]
- %conv2.1.conv2.1.layers.1.bn.running_mean[FLOAT, 64]
- %conv2.1.conv2.1.layers.1.bn.running_var[FLOAT, 64]
- %conv2.1.conv2.1.layers.1.bn.weight[FLOAT, 64]
- %conv2.1.conv2.1.layers.1.conv1.weight[FLOAT, 64x64x1x1]
- %conv2.1.conv2.1.layers.1.conv2.weight[FLOAT, 64x1x3x3]
- %conv2.1.conv2.2.layers.0.bn.bias[FLOAT, 64]
- %conv2.1.conv2.2.layers.0.bn.num_batches_tracked[INT64, scalar]
- %conv2.1.conv2.2.layers.0.bn.running_mean[FLOAT, 64]
- %conv2.1.conv2.2.layers.0.bn.running_var[FLOAT, 64]
- %conv2.1.conv2.2.layers.0.bn.weight[FLOAT, 64]
- %conv2.1.conv2.2.layers.0.conv1.weight[FLOAT, 64x64x1x1]
- %conv2.1.conv2.2.layers.0.conv2.weight[FLOAT, 64x1x3x3]
- %conv2.1.conv2.2.layers.1.bn.bias[FLOAT, 64]
- %conv2.1.conv2.2.layers.1.bn.num_batches_tracked[INT64, scalar]
- %conv2.1.conv2.2.layers.1.bn.running_mean[FLOAT, 64]
- %conv2.1.conv2.2.layers.1.bn.running_var[FLOAT, 64]
- %conv2.1.conv2.2.layers.1.bn.weight[FLOAT, 64]
- %conv2.1.conv2.2.layers.1.conv1.weight[FLOAT, 64x64x1x1]
- %conv2.1.conv2.2.layers.1.conv2.weight[FLOAT, 64x1x3x3]
- %conv2.1.conv2.2.layers.2.bn.bias[FLOAT, 64]
- %conv2.1.conv2.2.layers.2.bn.num_batches_tracked[INT64, scalar]
- %conv2.1.conv2.2.layers.2.bn.running_mean[FLOAT, 64]
- %conv2.1.conv2.2.layers.2.bn.running_var[FLOAT, 64]
- %conv2.1.conv2.2.layers.2.bn.weight[FLOAT, 64]
- %conv2.1.conv2.2.layers.2.conv1.weight[FLOAT, 64x64x1x1]
- %conv2.1.conv2.2.layers.2.conv2.weight[FLOAT, 64x1x3x3]
- %conv2.1.conv2.3.layers.0.bn.bias[FLOAT, 64]
- %conv2.1.conv2.3.layers.0.bn.num_batches_tracked[INT64, scalar]
- %conv2.1.conv2.3.layers.0.bn.running_mean[FLOAT, 64]
- %conv2.1.conv2.3.layers.0.bn.running_var[FLOAT, 64]
- %conv2.1.conv2.3.layers.0.bn.weight[FLOAT, 64]
- %conv2.1.conv2.3.layers.0.conv1.weight[FLOAT, 64x64x1x1]
- %conv2.1.conv2.3.layers.0.conv2.weight[FLOAT, 64x1x3x3]
- %conv2.1.conv2.3.layers.1.bn.bias[FLOAT, 64]
- %conv2.1.conv2.3.layers.1.bn.num_batches_tracked[INT64, scalar]
- %conv2.1.conv2.3.layers.1.bn.running_mean[FLOAT, 64]
- %conv2.1.conv2.3.layers.1.bn.running_var[FLOAT, 64]
- %conv2.1.conv2.3.layers.1.bn.weight[FLOAT, 64]
- %conv2.1.conv2.3.layers.1.conv1.weight[FLOAT, 64x64x1x1]
- %conv2.1.conv2.3.layers.1.conv2.weight[FLOAT, 64x1x3x3]
- %conv2.1.conv2.3.layers.2.bn.bias[FLOAT, 64]
- %conv2.1.conv2.3.layers.2.bn.num_batches_tracked[INT64, scalar]
- %conv2.1.conv2.3.layers.2.bn.running_mean[FLOAT, 64]
- %conv2.1.conv2.3.layers.2.bn.running_var[FLOAT, 64]
- %conv2.1.conv2.3.layers.2.bn.weight[FLOAT, 64]
- %conv2.1.conv2.3.layers.2.conv1.weight[FLOAT, 64x64x1x1]
- %conv2.1.conv2.3.layers.2.conv2.weight[FLOAT, 64x1x3x3]
- %conv2.1.conv2.3.layers.3.bn.bias[FLOAT, 64]
- %conv2.1.conv2.3.layers.3.bn.num_batches_tracked[INT64, scalar]
- %conv2.1.conv2.3.layers.3.bn.running_mean[FLOAT, 64]
- %conv2.1.conv2.3.layers.3.bn.running_var[FLOAT, 64]
- %conv2.1.conv2.3.layers.3.bn.weight[FLOAT, 64]
- %conv2.1.conv2.3.layers.3.conv1.weight[FLOAT, 64x64x1x1]
- %conv2.1.conv2.3.layers.3.conv2.weight[FLOAT, 64x1x3x3]
- %conv2.1.conv3.conv.weight[FLOAT, 256x64x1x1]
- %conv2.1.gate.fc1.bias[FLOAT, 4]
- %conv2.1.gate.fc1.weight[FLOAT, 4x64x1x1]
- %conv2.1.gate.fc2.bias[FLOAT, 64]
- %conv2.1.gate.fc2.weight[FLOAT, 64x4x1x1]
- %conv3.0.conv1.bn.bias[FLOAT, 96]
- %conv3.0.conv1.bn.num_batches_tracked[INT64, scalar]
- %conv3.0.conv1.bn.running_mean[FLOAT, 96]
- %conv3.0.conv1.bn.running_var[FLOAT, 96]
- %conv3.0.conv1.bn.weight[FLOAT, 96]
- %conv3.0.conv1.conv.weight[FLOAT, 96x256x1x1]
- %conv3.0.conv2.0.layers.0.bn.bias[FLOAT, 96]
- %conv3.0.conv2.0.layers.0.bn.num_batches_tracked[INT64, scalar]
- %conv3.0.conv2.0.layers.0.bn.running_mean[FLOAT, 96]
- %conv3.0.conv2.0.layers.0.bn.running_var[FLOAT, 96]
- %conv3.0.conv2.0.layers.0.bn.weight[FLOAT, 96]
- %conv3.0.conv2.0.layers.0.conv1.weight[FLOAT, 96x96x1x1]
- %conv3.0.conv2.0.layers.0.conv2.weight[FLOAT, 96x1x3x3]
- %conv3.0.conv2.1.layers.0.bn.bias[FLOAT, 96]
- %conv3.0.conv2.1.layers.0.bn.num_batches_tracked[INT64, scalar]
- %conv3.0.conv2.1.layers.0.bn.running_mean[FLOAT, 96]
- %conv3.0.conv2.1.layers.0.bn.running_var[FLOAT, 96]
- %conv3.0.conv2.1.layers.0.bn.weight[FLOAT, 96]
- %conv3.0.conv2.1.layers.0.conv1.weight[FLOAT, 96x96x1x1]
- %conv3.0.conv2.1.layers.0.conv2.weight[FLOAT, 96x1x3x3]
- %conv3.0.conv2.1.layers.1.bn.bias[FLOAT, 96]
- %conv3.0.conv2.1.layers.1.bn.num_batches_tracked[INT64, scalar]
- %conv3.0.conv2.1.layers.1.bn.running_mean[FLOAT, 96]
- %conv3.0.conv2.1.layers.1.bn.running_var[FLOAT, 96]
- %conv3.0.conv2.1.layers.1.bn.weight[FLOAT, 96]
- %conv3.0.conv2.1.layers.1.conv1.weight[FLOAT, 96x96x1x1]
- %conv3.0.conv2.1.layers.1.conv2.weight[FLOAT, 96x1x3x3]
- %conv3.0.conv2.2.layers.0.bn.bias[FLOAT, 96]
- %conv3.0.conv2.2.layers.0.bn.num_batches_tracked[INT64, scalar]
- %conv3.0.conv2.2.layers.0.bn.running_mean[FLOAT, 96]
- %conv3.0.conv2.2.layers.0.bn.running_var[FLOAT, 96]
- %conv3.0.conv2.2.layers.0.bn.weight[FLOAT, 96]
- %conv3.0.conv2.2.layers.0.conv1.weight[FLOAT, 96x96x1x1]
- %conv3.0.conv2.2.layers.0.conv2.weight[FLOAT, 96x1x3x3]
- %conv3.0.conv2.2.layers.1.bn.bias[FLOAT, 96]
- %conv3.0.conv2.2.layers.1.bn.num_batches_tracked[INT64, scalar]
- %conv3.0.conv2.2.layers.1.bn.running_mean[FLOAT, 96]
- %conv3.0.conv2.2.layers.1.bn.running_var[FLOAT, 96]
- %conv3.0.conv2.2.layers.1.bn.weight[FLOAT, 96]
- %conv3.0.conv2.2.layers.1.conv1.weight[FLOAT, 96x96x1x1]
- %conv3.0.conv2.2.layers.1.conv2.weight[FLOAT, 96x1x3x3]
- %conv3.0.conv2.2.layers.2.bn.bias[FLOAT, 96]
- %conv3.0.conv2.2.layers.2.bn.num_batches_tracked[INT64, scalar]
- %conv3.0.conv2.2.layers.2.bn.running_mean[FLOAT, 96]
- %conv3.0.conv2.2.layers.2.bn.running_var[FLOAT, 96]
- %conv3.0.conv2.2.layers.2.bn.weight[FLOAT, 96]
- %conv3.0.conv2.2.layers.2.conv1.weight[FLOAT, 96x96x1x1]
- %conv3.0.conv2.2.layers.2.conv2.weight[FLOAT, 96x1x3x3]
- %conv3.0.conv2.3.layers.0.bn.bias[FLOAT, 96]
- %conv3.0.conv2.3.layers.0.bn.num_batches_tracked[INT64, scalar]
- %conv3.0.conv2.3.layers.0.bn.running_mean[FLOAT, 96]
- %conv3.0.conv2.3.layers.0.bn.running_var[FLOAT, 96]
- %conv3.0.conv2.3.layers.0.bn.weight[FLOAT, 96]
- %conv3.0.conv2.3.layers.0.conv1.weight[FLOAT, 96x96x1x1]
- %conv3.0.conv2.3.layers.0.conv2.weight[FLOAT, 96x1x3x3]
- %conv3.0.conv2.3.layers.1.bn.bias[FLOAT, 96]
- %conv3.0.conv2.3.layers.1.bn.num_batches_tracked[INT64, scalar]
- %conv3.0.conv2.3.layers.1.bn.running_mean[FLOAT, 96]
- %conv3.0.conv2.3.layers.1.bn.running_var[FLOAT, 96]
- %conv3.0.conv2.3.layers.1.bn.weight[FLOAT, 96]
- %conv3.0.conv2.3.layers.1.conv1.weight[FLOAT, 96x96x1x1]
- %conv3.0.conv2.3.layers.1.conv2.weight[FLOAT, 96x1x3x3]
- %conv3.0.conv2.3.layers.2.bn.bias[FLOAT, 96]
- %conv3.0.conv2.3.layers.2.bn.num_batches_tracked[INT64, scalar]
- %conv3.0.conv2.3.layers.2.bn.running_mean[FLOAT, 96]
- %conv3.0.conv2.3.layers.2.bn.running_var[FLOAT, 96]
- %conv3.0.conv2.3.layers.2.bn.weight[FLOAT, 96]
- %conv3.0.conv2.3.layers.2.conv1.weight[FLOAT, 96x96x1x1]
- %conv3.0.conv2.3.layers.2.conv2.weight[FLOAT, 96x1x3x3]
- %conv3.0.conv2.3.layers.3.bn.bias[FLOAT, 96]
- %conv3.0.conv2.3.layers.3.bn.num_batches_tracked[INT64, scalar]
- %conv3.0.conv2.3.layers.3.bn.running_mean[FLOAT, 96]
- %conv3.0.conv2.3.layers.3.bn.running_var[FLOAT, 96]
- %conv3.0.conv2.3.layers.3.bn.weight[FLOAT, 96]
- %conv3.0.conv2.3.layers.3.conv1.weight[FLOAT, 96x96x1x1]
- %conv3.0.conv2.3.layers.3.conv2.weight[FLOAT, 96x1x3x3]
- %conv3.0.conv3.bn.bias[FLOAT, 384]
- %conv3.0.conv3.bn.num_batches_tracked[INT64, scalar]
- %conv3.0.conv3.bn.running_mean[FLOAT, 384]
- %conv3.0.conv3.bn.running_var[FLOAT, 384]
- %conv3.0.conv3.bn.weight[FLOAT, 384]
- %conv3.0.conv3.conv.weight[FLOAT, 384x96x1x1]
- %conv3.0.downsample.bn.bias[FLOAT, 384]
- %conv3.0.downsample.bn.num_batches_tracked[INT64, scalar]
- %conv3.0.downsample.bn.running_mean[FLOAT, 384]
- %conv3.0.downsample.bn.running_var[FLOAT, 384]
- %conv3.0.downsample.bn.weight[FLOAT, 384]
- %conv3.0.downsample.conv.weight[FLOAT, 384x256x1x1]
- %conv3.0.gate.fc1.bias[FLOAT, 6]
- %conv3.0.gate.fc1.weight[FLOAT, 6x96x1x1]
- %conv3.0.gate.fc2.bias[FLOAT, 96]
- %conv3.0.gate.fc2.weight[FLOAT, 96x6x1x1]
- %conv3.1.IN.bias[FLOAT, 384]
- %conv3.1.IN.weight[FLOAT, 384]
- %conv3.1.conv1.bn.bias[FLOAT, 96]
- %conv3.1.conv1.bn.num_batches_tracked[INT64, scalar]
- %conv3.1.conv1.bn.running_mean[FLOAT, 96]
- %conv3.1.conv1.bn.running_var[FLOAT, 96]
- %conv3.1.conv1.bn.weight[FLOAT, 96]
- %conv3.1.conv1.conv.weight[FLOAT, 96x384x1x1]
- %conv3.1.conv2.0.layers.0.bn.bias[FLOAT, 96]
- %conv3.1.conv2.0.layers.0.bn.num_batches_tracked[INT64, scalar]
- %conv3.1.conv2.0.layers.0.bn.running_mean[FLOAT, 96]
- %conv3.1.conv2.0.layers.0.bn.running_var[FLOAT, 96]
- %conv3.1.conv2.0.layers.0.bn.weight[FLOAT, 96]
- %conv3.1.conv2.0.layers.0.conv1.weight[FLOAT, 96x96x1x1]
- %conv3.1.conv2.0.layers.0.conv2.weight[FLOAT, 96x1x3x3]
- %conv3.1.conv2.1.layers.0.bn.bias[FLOAT, 96]
- %conv3.1.conv2.1.layers.0.bn.num_batches_tracked[INT64, scalar]
- %conv3.1.conv2.1.layers.0.bn.running_mean[FLOAT, 96]
- %conv3.1.conv2.1.layers.0.bn.running_var[FLOAT, 96]
- %conv3.1.conv2.1.layers.0.bn.weight[FLOAT, 96]
- %conv3.1.conv2.1.layers.0.conv1.weight[FLOAT, 96x96x1x1]
- %conv3.1.conv2.1.layers.0.conv2.weight[FLOAT, 96x1x3x3]
- %conv3.1.conv2.1.layers.1.bn.bias[FLOAT, 96]
- %conv3.1.conv2.1.layers.1.bn.num_batches_tracked[INT64, scalar]
- %conv3.1.conv2.1.layers.1.bn.running_mean[FLOAT, 96]
- %conv3.1.conv2.1.layers.1.bn.running_var[FLOAT, 96]
- %conv3.1.conv2.1.layers.1.bn.weight[FLOAT, 96]
- %conv3.1.conv2.1.layers.1.conv1.weight[FLOAT, 96x96x1x1]
- %conv3.1.conv2.1.layers.1.conv2.weight[FLOAT, 96x1x3x3]
- %conv3.1.conv2.2.layers.0.bn.bias[FLOAT, 96]
- %conv3.1.conv2.2.layers.0.bn.num_batches_tracked[INT64, scalar]
- %conv3.1.conv2.2.layers.0.bn.running_mean[FLOAT, 96]
- %conv3.1.conv2.2.layers.0.bn.running_var[FLOAT, 96]
- %conv3.1.conv2.2.layers.0.bn.weight[FLOAT, 96]
- %conv3.1.conv2.2.layers.0.conv1.weight[FLOAT, 96x96x1x1]
- %conv3.1.conv2.2.layers.0.conv2.weight[FLOAT, 96x1x3x3]
- %conv3.1.conv2.2.layers.1.bn.bias[FLOAT, 96]
- %conv3.1.conv2.2.layers.1.bn.num_batches_tracked[INT64, scalar]
- %conv3.1.conv2.2.layers.1.bn.running_mean[FLOAT, 96]
- %conv3.1.conv2.2.layers.1.bn.running_var[FLOAT, 96]
- %conv3.1.conv2.2.layers.1.bn.weight[FLOAT, 96]
- %conv3.1.conv2.2.layers.1.conv1.weight[FLOAT, 96x96x1x1]
- %conv3.1.conv2.2.layers.1.conv2.weight[FLOAT, 96x1x3x3]
- %conv3.1.conv2.2.layers.2.bn.bias[FLOAT, 96]
- %conv3.1.conv2.2.layers.2.bn.num_batches_tracked[INT64, scalar]
- %conv3.1.conv2.2.layers.2.bn.running_mean[FLOAT, 96]
- %conv3.1.conv2.2.layers.2.bn.running_var[FLOAT, 96]
- %conv3.1.conv2.2.layers.2.bn.weight[FLOAT, 96]
- %conv3.1.conv2.2.layers.2.conv1.weight[FLOAT, 96x96x1x1]
- %conv3.1.conv2.2.layers.2.conv2.weight[FLOAT, 96x1x3x3]
- %conv3.1.conv2.3.layers.0.bn.bias[FLOAT, 96]
- %conv3.1.conv2.3.layers.0.bn.num_batches_tracked[INT64, scalar]
- %conv3.1.conv2.3.layers.0.bn.running_mean[FLOAT, 96]
- %conv3.1.conv2.3.layers.0.bn.running_var[FLOAT, 96]
- %conv3.1.conv2.3.layers.0.bn.weight[FLOAT, 96]
- %conv3.1.conv2.3.layers.0.conv1.weight[FLOAT, 96x96x1x1]
- %conv3.1.conv2.3.layers.0.conv2.weight[FLOAT, 96x1x3x3]
- %conv3.1.conv2.3.layers.1.bn.bias[FLOAT, 96]
- %conv3.1.conv2.3.layers.1.bn.num_batches_tracked[INT64, scalar]
- %conv3.1.conv2.3.layers.1.bn.running_mean[FLOAT, 96]
- %conv3.1.conv2.3.layers.1.bn.running_var[FLOAT, 96]
- %conv3.1.conv2.3.layers.1.bn.weight[FLOAT, 96]
- %conv3.1.conv2.3.layers.1.conv1.weight[FLOAT, 96x96x1x1]
- %conv3.1.conv2.3.layers.1.conv2.weight[FLOAT, 96x1x3x3]
- %conv3.1.conv2.3.layers.2.bn.bias[FLOAT, 96]
- %conv3.1.conv2.3.layers.2.bn.num_batches_tracked[INT64, scalar]
- %conv3.1.conv2.3.layers.2.bn.running_mean[FLOAT, 96]
- %conv3.1.conv2.3.layers.2.bn.running_var[FLOAT, 96]
- %conv3.1.conv2.3.layers.2.bn.weight[FLOAT, 96]
- %conv3.1.conv2.3.layers.2.conv1.weight[FLOAT, 96x96x1x1]
- %conv3.1.conv2.3.layers.2.conv2.weight[FLOAT, 96x1x3x3]
- %conv3.1.conv2.3.layers.3.bn.bias[FLOAT, 96]
- %conv3.1.conv2.3.layers.3.bn.num_batches_tracked[INT64, scalar]
- %conv3.1.conv2.3.layers.3.bn.running_mean[FLOAT, 96]
- %conv3.1.conv2.3.layers.3.bn.running_var[FLOAT, 96]
- %conv3.1.conv2.3.layers.3.bn.weight[FLOAT, 96]
- %conv3.1.conv2.3.layers.3.conv1.weight[FLOAT, 96x96x1x1]
- %conv3.1.conv2.3.layers.3.conv2.weight[FLOAT, 96x1x3x3]
- %conv3.1.conv3.conv.weight[FLOAT, 384x96x1x1]
- %conv3.1.gate.fc1.bias[FLOAT, 6]
- %conv3.1.gate.fc1.weight[FLOAT, 6x96x1x1]
- %conv3.1.gate.fc2.bias[FLOAT, 96]
- %conv3.1.gate.fc2.weight[FLOAT, 96x6x1x1]
- %conv4.0.IN.bias[FLOAT, 512]
- %conv4.0.IN.weight[FLOAT, 512]
- %conv4.0.conv1.bn.bias[FLOAT, 128]
- %conv4.0.conv1.bn.num_batches_tracked[INT64, scalar]
- %conv4.0.conv1.bn.running_mean[FLOAT, 128]
- %conv4.0.conv1.bn.running_var[FLOAT, 128]
- %conv4.0.conv1.bn.weight[FLOAT, 128]
- %conv4.0.conv1.conv.weight[FLOAT, 128x384x1x1]
- %conv4.0.conv2.0.layers.0.bn.bias[FLOAT, 128]
- %conv4.0.conv2.0.layers.0.bn.num_batches_tracked[INT64, scalar]
- %conv4.0.conv2.0.layers.0.bn.running_mean[FLOAT, 128]
- %conv4.0.conv2.0.layers.0.bn.running_var[FLOAT, 128]
- %conv4.0.conv2.0.layers.0.bn.weight[FLOAT, 128]
- %conv4.0.conv2.0.layers.0.conv1.weight[FLOAT, 128x128x1x1]
- %conv4.0.conv2.0.layers.0.conv2.weight[FLOAT, 128x1x3x3]
- %conv4.0.conv2.1.layers.0.bn.bias[FLOAT, 128]
- %conv4.0.conv2.1.layers.0.bn.num_batches_tracked[INT64, scalar]
- %conv4.0.conv2.1.layers.0.bn.running_mean[FLOAT, 128]
- %conv4.0.conv2.1.layers.0.bn.running_var[FLOAT, 128]
- %conv4.0.conv2.1.layers.0.bn.weight[FLOAT, 128]
- %conv4.0.conv2.1.layers.0.conv1.weight[FLOAT, 128x128x1x1]
- %conv4.0.conv2.1.layers.0.conv2.weight[FLOAT, 128x1x3x3]
- %conv4.0.conv2.1.layers.1.bn.bias[FLOAT, 128]
- %conv4.0.conv2.1.layers.1.bn.num_batches_tracked[INT64, scalar]
- %conv4.0.conv2.1.layers.1.bn.running_mean[FLOAT, 128]
- %conv4.0.conv2.1.layers.1.bn.running_var[FLOAT, 128]
- %conv4.0.conv2.1.layers.1.bn.weight[FLOAT, 128]
- %conv4.0.conv2.1.layers.1.conv1.weight[FLOAT, 128x128x1x1]
- %conv4.0.conv2.1.layers.1.conv2.weight[FLOAT, 128x1x3x3]
- %conv4.0.conv2.2.layers.0.bn.bias[FLOAT, 128]
- %conv4.0.conv2.2.layers.0.bn.num_batches_tracked[INT64, scalar]
- %conv4.0.conv2.2.layers.0.bn.running_mean[FLOAT, 128]
- %conv4.0.conv2.2.layers.0.bn.running_var[FLOAT, 128]
- %conv4.0.conv2.2.layers.0.bn.weight[FLOAT, 128]
- %conv4.0.conv2.2.layers.0.conv1.weight[FLOAT, 128x128x1x1]
- %conv4.0.conv2.2.layers.0.conv2.weight[FLOAT, 128x1x3x3]
- %conv4.0.conv2.2.layers.1.bn.bias[FLOAT, 128]
- %conv4.0.conv2.2.layers.1.bn.num_batches_tracked[INT64, scalar]
- %conv4.0.conv2.2.layers.1.bn.running_mean[FLOAT, 128]
- %conv4.0.conv2.2.layers.1.bn.running_var[FLOAT, 128]
- %conv4.0.conv2.2.layers.1.bn.weight[FLOAT, 128]
- %conv4.0.conv2.2.layers.1.conv1.weight[FLOAT, 128x128x1x1]
- %conv4.0.conv2.2.layers.1.conv2.weight[FLOAT, 128x1x3x3]
- %conv4.0.conv2.2.layers.2.bn.bias[FLOAT, 128]
- %conv4.0.conv2.2.layers.2.bn.num_batches_tracked[INT64, scalar]
- %conv4.0.conv2.2.layers.2.bn.running_mean[FLOAT, 128]
- %conv4.0.conv2.2.layers.2.bn.running_var[FLOAT, 128]
- %conv4.0.conv2.2.layers.2.bn.weight[FLOAT, 128]
- %conv4.0.conv2.2.layers.2.conv1.weight[FLOAT, 128x128x1x1]
- %conv4.0.conv2.2.layers.2.conv2.weight[FLOAT, 128x1x3x3]
- %conv4.0.conv2.3.layers.0.bn.bias[FLOAT, 128]
- %conv4.0.conv2.3.layers.0.bn.num_batches_tracked[INT64, scalar]
- %conv4.0.conv2.3.layers.0.bn.running_mean[FLOAT, 128]
- %conv4.0.conv2.3.layers.0.bn.running_var[FLOAT, 128]
- %conv4.0.conv2.3.layers.0.bn.weight[FLOAT, 128]
- %conv4.0.conv2.3.layers.0.conv1.weight[FLOAT, 128x128x1x1]
- %conv4.0.conv2.3.layers.0.conv2.weight[FLOAT, 128x1x3x3]
- %conv4.0.conv2.3.layers.1.bn.bias[FLOAT, 128]
- %conv4.0.conv2.3.layers.1.bn.num_batches_tracked[INT64, scalar]
- %conv4.0.conv2.3.layers.1.bn.running_mean[FLOAT, 128]
- %conv4.0.conv2.3.layers.1.bn.running_var[FLOAT, 128]
- %conv4.0.conv2.3.layers.1.bn.weight[FLOAT, 128]
- %conv4.0.conv2.3.layers.1.conv1.weight[FLOAT, 128x128x1x1]
- %conv4.0.conv2.3.layers.1.conv2.weight[FLOAT, 128x1x3x3]
- %conv4.0.conv2.3.layers.2.bn.bias[FLOAT, 128]
- %conv4.0.conv2.3.layers.2.bn.num_batches_tracked[INT64, scalar]
- %conv4.0.conv2.3.layers.2.bn.running_mean[FLOAT, 128]
- %conv4.0.conv2.3.layers.2.bn.running_var[FLOAT, 128]
- %conv4.0.conv2.3.layers.2.bn.weight[FLOAT, 128]
- %conv4.0.conv2.3.layers.2.conv1.weight[FLOAT, 128x128x1x1]
- %conv4.0.conv2.3.layers.2.conv2.weight[FLOAT, 128x1x3x3]
- %conv4.0.conv2.3.layers.3.bn.bias[FLOAT, 128]
- %conv4.0.conv2.3.layers.3.bn.num_batches_tracked[INT64, scalar]
- %conv4.0.conv2.3.layers.3.bn.running_mean[FLOAT, 128]
- %conv4.0.conv2.3.layers.3.bn.running_var[FLOAT, 128]
- %conv4.0.conv2.3.layers.3.bn.weight[FLOAT, 128]
- %conv4.0.conv2.3.layers.3.conv1.weight[FLOAT, 128x128x1x1]
- %conv4.0.conv2.3.layers.3.conv2.weight[FLOAT, 128x1x3x3]
- %conv4.0.conv3.conv.weight[FLOAT, 512x128x1x1]
- %conv4.0.downsample.bn.bias[FLOAT, 512]
- %conv4.0.downsample.bn.num_batches_tracked[INT64, scalar]
- %conv4.0.downsample.bn.running_mean[FLOAT, 512]
- %conv4.0.downsample.bn.running_var[FLOAT, 512]
- %conv4.0.downsample.bn.weight[FLOAT, 512]
- %conv4.0.downsample.conv.weight[FLOAT, 512x384x1x1]
- %conv4.0.gate.fc1.bias[FLOAT, 8]
- %conv4.0.gate.fc1.weight[FLOAT, 8x128x1x1]
- %conv4.0.gate.fc2.bias[FLOAT, 128]
- %conv4.0.gate.fc2.weight[FLOAT, 128x8x1x1]
- %conv4.1.conv1.bn.bias[FLOAT, 128]
- %conv4.1.conv1.bn.num_batches_tracked[INT64, scalar]
- %conv4.1.conv1.bn.running_mean[FLOAT, 128]
- %conv4.1.conv1.bn.running_var[FLOAT, 128]
- %conv4.1.conv1.bn.weight[FLOAT, 128]
- %conv4.1.conv1.conv.weight[FLOAT, 128x512x1x1]
- %conv4.1.conv2.0.layers.0.bn.bias[FLOAT, 128]
- %conv4.1.conv2.0.layers.0.bn.num_batches_tracked[INT64, scalar]
- %conv4.1.conv2.0.layers.0.bn.running_mean[FLOAT, 128]
- %conv4.1.conv2.0.layers.0.bn.running_var[FLOAT, 128]
- %conv4.1.conv2.0.layers.0.bn.weight[FLOAT, 128]
- %conv4.1.conv2.0.layers.0.conv1.weight[FLOAT, 128x128x1x1]
- %conv4.1.conv2.0.layers.0.conv2.weight[FLOAT, 128x1x3x3]
- %conv4.1.conv2.1.layers.0.bn.bias[FLOAT, 128]
- %conv4.1.conv2.1.layers.0.bn.num_batches_tracked[INT64, scalar]
- %conv4.1.conv2.1.layers.0.bn.running_mean[FLOAT, 128]
- %conv4.1.conv2.1.layers.0.bn.running_var[FLOAT, 128]
- %conv4.1.conv2.1.layers.0.bn.weight[FLOAT, 128]
- %conv4.1.conv2.1.layers.0.conv1.weight[FLOAT, 128x128x1x1]
- %conv4.1.conv2.1.layers.0.conv2.weight[FLOAT, 128x1x3x3]
- %conv4.1.conv2.1.layers.1.bn.bias[FLOAT, 128]
- %conv4.1.conv2.1.layers.1.bn.num_batches_tracked[INT64, scalar]
- %conv4.1.conv2.1.layers.1.bn.running_mean[FLOAT, 128]
- %conv4.1.conv2.1.layers.1.bn.running_var[FLOAT, 128]
- %conv4.1.conv2.1.layers.1.bn.weight[FLOAT, 128]
- %conv4.1.conv2.1.layers.1.conv1.weight[FLOAT, 128x128x1x1]
- %conv4.1.conv2.1.layers.1.conv2.weight[FLOAT, 128x1x3x3]
- %conv4.1.conv2.2.layers.0.bn.bias[FLOAT, 128]
- %conv4.1.conv2.2.layers.0.bn.num_batches_tracked[INT64, scalar]
- %conv4.1.conv2.2.layers.0.bn.running_mean[FLOAT, 128]
- %conv4.1.conv2.2.layers.0.bn.running_var[FLOAT, 128]
- %conv4.1.conv2.2.layers.0.bn.weight[FLOAT, 128]
- %conv4.1.conv2.2.layers.0.conv1.weight[FLOAT, 128x128x1x1]
- %conv4.1.conv2.2.layers.0.conv2.weight[FLOAT, 128x1x3x3]
- %conv4.1.conv2.2.layers.1.bn.bias[FLOAT, 128]
- %conv4.1.conv2.2.layers.1.bn.num_batches_tracked[INT64, scalar]
- %conv4.1.conv2.2.layers.1.bn.running_mean[FLOAT, 128]
- %conv4.1.conv2.2.layers.1.bn.running_var[FLOAT, 128]
- %conv4.1.conv2.2.layers.1.bn.weight[FLOAT, 128]
- %conv4.1.conv2.2.layers.1.conv1.weight[FLOAT, 128x128x1x1]
- %conv4.1.conv2.2.layers.1.conv2.weight[FLOAT, 128x1x3x3]
- %conv4.1.conv2.2.layers.2.bn.bias[FLOAT, 128]
- %conv4.1.conv2.2.layers.2.bn.num_batches_tracked[INT64, scalar]
- %conv4.1.conv2.2.layers.2.bn.running_mean[FLOAT, 128]
- %conv4.1.conv2.2.layers.2.bn.running_var[FLOAT, 128]
- %conv4.1.conv2.2.layers.2.bn.weight[FLOAT, 128]
- %conv4.1.conv2.2.layers.2.conv1.weight[FLOAT, 128x128x1x1]
- %conv4.1.conv2.2.layers.2.conv2.weight[FLOAT, 128x1x3x3]
- %conv4.1.conv2.3.layers.0.bn.bias[FLOAT, 128]
- %conv4.1.conv2.3.layers.0.bn.num_batches_tracked[INT64, scalar]
- %conv4.1.conv2.3.layers.0.bn.running_mean[FLOAT, 128]
- %conv4.1.conv2.3.layers.0.bn.running_var[FLOAT, 128]
- %conv4.1.conv2.3.layers.0.bn.weight[FLOAT, 128]
- %conv4.1.conv2.3.layers.0.conv1.weight[FLOAT, 128x128x1x1]
- %conv4.1.conv2.3.layers.0.conv2.weight[FLOAT, 128x1x3x3]
- %conv4.1.conv2.3.layers.1.bn.bias[FLOAT, 128]
- %conv4.1.conv2.3.layers.1.bn.num_batches_tracked[INT64, scalar]
- %conv4.1.conv2.3.layers.1.bn.running_mean[FLOAT, 128]
- %conv4.1.conv2.3.layers.1.bn.running_var[FLOAT, 128]
- %conv4.1.conv2.3.layers.1.bn.weight[FLOAT, 128]
- %conv4.1.conv2.3.layers.1.conv1.weight[FLOAT, 128x128x1x1]
- %conv4.1.conv2.3.layers.1.conv2.weight[FLOAT, 128x1x3x3]
- %conv4.1.conv2.3.layers.2.bn.bias[FLOAT, 128]
- %conv4.1.conv2.3.layers.2.bn.num_batches_tracked[INT64, scalar]
- %conv4.1.conv2.3.layers.2.bn.running_mean[FLOAT, 128]
- %conv4.1.conv2.3.layers.2.bn.running_var[FLOAT, 128]
- %conv4.1.conv2.3.layers.2.bn.weight[FLOAT, 128]
- %conv4.1.conv2.3.layers.2.conv1.weight[FLOAT, 128x128x1x1]
- %conv4.1.conv2.3.layers.2.conv2.weight[FLOAT, 128x1x3x3]
- %conv4.1.conv2.3.layers.3.bn.bias[FLOAT, 128]
- %conv4.1.conv2.3.layers.3.bn.num_batches_tracked[INT64, scalar]
- %conv4.1.conv2.3.layers.3.bn.running_mean[FLOAT, 128]
- %conv4.1.conv2.3.layers.3.bn.running_var[FLOAT, 128]
- %conv4.1.conv2.3.layers.3.bn.weight[FLOAT, 128]
- %conv4.1.conv2.3.layers.3.conv1.weight[FLOAT, 128x128x1x1]
- %conv4.1.conv2.3.layers.3.conv2.weight[FLOAT, 128x1x3x3]
- %conv4.1.conv3.bn.bias[FLOAT, 512]
- %conv4.1.conv3.bn.num_batches_tracked[INT64, scalar]
- %conv4.1.conv3.bn.running_mean[FLOAT, 512]
- %conv4.1.conv3.bn.running_var[FLOAT, 512]
- %conv4.1.conv3.bn.weight[FLOAT, 512]
- %conv4.1.conv3.conv.weight[FLOAT, 512x128x1x1]
- %conv4.1.gate.fc1.bias[FLOAT, 8]
- %conv4.1.gate.fc1.weight[FLOAT, 8x128x1x1]
- %conv4.1.gate.fc2.bias[FLOAT, 128]
- %conv4.1.gate.fc2.weight[FLOAT, 128x8x1x1]
- %conv5.bn.bias[FLOAT, 512]
- %conv5.bn.num_batches_tracked[INT64, scalar]
- %conv5.bn.running_mean[FLOAT, 512]
- %conv5.bn.running_var[FLOAT, 512]
- %conv5.bn.weight[FLOAT, 512]
- %conv5.conv.weight[FLOAT, 512x512x1x1]
- %fc.0.bias[FLOAT, 512]
- %fc.0.weight[FLOAT, 512x512]
- %fc.1.bias[FLOAT, 512]
- %fc.1.num_batches_tracked[INT64, scalar]
- %fc.1.running_mean[FLOAT, 512]
- %fc.1.running_var[FLOAT, 512]
- %fc.1.weight[FLOAT, 512]
- %pool2.0.bn.bias[FLOAT, 256]
- %pool2.0.bn.num_batches_tracked[INT64, scalar]
- %pool2.0.bn.running_mean[FLOAT, 256]
- %pool2.0.bn.running_var[FLOAT, 256]
- %pool2.0.bn.weight[FLOAT, 256]
- %pool2.0.conv.weight[FLOAT, 256x256x1x1]
- %pool3.0.bn.bias[FLOAT, 384]
- %pool3.0.bn.num_batches_tracked[INT64, scalar]
- %pool3.0.bn.running_mean[FLOAT, 384]
- %pool3.0.bn.running_var[FLOAT, 384]
- %pool3.0.bn.weight[FLOAT, 384]
- %pool3.0.conv.weight[FLOAT, 384x384x1x1]
- ) {
- %553 = Conv[dilations = [1, 1], group = 1, kernel_shape = [7, 7], pads = [3, 3, 3, 3], strides = [2, 2]](%input.1, %conv1.conv.weight)
- %554 = InstanceNormalization[epsilon = 9.99999974737875e-06](%553, %conv1.bn.weight, %conv1.bn.bias)
- %555 = Relu(%554)
- %556 = MaxPool[kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [2, 2]](%555)
- %557 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%556, %conv2.0.conv1.conv.weight)
- %558 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%557, %conv2.0.conv1.bn.weight, %conv2.0.conv1.bn.bias, %conv2.0.conv1.bn.running_mean, %conv2.0.conv1.bn.running_var)
- %559 = Relu(%558)
- %560 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%559, %conv2.0.conv2.0.layers.0.conv1.weight)
- %561 = Conv[dilations = [1, 1], group = 64, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%560, %conv2.0.conv2.0.layers.0.conv2.weight)
- %562 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%561, %conv2.0.conv2.0.layers.0.bn.weight, %conv2.0.conv2.0.layers.0.bn.bias, %conv2.0.conv2.0.layers.0.bn.running_mean, %conv2.0.conv2.0.layers.0.bn.running_var)
- %563 = Relu(%562)
- %564 = GlobalAveragePool(%563)
- %565 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%564, %conv2.0.gate.fc1.weight, %conv2.0.gate.fc1.bias)
- %566 = Relu(%565)
- %567 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%566, %conv2.0.gate.fc2.weight, %conv2.0.gate.fc2.bias)
- %568 = Sigmoid(%567)
- %569 = Mul(%563, %568)
- %570 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%559, %conv2.0.conv2.1.layers.0.conv1.weight)
- %571 = Conv[dilations = [1, 1], group = 64, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%570, %conv2.0.conv2.1.layers.0.conv2.weight)
- %572 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%571, %conv2.0.conv2.1.layers.0.bn.weight, %conv2.0.conv2.1.layers.0.bn.bias, %conv2.0.conv2.1.layers.0.bn.running_mean, %conv2.0.conv2.1.layers.0.bn.running_var)
- %573 = Relu(%572)
- %574 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%573, %conv2.0.conv2.1.layers.1.conv1.weight)
- %575 = Conv[dilations = [1, 1], group = 64, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%574, %conv2.0.conv2.1.layers.1.conv2.weight)
- %576 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%575, %conv2.0.conv2.1.layers.1.bn.weight, %conv2.0.conv2.1.layers.1.bn.bias, %conv2.0.conv2.1.layers.1.bn.running_mean, %conv2.0.conv2.1.layers.1.bn.running_var)
- %577 = Relu(%576)
- %578 = GlobalAveragePool(%577)
- %579 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%578, %conv2.0.gate.fc1.weight, %conv2.0.gate.fc1.bias)
- %580 = Relu(%579)
- %581 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%580, %conv2.0.gate.fc2.weight, %conv2.0.gate.fc2.bias)
- %582 = Sigmoid(%581)
- %583 = Mul(%577, %582)
- %584 = Add(%569, %583)
- %585 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%559, %conv2.0.conv2.2.layers.0.conv1.weight)
- %586 = Conv[dilations = [1, 1], group = 64, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%585, %conv2.0.conv2.2.layers.0.conv2.weight)
- %587 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%586, %conv2.0.conv2.2.layers.0.bn.weight, %conv2.0.conv2.2.layers.0.bn.bias, %conv2.0.conv2.2.layers.0.bn.running_mean, %conv2.0.conv2.2.layers.0.bn.running_var)
- %588 = Relu(%587)
- %589 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%588, %conv2.0.conv2.2.layers.1.conv1.weight)
- %590 = Conv[dilations = [1, 1], group = 64, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%589, %conv2.0.conv2.2.layers.1.conv2.weight)
- %591 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%590, %conv2.0.conv2.2.layers.1.bn.weight, %conv2.0.conv2.2.layers.1.bn.bias, %conv2.0.conv2.2.layers.1.bn.running_mean, %conv2.0.conv2.2.layers.1.bn.running_var)
- %592 = Relu(%591)
- %593 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%592, %conv2.0.conv2.2.layers.2.conv1.weight)
- %594 = Conv[dilations = [1, 1], group = 64, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%593, %conv2.0.conv2.2.layers.2.conv2.weight)
- %595 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%594, %conv2.0.conv2.2.layers.2.bn.weight, %conv2.0.conv2.2.layers.2.bn.bias, %conv2.0.conv2.2.layers.2.bn.running_mean, %conv2.0.conv2.2.layers.2.bn.running_var)
- %596 = Relu(%595)
- %597 = GlobalAveragePool(%596)
- %598 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%597, %conv2.0.gate.fc1.weight, %conv2.0.gate.fc1.bias)
- %599 = Relu(%598)
- %600 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%599, %conv2.0.gate.fc2.weight, %conv2.0.gate.fc2.bias)
- %601 = Sigmoid(%600)
- %602 = Mul(%596, %601)
- %603 = Add(%584, %602)
- %604 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%559, %conv2.0.conv2.3.layers.0.conv1.weight)
- %605 = Conv[dilations = [1, 1], group = 64, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%604, %conv2.0.conv2.3.layers.0.conv2.weight)
- %606 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%605, %conv2.0.conv2.3.layers.0.bn.weight, %conv2.0.conv2.3.layers.0.bn.bias, %conv2.0.conv2.3.layers.0.bn.running_mean, %conv2.0.conv2.3.layers.0.bn.running_var)
- %607 = Relu(%606)
- %608 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%607, %conv2.0.conv2.3.layers.1.conv1.weight)
- %609 = Conv[dilations = [1, 1], group = 64, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%608, %conv2.0.conv2.3.layers.1.conv2.weight)
- %610 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%609, %conv2.0.conv2.3.layers.1.bn.weight, %conv2.0.conv2.3.layers.1.bn.bias, %conv2.0.conv2.3.layers.1.bn.running_mean, %conv2.0.conv2.3.layers.1.bn.running_var)
- %611 = Relu(%610)
- %612 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%611, %conv2.0.conv2.3.layers.2.conv1.weight)
- %613 = Conv[dilations = [1, 1], group = 64, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%612, %conv2.0.conv2.3.layers.2.conv2.weight)
- %614 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%613, %conv2.0.conv2.3.layers.2.bn.weight, %conv2.0.conv2.3.layers.2.bn.bias, %conv2.0.conv2.3.layers.2.bn.running_mean, %conv2.0.conv2.3.layers.2.bn.running_var)
- %615 = Relu(%614)
- %616 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%615, %conv2.0.conv2.3.layers.3.conv1.weight)
- %617 = Conv[dilations = [1, 1], group = 64, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%616, %conv2.0.conv2.3.layers.3.conv2.weight)
- %618 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%617, %conv2.0.conv2.3.layers.3.bn.weight, %conv2.0.conv2.3.layers.3.bn.bias, %conv2.0.conv2.3.layers.3.bn.running_mean, %conv2.0.conv2.3.layers.3.bn.running_var)
- %619 = Relu(%618)
- %620 = GlobalAveragePool(%619)
- %621 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%620, %conv2.0.gate.fc1.weight, %conv2.0.gate.fc1.bias)
- %622 = Relu(%621)
- %623 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%622, %conv2.0.gate.fc2.weight, %conv2.0.gate.fc2.bias)
- %624 = Sigmoid(%623)
- %625 = Mul(%619, %624)
- %626 = Add(%603, %625)
- %627 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%626, %conv2.0.conv3.conv.weight)
- %628 = InstanceNormalization[epsilon = 9.99999974737875e-06](%627, %conv2.0.IN.weight, %conv2.0.IN.bias)
- %629 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%556, %conv2.0.downsample.conv.weight)
- %630 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%629, %conv2.0.downsample.bn.weight, %conv2.0.downsample.bn.bias, %conv2.0.downsample.bn.running_mean, %conv2.0.downsample.bn.running_var)
- %631 = Add(%628, %630)
- %632 = Relu(%631)
- %633 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%632, %conv2.1.conv1.conv.weight)
- %634 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%633, %conv2.1.conv1.bn.weight, %conv2.1.conv1.bn.bias, %conv2.1.conv1.bn.running_mean, %conv2.1.conv1.bn.running_var)
- %635 = Relu(%634)
- %636 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%635, %conv2.1.conv2.0.layers.0.conv1.weight)
- %637 = Conv[dilations = [1, 1], group = 64, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%636, %conv2.1.conv2.0.layers.0.conv2.weight)
- %638 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%637, %conv2.1.conv2.0.layers.0.bn.weight, %conv2.1.conv2.0.layers.0.bn.bias, %conv2.1.conv2.0.layers.0.bn.running_mean, %conv2.1.conv2.0.layers.0.bn.running_var)
- %639 = Relu(%638)
- %640 = GlobalAveragePool(%639)
- %641 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%640, %conv2.1.gate.fc1.weight, %conv2.1.gate.fc1.bias)
- %642 = Relu(%641)
- %643 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%642, %conv2.1.gate.fc2.weight, %conv2.1.gate.fc2.bias)
- %644 = Sigmoid(%643)
- %645 = Mul(%639, %644)
- %646 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%635, %conv2.1.conv2.1.layers.0.conv1.weight)
- %647 = Conv[dilations = [1, 1], group = 64, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%646, %conv2.1.conv2.1.layers.0.conv2.weight)
- %648 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%647, %conv2.1.conv2.1.layers.0.bn.weight, %conv2.1.conv2.1.layers.0.bn.bias, %conv2.1.conv2.1.layers.0.bn.running_mean, %conv2.1.conv2.1.layers.0.bn.running_var)
- %649 = Relu(%648)
- %650 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%649, %conv2.1.conv2.1.layers.1.conv1.weight)
- %651 = Conv[dilations = [1, 1], group = 64, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%650, %conv2.1.conv2.1.layers.1.conv2.weight)
- %652 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%651, %conv2.1.conv2.1.layers.1.bn.weight, %conv2.1.conv2.1.layers.1.bn.bias, %conv2.1.conv2.1.layers.1.bn.running_mean, %conv2.1.conv2.1.layers.1.bn.running_var)
- %653 = Relu(%652)
- %654 = GlobalAveragePool(%653)
- %655 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%654, %conv2.1.gate.fc1.weight, %conv2.1.gate.fc1.bias)
- %656 = Relu(%655)
- %657 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%656, %conv2.1.gate.fc2.weight, %conv2.1.gate.fc2.bias)
- %658 = Sigmoid(%657)
- %659 = Mul(%653, %658)
- %660 = Add(%645, %659)
- %661 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%635, %conv2.1.conv2.2.layers.0.conv1.weight)
- %662 = Conv[dilations = [1, 1], group = 64, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%661, %conv2.1.conv2.2.layers.0.conv2.weight)
- %663 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%662, %conv2.1.conv2.2.layers.0.bn.weight, %conv2.1.conv2.2.layers.0.bn.bias, %conv2.1.conv2.2.layers.0.bn.running_mean, %conv2.1.conv2.2.layers.0.bn.running_var)
- %664 = Relu(%663)
- %665 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%664, %conv2.1.conv2.2.layers.1.conv1.weight)
- %666 = Conv[dilations = [1, 1], group = 64, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%665, %conv2.1.conv2.2.layers.1.conv2.weight)
- %667 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%666, %conv2.1.conv2.2.layers.1.bn.weight, %conv2.1.conv2.2.layers.1.bn.bias, %conv2.1.conv2.2.layers.1.bn.running_mean, %conv2.1.conv2.2.layers.1.bn.running_var)
- %668 = Relu(%667)
- %669 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%668, %conv2.1.conv2.2.layers.2.conv1.weight)
- %670 = Conv[dilations = [1, 1], group = 64, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%669, %conv2.1.conv2.2.layers.2.conv2.weight)
- %671 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%670, %conv2.1.conv2.2.layers.2.bn.weight, %conv2.1.conv2.2.layers.2.bn.bias, %conv2.1.conv2.2.layers.2.bn.running_mean, %conv2.1.conv2.2.layers.2.bn.running_var)
- %672 = Relu(%671)
- %673 = GlobalAveragePool(%672)
- %674 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%673, %conv2.1.gate.fc1.weight, %conv2.1.gate.fc1.bias)
- %675 = Relu(%674)
- %676 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%675, %conv2.1.gate.fc2.weight, %conv2.1.gate.fc2.bias)
- %677 = Sigmoid(%676)
- %678 = Mul(%672, %677)
- %679 = Add(%660, %678)
- %680 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%635, %conv2.1.conv2.3.layers.0.conv1.weight)
- %681 = Conv[dilations = [1, 1], group = 64, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%680, %conv2.1.conv2.3.layers.0.conv2.weight)
- %682 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%681, %conv2.1.conv2.3.layers.0.bn.weight, %conv2.1.conv2.3.layers.0.bn.bias, %conv2.1.conv2.3.layers.0.bn.running_mean, %conv2.1.conv2.3.layers.0.bn.running_var)
- %683 = Relu(%682)
- %684 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%683, %conv2.1.conv2.3.layers.1.conv1.weight)
- %685 = Conv[dilations = [1, 1], group = 64, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%684, %conv2.1.conv2.3.layers.1.conv2.weight)
- %686 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%685, %conv2.1.conv2.3.layers.1.bn.weight, %conv2.1.conv2.3.layers.1.bn.bias, %conv2.1.conv2.3.layers.1.bn.running_mean, %conv2.1.conv2.3.layers.1.bn.running_var)
- %687 = Relu(%686)
- %688 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%687, %conv2.1.conv2.3.layers.2.conv1.weight)
- %689 = Conv[dilations = [1, 1], group = 64, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%688, %conv2.1.conv2.3.layers.2.conv2.weight)
- %690 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%689, %conv2.1.conv2.3.layers.2.bn.weight, %conv2.1.conv2.3.layers.2.bn.bias, %conv2.1.conv2.3.layers.2.bn.running_mean, %conv2.1.conv2.3.layers.2.bn.running_var)
- %691 = Relu(%690)
- %692 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%691, %conv2.1.conv2.3.layers.3.conv1.weight)
- %693 = Conv[dilations = [1, 1], group = 64, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%692, %conv2.1.conv2.3.layers.3.conv2.weight)
- %694 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%693, %conv2.1.conv2.3.layers.3.bn.weight, %conv2.1.conv2.3.layers.3.bn.bias, %conv2.1.conv2.3.layers.3.bn.running_mean, %conv2.1.conv2.3.layers.3.bn.running_var)
- %695 = Relu(%694)
- %696 = GlobalAveragePool(%695)
- %697 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%696, %conv2.1.gate.fc1.weight, %conv2.1.gate.fc1.bias)
- %698 = Relu(%697)
- %699 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%698, %conv2.1.gate.fc2.weight, %conv2.1.gate.fc2.bias)
- %700 = Sigmoid(%699)
- %701 = Mul(%695, %700)
- %702 = Add(%679, %701)
- %703 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%702, %conv2.1.conv3.conv.weight)
- %704 = InstanceNormalization[epsilon = 9.99999974737875e-06](%703, %conv2.1.IN.weight, %conv2.1.IN.bias)
- %705 = Add(%704, %632)
- %706 = Relu(%705)
- %707 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%706, %pool2.0.conv.weight)
- %708 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%707, %pool2.0.bn.weight, %pool2.0.bn.bias, %pool2.0.bn.running_mean, %pool2.0.bn.running_var)
- %709 = Relu(%708)
- %710 = Pad[mode = 'constant', pads = [0, 0, 0, 0, 0, 0, 0, 0], value = 0](%709)
- %711 = AveragePool[kernel_shape = [2, 2], pads = [0, 0, 0, 0], strides = [2, 2]](%710)
- %712 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%711, %conv3.0.conv1.conv.weight)
- %713 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%712, %conv3.0.conv1.bn.weight, %conv3.0.conv1.bn.bias, %conv3.0.conv1.bn.running_mean, %conv3.0.conv1.bn.running_var)
- %714 = Relu(%713)
- %715 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%714, %conv3.0.conv2.0.layers.0.conv1.weight)
- %716 = Conv[dilations = [1, 1], group = 96, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%715, %conv3.0.conv2.0.layers.0.conv2.weight)
- %717 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%716, %conv3.0.conv2.0.layers.0.bn.weight, %conv3.0.conv2.0.layers.0.bn.bias, %conv3.0.conv2.0.layers.0.bn.running_mean, %conv3.0.conv2.0.layers.0.bn.running_var)
- %718 = Relu(%717)
- %719 = GlobalAveragePool(%718)
- %720 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%719, %conv3.0.gate.fc1.weight, %conv3.0.gate.fc1.bias)
- %721 = Relu(%720)
- %722 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%721, %conv3.0.gate.fc2.weight, %conv3.0.gate.fc2.bias)
- %723 = Sigmoid(%722)
- %724 = Mul(%718, %723)
- %725 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%714, %conv3.0.conv2.1.layers.0.conv1.weight)
- %726 = Conv[dilations = [1, 1], group = 96, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%725, %conv3.0.conv2.1.layers.0.conv2.weight)
- %727 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%726, %conv3.0.conv2.1.layers.0.bn.weight, %conv3.0.conv2.1.layers.0.bn.bias, %conv3.0.conv2.1.layers.0.bn.running_mean, %conv3.0.conv2.1.layers.0.bn.running_var)
- %728 = Relu(%727)
- %729 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%728, %conv3.0.conv2.1.layers.1.conv1.weight)
- %730 = Conv[dilations = [1, 1], group = 96, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%729, %conv3.0.conv2.1.layers.1.conv2.weight)
- %731 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%730, %conv3.0.conv2.1.layers.1.bn.weight, %conv3.0.conv2.1.layers.1.bn.bias, %conv3.0.conv2.1.layers.1.bn.running_mean, %conv3.0.conv2.1.layers.1.bn.running_var)
- %732 = Relu(%731)
- %733 = GlobalAveragePool(%732)
- %734 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%733, %conv3.0.gate.fc1.weight, %conv3.0.gate.fc1.bias)
- %735 = Relu(%734)
- %736 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%735, %conv3.0.gate.fc2.weight, %conv3.0.gate.fc2.bias)
- %737 = Sigmoid(%736)
- %738 = Mul(%732, %737)
- %739 = Add(%724, %738)
- %740 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%714, %conv3.0.conv2.2.layers.0.conv1.weight)
- %741 = Conv[dilations = [1, 1], group = 96, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%740, %conv3.0.conv2.2.layers.0.conv2.weight)
- %742 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%741, %conv3.0.conv2.2.layers.0.bn.weight, %conv3.0.conv2.2.layers.0.bn.bias, %conv3.0.conv2.2.layers.0.bn.running_mean, %conv3.0.conv2.2.layers.0.bn.running_var)
- %743 = Relu(%742)
- %744 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%743, %conv3.0.conv2.2.layers.1.conv1.weight)
- %745 = Conv[dilations = [1, 1], group = 96, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%744, %conv3.0.conv2.2.layers.1.conv2.weight)
- %746 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%745, %conv3.0.conv2.2.layers.1.bn.weight, %conv3.0.conv2.2.layers.1.bn.bias, %conv3.0.conv2.2.layers.1.bn.running_mean, %conv3.0.conv2.2.layers.1.bn.running_var)
- %747 = Relu(%746)
- %748 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%747, %conv3.0.conv2.2.layers.2.conv1.weight)
- %749 = Conv[dilations = [1, 1], group = 96, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%748, %conv3.0.conv2.2.layers.2.conv2.weight)
- %750 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%749, %conv3.0.conv2.2.layers.2.bn.weight, %conv3.0.conv2.2.layers.2.bn.bias, %conv3.0.conv2.2.layers.2.bn.running_mean, %conv3.0.conv2.2.layers.2.bn.running_var)
- %751 = Relu(%750)
- %752 = GlobalAveragePool(%751)
- %753 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%752, %conv3.0.gate.fc1.weight, %conv3.0.gate.fc1.bias)
- %754 = Relu(%753)
- %755 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%754, %conv3.0.gate.fc2.weight, %conv3.0.gate.fc2.bias)
- %756 = Sigmoid(%755)
- %757 = Mul(%751, %756)
- %758 = Add(%739, %757)
- %759 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%714, %conv3.0.conv2.3.layers.0.conv1.weight)
- %760 = Conv[dilations = [1, 1], group = 96, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%759, %conv3.0.conv2.3.layers.0.conv2.weight)
- %761 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%760, %conv3.0.conv2.3.layers.0.bn.weight, %conv3.0.conv2.3.layers.0.bn.bias, %conv3.0.conv2.3.layers.0.bn.running_mean, %conv3.0.conv2.3.layers.0.bn.running_var)
- %762 = Relu(%761)
- %763 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%762, %conv3.0.conv2.3.layers.1.conv1.weight)
- %764 = Conv[dilations = [1, 1], group = 96, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%763, %conv3.0.conv2.3.layers.1.conv2.weight)
- %765 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%764, %conv3.0.conv2.3.layers.1.bn.weight, %conv3.0.conv2.3.layers.1.bn.bias, %conv3.0.conv2.3.layers.1.bn.running_mean, %conv3.0.conv2.3.layers.1.bn.running_var)
- %766 = Relu(%765)
- %767 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%766, %conv3.0.conv2.3.layers.2.conv1.weight)
- %768 = Conv[dilations = [1, 1], group = 96, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%767, %conv3.0.conv2.3.layers.2.conv2.weight)
- %769 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%768, %conv3.0.conv2.3.layers.2.bn.weight, %conv3.0.conv2.3.layers.2.bn.bias, %conv3.0.conv2.3.layers.2.bn.running_mean, %conv3.0.conv2.3.layers.2.bn.running_var)
- %770 = Relu(%769)
- %771 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%770, %conv3.0.conv2.3.layers.3.conv1.weight)
- %772 = Conv[dilations = [1, 1], group = 96, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%771, %conv3.0.conv2.3.layers.3.conv2.weight)
- %773 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%772, %conv3.0.conv2.3.layers.3.bn.weight, %conv3.0.conv2.3.layers.3.bn.bias, %conv3.0.conv2.3.layers.3.bn.running_mean, %conv3.0.conv2.3.layers.3.bn.running_var)
- %774 = Relu(%773)
- %775 = GlobalAveragePool(%774)
- %776 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%775, %conv3.0.gate.fc1.weight, %conv3.0.gate.fc1.bias)
- %777 = Relu(%776)
- %778 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%777, %conv3.0.gate.fc2.weight, %conv3.0.gate.fc2.bias)
- %779 = Sigmoid(%778)
- %780 = Mul(%774, %779)
- %781 = Add(%758, %780)
- %782 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%781, %conv3.0.conv3.conv.weight)
- %783 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%782, %conv3.0.conv3.bn.weight, %conv3.0.conv3.bn.bias, %conv3.0.conv3.bn.running_mean, %conv3.0.conv3.bn.running_var)
- %784 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%711, %conv3.0.downsample.conv.weight)
- %785 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%784, %conv3.0.downsample.bn.weight, %conv3.0.downsample.bn.bias, %conv3.0.downsample.bn.running_mean, %conv3.0.downsample.bn.running_var)
- %786 = Add(%783, %785)
- %787 = Relu(%786)
- %788 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%787, %conv3.1.conv1.conv.weight)
- %789 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%788, %conv3.1.conv1.bn.weight, %conv3.1.conv1.bn.bias, %conv3.1.conv1.bn.running_mean, %conv3.1.conv1.bn.running_var)
- %790 = Relu(%789)
- %791 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%790, %conv3.1.conv2.0.layers.0.conv1.weight)
- %792 = Conv[dilations = [1, 1], group = 96, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%791, %conv3.1.conv2.0.layers.0.conv2.weight)
- %793 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%792, %conv3.1.conv2.0.layers.0.bn.weight, %conv3.1.conv2.0.layers.0.bn.bias, %conv3.1.conv2.0.layers.0.bn.running_mean, %conv3.1.conv2.0.layers.0.bn.running_var)
- %794 = Relu(%793)
- %795 = GlobalAveragePool(%794)
- %796 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%795, %conv3.1.gate.fc1.weight, %conv3.1.gate.fc1.bias)
- %797 = Relu(%796)
- %798 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%797, %conv3.1.gate.fc2.weight, %conv3.1.gate.fc2.bias)
- %799 = Sigmoid(%798)
- %800 = Mul(%794, %799)
- %801 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%790, %conv3.1.conv2.1.layers.0.conv1.weight)
- %802 = Conv[dilations = [1, 1], group = 96, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%801, %conv3.1.conv2.1.layers.0.conv2.weight)
- %803 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%802, %conv3.1.conv2.1.layers.0.bn.weight, %conv3.1.conv2.1.layers.0.bn.bias, %conv3.1.conv2.1.layers.0.bn.running_mean, %conv3.1.conv2.1.layers.0.bn.running_var)
- %804 = Relu(%803)
- %805 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%804, %conv3.1.conv2.1.layers.1.conv1.weight)
- %806 = Conv[dilations = [1, 1], group = 96, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%805, %conv3.1.conv2.1.layers.1.conv2.weight)
- %807 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%806, %conv3.1.conv2.1.layers.1.bn.weight, %conv3.1.conv2.1.layers.1.bn.bias, %conv3.1.conv2.1.layers.1.bn.running_mean, %conv3.1.conv2.1.layers.1.bn.running_var)
- %808 = Relu(%807)
- %809 = GlobalAveragePool(%808)
- %810 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%809, %conv3.1.gate.fc1.weight, %conv3.1.gate.fc1.bias)
- %811 = Relu(%810)
- %812 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%811, %conv3.1.gate.fc2.weight, %conv3.1.gate.fc2.bias)
- %813 = Sigmoid(%812)
- %814 = Mul(%808, %813)
- %815 = Add(%800, %814)
- %816 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%790, %conv3.1.conv2.2.layers.0.conv1.weight)
- %817 = Conv[dilations = [1, 1], group = 96, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%816, %conv3.1.conv2.2.layers.0.conv2.weight)
- %818 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%817, %conv3.1.conv2.2.layers.0.bn.weight, %conv3.1.conv2.2.layers.0.bn.bias, %conv3.1.conv2.2.layers.0.bn.running_mean, %conv3.1.conv2.2.layers.0.bn.running_var)
- %819 = Relu(%818)
- %820 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%819, %conv3.1.conv2.2.layers.1.conv1.weight)
- %821 = Conv[dilations = [1, 1], group = 96, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%820, %conv3.1.conv2.2.layers.1.conv2.weight)
- %822 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%821, %conv3.1.conv2.2.layers.1.bn.weight, %conv3.1.conv2.2.layers.1.bn.bias, %conv3.1.conv2.2.layers.1.bn.running_mean, %conv3.1.conv2.2.layers.1.bn.running_var)
- %823 = Relu(%822)
- %824 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%823, %conv3.1.conv2.2.layers.2.conv1.weight)
- %825 = Conv[dilations = [1, 1], group = 96, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%824, %conv3.1.conv2.2.layers.2.conv2.weight)
- %826 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%825, %conv3.1.conv2.2.layers.2.bn.weight, %conv3.1.conv2.2.layers.2.bn.bias, %conv3.1.conv2.2.layers.2.bn.running_mean, %conv3.1.conv2.2.layers.2.bn.running_var)
- %827 = Relu(%826)
- %828 = GlobalAveragePool(%827)
- %829 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%828, %conv3.1.gate.fc1.weight, %conv3.1.gate.fc1.bias)
- %830 = Relu(%829)
- %831 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%830, %conv3.1.gate.fc2.weight, %conv3.1.gate.fc2.bias)
- %832 = Sigmoid(%831)
- %833 = Mul(%827, %832)
- %834 = Add(%815, %833)
- %835 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%790, %conv3.1.conv2.3.layers.0.conv1.weight)
- %836 = Conv[dilations = [1, 1], group = 96, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%835, %conv3.1.conv2.3.layers.0.conv2.weight)
- %837 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%836, %conv3.1.conv2.3.layers.0.bn.weight, %conv3.1.conv2.3.layers.0.bn.bias, %conv3.1.conv2.3.layers.0.bn.running_mean, %conv3.1.conv2.3.layers.0.bn.running_var)
- %838 = Relu(%837)
- %839 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%838, %conv3.1.conv2.3.layers.1.conv1.weight)
- %840 = Conv[dilations = [1, 1], group = 96, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%839, %conv3.1.conv2.3.layers.1.conv2.weight)
- %841 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%840, %conv3.1.conv2.3.layers.1.bn.weight, %conv3.1.conv2.3.layers.1.bn.bias, %conv3.1.conv2.3.layers.1.bn.running_mean, %conv3.1.conv2.3.layers.1.bn.running_var)
- %842 = Relu(%841)
- %843 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%842, %conv3.1.conv2.3.layers.2.conv1.weight)
- %844 = Conv[dilations = [1, 1], group = 96, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%843, %conv3.1.conv2.3.layers.2.conv2.weight)
- %845 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%844, %conv3.1.conv2.3.layers.2.bn.weight, %conv3.1.conv2.3.layers.2.bn.bias, %conv3.1.conv2.3.layers.2.bn.running_mean, %conv3.1.conv2.3.layers.2.bn.running_var)
- %846 = Relu(%845)
- %847 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%846, %conv3.1.conv2.3.layers.3.conv1.weight)
- %848 = Conv[dilations = [1, 1], group = 96, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%847, %conv3.1.conv2.3.layers.3.conv2.weight)
- %849 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%848, %conv3.1.conv2.3.layers.3.bn.weight, %conv3.1.conv2.3.layers.3.bn.bias, %conv3.1.conv2.3.layers.3.bn.running_mean, %conv3.1.conv2.3.layers.3.bn.running_var)
- %850 = Relu(%849)
- %851 = GlobalAveragePool(%850)
- %852 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%851, %conv3.1.gate.fc1.weight, %conv3.1.gate.fc1.bias)
- %853 = Relu(%852)
- %854 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%853, %conv3.1.gate.fc2.weight, %conv3.1.gate.fc2.bias)
- %855 = Sigmoid(%854)
- %856 = Mul(%850, %855)
- %857 = Add(%834, %856)
- %858 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%857, %conv3.1.conv3.conv.weight)
- %859 = InstanceNormalization[epsilon = 9.99999974737875e-06](%858, %conv3.1.IN.weight, %conv3.1.IN.bias)
- %860 = Add(%859, %787)
- %861 = Relu(%860)
- %862 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%861, %pool3.0.conv.weight)
- %863 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%862, %pool3.0.bn.weight, %pool3.0.bn.bias, %pool3.0.bn.running_mean, %pool3.0.bn.running_var)
- %864 = Relu(%863)
- %865 = Pad[mode = 'constant', pads = [0, 0, 0, 0, 0, 0, 0, 0], value = 0](%864)
- %866 = AveragePool[kernel_shape = [2, 2], pads = [0, 0, 0, 0], strides = [2, 2]](%865)
- %867 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%866, %conv4.0.conv1.conv.weight)
- %868 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%867, %conv4.0.conv1.bn.weight, %conv4.0.conv1.bn.bias, %conv4.0.conv1.bn.running_mean, %conv4.0.conv1.bn.running_var)
- %869 = Relu(%868)
- %870 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%869, %conv4.0.conv2.0.layers.0.conv1.weight)
- %871 = Conv[dilations = [1, 1], group = 128, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%870, %conv4.0.conv2.0.layers.0.conv2.weight)
- %872 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%871, %conv4.0.conv2.0.layers.0.bn.weight, %conv4.0.conv2.0.layers.0.bn.bias, %conv4.0.conv2.0.layers.0.bn.running_mean, %conv4.0.conv2.0.layers.0.bn.running_var)
- %873 = Relu(%872)
- %874 = GlobalAveragePool(%873)
- %875 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%874, %conv4.0.gate.fc1.weight, %conv4.0.gate.fc1.bias)
- %876 = Relu(%875)
- %877 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%876, %conv4.0.gate.fc2.weight, %conv4.0.gate.fc2.bias)
- %878 = Sigmoid(%877)
- %879 = Mul(%873, %878)
- %880 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%869, %conv4.0.conv2.1.layers.0.conv1.weight)
- %881 = Conv[dilations = [1, 1], group = 128, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%880, %conv4.0.conv2.1.layers.0.conv2.weight)
- %882 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%881, %conv4.0.conv2.1.layers.0.bn.weight, %conv4.0.conv2.1.layers.0.bn.bias, %conv4.0.conv2.1.layers.0.bn.running_mean, %conv4.0.conv2.1.layers.0.bn.running_var)
- %883 = Relu(%882)
- %884 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%883, %conv4.0.conv2.1.layers.1.conv1.weight)
- %885 = Conv[dilations = [1, 1], group = 128, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%884, %conv4.0.conv2.1.layers.1.conv2.weight)
- %886 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%885, %conv4.0.conv2.1.layers.1.bn.weight, %conv4.0.conv2.1.layers.1.bn.bias, %conv4.0.conv2.1.layers.1.bn.running_mean, %conv4.0.conv2.1.layers.1.bn.running_var)
- %887 = Relu(%886)
- %888 = GlobalAveragePool(%887)
- %889 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%888, %conv4.0.gate.fc1.weight, %conv4.0.gate.fc1.bias)
- %890 = Relu(%889)
- %891 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%890, %conv4.0.gate.fc2.weight, %conv4.0.gate.fc2.bias)
- %892 = Sigmoid(%891)
- %893 = Mul(%887, %892)
- %894 = Add(%879, %893)
- %895 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%869, %conv4.0.conv2.2.layers.0.conv1.weight)
- %896 = Conv[dilations = [1, 1], group = 128, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%895, %conv4.0.conv2.2.layers.0.conv2.weight)
- %897 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%896, %conv4.0.conv2.2.layers.0.bn.weight, %conv4.0.conv2.2.layers.0.bn.bias, %conv4.0.conv2.2.layers.0.bn.running_mean, %conv4.0.conv2.2.layers.0.bn.running_var)
- %898 = Relu(%897)
- %899 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%898, %conv4.0.conv2.2.layers.1.conv1.weight)
- %900 = Conv[dilations = [1, 1], group = 128, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%899, %conv4.0.conv2.2.layers.1.conv2.weight)
- %901 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%900, %conv4.0.conv2.2.layers.1.bn.weight, %conv4.0.conv2.2.layers.1.bn.bias, %conv4.0.conv2.2.layers.1.bn.running_mean, %conv4.0.conv2.2.layers.1.bn.running_var)
- %902 = Relu(%901)
- %903 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%902, %conv4.0.conv2.2.layers.2.conv1.weight)
- %904 = Conv[dilations = [1, 1], group = 128, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%903, %conv4.0.conv2.2.layers.2.conv2.weight)
- %905 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%904, %conv4.0.conv2.2.layers.2.bn.weight, %conv4.0.conv2.2.layers.2.bn.bias, %conv4.0.conv2.2.layers.2.bn.running_mean, %conv4.0.conv2.2.layers.2.bn.running_var)
- %906 = Relu(%905)
- %907 = GlobalAveragePool(%906)
- %908 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%907, %conv4.0.gate.fc1.weight, %conv4.0.gate.fc1.bias)
- %909 = Relu(%908)
- %910 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%909, %conv4.0.gate.fc2.weight, %conv4.0.gate.fc2.bias)
- %911 = Sigmoid(%910)
- %912 = Mul(%906, %911)
- %913 = Add(%894, %912)
- %914 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%869, %conv4.0.conv2.3.layers.0.conv1.weight)
- %915 = Conv[dilations = [1, 1], group = 128, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%914, %conv4.0.conv2.3.layers.0.conv2.weight)
- %916 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%915, %conv4.0.conv2.3.layers.0.bn.weight, %conv4.0.conv2.3.layers.0.bn.bias, %conv4.0.conv2.3.layers.0.bn.running_mean, %conv4.0.conv2.3.layers.0.bn.running_var)
- %917 = Relu(%916)
- %918 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%917, %conv4.0.conv2.3.layers.1.conv1.weight)
- %919 = Conv[dilations = [1, 1], group = 128, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%918, %conv4.0.conv2.3.layers.1.conv2.weight)
- %920 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%919, %conv4.0.conv2.3.layers.1.bn.weight, %conv4.0.conv2.3.layers.1.bn.bias, %conv4.0.conv2.3.layers.1.bn.running_mean, %conv4.0.conv2.3.layers.1.bn.running_var)
- %921 = Relu(%920)
- %922 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%921, %conv4.0.conv2.3.layers.2.conv1.weight)
- %923 = Conv[dilations = [1, 1], group = 128, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%922, %conv4.0.conv2.3.layers.2.conv2.weight)
- %924 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%923, %conv4.0.conv2.3.layers.2.bn.weight, %conv4.0.conv2.3.layers.2.bn.bias, %conv4.0.conv2.3.layers.2.bn.running_mean, %conv4.0.conv2.3.layers.2.bn.running_var)
- %925 = Relu(%924)
- %926 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%925, %conv4.0.conv2.3.layers.3.conv1.weight)
- %927 = Conv[dilations = [1, 1], group = 128, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%926, %conv4.0.conv2.3.layers.3.conv2.weight)
- %928 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%927, %conv4.0.conv2.3.layers.3.bn.weight, %conv4.0.conv2.3.layers.3.bn.bias, %conv4.0.conv2.3.layers.3.bn.running_mean, %conv4.0.conv2.3.layers.3.bn.running_var)
- %929 = Relu(%928)
- %930 = GlobalAveragePool(%929)
- %931 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%930, %conv4.0.gate.fc1.weight, %conv4.0.gate.fc1.bias)
- %932 = Relu(%931)
- %933 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%932, %conv4.0.gate.fc2.weight, %conv4.0.gate.fc2.bias)
- %934 = Sigmoid(%933)
- %935 = Mul(%929, %934)
- %936 = Add(%913, %935)
- %937 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%936, %conv4.0.conv3.conv.weight)
- %938 = InstanceNormalization[epsilon = 9.99999974737875e-06](%937, %conv4.0.IN.weight, %conv4.0.IN.bias)
- %939 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%866, %conv4.0.downsample.conv.weight)
- %940 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%939, %conv4.0.downsample.bn.weight, %conv4.0.downsample.bn.bias, %conv4.0.downsample.bn.running_mean, %conv4.0.downsample.bn.running_var)
- %941 = Add(%938, %940)
- %942 = Relu(%941)
- %943 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%942, %conv4.1.conv1.conv.weight)
- %944 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%943, %conv4.1.conv1.bn.weight, %conv4.1.conv1.bn.bias, %conv4.1.conv1.bn.running_mean, %conv4.1.conv1.bn.running_var)
- %945 = Relu(%944)
- %946 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%945, %conv4.1.conv2.0.layers.0.conv1.weight)
- %947 = Conv[dilations = [1, 1], group = 128, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%946, %conv4.1.conv2.0.layers.0.conv2.weight)
- %948 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%947, %conv4.1.conv2.0.layers.0.bn.weight, %conv4.1.conv2.0.layers.0.bn.bias, %conv4.1.conv2.0.layers.0.bn.running_mean, %conv4.1.conv2.0.layers.0.bn.running_var)
- %949 = Relu(%948)
- %950 = GlobalAveragePool(%949)
- %951 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%950, %conv4.1.gate.fc1.weight, %conv4.1.gate.fc1.bias)
- %952 = Relu(%951)
- %953 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%952, %conv4.1.gate.fc2.weight, %conv4.1.gate.fc2.bias)
- %954 = Sigmoid(%953)
- %955 = Mul(%949, %954)
- %956 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%945, %conv4.1.conv2.1.layers.0.conv1.weight)
- %957 = Conv[dilations = [1, 1], group = 128, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%956, %conv4.1.conv2.1.layers.0.conv2.weight)
- %958 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%957, %conv4.1.conv2.1.layers.0.bn.weight, %conv4.1.conv2.1.layers.0.bn.bias, %conv4.1.conv2.1.layers.0.bn.running_mean, %conv4.1.conv2.1.layers.0.bn.running_var)
- %959 = Relu(%958)
- %960 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%959, %conv4.1.conv2.1.layers.1.conv1.weight)
- %961 = Conv[dilations = [1, 1], group = 128, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%960, %conv4.1.conv2.1.layers.1.conv2.weight)
- %962 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%961, %conv4.1.conv2.1.layers.1.bn.weight, %conv4.1.conv2.1.layers.1.bn.bias, %conv4.1.conv2.1.layers.1.bn.running_mean, %conv4.1.conv2.1.layers.1.bn.running_var)
- %963 = Relu(%962)
- %964 = GlobalAveragePool(%963)
- %965 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%964, %conv4.1.gate.fc1.weight, %conv4.1.gate.fc1.bias)
- %966 = Relu(%965)
- %967 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%966, %conv4.1.gate.fc2.weight, %conv4.1.gate.fc2.bias)
- %968 = Sigmoid(%967)
- %969 = Mul(%963, %968)
- %970 = Add(%955, %969)
- %971 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%945, %conv4.1.conv2.2.layers.0.conv1.weight)
- %972 = Conv[dilations = [1, 1], group = 128, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%971, %conv4.1.conv2.2.layers.0.conv2.weight)
- %973 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%972, %conv4.1.conv2.2.layers.0.bn.weight, %conv4.1.conv2.2.layers.0.bn.bias, %conv4.1.conv2.2.layers.0.bn.running_mean, %conv4.1.conv2.2.layers.0.bn.running_var)
- %974 = Relu(%973)
- %975 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%974, %conv4.1.conv2.2.layers.1.conv1.weight)
- %976 = Conv[dilations = [1, 1], group = 128, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%975, %conv4.1.conv2.2.layers.1.conv2.weight)
- %977 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%976, %conv4.1.conv2.2.layers.1.bn.weight, %conv4.1.conv2.2.layers.1.bn.bias, %conv4.1.conv2.2.layers.1.bn.running_mean, %conv4.1.conv2.2.layers.1.bn.running_var)
- %978 = Relu(%977)
- %979 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%978, %conv4.1.conv2.2.layers.2.conv1.weight)
- %980 = Conv[dilations = [1, 1], group = 128, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%979, %conv4.1.conv2.2.layers.2.conv2.weight)
- %981 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%980, %conv4.1.conv2.2.layers.2.bn.weight, %conv4.1.conv2.2.layers.2.bn.bias, %conv4.1.conv2.2.layers.2.bn.running_mean, %conv4.1.conv2.2.layers.2.bn.running_var)
- %982 = Relu(%981)
- %983 = GlobalAveragePool(%982)
- %984 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%983, %conv4.1.gate.fc1.weight, %conv4.1.gate.fc1.bias)
- %985 = Relu(%984)
- %986 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%985, %conv4.1.gate.fc2.weight, %conv4.1.gate.fc2.bias)
- %987 = Sigmoid(%986)
- %988 = Mul(%982, %987)
- %989 = Add(%970, %988)
- %990 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%945, %conv4.1.conv2.3.layers.0.conv1.weight)
- %991 = Conv[dilations = [1, 1], group = 128, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%990, %conv4.1.conv2.3.layers.0.conv2.weight)
- %992 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%991, %conv4.1.conv2.3.layers.0.bn.weight, %conv4.1.conv2.3.layers.0.bn.bias, %conv4.1.conv2.3.layers.0.bn.running_mean, %conv4.1.conv2.3.layers.0.bn.running_var)
- %993 = Relu(%992)
- %994 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%993, %conv4.1.conv2.3.layers.1.conv1.weight)
- %995 = Conv[dilations = [1, 1], group = 128, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%994, %conv4.1.conv2.3.layers.1.conv2.weight)
- %996 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%995, %conv4.1.conv2.3.layers.1.bn.weight, %conv4.1.conv2.3.layers.1.bn.bias, %conv4.1.conv2.3.layers.1.bn.running_mean, %conv4.1.conv2.3.layers.1.bn.running_var)
- %997 = Relu(%996)
- %998 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%997, %conv4.1.conv2.3.layers.2.conv1.weight)
- %999 = Conv[dilations = [1, 1], group = 128, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%998, %conv4.1.conv2.3.layers.2.conv2.weight)
- %1000 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%999, %conv4.1.conv2.3.layers.2.bn.weight, %conv4.1.conv2.3.layers.2.bn.bias, %conv4.1.conv2.3.layers.2.bn.running_mean, %conv4.1.conv2.3.layers.2.bn.running_var)
- %1001 = Relu(%1000)
- %1002 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%1001, %conv4.1.conv2.3.layers.3.conv1.weight)
- %1003 = Conv[dilations = [1, 1], group = 128, kernel_shape = [3, 3], pads = [1, 1, 1, 1], strides = [1, 1]](%1002, %conv4.1.conv2.3.layers.3.conv2.weight)
- %1004 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%1003, %conv4.1.conv2.3.layers.3.bn.weight, %conv4.1.conv2.3.layers.3.bn.bias, %conv4.1.conv2.3.layers.3.bn.running_mean, %conv4.1.conv2.3.layers.3.bn.running_var)
- %1005 = Relu(%1004)
- %1006 = GlobalAveragePool(%1005)
- %1007 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%1006, %conv4.1.gate.fc1.weight, %conv4.1.gate.fc1.bias)
- %1008 = Relu(%1007)
- %1009 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%1008, %conv4.1.gate.fc2.weight, %conv4.1.gate.fc2.bias)
- %1010 = Sigmoid(%1009)
- %1011 = Mul(%1005, %1010)
- %1012 = Add(%989, %1011)
- %1013 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%1012, %conv4.1.conv3.conv.weight)
- %1014 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%1013, %conv4.1.conv3.bn.weight, %conv4.1.conv3.bn.bias, %conv4.1.conv3.bn.running_mean, %conv4.1.conv3.bn.running_var)
- %1015 = Add(%1014, %942)
- %1016 = Relu(%1015)
- %1017 = Conv[dilations = [1, 1], group = 1, kernel_shape = [1, 1], pads = [0, 0, 0, 0], strides = [1, 1]](%1016, %conv5.conv.weight)
- %1018 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%1017, %conv5.bn.weight, %conv5.bn.bias, %conv5.bn.running_mean, %conv5.bn.running_var)
- %1019 = Relu(%1018)
- %1020 = GlobalAveragePool(%1019)
- %1021 = Shape(%1020)
- %1022 = Constant[value = <Scalar Tensor []>]()
- %1023 = Gather[axis = 0](%1021, %1022)
- %1024 = Constant[value = <Scalar Tensor []>]()
- %1025 = Unsqueeze[axes = [0]](%1023)
- %1026 = Unsqueeze[axes = [0]](%1024)
- %1027 = Concat[axis = 0](%1025, %1026)
- %1028 = Reshape(%1020, %1027)
- %1029 = Gemm[alpha = 1, beta = 1, transB = 1](%1028, %fc.0.weight, %fc.0.bias)
- %1030 = BatchNormalization[epsilon = 9.99999974737875e-06, momentum = 0.899999976158142](%1029, %fc.1.weight, %fc.1.bias, %fc.1.running_mean, %fc.1.running_var)
- %1031 = Relu(%1030)
- return %1031
- }
Advertisement
Add Comment
Please, Sign In to add comment