Hi,
I follow the documentation for the semantic segmentation example. but when I do the cmd './ train_cityscapes_segmentation.sh'.
Logging output to training/cityscapes5_jsegnet21v2_2020-08-17_11-30-06/train-log_2020-08-17_11-30-06.txt
training/cityscapes5_jsegnet21v2_2020-08-17_11-30-06/initial
num_gpus: 1 gpulist: ['0']
I0817 11:30:08.093606 25692 caffe.cpp:902] This is NVCaffe 0.17.0 started at Mon Aug 17 11:30:07 2020
I0817 11:30:08.093741 25692 caffe.cpp:904] CuDNN version: 7605
I0817 11:30:08.093746 25692 caffe.cpp:905] CuBLAS version: 9000
I0817 11:30:08.093749 25692 caffe.cpp:906] CUDA version: 9000
I0817 11:30:08.093750 25692 caffe.cpp:907] CUDA driver version: 10000
I0817 11:30:08.093755 25692 caffe.cpp:908] Arguments:
[0]: /home/jiandong/Project/caffe-jacinto/build/tools/caffe
[1]: train
[2]: --solver=training/cityscapes5_jsegnet21v2_2020-08-17_11-30-06/initial/solver.prototxt
[3]: --weights=../trained/image_classification/imagenet_jacintonet11v2/initial/imagenet_jacintonet11v2_iter_320000.caffemodel
[4]: --gpu
[5]: 0
I0817 11:30:08.114403 25692 gpu_memory.cpp:105] GPUMemory::Manager initialized
I0817 11:30:08.114796 25692 gpu_memory.cpp:107] Total memory: 4236312576, Free: 3315728384, dev_info[0]: total=4236312576 free=3315728384
I0817 11:30:08.114804 25692 caffe.cpp:226] Using GPUs 0
I0817 11:30:08.115073 25692 caffe.cpp:230] GPU 0: GeForce GTX 1050 Ti
I0817 11:30:08.115137 25692 solver.cpp:41] Solver data type: FLOAT
I0817 11:30:08.121004 25692 solver.cpp:44] Initializing solver from parameters:
train_net: "training/cityscapes5_jsegnet21v2_2020-08-17_11-30-06/initial/train.prototxt"
test_net: "training/cityscapes5_jsegnet21v2_2020-08-17_11-30-06/initial/test.prototxt"
test_iter: 500
test_interval: 2000
base_lr: 0.01
display: 100
max_iter: 120000
lr_policy: "multistep"
gamma: 0.1
power: 1
momentum: 0.9
weight_decay: 0.0001
snapshot: 10000
snapshot_prefix: "training/cityscapes5_jsegnet21v2_2020-08-17_11-30-06/initial/cityscapes5_jsegnet21v2"
solver_mode: GPU
device_id: 0
random_seed: 33
debug_info: false
train_state {
level: 0
stage: ""
}
snapshot_after_train: true
test_initialization: false
stepvalue: 60000
stepvalue: 90000
iter_size: 4
type: "SGD"
I0817 11:30:08.121160 25692 solver.cpp:76] Creating training net from train_net file: training/cityscapes5_jsegnet21v2_2020-08-17_11-30-06/initial/train.prototxt
I0817 11:30:08.121774 25692 net.cpp:457] The NetState phase (0) differed from the phase (1) specified by a rule in layer accuracy/top1
I0817 11:30:08.121781 25692 net.cpp:457] The NetState phase (0) differed from the phase (1) specified by a rule in layer accuracy/top5
I0817 11:30:08.122068 25692 net.cpp:80] Initializing net from parameters:
name: "jsegnet21v2_train"
state {
phase: TRAIN
level: 0
stage: ""
}
layer {
name: "data"
type: "ImageLabelData"
top: "data"
top: "label"
transform_param {
mirror: true
crop_size: 640
mean_value: 0
}
image_label_data_param {
image_list_path: "data/train-image-lmdb"
label_list_path: "data/train-label-lmdb"
batch_size: 4
shuffle: true
threads: 1
backend: LMDB
}
}
layer {
name: "data/bias"
type: "Bias"
bottom: "data"
top: "data/bias"
param {
lr_mult: 0
decay_mult: 0
}
bias_param {
filler {
type: "constant"
value: -128
}
}
}
layer {
name: "conv1a"
type: "Convolution"
bottom: "data/bias"
top: "conv1a"
param {
lr_mult: 1
decay_mult: 1
}
param {
lr_mult: 2
decay_mult: 0
}
convolution_param {
num_output: 32
bias_term: true
pad: 2
kernel_size: 5
group: 1
stride: 2
weight_filler {
type: "msra"
}
bias_filler {
type: "constant"
value: 0
}
dilation: 1
}
}
layer {
name: "conv1a/bn"
type: "BatchNorm"
bottom: "conv1a"
top: "conv1a"
batch_norm_param {
moving_average_fraction: 0.99
eps: 0.0001
scale_bias: true
}
}
layer {
name: "conv1a/relu"
type: "ReLU"
bottom: "conv1a"
top: "conv1a"
}
layer {
name: "conv1b"
type: "Convolution"
bottom: "conv1a"
top: "conv1b"
param {
lr_mult: 1
decay_mult: 1
}
param {
lr_mult: 2
decay_mult: 0
}
convolution_param {
num_output: 32
bias_term: true
pad: 1
kernel_size: 3
group: 4
stride: 1
weight_filler {
type: "msra"
}
bias_filler {
type: "constant"
value: 0
}
dilation: 1
}
}
layer {
name: "conv1b/bn"
type: "BatchNorm"
bottom: "conv1b"
top: "conv1b"
batch_norm_param {
moving_average_fraction: 0.99
eps: 0.0001
scale_bias: true
}
}
layer {
name: "conv1b/relu"
type: "ReLU"
bottom: "conv1b"
top: "conv1b"
}
layer {
name: "pool1"
type: "Pooling"
bottom: "conv1b"
top: "pool1"
pooling_param {
pool: MAX
kernel_size: 2
stride: 2
}
}
layer {
name: "res2a_branch2a"
type: "Convolution"
bottom: "pool1"
top: "res2a_branch2a"
param {
lr_mult: 1
decay_mult: 1
}
param {
lr_mult: 2
decay_mult: 0
}
convolution_param {
num_output: 64
bias_term: true
pad: 1
kernel_size: 3
group: 1
stride: 1
weight_filler {
type: "msra"
}
bias_filler {
type: "constant"
value: 0
}
dilation: 1
}
}
layer {
name: "res2a_branch2a/bn"
type: "BatchNorm"
bottom: "res2a_branch2a"
top: "res2a_branch2a"
batch_norm_param {
moving_average_fraction: 0.99
eps: 0.0001
scale_bias: true
}
}
layer {
name: "res2a_branch2a/relu"
type: "ReLU"
bottom: "res2a_branch2a"
top: "res2a_branch2a"
}
layer {
name: "res2a_branch2b"
type: "Convolution"
bottom: "res2a_branch2a"
top: "res2a_branch2b"
param {
lr_mult: 1
decay_mult:1 }decay_mult: 0lr_mult: 2param {
}
convolution_param {
num_output: 64
bias_term: true
pad: 1
kernel_size: 3
group: 4
stride: 1
weight_filler {
type: "msra"
}
bias_filler {
type: "constant"
value: 0
}
dilation: 1
}
}
layer {
name: "res2a_branch2b/bn"
type: "BatchNorm"
bottom: "res2a_branch2b"
top: "res2a_branch2b"
batch_norm_param {
moving_average_fraction: 0.99
eps: 0.0001
scale_bias: true
}
}
layer {
name: "res2a_branch2b/relu"
type: "ReLU"
bottom: "res2a_branch2b"
top: "res2a_branch2b"
}
layer {
name: "pool2"
type: "Pooling"
bottom: "res2a_branch2b"
top: "pool2"
pooling_param {
pool: MAX
kernel_size: 2
stride: 2
}
}
layer {
name: "res3a_branch2a"
type: "Convolution"
bottom: "pool2"
top: "res3a_branch2a"
param {
lr_mult: 1
decay_mult: 1
}
param {
lr_mult: 2
decay_mult: 0
}
convolution_param {
num_output: 128
bias_term: true
pad: 1
kernel_size: 3
group: 1
stride: 1
weight_filler {
type: "msra"
}
bias_filler {
type: "constant"
value: 0
}
dilation: 1
}
}
layer {
name: "res3a_branch2a/bn"
type: "BatchNorm"
bottom: "res3a_branch2a"
top: "res3a_branch2a"
batch_norm_param {
moving_average_fraction: 0.99
eps: 0.0001
scale_bias: true
}
}
layer {
name: "res3a_branch2a/relu"
type: "ReLU"
bottom: "res3a_branch2a"
top: "res3a_branch2a"
}
layer {
name: "res3a_branch2b"
type: "Convolution"
bottom: "res3a_branch2a"
top: "res3a_branch2b"
param {
lr_mult: 1
decay_mult: 1
}
param {
lr_mult: 2
decay_mult: 0
}
convolution_param {
num_output: 128
bias_term: true
pad: 1
kernel_size: 3
group: 4
stride: 1
weight_filler {
type: "msra"
}
bias_filler {
type: "constant"
value: 0
}
dilation: 1
}
}
layer {
name: "res3a_branch2b/bn"
type: "BatchNorm"
bottom: "res3a_branch2b"
top: "res3a_branch2b"
batch_norm_param {
moving_average_fraction: 0.99
eps: 0.0001
scale_bias: true
}
}
layer {
name: "res3a_branch2b/relu"
type: "ReLU"
bottom: "res3a_branch2b"
top: "res3a_branch2b"
}
layer {
name: "pool3"
type: "Pooling"
bottom: "res3a_branch2b"
top: "pool3"
pooling_param {
pool: MAX
kernel_size: 2
stride: 2
}
}
layer {
name: "res4a_branch2a"
type: "Convolution"
bottom: "pool3"
top: "res4a_branch2a"
param {
lr_mult: 1
decay_mult:1 }decay_mult: 0lr_mult: 2param {
}
convolution_param {
num_output: 256
bias_term: true
pad: 1
kernel_size: 3
group: 1
stride: 1
weight_filler {
type: "msra"
}
bias_filler {
type: "constant"
value: 0
}
dilation: 1
}
}
layer {
name: "res4a_branch2a/bn"
type: "BatchNorm"
bottom: "res4a_branch2a"
top: "res4a_branch2a"
batch_norm_param {
moving_average_fraction: 0.99
eps: 0.0001
scale_bias: true
}
}
layer {
name: "res4a_branch2a/relu"
type: "ReLU"
bottom: "res4a_branch2a"
top: "res4a_branch2a"
}
layer {
name: "res4a_branch2b"
type: "Convolution"
bottom: "res4a_branch2a"
top: "res4a_branch2b"
param {
lr_mult: 1
decay_mult: 1
}
param {
lr_mult: 2
decay_mult: 0
}
convolution_param {
num_output: 256
bias_term: true
pad: 1
kernel_size: 3
group: 4
stride: 1
weight_filler {
type: "msra"
}
bias_filler {
type: "constant"
value: 0
}
dilation: 1
}
}
layer {
name: "res4a_branch2b/bn"
type: "BatchNorm"
bottom: "res4a_branch2b"
top: "res4a_branch2b"
batch_norm_param {
moving_average_fraction: 0.99
eps: 0.0001
scale_bias: true
}
}
layer {
name: "res4a_branch2b/relu"
type: "ReLU"
bottom: "res4a_branch2b"
top: "res4a_branch2b"
}
layer {
name: "pool4"
type: "Pooling"
bottom: "res4a_branch2b"
top: "pool4"
pooling_param {
pool: MAX
kernel_size: 1
stride: 1
}
}
layer {
name: "res5a_branch2a"
type: "Convolution"
bottom: "pool4"
top: "res5a_branch2a"
param {
lr_mult: 1
decay_mult: 1
}
param {
lr_mult: 2
decay_mult: 0
}
convolution_param {
num_output: 512
bias_term: true
pad: 2
kernel_size: 3
group: 1
stride: 1
weight_filler {
type: "msra"
}
bias_filler {
type: "constant"
value: 0
}
dilation: 2
}
}
layer {
name: "res5a_branch2a/bn"
type: "BatchNorm"
bottom: "res5a_branch2a"
top: "res5a_branch2a"
batch_norm_param {
moving_average_fraction: 0.99
eps: 0.0001
scale_bias: true
}
}
layer {
name: "res5a_branch2a/relu"
type: "ReLU"
bottom: "res5a_branch2a"
top: "res5a_branch2a"
}
layer {
name: "res5a_branch2b"
type: "Convolution"
bottom: "res5a_branch2a"
top: "res5a_branch2b"
param {
lr_mult: 1
decay_mult: 1
}
param {
lr_mult: 2
decay_mult: 0
}
convolution_param {
num_output: 512
bias_term: true
pad: 2
kernel_size: 3
group: 4
stride: 1
weight_filler {
type: "msra"
}
bias_filler {
type: "constant"
value: 0
}
dilation: 2
}
}
layer {
name: "res5a_branch2b/bn"
type: "BatchNorm"
bottom: "res5a_branch2b"
top: "res5a_branch2b"
batch_norm_param {
moving_average_fraction: 0.99
eps: 0.0001
scale_bias: true
}
}
layer {
name: "res5a_branch2b/relu"
type: "ReLU"
bottom: "res5a_branch2b"
top: "res5a_branch2b"
}
layer {
name: "out5a"
type: "Convolution"
bottom: "res5a_branch2b"
top: "out5a"
param {
lr_mult: 1
decay_mult: 1
}
param {
lr_mult: 2
decay_mult: 0
}
convolution_param {
num_output: 64
bias_term: true
pad: 4
kernel_size: 3
group: 2
stride: 1
weight_filler {
type: "msra"
}
bias_filler {
type: "constant"
value: 0
}
dilation: 4
}
}
layer {
name: "out5a/bn"
type: "BatchNorm"
bottom: "out5a"
top: "out5a"
batch_norm_param {
moving_average_fraction: 0.99
eps: 0.0001
scale_bias: true
}
}
layer {
name: "out5a/relu"
type: "ReLU"
bottom: "out5a"
top: "out5a"
}
layer {
name: "out5a_up2"
type: "Deconvolution"
bottom: "out5a"
top: "out5a_up2"
param {
lr_mult: 0
decay_mult: 0
}
convolution_param {
num_output: 64
bias_term: false
pad: 1
kernel_size: 4
group: 64
stride: 2
weight_filler {
type: "bilinear"
}
}
}
layer {
name: "out3a"
type: "Convolution"
bottom: "res3a_branch2b"
top: "out3a"}decay_mult: 1lr_mult: 1
param {
param {
lr_mult: 2
decay_mult: 0
}
convolution_param {
num_output: 64
bias_term: true
pad: 1
kernel_size: 3
group: 2
stride: 1
weight_filler {
type: "msra"
}
bias_filler {
type: "constant"
value: 0
}
dilation: 1
}
}
layer {
name: "out3a/bn"
type: "BatchNorm"
bottom: "out3a"
top: "out3a"
batch_norm_param {
moving_average_fraction: 0.99
eps: 0.0001
scale_bias: true
}
}
layer {
name: "out3a/relu"
type: "ReLU"
bottom: "out3a"
top: "out3a"
}
layer {
name: "out3_out5_combined"
type: "Eltwise"
bottom: "out5a_up2"
bottom: "out3a"
top: "out3_out5_combined"
}
layer {
name: "ctx_conv1"
type: "Convolution"
bottom: "out3_out5_combined"
top: "ctx_conv1"
param {
lr_mult: 1
decay_mult: 1
}
param {
lr_mult: 2
decay_mult: 0
}
convolution_param {
num_output: 64
bias_term: true
pad: 1
kernel_size: 3
group: 1
stride: 1
weight_filler {
type: "msra"
}
bias_filler {
type: "constant"
value: 0
}
dilation: 1
}
}
layer {
name: "ctx_conv1/bn"
type: "BatchNorm"
bottom: "ctx_conv1"
top: "ctx_conv1"
batch_norm_param {
moving_average_fraction: 0.99
eps: 0.0001
scale_bias: true
}
}
layer {
name: "ctx_conv1/relu"
type: "ReLU"
bottom: "ctx_conv1"
top: "ctx_conv1"
}
layer {
name: "ctx_conv2"
type: "Convolution"
bottom: "ctx_conv1"
top: "ctx_conv2"
param {
lr_mult: 1
decay_mult: 1
}
param {
lr_mult: 2
decay_mult: 0
}
convolution_param {
num_output: 64
bias_term: true
pad: 4
kernel_size: 3
group: 1
stride: 1
weight_filler {
type: "msra"
}
bias_filler {
type: "constant"
value: 0
}
dilation: 4
}
}
layer {
name: "ctx_conv2/bn"
type: "BatchNorm"
bottom: "ctx_conv2"
top: "ctx_conv2"
batch_norm_param {
moving_average_fraction: 0.99
eps: 0.0001
scale_bias: true
}
}
layer {
name: "ctx_conv2/relu"
type: "ReLU"
bottom: "ctx_conv2"
top: "ctx_conv2"
}
layer {
name: "ctx_conv3"
type: "Convolution"
bottom: "ctx_conv2"
top: "ctx_conv3"
param {
lr_mult: 1
decay_mult: 1
}
param {
lr_mult: 2
decay_mult: 0
}
convolution_param {
num_output: 64
bias_term: true
pad: 4
kernel_size: 3
group: 1
stride: 1
weight_filler {
type: "msra"
}
bias_filler {
type: "constant"
value: 0
}
dilation: 4
}
}
layer {
name: "ctx_conv3/bn"
type: "BatchNorm"
bottom: "ctx_conv3"
top: "ctx_conv3"
batch_norm_param {
moving_average_fraction: 0.99
eps: 0.0001
scale_bias: true
}
}
layer {
name: "ctx_conv3/relu"
type: "ReLU"
bottom: "ctx_conv3"
top: "ctx_conv3"
}
layer {
name: "ctx_conv4"
type: "Convolution"
bottom: "ctx_conv3"
top: "ctx_conv4"
param {
lr_mult: 1
decay_mult: 1
}
param {
lr_mult: 2
decay_mult: 0
}
convolution_param {
num_output: 64
bias_term: true
pad: 4
kernel_size: 3
group: 1
stride: 1
weight_filler {
type: "msra"
}
bias_filler {
type: "constant"
value: 0
}
dilation: 4
}
}
layer {
name: "ctx_conv4/bn"
type: "BatchNorm"
bottom: "ctx_conv4"
top: "ctx_conv4"
batch_norm_param {
moving_average_fraction: 0.99
eps: 0.0001
scale_bias: true
}
}
layer {
name: "ctx_conv4/relu"
type: "ReLU"
bottom: "ctx_conv4"
top: "ctx_conv4"
}
layer {
name: "ctx_final"
type: "Convolution"
bottom: "ctx_conv4"
top: "ctx_final"
param {
lr_mult: 1
decay_mult: 1
}
param {
lr_mult: 2
decay_mult: 0
}
convolution_param {
num_output: 8
bias_term: true
pad: 1
kernel_size: 3
kernel_size: 3
group: 1
stride: 1
weight_filler {
type: "msra"
}
bias_filler {
type: "constant"
value: 0
}
dilation: 1
}
}
layer {
name: "ctx_final/relu"
type: "ReLU"
bottom: "ctx_final"
top: "ctx_final"
}
layer {
name: "out_deconv_final_up2"
type: "Deconvolution"
bottom: "ctx_final"
top: "out_deconv_final_up2"
param {
lr_mult: 0
decay_mult: 0
}
convolution_param {
num_output: 8
bias_term: false
pad: 1
kernel_size: 4
group: 8
stride: 2
weight_filler {
type: "bilinear"
}
}
}
layer {
name: "out_deconv_final_up4"
type: "Deconvolution"
bottom: "out_deconv_final_up2"
top: "out_deconv_final_up4"}decay_mult: 0lr_mult: 0
param {
convolution_param {
num_output: 8
bias_term: false
pad: 1
kernel_size: 4
group: 8
stride: 2
weight_filler {
type: "bilinear"
}
}
}
layer {
name: "out_deconv_final_up8"
type: "Deconvolution"
bottom: "out_deconv_final_up4"
top: "out_deconv_final_up8"
param {
lr_mult: 0
decay_mult: 0
}
convolution_param {
num_output: 8
bias_term: false
pad: 1
kernel_size: 4
group: 8
stride: 2
weight_filler {
type: "bilinear"
}
}
}
layer {
name: "loss"
type: "SoftmaxWithLoss"
bottom: "out_deconv_final_up8"
bottom: "label"
top: "loss"
propagate_down: true
propagate_down: false
loss_param {
ignore_label: 255
normalization: VALID
}
}
I0817 11:30:08.122395 25692 net.cpp:110] Using FLOAT as default forward math type
I0817 11:30:08.122406 25692 net.cpp:116] Using FLOAT as default backward math type
I0817 11:30:08.122413 25692 layer_factory.hpp:172] Creating layer 'data' of type 'ImageLabelData'
I0817 11:30:08.122421 25692 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT
I0817 11:30:08.122437 25692 net.cpp:200] Created Layer data (0)
I0817 11:30:08.122442 25692 net.cpp:542] data -> data
I0817 11:30:08.122463 25692 net.cpp:542] data -> label
I0817 11:30:08.122532 25692 internal_thread.cpp:19] Starting 1 internal thread(s) on device 0
I0817 11:30:08.123039 25692 internal_thread.cpp:19] Starting 1 internal thread(s) on device 0
I0817 11:30:08.123042 25712 blocking_queue.cpp:40] Data layer prefetch queue empty
I0817 11:30:08.123098 25692 data_reader.cpp:58] Data Reader threads: 1, out queues: 1, depth: 4
I0817 11:30:08.123612 25692 internal_thread.cpp:19] Starting 1 internal thread(s) on device 0
I0817 11:30:08.124177 25714 db_lmdb.cpp:36] Opened lmdb data/train-image-lmdb
*** Aborted at 1597635008 (unix time) try "date -d @1597635008" if you are using GNU date ***
PC: @ 0x7fa63a96cecf caffe::C2TensorProtos::MergePartialFromCodedStream()
*** SIGSEGV (@0x0) received by PID 25692 (TID 0x7fa5a2263700) from PID 0; stack trace: ***
@ 0x7fa63844a4c0 (unknown)
@ 0x7fa63a96cecf caffe::C2TensorProtos::MergePartialFromCodedStream()
@ 0x7fa63985d892 google::protobuf::MessageLite::ParseFromArray()
@ 0x7fa63a9aa4e1 caffe::DataReader<>::CursorManager::fetch()
@ 0x7fa63a9b71f0 caffe::DataReader<>::InternalThreadEntryN()
@ 0x7fa63ae3f640 caffe::InternalThread::entry()
@ 0x7fa63ae415db boost::detail::thread_data<>::run()
@ 0x7fa62e8045d5 (unknown)
@ 0x7fa61780f6ba start_thread
@ 0x7fa63851c4dd clone