Git Product home page Git Product logo

deeplab-context's Introduction

DeepLab

Introduction

DeepLab is a state-of-art deep learning system for semantic image segmentation built on top of Caffe.

It combines densely-computed deep convolutional neural network (CNN) responses with densely connected conditional random fields (CRF).

This distribution provides a publicly available implementation for the key model ingredients first reported in an arXiv paper, accepted in revised form as conference publication to the ICLR-2015 conference. It also contains implementations for methods supporting model learning using only weakly labeled examples, described in a second follow-up arXiv paper. Please consult and consider citing the following papers:

@inproceedings{chen14semantic,
  title={Semantic Image Segmentation with Deep Convolutional Nets and Fully Connected CRFs},
  author={Liang-Chieh Chen and George Papandreou and Iasonas Kokkinos and Kevin Murphy and Alan L Yuille},
  booktitle={ICLR},
  url={http://arxiv.org/abs/1412.7062},
  year={2015}
}

@article{papandreou15weak,
  title={Weakly- and Semi-Supervised Learning of a DCNN for Semantic Image Segmentation},
  author={George Papandreou and Liang-Chieh Chen and Kevin Murphy and Alan L Yuille},
  journal={arxiv:1502.02734},
  year={2015}
}

Note that if you use the densecrf implementation, please consult and cite the following paper:

@inproceedings{KrahenbuhlK11,
  title={Efficient Inference in Fully Connected CRFs with Gaussian Edge Potentials},
  author={Philipp Kr{\"{a}}henb{\"{u}}hl and Vladlen Koltun},
  booktitle={NIPS},      
  year={2011}
}

Performance

DeepLab currently achieves 73.9% on the challenging PASCAL VOC 2012 image segmentation task -- see the leaderboard.

Pre-trained models

We have released several trained models and corresponding prototxt files at here. Please check it for more model details.

The best model among the released ones yields 73.6% on PASCAL VOC 2012 test set.

Python wrapper requirements

  1. Install wget library for python
sudo pip install wget
  1. Change DATA_ROOT to point to the PASCAL images

  2. To use the mat_read_layer and mat_write_layer, please download and install matio.

Running the code

python run.py

FAQ

Check FAQ if you have some problems while using the code.

deeplab-context's People

Contributors

gpapan avatar jay-lcchen avatar thelegendali avatar vittalp avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

deeplab-context's Issues

What do pointers to CAFFE_BIN and CAFFE_DIR mean?

In DeepLab-Context/python/my_script/tools.py lines 39-40

CAFFE_DIR='./'
CAFFE_BIN='.build_release/tools/caffe.bin'

Do they just point to where caffe is installed in the system, or is it to create local directories considering it is a part of mkdir function?

License info

Hi,
I have a dataset that is weakly annotated (i.e., bounding boxes) and was looking into trying to utilize this code to run the algorithm described in http://arxiv.org/abs/1502.02734, and had a couple questions:
1.) As someone who can't feasibly modify the C++ code that's being wrapped/interfaced with, is repurposing this code for datasets other than PASCAL/VOC doable? If so, are there any examples I could look at?
2.) What sort of license, if any, does this code fall under?

Thanks for any help, and for the python bindings to deeplab!

issue with matio

Hello @TheLegendAli,

While building caffe in DeepLab I am facing the following issue

ubuntu@ip-172-31-21-77:~/workspace/DeepLab-Context2/build$ cmake ..
-- Boost version: 1.54.0
-- Found the following Boost libraries:
-- system
-- thread
-- filesystem
-- Found gflags (include: /usr/include, library: /usr/lib/x86_64-linux-gnu/libgflags.so)
-- Found glog (include: /usr/include, library: /usr/lib/x86_64-linux-gnu/libglog.so)
-- Found PROTOBUF Compiler: /usr/bin/protoc
-- Found lmdb (include: /usr/include, library: /usr/lib/x86_64-linux-gnu/liblmdb.so)
-- Found LevelDB (include: /usr/include, library: /usr/lib/x86_64-linux-gnu/libleveldb.so)
-- Found Snappy (include: /usr/include, library: /usr/lib/libsnappy.so)
-- CUDA detected: 7.0
-- Found cuDNN: ver. 4.0.7 found (include: /usr/local/cuda-7.0/include, library: /usr/local/cuda-7.0/lib64/libcudnn.so)
-- Added CUDA NVCC flags for: sm_30
-- OpenCV found (/usr/local/share/OpenCV)
-- Found Atlas (include: /usr/include, library: /usr/lib/libatlas.so)
-- NumPy ver. 1.8.2 found (include: /usr/lib/python2.7/dist-packages/numpy/core/include)
-- Boost version: 1.54.0
-- Found the following Boost libraries:
-- python

-- Could NOT find Doxygen (missing: DOXYGEN_EXECUTABLE)

-- ******************* Caffe Configuration Summary *******************
-- General:
-- Version : 1.0.0-rc3
-- Git : 6abd64d-dirty
-- System : Linux
-- C++ compiler : /usr/bin/c++
-- Release CXX flags : -O3 -DNDEBUG -fPIC -Wall -Wno-sign-compare -Wno-uninitialized
-- Debug CXX flags : -g -fPIC -Wall -Wno-sign-compare -Wno-uninitialized

-- Build type : Release

-- BUILD_SHARED_LIBS : ON
-- BUILD_python : ON
-- BUILD_matlab : OFF
-- BUILD_docs : ON
-- CPU_ONLY : OFF
-- USE_OPENCV : ON
-- USE_LEVELDB : ON
-- USE_LMDB : ON

-- ALLOW_LMDB_NOLOCK : OFF

-- Dependencies:
-- BLAS : Yes (Atlas)
-- Boost : Yes (ver. 1.54)
-- glog : Yes
-- gflags : Yes
-- protobuf : Yes (ver. 2.5.0)
-- lmdb : Yes (ver. 0.9.10)
-- LevelDB : Yes (ver. 1.15)
-- Snappy : Yes (ver. 1.1.0)
-- OpenCV : Yes (ver. 3.1.0)

-- CUDA : Yes (ver. 7.0)

-- NVIDIA CUDA:
-- Target GPU(s) : Auto
-- GPU arch(s) : sm_30

-- cuDNN : Yes (ver. 4.0.7)

-- Python:
-- Interpreter : /usr/bin/python2.7 (ver. 2.7.6)
-- Libraries : /usr/lib/x86_64-linux-gnu/libpython2.7.so (ver 2.7.6)

-- NumPy : /usr/lib/python2.7/dist-packages/numpy/core/include (ver 1.8.2)

-- Documentaion:
-- Doxygen : No

-- config_file :

-- Install:

-- Install path : /home/ubuntu/workspace/DeepLab-Context2/build/install

-- Configuring done
-- Generating done
-- Build files have been written to: /home/ubuntu/workspace/DeepLab-Context2/build
ubuntu@ip-172-31-21-77:/workspace/DeepLab-Context2/build$ make -j32 all
[ 1%] Running C++/Python protocol buffer compiler on /home/ubuntu/workspace/DeepLab-Context2/src/caffe/proto/caffe.proto
Scanning dependencies of target proto
[ 1%] Building CXX object src/caffe/CMakeFiles/proto.dir///include/caffe/proto/caffe.pb.cc.o
Linking CXX static library ../../lib/libproto.a
[ 1%] Built target proto
[ 2%] [ 2%] [ 4%] [ 4%] [ 4%] [ 5%] [ 6%] [ 8%] [ 8%] [ 8%] [ 8%] [ 9%] [ 9%] [ 9%] [ 9%] [ 11%] [ 13%] [ 13%] [ 13%] [ 15%] [ 15%] [ 16%] [ 16%] [ 16%] [ 18%] [ 19%] [ 19%] [ 19%] [ 19%] [ 19%] [ 19%] [ 19%] Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/solvers/./cuda_compile_generated_sgd_solver.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_silence_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_embed_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_batch_reindex_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_softmax_loss_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_log_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_threshold_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_spatial_product_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_cudnn_sigmoid_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_bnll_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_sigmoid_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_inner_product_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_domain_transform_forward_only_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_absval_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_dropout_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_mvn_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_cudnn_tanh_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_pooling_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_domain_transform_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_cudnn_lcn_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_cudnn_pooling_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_cudnn_lrn_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_eltwise_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_tile_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_filter_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_deconv_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_power_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_reduction_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_concat_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_im2col_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_hdf5_output_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_relu_layer.cu.o
[ 19%] Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_hdf5_data_layer.cu.o
[ 19%] Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_euclidean_loss_layer.cu.o
[ 20%] Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_tanh_layer.cu.o
[ 20%] Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_sigmoid_cross_entropy_loss_layer.cu.o
[ 20%] Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_bias_layer.cu.o
[ 22%] Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_scale_layer.cu.o
[ 22%] Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_base_data_layer.cu.o
[ 23%] Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_contrastive_loss_layer.cu.o
[ 23%] [ 23%] [ 25%] Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_exp_layer.cu.o
[ 25%] Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_lrn_layer.cu.o
[ 25%] Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_cudnn_conv_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_elu_layer.cu.o
[ 26%] Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_slice_layer.cu.o
[ 26%] [ 27%] [ 27%] Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_conv_layer.cu.o
[ 27%] [ 29%] [ 29%] Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_split_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_prelu_layer.cu.o
[ 30%] [ 30%] Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_cudnn_relu_layer.cu.o
[ 30%] Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_softmax_layer.cu.o
[ 31%] [ 31%] Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_batch_norm_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/layers/./cuda_compile_generated_cudnn_softmax_layer.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/util/./cuda_compile_generated_math_functions.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/util/./cuda_compile_generated_im2col.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/util/./cuda_compile_generated_interp.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/solvers/./cuda_compile_generated_adadelta_solver.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/solvers/./cuda_compile_generated_adagrad_solver.cu.o
[ 31%] [ 33%] Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/solvers/./cuda_compile_generated_rmsprop_solver.cu.o
[ 33%] Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/solvers/./cuda_compile_generated_nesterov_solver.cu.o
Building NVCC (Device) object src/caffe/CMakeFiles/cuda_compile.dir/solvers/./cuda_compile_generated_adam_solver.cu.o
Scanning dependencies of target caffe
[ 33%] [ 33%] [ 34%] [ 34%] [ 34%] [ 36%] [ 37%] [ 37%] [ 37%] [ 38%] [ 38%] [ 38%] [ 40%] [ 40%] [ 40%] [ 43%] [ 43%] [ 43%] [ 44%] [ 45%] [ 47%] [ 47%] [ 47%] [ 47%] [ 48%] [ 50%] [ 50%] [ 50%] [ 50%] [ 50%] [ 50%] Building CXX object src/caffe/CMakeFiles/caffe.dir/common.cpp.o
[ 50%] Building CXX object src/caffe/CMakeFiles/caffe.dir/net.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layer_factory.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/blob.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/data_transformer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/mat_write_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/data_reader.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/infogain_loss_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/seg_accuracy_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/batch_norm_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/spp_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/window_data_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/elu_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/cudnn_relu_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/bias_channel_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/bnll_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/hdf5_output_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/cudnn_conv_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/base_conv_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/contrastive_loss_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/dummy_data_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/inner_product_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/silence_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/cudnn_pooling_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/multinomial_logistic_loss_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/pooling_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/euclidean_loss_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/tanh_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/mat_read_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/softmax_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/adaptive_bias_channel_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/split_layer.cpp.o
[ 50%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/accuracy_layer.cpp.o
[ 50%] [ 51%] [ 51%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/hinge_loss_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/mvn_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/reduction_layer.cpp.o
[ 51%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/dropout_layer.cpp.o
[ 52%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/threshold_layer.cpp.o
[ 52%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/scale_layer.cpp.o
[ 54%] [ 54%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/cudnn_softmax_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/eltwise_layer.cpp.o
[ 54%] [ 55%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/cudnn_tanh_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/data_layer.cpp.o
[ 55%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/sigmoid_layer.cpp.o
[ 56%] [ 56%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/memory_data_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/conv_layer.cpp.o
[ 56%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/exp_layer.cpp.o
[ 58%] [ 58%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/log_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/spatial_product_layer.cpp.o
[ 58%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/cudnn_lcn_layer.cpp.o
[ 59%] [ 59%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/batch_reindex_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/embed_layer.cpp.o
[ 61%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/densecrf_layer.cpp.o
[ 61%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/concat_layer.cpp.o
[ 61%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/domain_transform_forward_only_layer.cpp.o
[ 62%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/domain_transform_layer.cpp.o
[ 62%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/im2col_layer.cpp.o
[ 62%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/bias_layer.cpp.o
[ 63%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/filter_layer.cpp.o
[ 63%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/image_data_layer.cpp.o
[ 65%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/reshape_layer.cpp.o
[ 65%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/power_layer.cpp.o
[ 65%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/absval_layer.cpp.o
[ 66%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/argmax_layer.cpp.o
[ 66%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/cudnn_sigmoid_layer.cpp.o
[ 66%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/flatten_layer.cpp.o
[ 68%] [ 68%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/hdf5_data_layer.cpp.o
[ 69%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/cudnn_lrn_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/softmax_loss_layer.cpp.o
[ 69%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/image_seg_data_layer.cpp.o
[ 69%] [ 70%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/prelu_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/slice_layer.cpp.o
[ 70%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/sigmoid_cross_entropy_loss_layer.cpp.o
[ 70%] [ 72%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/relu_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/loss_layer.cpp.o
[ 72%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/deconv_layer.cpp.o
[ 73%] [ 73%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/interp_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/lrn_layer.cpp.o
[ 73%] [ 75%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/base_data_layer.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/tile_layer.cpp.o
[ 75%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/neuron_layer.cpp.o
[ 75%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/unique_label_layer.cpp.o
[ 76%] Building CXX object src/caffe/CMakeFiles/caffe.dir/util/matio_io.cpp.o
[ 76%] Building CXX object src/caffe/CMakeFiles/caffe.dir/util/confusion_matrix.cpp.o
[ 77%] Building CXX object src/caffe/CMakeFiles/caffe.dir/util/insert_splits.cpp.o
[ 77%] Building CXX object src/caffe/CMakeFiles/caffe.dir/util/permutohedral.cpp.o
[ 77%] [ 79%] Building CXX object src/caffe/CMakeFiles/caffe.dir/util/benchmark.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/util/db.cpp.o
[ 79%] Building CXX object src/caffe/CMakeFiles/caffe.dir/util/cudnn.cpp.o
[ 79%] Building CXX object src/caffe/CMakeFiles/caffe.dir/util/upgrade_proto.cpp.o
[ 80%] Building CXX object src/caffe/CMakeFiles/caffe.dir/util/io.cpp.o
[ 80%] Building CXX object src/caffe/CMakeFiles/caffe.dir/util/blocking_queue.cpp.o
[ 81%] Building CXX object src/caffe/CMakeFiles/caffe.dir/util/im2col.cpp.o
[ 81%] Building CXX object src/caffe/CMakeFiles/caffe.dir/util/db_lmdb.cpp.o
[ 81%] Building CXX object src/caffe/CMakeFiles/caffe.dir/util/db_leveldb.cpp.o
[ 83%] Building CXX object src/caffe/CMakeFiles/caffe.dir/util/interp.cpp.o
[ 83%] Building CXX object src/caffe/CMakeFiles/caffe.dir/util/densecrf_util.cpp.o
[ 84%] [ 84%] Building CXX object src/caffe/CMakeFiles/caffe.dir/util/densecrf_pairwise.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/util/math_functions.cpp.o
[ 84%] Building CXX object src/caffe/CMakeFiles/caffe.dir/util/signal_handler.cpp.o
[ 86%] Building CXX object src/caffe/CMakeFiles/caffe.dir/util/hdf5.cpp.o
[ 86%] Building CXX object src/caffe/CMakeFiles/caffe.dir/syncedmem.cpp.o
[ 86%] [ 87%] Building CXX object src/caffe/CMakeFiles/caffe.dir/internal_thread.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/layer.cpp.o
[ 87%] Building CXX object src/caffe/CMakeFiles/caffe.dir/parallel.cpp.o
[ 88%] Building CXX object src/caffe/CMakeFiles/caffe.dir/solver.cpp.o
[ 88%] Building CXX object src/caffe/CMakeFiles/caffe.dir/solvers/sgd_solver.cpp.o
[ 88%] [ 90%] [ 90%] Building CXX object src/caffe/CMakeFiles/caffe.dir/solvers/adam_solver.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/solvers/adagrad_solver.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/solvers/nesterov_solver.cpp.o
[ 90%] [ 91%] Building CXX object src/caffe/CMakeFiles/caffe.dir/solvers/adadelta_solver.cpp.o
Building CXX object src/caffe/CMakeFiles/caffe.dir/solvers/rmsprop_solver.cpp.o
Linking CXX shared library ../../lib/libcaffe.so
[ 91%] Built target caffe
Scanning dependencies of target CMakeCXXCompilerId
Scanning dependencies of target caffe.bin
Scanning dependencies of target convert_imageset
Scanning dependencies of target compute_image_mean
Scanning dependencies of target extract_features
Scanning dependencies of target device_query
Scanning dependencies of target test_net
Scanning dependencies of target finetune_net
Scanning dependencies of target upgrade_net_proto_binary
Scanning dependencies of target convert_mnist_data
Scanning dependencies of target upgrade_solver_proto_text
Scanning dependencies of target convert_mnist_siamese_data
Scanning dependencies of target train_net
Scanning dependencies of target classification
Scanning dependencies of target net_speed_benchmark
Scanning dependencies of target upgrade_net_proto_text
Scanning dependencies of target convert_cifar_data
Scanning dependencies of target pycaffe
[ 91%] [ 91%] Building CXX object tools/CMakeFiles/CMakeCXXCompilerId.dir/CMakeFiles/2.8.12.2/CompilerIdCXX/CMakeCXXCompilerId.cpp.o
Building CXX object examples/CMakeFiles/convert_mnist_data.dir/mnist/convert_mnist_data.cpp.o
Linking CXX executable CMakeCXXCompilerId
[ 93%] [ 93%] [ 94%] [ 94%] [ 94%] [ 94%] Building CXX object tools/CMakeFiles/device_query.dir/device_query.cpp.o
Building CXX object tools/CMakeFiles/compute_image_mean.dir/compute_image_mean.cpp.o
Building CXX object tools/CMakeFiles/convert_imageset.dir/convert_imageset.cpp.o
Building CXX object examples/CMakeFiles/convert_mnist_siamese_data.dir/siamese/convert_mnist_siamese_data.cpp.o
Building CXX object examples/CMakeFiles/convert_cifar_data.dir/cifar10/convert_cifar_data.cpp.o
[ 95%] [ 95%] Building CXX object tools/CMakeFiles/extract_features.dir/extract_features.cpp.o
[ 97%] [ 97%] [ 97%] [ 97%] [ 98%] [ 98%] Building CXX object tools/CMakeFiles/test_net.dir/test_net.cpp.o
Building CXX object tools/CMakeFiles/finetune_net.dir/finetune_net.cpp.o
Building CXX object tools/CMakeFiles/upgrade_net_proto_binary.dir/upgrade_net_proto_binary.cpp.o
Building CXX object tools/CMakeFiles/caffe.bin.dir/caffe.cpp.o
Building CXX object tools/CMakeFiles/train_net.dir/train_net.cpp.o
Building CXX object tools/CMakeFiles/upgrade_net_proto_text.dir/upgrade_net_proto_text.cpp.o
Building CXX object tools/CMakeFiles/net_speed_benchmark.dir/net_speed_benchmark.cpp.o
Building CXX object tools/CMakeFiles/upgrade_solver_proto_text.dir/upgrade_solver_proto_text.cpp.o
[ 98%] Built target CMakeCXXCompilerId
[100%] [100%] Building CXX object examples/CMakeFiles/classification.dir/cpp_classification/classification.cpp.o
Building CXX object python/CMakeFiles/pycaffe.dir/caffe/_caffe.cpp.o
Linking CXX executable mnist/convert_mnist_data
Linking CXX executable device_query
[100%] Built target device_query
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarCreate' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_CreateVer'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarWrite' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_VarFree'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarReadInfo' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_Close'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarReadDataLinear' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_Open'
collect2: error: ld returned 1 exit status
make[2]: *** [examples/mnist/convert_mnist_data] Error 1
make[1]: *** [examples/CMakeFiles/convert_mnist_data.dir/all] Error 2
make[1]: *** Waiting for unfinished jobs....
Linking CXX executable siamese/convert_mnist_siamese_data
Linking CXX executable cifar10/convert_cifar_data
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarCreate' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_CreateVer'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarWrite' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_VarFree'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarReadInfo' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_Close'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarReadDataLinear' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_Open'
collect2: error: ld returned 1 exit status
make[2]: *** [examples/siamese/convert_mnist_siamese_data] Error 1
make[1]: *** [examples/CMakeFiles/convert_mnist_siamese_data.dir/all] Error 2
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarCreate' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_CreateVer'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarWrite' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_VarFree'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarReadInfo' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_Close'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarReadDataLinear' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_Open'
collect2: error: ld returned 1 exit status
make[2]: *** [examples/cifar10/convert_cifar_data] Error 1
make[1]: *** [examples/CMakeFiles/convert_cifar_data.dir/all] Error 2
Linking CXX executable compute_image_mean
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarCreate' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_CreateVer'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarWrite' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_VarFree'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarReadInfo' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_Close'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarReadDataLinear' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_Open'
collect2: error: ld returned 1 exit status
make[2]: *** [tools/compute_image_mean] Error 1
make[1]: *** [tools/CMakeFiles/compute_image_mean.dir/all] Error 2
Linking CXX executable convert_imageset
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarCreate' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_CreateVer'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarWrite' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_VarFree'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarReadInfo' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_Close'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarReadDataLinear' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_Open'
collect2: error: ld returned 1 exit status
make[2]: *** [tools/convert_imageset] Error 1
make[1]: *** [tools/CMakeFiles/convert_imageset.dir/all] Error 2
Linking CXX executable test_net
Linking CXX executable train_net
Linking CXX executable net_speed_benchmark
Linking CXX executable upgrade_net_proto_text
[100%] Built target test_net
Linking CXX executable upgrade_solver_proto_text
[100%] Built target train_net
Linking CXX executable finetune_net
[100%] Built target net_speed_benchmark
Linking CXX executable upgrade_net_proto_binary
[100%] Built target finetune_net
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarCreate' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_CreateVer'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarWrite' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_VarFree'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarReadInfo' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_Close'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarReadDataLinear' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_Open'
collect2: error: ld returned 1 exit status
make[2]: *** [tools/upgrade_net_proto_text] Error 1
make[1]: *** [tools/CMakeFiles/upgrade_net_proto_text.dir/all] Error 2
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarCreate' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_CreateVer'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarWrite' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_VarFree'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarReadInfo' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_Close'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarReadDataLinear' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_Open'
collect2: error: ld returned 1 exit status
make[2]: *** [tools/upgrade_solver_proto_text] Error 1
make[1]: *** [tools/CMakeFiles/upgrade_solver_proto_text.dir/all] Error 2
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarCreate' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_CreateVer'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarWrite' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_VarFree'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarReadInfo' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_Close'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarReadDataLinear' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_Open'
collect2: error: ld returned 1 exit status
make[2]: *** [tools/upgrade_net_proto_binary] Error 1
make[1]: *** [tools/CMakeFiles/upgrade_net_proto_binary.dir/all] Error 2
Linking CXX executable extract_features
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarCreate' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_CreateVer'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarWrite' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_VarFree'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarReadInfo' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_Close'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarReadDataLinear' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_Open'
collect2: error: ld returned 1 exit status
make[2]: *** [tools/extract_features] Error 1
make[1]: *** [tools/CMakeFiles/extract_features.dir/all] Error 2
Linking CXX executable cpp_classification/classification
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarCreate' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_CreateVer'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarWrite' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_VarFree'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarReadInfo' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_Close'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarReadDataLinear' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_Open'
collect2: error: ld returned 1 exit status
make[2]: *** [examples/cpp_classification/classification] Error 1
make[1]: *** [examples/CMakeFiles/classification.dir/all] Error 2
Linking CXX executable caffe
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarCreate' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_CreateVer'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarWrite' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_VarFree'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarReadInfo' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_Close'
../lib/libcaffe.so.1.0.0-rc3: undefined reference to Mat_VarReadDataLinear' ../lib/libcaffe.so.1.0.0-rc3: undefined reference toMat_Open'
collect2: error: ld returned 1 exit status
make[2]: *** [tools/caffe] Error 1
make[1]: *** [tools/CMakeFiles/caffe.bin.dir/all] Error 2
Linking CXX shared library ../lib/_caffe.so
Creating symlink /home/ubuntu/workspace/DeepLab-Context2/python/caffe/_caffe.so -> /home/ubuntu/workspace/DeepLab-Context2/build/lib/_caffe.so
[100%] Built target pycaffe
make: *** [all] Error 2
ubuntu@ip-172-31-21-77:
/workspace/DeepLab-Context2/build$

I have installed matio, Could you please help me.

Documentation for DeepLab

Hi,

Is there a DeepLab documentation/layer catalogue available online for the new implemented layers?

DenseCRF compilation problem

Hi! First of all, thanks for your work! I have an issue when run make:

In file included from test_densecrf/simple_dense_inference.cpp:31:0: test_densecrf/../libDenseCRF/densecrf.h:116:23: warning: unused parameter ‘o’ [-Wunused-parameter] DenseCRF(DenseCRF & o) {} ^ In file included from test_densecrf/simple_dense_inference.cpp:31:0: test_densecrf/../libDenseCRF/densecrf.h:191:41: warning: unused parameter ‘o’ [-Wunused-parameter] BipartiteDenseCRF(BipartiteDenseCRF & o){} ^ test_densecrf/../libDenseCRF/densecrf.h:234:25: warning: unused parameter ‘filter’ [-Wunused-parameter] Filter( const Filter& filter ){} ^ In file included from test_densecrf/../libDenseCRF/util.h:31:0, from test_densecrf/simple_dense_inference.cpp:32: test_densecrf/../libDenseCRF/permutohedral.h: In member function ‘void HashTable::grow()’: test_densecrf/../libDenseCRF/permutohedral.h:83:20: warning: comparison between signed and unsigned integer expressions [-Wsign-compare] for( int i=0; i<old_capacity; i++ ) ^ make[1]: se sale del directorio '~/densecrf' make prog_refine_pascal_nat make[1]: se entra en el directorio '~/densecrf' g++ -w refine_pascal_nat/dense_inference.cpp -o prog_refine_pascal_nat -W -Wall -O2 -L. -lDenseCRF -lmatio -lhdf5 -fopenmp -lpthread -I./util/ make[1]: se sale del directorio '~/densecrf'
Sorry, I have the OS in spanish...do you know how I can fiix this? Thanks in advance!

DeepLabV2-ResNet101 - loss going up during training

I set up the DeepLabV2-ResNet101 model from the bitbucket code (so not using this python implementation) and during training I notice that the loss is going up. As I run this program for longer, the loss approaches 350. I was wondering if anyone would have any idea what could be the cause of this. Here is my run_pascal.sh script, I did not modify anything past the ## Training #1 (on train_aug) comment:

#/bin/sh
 
## MODIFY PATH for YOUR SETTING
ROOT_DIR=/deeplab
 
CAFFE_DIR=../code
CAFFE_BIN=${CAFFE_DIR}/.build_release/tools/caffe.bin
 
EXP=voc12
 
if [ "${EXP}" = "voc12" ]; then
    NUM_LABELS=21
    DATA_ROOT=${ROOT_DIR}/data/VOCdevkit/VOC2012
else
    NUM_LABELS=0
    echo "Wrong exp name"
fi
 
 
## Specify which model to train
########### voc12 ################
NET_ID=deeplabv2_resnet101
 
 
## Variables used for weakly or semi-supervisedly training
#TRAIN_SET_SUFFIX=
TRAIN_SET_SUFFIX=_aug
 
TRAIN_SET_STRONG=train
#TRAIN_SET_STRONG=train200
#TRAIN_SET_STRONG=train500
#TRAIN_SET_STRONG=train1000
#TRAIN_SET_STRONG=train750
 
TRAIN_SET_WEAK_LEN=0 #5000
 
DEV_ID=0
 
#####
 
## Create dirs
 
CONFIG_DIR=${EXP}/config/${NET_ID}
MODEL_DIR=${EXP}/model/${NET_ID}
mkdir -p ${MODEL_DIR}
LOG_DIR=${EXP}/log/${NET_ID}
mkdir -p ${LOG_DIR}
export GLOG_log_dir=${LOG_DIR}
 
## Run
 
RUN_TRAIN=1
RUN_TEST=0
RUN_TRAIN2=0
RUN_TEST2=0

Furthermore, I have the SegmentationClassAug ground truth images from your dropbox link (https://www.dropbox.com/s/oeu149j8qtbs1x0/SegmentationClassAug.zip?dl=0) in the data/VOCdevkit/VOC2012/ folder.

Here is what the training looks like:

I1127 01:09:18.232293 31928 net.cpp:270] This network produces output accuracy
I1127 01:09:18.232307 31928 net.cpp:270] This network produces output accuracy_res05
I1127 01:09:18.232322 31928 net.cpp:270] This network produces output accuracy_res075
I1127 01:09:18.232336 31928 net.cpp:270] This network produces output accuracy_res1
I1127 01:09:18.348408 31928 net.cpp:283] Network initialization done.
I1127 01:09:18.355520 31928 solver.cpp:60] Solver scaffolding done.
I1127 01:09:18.399452 31928 caffe.cpp:129] Finetuning from voc12/model/deeplabv2_resnet101/init.caffemodel
I1127 01:09:19.118443 31928 net.cpp:816] Ignoring source layer fc1_coco
I1127 01:09:19.118538 31928 net.cpp:816] Ignoring source layer fc1_coco_fc1_coco_0_split
I1127 01:09:19.132285 31928 caffe.cpp:219] Starting Optimization
I1127 01:09:19.132372 31928 solver.cpp:280] Solving deeplabv2_resnet101
I1127 01:09:19.132383 31928 solver.cpp:281] Learning Rate Policy: poly
I1127 01:09:25.117411 31928 solver.cpp:229] Iteration 0, loss = 261.803
I1127 01:09:25.117547 31928 solver.cpp:245]     Train net output #0: accuracy = 0.0291829
I1127 01:09:25.117589 31928 solver.cpp:245]     Train net output #1: accuracy = 0.0872093
I1127 01:09:25.117614 31928 solver.cpp:245]     Train net output #2: accuracy = 0.66929
I1127 01:09:25.117645 31928 solver.cpp:245]     Train net output #3: accuracy_res05 = 0.046683
I1127 01:09:25.117672 31928 solver.cpp:245]     Train net output #4: accuracy_res05 = 0.140741
I1127 01:09:25.117697 31928 solver.cpp:245]     Train net output #5: accuracy_res05 = 0.669494
I1127 01:09:25.117720 31928 solver.cpp:245]     Train net output #6: accuracy_res075 = 0.035668
I1127 01:09:25.117766 31928 solver.cpp:245]     Train net output #7: accuracy_res075 = 0.106589
I1127 01:09:25.117786 31928 solver.cpp:245]     Train net output #8: accuracy_res075 = 0.669965
I1127 01:09:25.117815 31928 solver.cpp:245]     Train net output #9: accuracy_res1 = 0.0486381
I1127 01:09:25.117841 31928 solver.cpp:245]     Train net output #10: accuracy_res1 = 0.145349
I1127 01:09:25.117866 31928 solver.cpp:245]     Train net output #11: accuracy_res1 = 0.627653
I1127 01:09:25.117985 31928 sgd_solver.cpp:106] Iteration 0, lr = 0.00025
I1127 01:10:54.876021 31928 solver.cpp:229] Iteration 20, loss = 336.749
I1127 01:10:54.877192 31928 solver.cpp:245]     Train net output #0: accuracy = 0
I1127 01:10:54.877221 31928 solver.cpp:245]     Train net output #1: accuracy = 0
I1127 01:10:54.877240 31928 solver.cpp:245]     Train net output #2: accuracy = 0.857143
I1127 01:10:54.877261 31928 solver.cpp:245]     Train net output #3: accuracy_res05 = 0
I1127 01:10:54.877290 31928 solver.cpp:245]     Train net output #4: accuracy_res05 = 0
I1127 01:10:54.877307 31928 solver.cpp:245]     Train net output #5: accuracy_res05 = 0.857143
I1127 01:10:54.877341 31928 solver.cpp:245]     Train net output #6: accuracy_res075 = 0
I1127 01:10:54.877368 31928 solver.cpp:245]     Train net output #7: accuracy_res075 = 0
I1127 01:10:54.877393 31928 solver.cpp:245]     Train net output #8: accuracy_res075 = 0.857143
I1127 01:10:54.877419 31928 solver.cpp:245]     Train net output #9: accuracy_res1 = 0
I1127 01:10:54.877435 31928 solver.cpp:245]     Train net output #10: accuracy_res1 = 0
I1127 01:10:54.877454 31928 solver.cpp:245]     Train net output #11: accuracy_res1 = 0.857143
I1127 01:10:54.877490 31928 sgd_solver.cpp:106] Iteration 20, lr = 0.000249775

Code about Attention to Scale

Hi , I am interesting in the paper "Attention to Scale: Scale-aware Semantic Image Segmentation",you have given the address of the code ,but I can't find it . Would you please give me some help?
Thanks a lot

How to specify the validation data during training and output test accuracy

Hello,
There is no test net accuracy in the log file even I specified the test data layer in the train.prototxt file:

layers {
  name: "data"
  type: IMAGE_SEG_DATA
  top: "data"
  top: "label"
  image_data_param {
    root_folder: ""
    source: "./list/test_aug.txt"
    label_type: PIXEL
    batch_size: 100
  }
  transform_param {
    mean_value: 75.209
    mean_value: 85.950
    mean_value: 95.685
    crop_size: 663
    mirror: false
  }
  include: { phase: TEST }
}

Only loss and train net accuracy in the log file, such as:

I0424 09:32:20.303715 32660 solver.cpp:209] Iteration 50, loss = 0.270298
I0424 09:32:20.303766 32660 solver.cpp:224]     Train net output #0: accuracy = 0.883158
I0424 09:32:20.303774 32660 solver.cpp:224]     Train net output #1: accuracy = 0.780627
I0424 09:32:20.303809 32660 solver.cpp:224]     Train net output #2: accuracy = 0.70205

How do I get the test accuracy during training?
What the meaning of Train net output #0, Train net output "#1", Train net output "#2"? "#0", "#1", "#2" means something?

Thanks for any suggestion!

Did anyone meet this question? What should I do?

Training1 net voc12/vgg128_noup
Running C:/Users/v-kexzha/caffe-deeplab-ckc/caffe/bin/caffe train --solver=voc12/config/vgg128_noup/solver_train_aug.pro
totxt --weights=voc12/model/vgg128_noup/init.caffemodel --gpu=0
I0223 14:17:55.440616 40844 caffe.cpp:135] Use GPU with device ID 0
I0223 14:17:56.575695 40844 common.cpp:22] System entropy source not available, using fallback algorithm to generate see
d instead.
I0223 14:17:56.576696 40844 caffe.cpp:144] Starting Optimization
I0223 14:17:56.576696 40844 solver.cpp:32] Initializing solver from parameters:
train_net: "voc12/config/vgg128_noup/train_train_aug.prototxt"
base_lr: 0.001
display: 10
max_iter: 6000
lr_policy: "step"
gamma: 0.1
momentum: 0.9
weight_decay: 0.0005
stepsize: 2000
snapshot: 2000
snapshot_prefix: "voc12/model/vgg128_noup/train"
solver_mode: GPU
I0223 14:17:56.577698 40844 solver.cpp:58] Creating training net from train_net file: voc12/config/vgg128_noup/train_tra
in_aug.prototxt
I0223 14:17:56.579700 40844 net.cpp:39] Initializing net from parameters:
name: "vgg128_noup"
layers {
top: "data"
top: "label"
name: "data"
type: IMAGE_SEG_DATA
image_data_param {
source: "voc12/list/train_aug.txt"
batch_size: 30
shuffle: true
root_folder: "D:/v-kexzha/voc12/VOCdevkit/VOC2012"
label_type: PIXEL
}
include {
phase: TRAIN
}
transform_param {
mirror: true
crop_size: 321
mean_value: 104.008
mean_value: 116.669
mean_value: 122.675
}
}
layers {
bottom: "data"
top: "conv1_1"
name: "conv1_1"
type: CONVOLUTION
blobs_lr: 1
blobs_lr: 2
weight_decay: 1
weight_decay: 0
convolution_param {
num_output: 64
pad: 1
kernel_size: 3
}
}
layers {
bottom: "conv1_1"
top: "conv1_1"
name: "relu1_1"
type: RELU
}
layers {
bottom: "conv1_1"
top: "conv1_2"
name: "conv1_2"
type: CONVOLUTION
blobs_lr: 1
blobs_lr: 2
weight_decay: 1
weight_decay: 0
convolution_param {
num_output: 64
pad: 1
kernel_size: 3
}
}
layers {
bottom: "conv1_2"
top: "conv1_2"
name: "relu1_2"
type: RELU
}
layers {
bottom: "conv1_2"
top: "pool1"
name: "pool1"
type: POOLING
pooling_param {
pool: MAX
kernel_size: 3
stride: 2
pad: 1
}
}
layers {
bottom: "pool1"
top: "conv2_1"
name: "conv2_1"
type: CONVOLUTION
blobs_lr: 1
blobs_lr: 2
weight_decay: 1
weight_decay: 0
convolution_param {
num_output: 128
pad: 1
kernel_size: 3
}
}
layers {
bottom: "conv2_1"
top: "conv2_1"
name: "relu2_1"
type: RELU
}
layers {
bottom: "conv2_1"
top: "conv2_2"
name: "conv2_2"
type: CONVOLUTION
blobs_lr: 1
blobs_lr: 2
weight_decay: 1
weight_decay: 0
convolution_param {
num_output: 128
pad: 1
kernel_size: 3
}
}
layers {
bottom: "conv2_2"
top: "conv2_2"
name: "relu2_2"
type: RELU
}
layers {
bottom: "conv2_2"
top: "pool2"
name: "pool2"
type: POOLING
pooling_param {
pool: MAX
kernel_size: 3
stride: 2
pad: 1
}
}
layers {
bottom: "pool2"
top: "conv3_1"
name: "conv3_1"
type: CONVOLUTION
blobs_lr: 1
blobs_lr: 2
weight_decay: 1
weight_decay: 0
convolution_param {
num_output: 256
pad: 1
kernel_size: 3
}
}
layers {
bottom: "conv3_1"
top: "conv3_1"
name: "relu3_1"
type: RELU
}
layers {
bottom: "conv3_1"
top: "conv3_2"
name: "conv3_2"
type: CONVOLUTION
blobs_lr: 1
blobs_lr: 2
weight_decay: 1
weight_decay: 0
convolution_param {
num_output: 256
pad: 1
kernel_size: 3
}
}
layers {
bottom: "conv3_2"
top: "conv3_2"
name: "relu3_2"
type: RELU
}
layers {
bottom: "conv3_2"
top: "conv3_3"
name: "conv3_3"
type: CONVOLUTION
blobs_lr: 1
blobs_lr: 2
weight_decay: 1
weight_decay: 0
convolution_param {
num_output: 256
pad: 1
kernel_size: 3
}
}
layers {
bottom: "conv3_3"
top: "conv3_3"
name: "relu3_3"
type: RELU
}
layers {
bottom: "conv3_3"
top: "pool3"
name: "pool3"
type: POOLING
pooling_param {
pool: MAX
kernel_size: 3
stride: 2
pad: 1
}
}
layers {
bottom: "pool3"
top: "conv4_1"
name: "conv4_1"
type: CONVOLUTION
blobs_lr: 1
blobs_lr: 2
weight_decay: 1
weight_decay: 0
convolution_param {
num_output: 512
pad: 1
kernel_size: 3
}
}
layers {
bottom: "conv4_1"
top: "conv4_1"
name: "relu4_1"
type: RELU
}
layers {
bottom: "conv4_1"
top: "conv4_2"
name: "conv4_2"
type: CONVOLUTION
blobs_lr: 1
blobs_lr: 2
weight_decay: 1
weight_decay: 0
convolution_param {
num_output: 512
pad: 1
kernel_size: 3
}
}
layers {
bottom: "conv4_2"
top: "conv4_2"
name: "relu4_2"
type: RELU
}
layers {
bottom: "conv4_2"
top: "conv4_3"
name: "conv4_3"
type: CONVOLUTION
blobs_lr: 1
blobs_lr: 2
weight_decay: 1
weight_decay: 0
convolution_param {
num_output: 512
pad: 1
kernel_size: 3
}
}
layers {
bottom: "conv4_3"
top: "conv4_3"
name: "relu4_3"
type: RELU
}
layers {
bottom: "conv4_3"
top: "pool4"
name: "pool4"
type: POOLING
pooling_param {
pool: MAX
kernel_size: 3
stride: 1
pad: 1
}
}
layers {
bottom: "pool4"
top: "conv5_1"
name: "conv5_1"
type: CONVOLUTION
blobs_lr: 1
blobs_lr: 2
weight_decay: 1
weight_decay: 0
convolution_param {
num_output: 512
pad: 2
kernel_size: 3
hole: 2
}
}
layers {
bottom: "conv5_1"
top: "conv5_1"
name: "relu5_1"
type: RELU
}
layers {
bottom: "conv5_1"
top: "conv5_2"
name: "conv5_2"
type: CONVOLUTION
blobs_lr: 1
blobs_lr: 2
weight_decay: 1
weight_decay: 0
convolution_param {
num_output: 512
pad: 2
kernel_size: 3
hole: 2
}
}
layers {
bottom: "conv5_2"
top: "conv5_2"
name: "relu5_2"
type: RELU
}
layers {
bottom: "conv5_2"
top: "conv5_3"
name: "conv5_3"
type: CONVOLUTION
blobs_lr: 1
blobs_lr: 2
weight_decay: 1
weight_decay: 0
convolution_param {
num_output: 512
pad: 2
kernel_size: 3
hole: 2
}
}
layers {
bottom: "conv5_3"
top: "conv5_3"
name: "relu5_3"
type: RELU
}
layers {
bottom: "conv5_3"
top: "pool5"
name: "pool5"
type: POOLING
pooling_param {
pool: MAX
kernel_size: 3
stride: 1
pad: 1
}
}
layers {
bottom: "pool5"
top: "pool5a"
name: "pool5a"
type: POOLING
pooling_param {
pool: AVE
kernel_size: 3
stride: 1
pad: 1
}
}
layers {
bottom: "pool5a"
top: "fc6"
name: "fc6"
type: CONVOLUTION
blobs_lr: 1
blobs_lr: 2
weight_decay: 1
weight_decay: 0
convolution_param {
num_output: 1024
pad: 12
kernel_size: 3
hole: 12
}
}
layers {
bottom: "fc6"
top: "fc6"
name: "relu6"
type: RELU
}
layers {
bottom: "fc6"
top: "fc6"
name: "drop6"
type: DROPOUT
dropout_param {
dropout_ratio: 0.5
}
}
layers {
bottom: "fc6"
top: "fc7"
name: "fc7"
type: CONVOLUTION
blobs_lr: 1
blobs_lr: 2
weight_decay: 1
weight_decay: 0
convolution_param {
num_output: 1024
kernel_size: 1
}
}
layers {
bottom: "fc7"
top: "fc7"
name: "relu7"
type: RELU
}
layers {
bottom: "fc7"
top: "fc7"
name: "drop7"
type: DROPOUT
dropout_param {
dropout_ratio: 0.5
}
}
layers {
bottom: "fc7"
top: "fc8_voc12"
name: "fc8_voc12"
type: CONVOLUTION
blobs_lr: 10
blobs_lr: 20
weight_decay: 1
weight_decay: 0
convolution_param {
num_output: 21
kernel_size: 1
weight_filler {
type: "gaussian"
std: 0.01
}
bias_filler {
type: "constant"
value: 0
}
}
}
layers {
bottom: "label"
top: "label_shrink"
name: "label_shrink"
type: INTERP
interp_param {
shrink_factor: 8
pad_beg: 0
pad_end: 0
}
}
layers {
bottom: "fc8_voc12"
bottom: "label_shrink"
name: "loss"
type: SOFTMAX_LOSS
include {
phase: TRAIN
}
softmaxloss_param {
ignore_label: 255
}
}
layers {
bottom: "fc8_voc12"
bottom: "label_shrink"
top: "accuracy"
name: "accuracy"
type: SEG_ACCURACY
seg_accuracy_param {
ignore_label: 255
}
}
state {
phase: TRAIN
}
I0223 14:17:56.708823 40844 layer_factory.hpp:78] Creating layer data
I0223 14:17:56.708823 40844 net.cpp:67] Creating Layer data
I0223 14:17:56.709825 40844 net.cpp:356] data -> data
I0223 14:17:56.710824 40844 net.cpp:356] data -> label
I0223 14:17:56.710824 40844 net.cpp:356] data -> (automatic)
I0223 14:17:56.711825 40844 net.cpp:96] Setting up data
I0223 14:17:56.712826 40844 image_seg_data_layer.cpp:45] Opening file voc12/list/train_aug.txt
I0223 14:17:56.772883 40844 image_seg_data_layer.cpp:62] Shuffling data
I0223 14:17:56.773885 40844 common.cpp:22] System entropy source not available, using fallback algorithm to generate see
d instead.
I0223 14:17:56.775887 40844 image_seg_data_layer.cpp:67] A total of 10582 images.
I0223 14:17:56.788899 40844 image_seg_data_layer.cpp:113] output data size: 30,3,321,321
I0223 14:17:56.789899 40844 image_seg_data_layer.cpp:117] output label size: 30,1,321,321
I0223 14:17:56.790902 40844 image_seg_data_layer.cpp:121] output data_dim size: 30,1,1,2
I0223 14:17:56.818928 40844 net.cpp:103] Top shape: 30 3 321 321 (9273690)
I0223 14:17:56.819928 40844 net.cpp:103] Top shape: 30 1 321 321 (3091230)
I0223 14:17:56.820930 40844 net.cpp:103] Top shape: 30 1 1 2 (60)
I0223 14:17:56.821930 40844 layer_factory.hpp:78] Creating layer conv1_1
I0223 14:17:56.821930 40844 net.cpp:67] Creating Layer conv1_1
I0223 14:17:56.823935 40844 net.cpp:394] conv1_1 <- data
I0223 14:17:56.825937 40844 net.cpp:356] conv1_1 -> conv1_1
I0223 14:17:56.825937 40844 net.cpp:96] Setting up conv1_1
I0223 14:17:56.826934 40844 net.cpp:103] Top shape: 30 64 321 321 (197838720)
I0223 14:17:56.827937 40844 layer_factory.hpp:78] Creating layer relu1_1
I0223 14:17:56.828938 40844 net.cpp:67] Creating Layer relu1_1
I0223 14:17:56.829938 40844 net.cpp:394] relu1_1 <- conv1_1
I0223 14:17:56.830940 40844 net.cpp:345] relu1_1 -> conv1_1 (in-place)
I0223 14:17:56.830940 40844 net.cpp:96] Setting up relu1_1
I0223 14:17:56.831948 40844 net.cpp:103] Top shape: 30 64 321 321 (197838720)
I0223 14:17:56.832940 40844 layer_factory.hpp:78] Creating layer conv1_2
I0223 14:17:56.833941 40844 net.cpp:67] Creating Layer conv1_2
I0223 14:17:56.834944 40844 net.cpp:394] conv1_2 <- conv1_1
I0223 14:17:56.835944 40844 net.cpp:356] conv1_2 -> conv1_2
I0223 14:17:56.835944 40844 net.cpp:96] Setting up conv1_2
I0223 14:17:56.837946 40844 net.cpp:103] Top shape: 30 64 321 321 (197838720)
I0223 14:17:56.837946 40844 layer_factory.hpp:78] Creating layer relu1_2
I0223 14:17:56.838946 40844 net.cpp:67] Creating Layer relu1_2
I0223 14:17:56.839947 40844 net.cpp:394] relu1_2 <- conv1_2
I0223 14:17:56.840948 40844 net.cpp:345] relu1_2 -> conv1_2 (in-place)
I0223 14:17:56.841950 40844 net.cpp:96] Setting up relu1_2
I0223 14:17:56.842952 40844 net.cpp:103] Top shape: 30 64 321 321 (197838720)
I0223 14:17:56.843952 40844 layer_factory.hpp:78] Creating layer pool1
I0223 14:17:56.844952 40844 net.cpp:67] Creating Layer pool1
I0223 14:17:56.844952 40844 net.cpp:394] pool1 <- conv1_2
I0223 14:17:56.845952 40844 net.cpp:356] pool1 -> pool1
I0223 14:17:56.846953 40844 net.cpp:96] Setting up pool1
I0223 14:17:56.846953 40844 net.cpp:103] Top shape: 30 64 161 161 (49768320)
I0223 14:17:56.847955 40844 layer_factory.hpp:78] Creating layer conv2_1
I0223 14:17:56.848961 40844 net.cpp:67] Creating Layer conv2_1
I0223 14:17:56.848961 40844 net.cpp:394] conv2_1 <- pool1
I0223 14:17:56.849959 40844 net.cpp:356] conv2_1 -> conv2_1
I0223 14:17:56.850958 40844 net.cpp:96] Setting up conv2_1
I0223 14:17:56.851959 40844 net.cpp:103] Top shape: 30 128 161 161 (99536640)
I0223 14:17:56.852960 40844 layer_factory.hpp:78] Creating layer relu2_1
I0223 14:17:56.854961 40844 net.cpp:67] Creating Layer relu2_1
I0223 14:17:56.855963 40844 net.cpp:394] relu2_1 <- conv2_1
I0223 14:17:56.856963 40844 net.cpp:345] relu2_1 -> conv2_1 (in-place)
I0223 14:17:56.857964 40844 net.cpp:96] Setting up relu2_1
I0223 14:17:56.858965 40844 net.cpp:103] Top shape: 30 128 161 161 (99536640)
I0223 14:17:56.859971 40844 layer_factory.hpp:78] Creating layer conv2_2
I0223 14:17:56.860968 40844 net.cpp:67] Creating Layer conv2_2
I0223 14:17:56.861968 40844 net.cpp:394] conv2_2 <- conv2_1
I0223 14:17:56.862969 40844 net.cpp:356] conv2_2 -> conv2_2
I0223 14:17:56.862969 40844 net.cpp:96] Setting up conv2_2
I0223 14:17:56.864971 40844 net.cpp:103] Top shape: 30 128 161 161 (99536640)
I0223 14:17:56.864971 40844 layer_factory.hpp:78] Creating layer relu2_2
I0223 14:17:56.865972 40844 net.cpp:67] Creating Layer relu2_2
I0223 14:17:56.866973 40844 net.cpp:394] relu2_2 <- conv2_2
I0223 14:17:56.867974 40844 net.cpp:345] relu2_2 -> conv2_2 (in-place)
I0223 14:17:56.868975 40844 net.cpp:96] Setting up relu2_2
I0223 14:17:56.869977 40844 net.cpp:103] Top shape: 30 128 161 161 (99536640)
I0223 14:17:56.870978 40844 layer_factory.hpp:78] Creating layer pool2
I0223 14:17:56.871978 40844 net.cpp:67] Creating Layer pool2
I0223 14:17:56.871978 40844 net.cpp:394] pool2 <- conv2_2
I0223 14:17:56.872978 40844 net.cpp:356] pool2 -> pool2
I0223 14:17:56.873980 40844 net.cpp:96] Setting up pool2
I0223 14:17:56.874982 40844 net.cpp:103] Top shape: 30 128 81 81 (25194240)
I0223 14:17:56.875984 40844 layer_factory.hpp:78] Creating layer conv3_1
I0223 14:17:56.876982 40844 net.cpp:67] Creating Layer conv3_1
I0223 14:17:56.877987 40844 net.cpp:394] conv3_1 <- pool2
I0223 14:17:56.877987 40844 net.cpp:356] conv3_1 -> conv3_1
I0223 14:17:56.878985 40844 net.cpp:96] Setting up conv3_1
I0223 14:17:56.880987 40844 net.cpp:103] Top shape: 30 256 81 81 (50388480)
I0223 14:17:56.880987 40844 layer_factory.hpp:78] Creating layer relu3_1
I0223 14:17:56.881988 40844 net.cpp:67] Creating Layer relu3_1
I0223 14:17:56.886992 40844 net.cpp:394] relu3_1 <- conv3_1
I0223 14:17:56.888994 40844 net.cpp:345] relu3_1 -> conv3_1 (in-place)
I0223 14:17:56.889996 40844 net.cpp:96] Setting up relu3_1
I0223 14:17:56.890996 40844 net.cpp:103] Top shape: 30 256 81 81 (50388480)
I0223 14:17:56.891997 40844 layer_factory.hpp:78] Creating layer conv3_2
I0223 14:17:56.892998 40844 net.cpp:67] Creating Layer conv3_2
I0223 14:17:56.893998 40844 net.cpp:394] conv3_2 <- conv3_1
I0223 14:17:56.895000 40844 net.cpp:356] conv3_2 -> conv3_2
I0223 14:17:56.896000 40844 net.cpp:96] Setting up conv3_2
I0223 14:17:56.898002 40844 net.cpp:103] Top shape: 30 256 81 81 (50388480)
I0223 14:17:56.899003 40844 layer_factory.hpp:78] Creating layer relu3_2
I0223 14:17:56.900005 40844 net.cpp:67] Creating Layer relu3_2
I0223 14:17:56.901006 40844 net.cpp:394] relu3_2 <- conv3_2
I0223 14:17:56.902006 40844 net.cpp:345] relu3_2 -> conv3_2 (in-place)
I0223 14:17:56.903007 40844 net.cpp:96] Setting up relu3_2
I0223 14:17:56.904008 40844 net.cpp:103] Top shape: 30 256 81 81 (50388480)
I0223 14:17:56.905009 40844 layer_factory.hpp:78] Creating layer conv3_3
I0223 14:17:56.906010 40844 net.cpp:67] Creating Layer conv3_3
I0223 14:17:56.907011 40844 net.cpp:394] conv3_3 <- conv3_2
I0223 14:17:56.908012 40844 net.cpp:356] conv3_3 -> conv3_3
I0223 14:17:56.909013 40844 net.cpp:96] Setting up conv3_3
I0223 14:17:56.911015 40844 net.cpp:103] Top shape: 30 256 81 81 (50388480)
I0223 14:17:56.912015 40844 layer_factory.hpp:78] Creating layer relu3_3
I0223 14:17:56.913017 40844 net.cpp:67] Creating Layer relu3_3
I0223 14:17:56.915020 40844 net.cpp:394] relu3_3 <- conv3_3
I0223 14:17:56.916020 40844 net.cpp:345] relu3_3 -> conv3_3 (in-place)
I0223 14:17:56.917021 40844 net.cpp:96] Setting up relu3_3
I0223 14:17:56.918021 40844 net.cpp:103] Top shape: 30 256 81 81 (50388480)
I0223 14:17:56.919023 40844 layer_factory.hpp:78] Creating layer pool3
I0223 14:17:56.919023 40844 net.cpp:67] Creating Layer pool3
I0223 14:17:56.920024 40844 net.cpp:394] pool3 <- conv3_3
I0223 14:17:56.921025 40844 net.cpp:356] pool3 -> pool3
I0223 14:17:56.922025 40844 net.cpp:96] Setting up pool3
I0223 14:17:56.923027 40844 net.cpp:103] Top shape: 30 256 41 41 (12910080)
I0223 14:17:56.924027 40844 layer_factory.hpp:78] Creating layer conv4_1
I0223 14:17:56.924027 40844 net.cpp:67] Creating Layer conv4_1
I0223 14:17:56.925029 40844 net.cpp:394] conv4_1 <- pool3
I0223 14:17:56.926029 40844 net.cpp:356] conv4_1 -> conv4_1
I0223 14:17:56.926029 40844 net.cpp:96] Setting up conv4_1
I0223 14:17:56.931033 40844 net.cpp:103] Top shape: 30 512 41 41 (25820160)
I0223 14:17:56.932035 40844 layer_factory.hpp:78] Creating layer relu4_1
I0223 14:17:56.933037 40844 net.cpp:67] Creating Layer relu4_1
I0223 14:17:56.934037 40844 net.cpp:394] relu4_1 <- conv4_1
I0223 14:17:56.934037 40844 net.cpp:345] relu4_1 -> conv4_1 (in-place)
I0223 14:17:56.935037 40844 net.cpp:96] Setting up relu4_1
I0223 14:17:56.936038 40844 net.cpp:103] Top shape: 30 512 41 41 (25820160)
I0223 14:17:56.936038 40844 layer_factory.hpp:78] Creating layer conv4_2
I0223 14:17:56.937039 40844 net.cpp:67] Creating Layer conv4_2
I0223 14:17:56.938040 40844 net.cpp:394] conv4_2 <- conv4_1
I0223 14:17:56.939041 40844 net.cpp:356] conv4_2 -> conv4_2
I0223 14:17:56.940042 40844 net.cpp:96] Setting up conv4_2
I0223 14:17:56.949051 40844 net.cpp:103] Top shape: 30 512 41 41 (25820160)
I0223 14:17:56.950052 40844 layer_factory.hpp:78] Creating layer relu4_2
I0223 14:17:56.951053 40844 net.cpp:67] Creating Layer relu4_2
I0223 14:17:56.952054 40844 net.cpp:394] relu4_2 <- conv4_2
I0223 14:17:56.953054 40844 net.cpp:345] relu4_2 -> conv4_2 (in-place)
I0223 14:17:56.954056 40844 net.cpp:96] Setting up relu4_2
I0223 14:17:56.955056 40844 net.cpp:103] Top shape: 30 512 41 41 (25820160)
I0223 14:17:56.956058 40844 layer_factory.hpp:78] Creating layer conv4_3
I0223 14:17:56.957059 40844 net.cpp:67] Creating Layer conv4_3
I0223 14:17:56.958060 40844 net.cpp:394] conv4_3 <- conv4_2
I0223 14:17:56.959060 40844 net.cpp:356] conv4_3 -> conv4_3
I0223 14:17:56.960062 40844 net.cpp:96] Setting up conv4_3
I0223 14:17:56.969070 40844 net.cpp:103] Top shape: 30 512 41 41 (25820160)
I0223 14:17:56.970072 40844 layer_factory.hpp:78] Creating layer relu4_3
I0223 14:17:56.971073 40844 net.cpp:67] Creating Layer relu4_3
I0223 14:17:56.972074 40844 net.cpp:394] relu4_3 <- conv4_3
I0223 14:17:56.973075 40844 net.cpp:345] relu4_3 -> conv4_3 (in-place)
I0223 14:17:56.975075 40844 net.cpp:96] Setting up relu4_3
I0223 14:17:56.976078 40844 net.cpp:103] Top shape: 30 512 41 41 (25820160)
I0223 14:17:56.977078 40844 layer_factory.hpp:78] Creating layer pool4
I0223 14:17:56.978080 40844 net.cpp:67] Creating Layer pool4
I0223 14:17:56.978080 40844 net.cpp:394] pool4 <- conv4_3
I0223 14:17:56.979080 40844 net.cpp:356] pool4 -> pool4
I0223 14:17:56.981081 40844 net.cpp:96] Setting up pool4
I0223 14:17:56.982082 40844 net.cpp:103] Top shape: 30 512 41 41 (25820160)
I0223 14:17:56.983084 40844 layer_factory.hpp:78] Creating layer conv5_1
I0223 14:17:56.984086 40844 net.cpp:67] Creating Layer conv5_1
I0223 14:17:56.985086 40844 net.cpp:394] conv5_1 <- pool4
I0223 14:17:56.986086 40844 net.cpp:356] conv5_1 -> conv5_1
I0223 14:17:56.987087 40844 net.cpp:96] Setting up conv5_1
I0223 14:17:56.997097 40844 net.cpp:103] Top shape: 30 512 41 41 (25820160)
I0223 14:17:56.998097 40844 layer_factory.hpp:78] Creating layer relu5_1
I0223 14:17:56.999099 40844 net.cpp:67] Creating Layer relu5_1
I0223 14:17:57.001102 40844 net.cpp:394] relu5_1 <- conv5_1
I0223 14:17:57.002101 40844 net.cpp:345] relu5_1 -> conv5_1 (in-place)
I0223 14:17:57.003103 40844 net.cpp:96] Setting up relu5_1
I0223 14:17:57.004104 40844 net.cpp:103] Top shape: 30 512 41 41 (25820160)
I0223 14:17:57.005105 40844 layer_factory.hpp:78] Creating layer conv5_2
I0223 14:17:57.008107 40844 net.cpp:67] Creating Layer conv5_2
I0223 14:17:57.008107 40844 net.cpp:394] conv5_2 <- conv5_1
I0223 14:17:57.010109 40844 net.cpp:356] conv5_2 -> conv5_2
I0223 14:17:57.010109 40844 net.cpp:96] Setting up conv5_2
I0223 14:17:57.023121 40844 net.cpp:103] Top shape: 30 512 41 41 (25820160)
I0223 14:17:57.024122 40844 layer_factory.hpp:78] Creating layer relu5_2
I0223 14:17:57.026124 40844 net.cpp:67] Creating Layer relu5_2
I0223 14:17:57.027127 40844 net.cpp:394] relu5_2 <- conv5_2
I0223 14:17:57.028126 40844 net.cpp:345] relu5_2 -> conv5_2 (in-place)
I0223 14:17:57.029127 40844 net.cpp:96] Setting up relu5_2
I0223 14:17:57.030128 40844 net.cpp:103] Top shape: 30 512 41 41 (25820160)
I0223 14:17:57.031129 40844 layer_factory.hpp:78] Creating layer conv5_3
I0223 14:17:57.031129 40844 net.cpp:67] Creating Layer conv5_3
I0223 14:17:57.032130 40844 net.cpp:394] conv5_3 <- conv5_2
I0223 14:17:57.033131 40844 net.cpp:356] conv5_3 -> conv5_3
I0223 14:17:57.034132 40844 net.cpp:96] Setting up conv5_3
I0223 14:17:57.049146 40844 net.cpp:103] Top shape: 30 512 41 41 (25820160)
I0223 14:17:57.050148 40844 layer_factory.hpp:78] Creating layer relu5_3
I0223 14:17:57.052150 40844 net.cpp:67] Creating Layer relu5_3
I0223 14:17:57.053150 40844 net.cpp:394] relu5_3 <- conv5_3
I0223 14:17:57.054152 40844 net.cpp:345] relu5_3 -> conv5_3 (in-place)
I0223 14:17:57.055152 40844 net.cpp:96] Setting up relu5_3
I0223 14:17:57.057153 40844 net.cpp:103] Top shape: 30 512 41 41 (25820160)
I0223 14:17:57.057153 40844 layer_factory.hpp:78] Creating layer pool5
I0223 14:17:57.059155 40844 net.cpp:67] Creating Layer pool5
I0223 14:17:57.060158 40844 net.cpp:394] pool5 <- conv5_3
I0223 14:17:57.061158 40844 net.cpp:356] pool5 -> pool5
I0223 14:17:57.062160 40844 net.cpp:96] Setting up pool5
I0223 14:17:57.063159 40844 net.cpp:103] Top shape: 30 512 41 41 (25820160)
I0223 14:17:57.065161 40844 layer_factory.hpp:78] Creating layer pool5a
I0223 14:17:57.067163 40844 net.cpp:67] Creating Layer pool5a
I0223 14:17:57.068164 40844 net.cpp:394] pool5a <- pool5
I0223 14:17:57.069165 40844 net.cpp:356] pool5a -> pool5a
I0223 14:17:57.070166 40844 net.cpp:96] Setting up pool5a
I0223 14:17:57.071167 40844 net.cpp:103] Top shape: 30 512 41 41 (25820160)
I0223 14:17:57.072168 40844 layer_factory.hpp:78] Creating layer fc6
I0223 14:17:57.073169 40844 net.cpp:67] Creating Layer fc6
I0223 14:17:57.074170 40844 net.cpp:394] fc6 <- pool5a
I0223 14:17:57.074170 40844 net.cpp:356] fc6 -> fc6
I0223 14:17:57.075171 40844 net.cpp:96] Setting up fc6
I0223 14:17:57.094190 40844 net.cpp:103] Top shape: 30 1024 41 41 (51640320)
I0223 14:17:57.099195 40844 layer_factory.hpp:78] Creating layer relu6
I0223 14:17:57.100195 40844 net.cpp:67] Creating Layer relu6
I0223 14:17:57.101197 40844 net.cpp:394] relu6 <- fc6
I0223 14:17:57.102196 40844 net.cpp:345] relu6 -> fc6 (in-place)
I0223 14:17:57.103199 40844 net.cpp:96] Setting up relu6
I0223 14:17:57.104198 40844 net.cpp:103] Top shape: 30 1024 41 41 (51640320)
I0223 14:17:57.106200 40844 layer_factory.hpp:78] Creating layer drop6
I0223 14:17:57.107201 40844 net.cpp:67] Creating Layer drop6
I0223 14:17:57.108202 40844 net.cpp:394] drop6 <- fc6
I0223 14:17:57.110204 40844 net.cpp:345] drop6 -> fc6 (in-place)
I0223 14:17:57.111205 40844 net.cpp:96] Setting up drop6
I0223 14:17:57.111205 40844 net.cpp:103] Top shape: 30 1024 41 41 (51640320)
I0223 14:17:57.112206 40844 layer_factory.hpp:78] Creating layer fc7
I0223 14:17:57.113207 40844 net.cpp:67] Creating Layer fc7
I0223 14:17:57.114209 40844 net.cpp:394] fc7 <- fc6
I0223 14:17:57.115211 40844 net.cpp:356] fc7 -> fc7
I0223 14:17:57.116211 40844 net.cpp:96] Setting up fc7
I0223 14:17:57.120214 40844 net.cpp:103] Top shape: 30 1024 41 41 (51640320)
I0223 14:17:57.121214 40844 layer_factory.hpp:78] Creating layer relu7
I0223 14:17:57.122215 40844 net.cpp:67] Creating Layer relu7
I0223 14:17:57.123217 40844 net.cpp:394] relu7 <- fc7
I0223 14:17:57.126220 40844 net.cpp:345] relu7 -> fc7 (in-place)
I0223 14:17:57.130223 40844 net.cpp:96] Setting up relu7
I0223 14:17:57.131225 40844 net.cpp:103] Top shape: 30 1024 41 41 (51640320)
I0223 14:17:57.135228 40844 layer_factory.hpp:78] Creating layer drop7
I0223 14:17:57.137230 40844 net.cpp:67] Creating Layer drop7
I0223 14:17:57.138231 40844 net.cpp:394] drop7 <- fc7
I0223 14:17:57.139231 40844 net.cpp:345] drop7 -> fc7 (in-place)
I0223 14:17:57.140233 40844 net.cpp:96] Setting up drop7
I0223 14:17:57.141233 40844 net.cpp:103] Top shape: 30 1024 41 41 (51640320)
I0223 14:17:57.142235 40844 layer_factory.hpp:78] Creating layer fc8_voc12
I0223 14:17:57.142235 40844 net.cpp:67] Creating Layer fc8_voc12
I0223 14:17:57.143235 40844 net.cpp:394] fc8_voc12 <- fc7
I0223 14:17:57.144237 40844 net.cpp:356] fc8_voc12 -> fc8_voc12
I0223 14:17:57.144237 40844 net.cpp:96] Setting up fc8_voc12
I0223 14:17:57.146239 40844 net.cpp:103] Top shape: 30 21 41 41 (1059030)
I0223 14:17:57.147239 40844 layer_factory.hpp:78] Creating layer fc8_voc12_fc8_voc12_0_split
I0223 14:17:57.149241 40844 net.cpp:67] Creating Layer fc8_voc12_fc8_voc12_0_split
I0223 14:17:57.150243 40844 net.cpp:394] fc8_voc12_fc8_voc12_0_split <- fc8_voc12
I0223 14:17:57.152245 40844 net.cpp:356] fc8_voc12_fc8_voc12_0_split -> fc8_voc12_fc8_voc12_0_split_0
I0223 14:17:57.153245 40844 net.cpp:356] fc8_voc12_fc8_voc12_0_split -> fc8_voc12_fc8_voc12_0_split_1
I0223 14:17:57.154247 40844 net.cpp:96] Setting up fc8_voc12_fc8_voc12_0_split
I0223 14:17:57.156250 40844 net.cpp:103] Top shape: 30 21 41 41 (1059030)
I0223 14:17:57.157249 40844 net.cpp:103] Top shape: 30 21 41 41 (1059030)
I0223 14:17:57.157249 40844 layer_factory.hpp:78] Creating layer label_shrink
I0223 14:17:57.158251 40844 net.cpp:67] Creating Layer label_shrink
I0223 14:17:57.158251 40844 net.cpp:394] label_shrink <- label
I0223 14:17:57.159252 40844 net.cpp:356] label_shrink -> label_shrink
I0223 14:17:57.159252 40844 net.cpp:96] Setting up label_shrink
I0223 14:17:57.160253 40844 net.cpp:103] Top shape: 30 1 41 41 (50430)
I0223 14:17:57.160253 40844 layer_factory.hpp:78] Creating layer label_shrink_label_shrink_0_split
I0223 14:17:57.162255 40844 net.cpp:67] Creating Layer label_shrink_label_shrink_0_split
I0223 14:17:57.164255 40844 net.cpp:394] label_shrink_label_shrink_0_split <- label_shrink
I0223 14:17:57.165256 40844 net.cpp:356] label_shrink_label_shrink_0_split -> label_shrink_label_shrink_0_split_0
I0223 14:17:57.166257 40844 net.cpp:356] label_shrink_label_shrink_0_split -> label_shrink_label_shrink_0_split_1
I0223 14:17:57.168259 40844 net.cpp:96] Setting up label_shrink_label_shrink_0_split
I0223 14:17:57.169260 40844 net.cpp:103] Top shape: 30 1 41 41 (50430)
I0223 14:17:57.170262 40844 net.cpp:103] Top shape: 30 1 41 41 (50430)
I0223 14:17:57.171262 40844 layer_factory.hpp:78] Creating layer loss
I0223 14:17:57.172264 40844 net.cpp:67] Creating Layer loss
I0223 14:17:57.173264 40844 net.cpp:394] loss <- fc8_voc12_fc8_voc12_0_split_0
I0223 14:17:57.174266 40844 net.cpp:394] loss <- label_shrink_label_shrink_0_split_0
I0223 14:17:57.175266 40844 net.cpp:356] loss -> (automatic)
I0223 14:17:57.175266 40844 net.cpp:96] Setting up loss
I0223 14:17:57.176267 40844 softmax_loss_layer.cpp:40] Weight_Loss file is not provided. Assign all one to it.
I0223 14:17:57.177268 40844 net.cpp:103] Top shape: 1 1 1 1 (1)
I0223 14:17:57.178269 40844 net.cpp:109] with loss weight 1
I0223 14:17:57.179270 40844 layer_factory.hpp:78] Creating layer accuracy
I0223 14:17:57.180271 40844 net.cpp:67] Creating Layer accuracy
I0223 14:17:57.181272 40844 net.cpp:394] accuracy <- fc8_voc12_fc8_voc12_0_split_1
I0223 14:17:57.183274 40844 net.cpp:394] accuracy <- label_shrink_label_shrink_0_split_1
I0223 14:17:57.183274 40844 net.cpp:356] accuracy -> accuracy
I0223 14:17:57.185276 40844 net.cpp:96] Setting up accuracy
I0223 14:17:57.186276 40844 net.cpp:103] Top shape: 1 1 1 3 (3)
I0223 14:17:57.187278 40844 net.cpp:172] accuracy does not need backward computation.
I0223 14:17:57.189280 40844 net.cpp:170] loss needs backward computation.
I0223 14:17:57.191282 40844 net.cpp:172] label_shrink_label_shrink_0_split does not need backward computation.
I0223 14:17:57.192282 40844 net.cpp:172] label_shrink does not need backward computation.
I0223 14:17:57.193284 40844 net.cpp:170] fc8_voc12_fc8_voc12_0_split needs backward computation.
I0223 14:17:57.194285 40844 net.cpp:170] fc8_voc12 needs backward computation.
I0223 14:17:57.195286 40844 net.cpp:170] drop7 needs backward computation.
I0223 14:17:57.196287 40844 net.cpp:170] relu7 needs backward computation.
I0223 14:17:57.197288 40844 net.cpp:170] fc7 needs backward computation.
I0223 14:17:57.198288 40844 net.cpp:170] drop6 needs backward computation.
I0223 14:17:57.199290 40844 net.cpp:170] relu6 needs backward computation.
I0223 14:17:57.200289 40844 net.cpp:170] fc6 needs backward computation.
I0223 14:17:57.201292 40844 net.cpp:170] pool5a needs backward computation.
I0223 14:17:57.202292 40844 net.cpp:170] pool5 needs backward computation.
I0223 14:17:57.203292 40844 net.cpp:170] relu5_3 needs backward computation.
I0223 14:17:57.204294 40844 net.cpp:170] conv5_3 needs backward computation.
I0223 14:17:57.204294 40844 net.cpp:170] relu5_2 needs backward computation.
I0223 14:17:57.205296 40844 net.cpp:170] conv5_2 needs backward computation.
I0223 14:17:57.206295 40844 net.cpp:170] relu5_1 needs backward computation.
I0223 14:17:57.207298 40844 net.cpp:170] conv5_1 needs backward computation.
I0223 14:17:57.207298 40844 net.cpp:170] pool4 needs backward computation.
I0223 14:17:57.208297 40844 net.cpp:170] relu4_3 needs backward computation.
I0223 14:17:57.208297 40844 net.cpp:170] conv4_3 needs backward computation.
I0223 14:17:57.209298 40844 net.cpp:170] relu4_2 needs backward computation.
I0223 14:17:57.210300 40844 net.cpp:170] conv4_2 needs backward computation.
I0223 14:17:57.211302 40844 net.cpp:170] relu4_1 needs backward computation.
I0223 14:17:57.212301 40844 net.cpp:170] conv4_1 needs backward computation.
I0223 14:17:57.213302 40844 net.cpp:170] pool3 needs backward computation.
I0223 14:17:57.214303 40844 net.cpp:170] relu3_3 needs backward computation.
I0223 14:17:57.216305 40844 net.cpp:170] conv3_3 needs backward computation.
I0223 14:17:57.217306 40844 net.cpp:170] relu3_2 needs backward computation.
I0223 14:17:57.217306 40844 net.cpp:170] conv3_2 needs backward computation.
I0223 14:17:57.219311 40844 net.cpp:170] relu3_1 needs backward computation.
I0223 14:17:57.220309 40844 net.cpp:170] conv3_1 needs backward computation.
I0223 14:17:57.220309 40844 net.cpp:170] pool2 needs backward computation.
I0223 14:17:57.221310 40844 net.cpp:170] relu2_2 needs backward computation.
I0223 14:17:57.222311 40844 net.cpp:170] conv2_2 needs backward computation.
I0223 14:17:57.223312 40844 net.cpp:170] relu2_1 needs backward computation.
I0223 14:17:57.224313 40844 net.cpp:170] conv2_1 needs backward computation.
I0223 14:17:57.224313 40844 net.cpp:170] pool1 needs backward computation.
I0223 14:17:57.225314 40844 net.cpp:170] relu1_2 needs backward computation.
I0223 14:17:57.226315 40844 net.cpp:170] conv1_2 needs backward computation.
I0223 14:17:57.227315 40844 net.cpp:170] relu1_1 needs backward computation.
I0223 14:17:57.228317 40844 net.cpp:170] conv1_1 needs backward computation.
I0223 14:17:57.229318 40844 net.cpp:172] data does not need backward computation.
I0223 14:17:57.229318 40844 net.cpp:208] This network produces output accuracy
I0223 14:17:57.230319 40844 net.cpp:467] Collecting Learning Rate and Weight Decay.
I0223 14:17:57.231319 40844 net.cpp:219] Network initialization done.
I0223 14:17:57.232321 40844 net.cpp:220] Memory required for data: 9170170576
I0223 14:17:57.233321 40844 solver.cpp:41] Solver scaffolding done.
I0223 14:17:57.234323 40844 caffe.cpp:153] Finetuning from voc12/model/vgg128_noup/init.caffemodel
I0223 14:17:58.057106 40844 net.cpp:740] Target layer fc8_voc12 not initialized.
I0223 14:17:58.067114 40844 solver.cpp:160] Solving vgg128_noup
I0223 14:17:58.068115 40844 solver.cpp:161] Learning Rate Policy: step
F0223 14:17:58.070117 40844 blob.hpp:55] Check failed: n <= num_ (0 vs. -335549216)
*** Check failure stack trace: ***

Unable to deploy the Deeplab-Coco and VOC2012 finetuned dataset using python

When I tried deploying the net using caffe.Net and caffe.classifier interface using the prototxt and caffemodel given in the link
http://web.cs.ucla.edu/~lcchen/deeplab-public/vgg128_noup_pool3_cocomix/

I am getting an error message saying that

F0212 12:52:56.985724 7144 insert_splits.cpp:35] Unknown blob input data to layer 0
*** Check failure stack trace: ***
Aborted (core dumped)

Am I supposed to use a different prototxt file ?

Check failed: error == cudaSuccess (8 vs. 0) invalid device function, out of memory

I run the codes on Ubuntu 16.04, and caffe has GPU support (Quadro M4000, 8GB ), but one error occurs, anyone has ideas what's going on? how can I solve this problem? Many thanks for your help.

I0409 20:53:47.976828 21105 net.cpp:170] pool2 needs backward computation.
I0409 20:53:47.976835 21105 net.cpp:170] relu2_2 needs backward computation.
I0409 20:53:47.976843 21105 net.cpp:170] conv2_2 needs backward computation.
I0409 20:53:47.976853 21105 net.cpp:170] relu2_1 needs backward computation.
I0409 20:53:47.976861 21105 net.cpp:170] conv2_1 needs backward computation.
I0409 20:53:47.976869 21105 net.cpp:170] pool1 needs backward computation.
I0409 20:53:47.976878 21105 net.cpp:170] relu1_2 needs backward computation.
I0409 20:53:47.976887 21105 net.cpp:170] conv1_2 needs backward computation.
I0409 20:53:47.976897 21105 net.cpp:170] relu1_1 needs backward computation.
I0409 20:53:47.976907 21105 net.cpp:170] conv1_1 needs backward computation.
I0409 20:53:47.976915 21105 net.cpp:172] data does not need backward computation.
I0409 20:53:47.976925 21105 net.cpp:208] This network produces output accuracy
I0409 20:53:47.976969 21105 net.cpp:467] Collecting Learning Rate and Weight Decay.
I0409 20:53:47.976986 21105 net.cpp:219] Network initialization done.
I0409 20:53:47.976995 21105 net.cpp:220] Memory required for data: 9170170576
I0409 20:53:47.977139 21105 solver.cpp:41] Solver scaffolding done.
I0409 20:53:47.977152 21105 caffe.cpp:118] Finetuning from voc12/model/vgg128_noup/init.caffemodel
I0409 20:53:48.125612 21105 net.cpp:740] Target layer fc6 not initialized.
I0409 20:53:48.125658 21105 net.cpp:740] Target layer fc7 not initialized.
I0409 20:53:48.125666 21105 net.cpp:740] Target layer fc8_voc12 not initialized.
I0409 20:53:48.127007 21105 solver.cpp:160] Solving vgg128_noup
I0409 20:53:48.127018 21105 solver.cpp:161] Learning Rate Policy: step
F0409 20:53:48.199013 21105 im2col.cu:68] Check failed: error == cudaSuccess (8 vs. 0) invalid device function
*** Check failure stack trace: ***
@ 0x7fd71f3715cd google::LogMessage::Fail()
@ 0x7fd71f373433 google::LogMessage::SendToLog()
@ 0x7fd71f37115b google::LogMessage::Flush()
@ 0x7fd71f373e1e google::LogMessageFatal::~LogMessageFatal()
@ 0x5c878a caffe::im2col_gpu<>()
@ 0x5bd086 caffe::ConvolutionLayer<>::Forward_gpu()
@ 0x55aaba caffe::Net<>::ForwardFromTo()
@ 0x55aca7 caffe::Net<>::ForwardPrefilled()
@ 0x5aef94 caffe::Solver<>::Solve()
@ 0x41be0a train()
@ 0x415708 main
@ 0x7fd71c112830 __libc_start_main
@ 0x41a769 _start
@ (nil) (unknown)
Aborted (core dumped)

Reading .mat files

Hi,

I get .mat files as output but I don't have matlab. I heard we can use matio or scipy to read them but the shape of the array is like (579,579,21,1). I can't figure how to see them as images. Any help, pointers or suggestions are appreciated.

Thanks for sharing the wonderful project.

F1120 22:05:28.240000 34447 net.cpp:717] Check failed: target_blobs[j]->channels() == source_layer.blobs(j).channels() (1 vs. 3) Incompatible parameter size for layer conv1_1

Training1 net voc12/vgg128_noup
Running ./build/tools/caffe train --solver=voc12/config/vgg128_noup/solver_train_aug.prototxt --weights=voc12/model/vgg128_noup/init.caffemodel --gpu=0
I1120 22:05:22.482954 34447 caffe.cpp:102] Use GPU with device ID 0
I1120 22:05:26.820761 34447 caffe.cpp:110] Starting Optimization
I1120 22:05:26.820894 34447 solver.cpp:32] Initializing solver from parameters:
train_net: "voc12/config/vgg128_noup/train_train_aug.prototxt"
....
....
....
....
....
I1120 22:05:27.910971 34447 solver.cpp:41] Solver scaffolding done.
I1120 22:05:27.910979 34447 caffe.cpp:118] Finetuning from voc12/model/vgg128_noup/init.caffemodel
I1120 22:05:28.239959 34447 net.cpp:707] Copying source layer conv1_1
F1120 22:05:28.240000 34447 net.cpp:717] Check failed: target_blobs[j]->channels() == source_layer.blobs(j).channels() (1 vs. 3) Incompatible parameter size for layer conv1_1
*** Check failure stack trace: ***
@ 0x7fe28b87aa5d google::LogMessage::Fail()
@ 0x7fe28b87eef7 google::LogMessage::SendToLog()
@ 0x7fe28b87cd59 google::LogMessage::Flush()
@ 0x7fe28b87d05d google::LogMessageFatal::~LogMessageFatal()
@ 0x5946d0 caffe::Net<>::CopyTrainedLayersFrom()
@ 0x594df3 caffe::Net<>::CopyTrainedLayersFrom()
@ 0x56c301 train()
@ 0x56e88f main
@ 0x30f8a1ed1d (unknown)
@ 0x56b8a9 (unknown)

Process finished with exit code 0

please help me ,thanks

Questions on softmax layer

Hi!

I am reproducing the results of DeepLab, and I met a problem.

I used the method mentioned here, and when I run run.py, it told me that there are unexpected labels in softmax layer (such as 55, 150 and so on). I have to force NUM_LABELS=255 to make it run.

How could I fix this problem? Thanks very much!

Shangxuan

Test problem with DEEPLABV2-RESNET101

Dear Ali,

I tried to evaluate the segmentation results with DeepLabV2-RESNET101. However, when I used "MatWrite" layer to save the segmentation results to mat files. I found that inferred segmentation results have been cropped to 513 x 513. So when I use 'EvalRun.m' to evaluate the results, errors occurred in 'MyVOCevalseg.m'.

"Results image '2007_000033' is the wrong size, was 513 x 513, should be 366 x 500."

In this case, should I save the cropped groundtruth label images(513x513) in order to evaluate the results? Or is there any other solutions (such as saving inferred segmentation results to original size--366x500 with 'MatWrite' layer)?

Thank you very much for your time.
Kind regards,

Ruihao

Data augmentation has no effect?

when I add "scale_factors" in transform_param, I can not get better result,why?

transform_param {
mean_value: 104.008
mean_value: 116.669
mean_value: 122.675
crop_size: 321
mirror: true
scale_factors: 0.5
scale_factors: 0.75
scale_factors: 1
scale_factors: 1.25
scale_factors: 1.5
}

error with src/caffe/layers/window_data_layer.cpp when building.

My environment: Ubuntu 14.04, Opencv3.0, cuda8, cudnn5. When make all in directory /DeepLab-Context/, error occures:
src/caffe/layers/window_data_layer.cpp:26:11: error: ‘const int CV_LOAD_IMAGE_COLOR’ redeclared as different kind of symbol
const int CV_LOAD_IMAGE_COLOR = cv::IMREAD_COLOR;

In file included from /usr/local/include/opencv2/highgui/highgui_c.h:47:0,
from /usr/local/include/opencv2/highgui.hpp:701,
from /usr/local/include/opencv2/highgui/highgui.hpp:48,
from src/caffe/layers/window_data_layer.cpp:10:
/usr/local/include/opencv2/imgcodecs/imgcodecs_c.h:62:5: error: previous declaration of ‘ CV_LOAD_IMAGE_COLOR’
CV_LOAD_IMAGE_COLOR =1,
^
make: *** [.build_release/src/caffe/layers/window_data_layer.o] error 1

So, what should i do? thanks for help!

the version of caffe

hi, @TheLegendAli
when I python run.py ,ERROR: [libprotobuf ERROR google/protobuf/text_format.cc:245] Error parsing text-format caffe.NetParameter: 24:3: Unknown enumeration value of "IMAGE_SEG_DATA" for field "type".
It look like a version issue. my caffe version is 1.0.0, could you tell which version of yours.
Or how do I get rid of this error?
Help is greatly appreciated!

database link

Link to additional dataset Bharath Hariharan et al. is not valid. where can I get this dataset.

shape mismatch at layer 'fc6'

Hi,

I'm getting the following error.

Cannot copy param 0 weights from layer 'fc6'; shape mismatch. Source param shape is 1024 512 3 3 (4718592); target param shape is 4096 512 4 4 (33554432). To learn this layer's parameters from scratch rather than copying from a saved net, rename the layer.

Any idea what is wrong here?
Thanks in advance.

Can not load the pre-trained models.

Hi Ali. Thanks for this program. I have one issue - can not load model:
Downloading init caffemodel from http://ccvl.stat.ucla.edu/ccvl/init_models/vgg16_20M.caffemodel
Traceback (most recent call last):
File "run.py", line 67, in
run(RUN_TRAIN, RUN_TEST, RUN_TRAIN2, RUN_TEST2, RUN_SAVE)
File "run.py", line 55, in run
if RUN_TRAIN : trainer()
File "/home/lumius/Dropbox/DeepLab-Context/python/my_script/trainer.py", line 52, in trainer
model = train_prototxt_maker(train, init, solver, train_set)
File "/home/lumius/Dropbox/DeepLab-Context/python/my_script/trainer.py", line 32, in train_prototxt_maker
if not os.path.isfile(model): model=model_finder(os.environ['EXP']+ '/model/' + os.environ['NET_ID'])
File "/home/lumius/Dropbox/DeepLab-Context/python/my_script/tools.py", line 16, in model_finder
filename = wget.download(url)
File "/usr/local/lib/python2.7/dist-packages/wget.py", line 526, in download
(tmpfile, headers) = ulib.urlretrieve(binurl, tmpfile, callback)
File "/usr/lib/python2.7/urllib.py", line 98, in urlretrieve
return opener.retrieve(url, filename, reporthook, data)
File "/usr/lib/python2.7/urllib.py", line 245, in retrieve
fp = self.open(url, data)
File "/usr/lib/python2.7/urllib.py", line 213, in open
return getattr(self, name)(url)
File "/usr/lib/python2.7/urllib.py", line 364, in open_http
return self.http_error(url, fp, errcode, errmsg, headers)
File "/usr/lib/python2.7/urllib.py", line 377, in http_error
result = method(url, fp, errcode, errmsg, headers)
File "/usr/lib/python2.7/urllib.py", line 673, in http_error_301
return self.http_error_302(url, fp, errcode, errmsg, headers, data)
File "/usr/lib/python2.7/urllib.py", line 642, in http_error_302
headers, data)
File "/usr/lib/python2.7/urllib.py", line 669, in redirect_internal
return self.open(newurl)
File "/usr/lib/python2.7/urllib.py", line 213, in open
return getattr(self, name)(url)
File "/usr/lib/python2.7/urllib.py", line 443, in open_https
h.endheaders(data)
File "/usr/lib/python2.7/httplib.py", line 1053, in endheaders
self._send_output(message_body)
File "/usr/lib/python2.7/httplib.py", line 897, in _send_output
self.send(msg)
File "/usr/lib/python2.7/httplib.py", line 859, in send
self.connect()
File "/usr/lib/python2.7/httplib.py", line 1278, in connect
server_hostname=server_hostname)
File "/usr/lib/python2.7/ssl.py", line 353, in wrap_socket
_context=self)
File "/usr/lib/python2.7/ssl.py", line 601, in init
self.do_handshake()
File "/usr/lib/python2.7/ssl.py", line 830, in do_handshake
self._sslobj.do_handshake()
IOError: [Errno socket error] [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed (_ssl.c:590)
then run run.py

Seems that there's no use of any 'crf' folders.

Seems that there's no use of any 'crf' folders. Does it mean that there are no crf features generated?
Also curious about why using tool.saver() to save the model and there's no deploy.prototxt.

How to pretrain model on MS-COCO?

I'm the new to semantic segmentation, the deeplab paper shows that pre-train models on coco dataset can much improve the results, but the coco dataset just give the instance level labels, how to transfer data to satisfy the category level requirement, or anyone can give me some examples on how to pre-train models on coco datasets?
Thanks!

error while loading shared libraries: libmatio.so.2

Hi,

I'm trying to run run.py script but encounter this issue. I've installed libmatio and I can see libmatio.so.2 located in "/usr/local/lib".
I've tried reinstall libmatio and caffe several times but still failed, can anyone help me? Thanks!

How to train on my data

HI @TheLegendAli

I'm try to train Deeplab v2 with my data. But i don't know how create data for training.
In file train_aug.txt, only mapping between original image (JPEGImages/2007_000364.jpg) with contour image (/SegmentationClassAug/2007_000364.png) --> That seem all image in SegmentationClassAug folder is edge(contour). So how program can mapping color label for train?

Thanks

net->copytrainedlayer failed.

I have successfully compiled the code and trained a model using my own data.Then I wrote a c++ script to test a single image .But unfortunately, net->copytrainedlayer(model) seems not work ,weights of layers are still 0.Can you help me or just give me some clues ?
ps, I highly doubt that if this function can not work well,the training process may be also incorrect.

I0413 11:50:28.412251 9144 net.cpp:170] relu1_2 needs backward computation.
I0413 11:50:28.412251 9144 net.cpp:170] conv1_2 needs backward computation.
I0413 11:50:28.413254 9144 net.cpp:170] relu1_1 needs backward computation.
I0413 11:50:28.413254 9144 net.cpp:170] conv1_1 needs backward computation.
I0413 11:50:28.414253 9144 net.cpp:172] data does not need backward computation.
I0413 11:50:28.414253 9144 net.cpp:208] This network produces output accuracy
I0413 11:50:28.414253 9144 net.cpp:467] Collecting Learning Rate and Weight Decay.
I0413 11:50:28.416255 9144 net.cpp:219] Network initialization done.
I0413 11:50:28.416255 9144 net.cpp:220] Memory required for data: 2445378832
I0413 11:50:28.417255 9144 solver.cpp:41] Solver scaffolding done.
I0413 11:50:28.417255 9144 caffe.cpp:118] Finetuning from voc12/model/vgg128_noup/init.caffemodel
I0413 11:50:28.417255 9144 net.cpp:740] Target layer conv1_1 not initialized.
I0413 11:50:28.418267 9144 net.cpp:740] Target layer conv1_2 not initialized.
I0413 11:50:28.418267 9144 net.cpp:740] Target layer conv2_1 not initialized.
I0413 11:50:28.419257 9144 net.cpp:740] Target layer conv2_2 not initialized.
I0413 11:50:28.419257 9144 net.cpp:740] Target layer conv3_1 not initialized.
I0413 11:50:28.420267 9144 net.cpp:740] Target layer conv3_2 not initialized.
I0413 11:50:28.420267 9144 net.cpp:740] Target layer conv3_3 not initialized.
Could not create logging file: File exists
COULD NOT CREATE A LOGGINGFILE 20170413-115028.4828!I0413 11:50:28.421259 9144 net.cpp:740] Target layer conv4_1 not initialized.
I0413 11:50:28.422260 9144 net.cpp:740] Target layer conv4_2 not initialized.
I0413 11:50:28.422260 9144 net.cpp:740] Target layer conv4_3 not initialized.
I0413 11:50:28.422260 9144 net.cpp:740] Target layer conv5_1 not initialized.
I0413 11:50:28.423260 9144 net.cpp:740] Target layer conv5_2 not initialized.
I0413 11:50:28.423260 9144 net.cpp:740] Target layer conv5_3 not initialized.
I0413 11:50:28.424262 9144 net.cpp:740] Target layer fc6 not initialized.
I0413 11:50:28.424262 9144 net.cpp:740] Target layer fc7 not initialized.
I0413 11:50:28.425261 9144 net.cpp:740] Target layer fc8_voc12 not initialized.
I0413 11:50:28.426262 9144 solver.cpp:160] Solving vgg128_noup
I0413 11:50:28.426262 9144 solver.cpp:161] Learning Rate Policy: step
I0413 11:50:29.027125 9144 solver.cpp:209] Iteration 0, loss = 3.04452
I0413 11:50:35.966346 9144 solver.cpp:224] Train net output #0: accuracy = 0
I0413 11:50:35.967347 9144 solver.cpp:224] Train net output #1: accuracy = 0
I0413 11:50:35.967347 9144 solver.cpp:224] Train net output #2: accuracy = 0.571429

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.