diff --git a/examples/cifar10/cifar10_full_train_test.prototxt b/examples/cifar10/cifar10_full_train_test.prototxt index 8fde19f046e..936920b3a24 100644 --- a/examples/cifar10/cifar10_full_train_test.prototxt +++ b/examples/cifar10/cifar10_full_train_test.prototxt @@ -6,6 +6,7 @@ layers { top: "label" data_param { source: "examples/cifar10/cifar10_train_leveldb" + backend: LEVELDB batch_size: 100 } transform_param { @@ -20,6 +21,7 @@ layers { top: "label" data_param { source: "examples/cifar10/cifar10_test_leveldb" + backend: LEVELDB batch_size: 100 } transform_param { diff --git a/examples/cifar10/cifar10_quick_train_test.prototxt b/examples/cifar10/cifar10_quick_train_test.prototxt index 409cfe809f4..30f801a580d 100644 --- a/examples/cifar10/cifar10_quick_train_test.prototxt +++ b/examples/cifar10/cifar10_quick_train_test.prototxt @@ -6,6 +6,7 @@ layers { top: "label" data_param { source: "examples/cifar10/cifar10_train_leveldb" + backend: LEVELDB batch_size: 100 } transform_param { @@ -20,6 +21,7 @@ layers { top: "label" data_param { source: "examples/cifar10/cifar10_test_leveldb" + backend: LEVELDB batch_size: 100 } transform_param { diff --git a/examples/cifar10/create_cifar10.sh b/examples/cifar10/create_cifar10.sh index ad5038e0c3e..dfba7cca48a 100755 --- a/examples/cifar10/create_cifar10.sh +++ b/examples/cifar10/create_cifar10.sh @@ -13,6 +13,6 @@ rm -rf $EXAMPLE/cifar10_train_leveldb $EXAMPLE/cifar10_test_leveldb echo "Computing image mean..." ./build/tools/compute_image_mean $EXAMPLE/cifar10_train_leveldb \ - $EXAMPLE/mean.binaryproto + $EXAMPLE/mean.binaryproto leveldb echo "Done." diff --git a/examples/siamese/mnist_siamese_train_test.prototxt b/examples/siamese/mnist_siamese_train_test.prototxt index 92361c31dc7..b9c08e6c179 100644 --- a/examples/siamese/mnist_siamese_train_test.prototxt +++ b/examples/siamese/mnist_siamese_train_test.prototxt @@ -6,6 +6,7 @@ layers { top: "sim" data_param { source: "examples/siamese/mnist_siamese_train_leveldb" + backend: LEVELDB scale: 0.00390625 batch_size: 64 } @@ -18,6 +19,7 @@ layers { top: "sim" data_param { source: "examples/siamese/mnist_siamese_test_leveldb" + backend: LEVELDB scale: 0.00390625 batch_size: 100 } diff --git a/examples/siamese/readme.md b/examples/siamese/readme.md index ce98ec10819..07ef89c51b8 100644 --- a/examples/siamese/readme.md +++ b/examples/siamese/readme.md @@ -71,6 +71,7 @@ or different classes (`sim`). top: "sim" data_param { source: "examples/siamese/mnist-siamese-train-leveldb" + backend: LEVELDB scale: 0.00390625 batch_size: 64 } diff --git a/src/caffe/proto/caffe.proto b/src/caffe/proto/caffe.proto index 9395c38f3e9..b07ba83fa82 100644 --- a/src/caffe/proto/caffe.proto +++ b/src/caffe/proto/caffe.proto @@ -414,9 +414,9 @@ message DataParameter { // The rand_skip variable is for the data layer to skip a few data points // to avoid all asynchronous sgd clients to start at the same point. The skip // point would be set as rand_skip * rand(0,1). Note that rand_skip should not - // be larger than the number of keys in the leveldb. + // be larger than the number of keys in the db. optional uint32 rand_skip = 7 [default = 0]; - optional DB backend = 8 [default = LEVELDB]; + optional DB backend = 8 [default = LMDB]; // DEPRECATED. See TransformationParameter. For data pre-processing, we can do // simple scaling and subtracting the data mean, if provided. Note that the // mean subtraction is always carried out before scaling. @@ -504,7 +504,7 @@ message ImageDataParameter { // The rand_skip variable is for the data layer to skip a few data points // to avoid all asynchronous sgd clients to start at the same point. The skip // point would be set as rand_skip * rand(0,1). Note that rand_skip should not - // be larger than the number of keys in the leveldb. + // be larger than the number of keys in the db. optional uint32 rand_skip = 7 [default = 0]; // Whether or not ImageLayer should shuffle the list of files at every epoch. optional bool shuffle = 8 [default = false]; @@ -741,7 +741,7 @@ message V0LayerParameter { // The rand_skip variable is for the data layer to skip a few data points // to avoid all asynchronous sgd clients to start at the same point. The skip // point would be set as rand_skip * rand(0,1). Note that rand_skip should not - // be larger than the number of keys in the leveldb. + // be larger than the number of keys in the db. optional uint32 rand_skip = 53 [default = 0]; // Fields related to detection (det_*)