I am trying to use the deep word code that Caffe uses. They use the GoogLeNet model, pre-prepared on ImageNet, as envisioned by ModelZoo. This means that the network has been trained in images cropped to 224x224 pixels. From train_val.prototext :
layer {
name: "data"
type: "Data"
...
transform_param {
mirror: true
crop_size: 224
...
deploy.prototext , used for processing, also defines an input level of size 224x224x3x10 (RGB images of size 224x224, batchsize 10).
name: "GoogleNet"
input: "data"
input_shape {
dim: 10
dim: 3
dim: 224
dim: 224
}
However, I can use this network to process images of any size (the example above used one of 1024x574 pixels).
, , ?