I am working on a limited number of large images, each of which may have 3072*3072pixels. To train the semantic segmentation model using FCN or U-net, I will build a large selection of training sets, each training image 128*128.
At the forecasting stage, I do this in order to cut a large image into small pieces, just like a training kit 128*128, and feed these small pieces into a trained model, to get a predicted mask. Subsequently, I simply stitch these small patches to get a mask for the whole image. Is this the right mechanism to perform semantic segmentation against large images?
source
share