I am new to CNNs and am building a model using Keras to combine inputs from multiple sources. Two of my sources have different dimensions and cannot be scaled by an integer number (i.e., x2 or x3 smaller). Therefore, simply max-pooling will not work. I am having trouble figuring out how to downsample the larger image. Here are the exact dimensions:
Image1: 7000 x 4000
Image2: 2607 x 1370
Is there a best practice for dealing with non-conventional downsampling?
I am applying a Conv2D layer and am thinking that combing the appropriately sized filter (1787x1261 with stride=1) with a max pooling (2x2 and stride=2) would give me the correct dimensions. Any reason why that is a bad idea? This does seem like a large filter compared to the total size of the image.
Somewhat related, would it be better to run the model on smaller chunks of the full image? That way I could control the size of each chunk?